J Chapter quantity quantity.

advertisement
J
Chapter
LINEAR FUNCTIONS
You
probably recall from calculus that a function is a rule which associates
particular values of one variable quantity to particular values of another
variable quantity. Analysis is that branch of mathematics devoted to the
study, or analysis, of functions. The main kind of analysis that goes on is
this: for small changes in the first variable, we try to determine an approxi
mate value to the corresponding change in the second.
Now, we ask, for
in
the
first
variable
to
what
extent
can
we
large changes
predict, from such
the
in
the
second?
The primary
approximations,
corresponding change
involved
in
this
kind
of
is
of
the problem.
technique
analysis
simplification
That is, we replace the given function by a suitable very simple and more
easily calculable function and work with this simple function instead (making
sure to keep in mind the effect of that replacement).
The simplest possible functions are those which behave linearly. This
means that they have a straight line as graph.
Such a function has the
in
the
value
of
the function correspond
The
increment
following property.
to
an
in
is
a
constant
increment
the
variable
ing
multiple of that increment :
f(x
for
some
+
C.
t)-f(x)
=
Ct
Now, when
(1.1)
one moves
to the consideration of
functions of
simplest functions becomes
study
quantities
mathematical
it
forms
a
that
discipline, called linear
special
complex enough
with
the concepts and
of
one
calculus
The
variable,
coupled
algebra.
several variable
the
of
even
these
1
2
1.
Linear Functions
algebra constitute the basic tools of analysis of functions
of several variables.
It is our purpose in this text to study this subject.
First then, we must study the notions and methods of linear algebra.
We begin our study in a familiar context: that of the solution of a system
of simultaneous equations in more than one unknown. We shall develop
a standard technique for discovering solutions (if they exist), called row
reduction.
This is the foundation pillar of the theory of linear algebra.
After a brief section on notational conventions, we look at the system of
equations from another point of view. Instead of seeking particular solu
tions of a system, we analyze the system itself. This leads us to consider the
fundamental concept of linear algebra: that of a linear transformation. In
this larger context we can resolve the question of existence of solutions and
effectively describe the totality of all solutions to a given linear problem.
We proceed then to analyze the totality of linear transformations as an
object of interest in itself. This chapter ends with the study of several
important topics allied with linear algebra. We study the plane as the system
of complex numbers and the inner and vector products in three space.
techniques
1.1
Let
of linear
Simultaneous
us
Equations
begin by considering
a
well-known
to systems of simultaneous linear equations.
is that of two equations in two unknowns.
problem : that of finding solutions
The simplest nontrivial example
Examples
The
technique
ables.
for solution is that of elimination of
This is
one
of the vari
accomplished by multiplying the equations by appro
priate nonzero numbers and adding or subtracting the resulting
equations. This is quite legitimate, for the set of solutions of the
system will not be changed by any such operations. It is our intention
to select such operations so that we eventually obtain as
equations:
x
something, y
something. In the present case this is quite
easy: if we add five times the second equation to the first, y will
conveniently disappear:
=
=
8* +
35x
43x
-
5y
5y
=
3
=
40
^43
and
obtain the
we
equation gives
solution. Let
2.
5y
try
us
a
3,
=
few
x
=
y
=
or
more
1
that in the first
Substituting
.
Then
1.
illustrative
x
=
1,
y
=
1 is the
examples.
3x-2y= 9
11
+ 3y
-x
We
equation
8 +
3
Simultaneous Equations
1.1
=
eliminate
can
as
x
follows:
multiply
the second
equation by
3
and add :
3x-2y= 9
33
9y
-3x +
=
7y
We obtain
3.
v
3x +
6x +
If
we
42
=
6 and
=
7.
x
1
4y=
8y
n~
'
v
'
15
=
subtract twice the first
equation
from the second,
we
obtain
a mess:
6x + Sy= 15
14
-6x-8y
^ 4x
=
0=
Thus there
because
can
be
they imply
if the second
6x +
Sy
then
our
but
1
=
equation
x
+
3x-2y
2x +
y
z
=
equation
information.
We
causes
0
can
0 which is true,
conclude that our
=
always produce results.
for this in Section 1.3.
z
=
more
5
+ 5z= -I
-
new
generalize our technique to systems involving
Consider, for example, the system
us now
y +
would lead to the
of elimination does not
We shall go into the
variables.
were
offers much
simple technique
4. Let
Equation
and y satisfying Equations (1.3),
(1.4) which is patently false. Notice,
x
14
technique
hardly
numbers
no
the
0
(1.5)
4
1.
Linear Functions
equation expresses x in terms of y and z; if the second and
equation were free of x we could solve as above for the two
The first
third
unknowns y, z and then use the first to find x. But now it is easy to
replace those last two equations by another two which must also be
satisfied and which are free of the variable x. We use the first to
eliminate
from the latter two.
x
Namely,
subtract three times the
first from the second:
3x
-
3x +
2y
3y
+ 5z
=
+ 3z
=
1
-
15
5v + 2z= -16
-
and twice the first from the third :
2x +
2x +
=
0
+ 2z=
10
z
y
2y
y-3z
-
The system
x
+
y +
-5y
-
y
and
-10
=
has been
(1.5)
z
replaced by
this
new
5
=
+ 2z=-16
-
3z
=
(1.6)
10
-
way clear to the end.
in
two
unknowns
:
system
we can now see our
as a
system:
We solve the last two
-5y+ 2z= -16
50
5y+15z
=
34~
17z=
2
z=
Then, substituting this value in the last equation,
10
first
z
=
5.
or
y
=
4.
equation,
Finally, substituting
we
find
x
=
2.
x
y
2x + y
-
-4x-y+
z
=
5
3z
=
0
z=
10
-
1.
we
obtain
these values for y and
Thus the solution is
x
=
6
y
-
z
1,
=
in the
y
=
4,
1.1
Eliminate
x
multiples
of the
2x +
y
from the second and third
3z
-
-4x
4x
4y
5y
-
The
x
-
-
-
3y5y
-
z
=
10
=
20
3z
=
30
has been
=
5
-10
3z
10
-
z=
z
We solve the last two
Of course,
we
replaced by these equations :
30
=
into the first
equations by adding appropriate
10
4z
given system
y
5
0
=
=
y +
-
-
Equations
first;
2x-2y-2z=
z
3y
-
Simultaneous
easily: y
equation completes
can
run
=
-30/7,
z
=
the solution:
into difficulties
as we
-20/7. Substitutions
15/7.
x
=
did in the two unknown
equations of Example 3. We should be prepared for such occurrences
and perhaps even more mysterious ones.
Nevertheless, our technique is
productive : if there is a solution to be had we can locate it by this process
of successive eliminations. Furthermore, it easily generalizes to systems
with more unknowns. This is the technique stated for the case of n un
knowns. Eliminate the first variable from all the equations except the first
by adding appropriate multiples of the first. Then, we handle the resulting
1 unknowns.
That is, using the second equation
equations as a system in n
we can eliminate the second variable from all but the second equation, using
the new third equation we can eliminate the third variable from the remain
ing equations, and so forth. Eventually we run out of equations and we
ought to be able to find the desired solution by a succession of substitutions.
We shall want to do
to be able to
systems, and
predict
we
more
than discover solutions if
the existence of solutions ;
want to know in
some sense
we want
they exist.
We want
to be able to compare
how many solutions there
we should come to understand the nature of
are.
given system
words,
equations. In order to do that we have to analyze this technique and
develop a notation and theory which do so. That is where linear algebra
begins. Before going into this, we study another pair of more complicated
examples.
In other
of
a
6
1.
Linear Functions
Examples
6.
x
5x
2y
+
-
-
y
y +
3x +
According
13
z
3w
=
z
+ 2w
=
z
+
=
4
=
-7
w
2y-2z
technique,
to our
14
-
adding the suitable multiple
(-5)
(first)
(first)
(first)
x
0
x
(-3)
x
Now
we
+ second:
+ third
+ 4z + 17w
+ fourth:
4y
-
solve this set
+
=
-79
+
w
=
4
z+
9w
=
-46
z
by applying the
equations by a
by
We do this
x
y +
:
the last three
does not appear.
of the first equation:
lly
-
replace
we
set in which the variable
new
same
procedure:
we
now
eliminate y.
Of course the order of the equations is not relevant;
we could have listed them some other way.
Since we can avoid
fractions by adding multiples of the second equation to the first and
third, let's do it that
(11)
+ first:
(second)
(second)
x
4
x
way.
15z + 28 w
=
-35
5z + 13w
=
-30
+ third:
Finally, of this set, (-3) x (second) + first gives
the original set of four equations is replaced by this
x
+
2y
y +
z
3w
=
13
+
w
=
4
z
-
live
=
55.
Thus
set:
5z+ 13w= -30
-llw
The solutions
w
5
=
7.
x
z
2y
easily found,
are now
=
Now, let
+
55
=
7
us
+ 3z +
y
=
u-
v
4u + 3v
3z
x=l
consider this set:
-5x+ y + lz
2y+
2
u
v
=
2
=
-5
=
18
=
5
t1,7)
1.1
x
is
eliminated from the last two
already
to eliminate
from the second,
of the last three above,
place
1
x
22z + 5m
\y +
2y+
Simultaneous Equations
5v
=
5
4u + 3v=
18
3z
-
u
v
we
7
Using the first
equations in
equations.
obtain these three
(1.8)
5
=
Now y is already eliminated from the last. We eliminate it from the
second (without getting involved with fractions) in this way :
(-2)
x
(first)
Now this
+
(11)
(second):
x
-44z + 34u + 43y
188
=
with the last of the set
equation together
(1.8) gives this
system
-44z + 34u + 43v= 188
3z
We
can
eliminate
the system
x
2y
lly
+
v
u
+
(1.7)
5
=
v
from the first to obtain 129z
3z +
v
=
+ 22z + 5w-5i>
=
3z-
u
v=
u-
we can
=
27.
Thus
2
5
K
5
-
'
=-27
\29z-9u
Now
9u
has been transformed into this:
solve for
by
x
the first
equation
once we
know y, z, u, v;
solve for y by the second once we know z, u, v; we can solve
in the third once we know z and u; and we can use any z, u which
we can
for
v
make the last
For
true.
equation
example,
if
z
=
0,
we
must have
1
Notice that for
0, x
2, y
up the line : v
that
make
these equations
find
can
of
z
we
value
u,
v,
x,
y
always
any
all hold. Thus in this case there is more than one solution.
u
=
3, and
Formulation
Now, let
=
so on
of the Procedure:
us
=
=
.
Row Reduction
turn to the abstract
formulation of this procedure.
In the
general case we will have some, say m, equations in n unknowns. Let
x". These m equations may be written
refer to the unknowns as x1,
.
.
.
,
+
a21x2
a22x2
+
+
a1x"
an2x"
+
a2mx2
+
+
a,"*"
a^x1
a^x1
+
aimxl
+
---
+
=
b1
=
b2
=bn
us
as
(1 10)
8
1.
We
Linear Functions
proceed
to solve this
a2\a^
x
al3/a1 and
system
multiply the first equation
equation; multiply the first equation by
follows:
as
and add it to the second
by
which
a^x1
We
now
=
+
a21x2
+
+
<x2x
+
+ aV
tx2mx2
+
+
continue with the
equations
equation).
computed
a/
add it to the third and
The result will be
so on.
in
n
-
ajxr
ocnmx"
=
=
p2
(1.11)
P"
technique applied
same
to the
This is
now
an
introduce
equation
must be
a
x" are known.
x2,
modified.
We just
slightly
once
1
Of course, if
...,
formalism which allows
It is clear that the
Equations (1.10)
m
given by the system (1.11) (except for the first
effective reduction of the problem, because x1 can be
from the first
procedure.
system of
1 unknowns
0, this technique
us
to
renumber the
nonzero
in
no
keep
and then
equation
so
track of this
of the left side of the system of
is embodied in the array of numbers.
/V 021
a
A
system,
b1
=
equations so that the coefficient of x1 in the first one is
proceed as above. If that is impossible then x1 appears
we can disregard it and work with x2 instead.
We
a new
may write this way:
we
essence
2\
=
(1.12)
amJ
This array is called a matrix: the
upper index of the general term is the
index and the lower index is the column index.
Thus, a53 is the number
in the third row and fifth column,
a\2 is in the seventh row and
row
column, akJ is the number in the jth row
(1.12) is an m x n matrix: it has m rows and
b
is
forty-second
and the kth column.
n
columns.
Symbol
The matrix
=
an m x
Ax
1 matrix.
=
b
Equations (1.10)
can now
be written
symbolically
as
(1.13)
Simultaneous Equations
1.1
Now the
technique
for
solving
the
equation
described above consists of
sequence of such equations with new matrices A and b, ending with
solution is obvious. The step from one equation to the next is
by
operation (remember, the
of these particular steps :
a row
one
rows are
Step 1 Multiply a row by a nonzero
Step 2. Add one row to another.
Step 3. Interchange two rows.
.
It is clear
change
not
(and
a
/l 2*
0
0
\
a
whose
performed
equations) ; that is,
1.3) that any such operation does
Finally, the end result desired is a
will be verified in Section
matrix of this form, called
1
the separate
one
constant.
the collection of solutions.
0
9
row-reduced matrix:
^
(1.14)
/
:
Descriptively : the first nonzero entry of any row is a 1 and this 1 in any
is to the right of the 1 in any previous row. This is the kind of matrix
the above procedure leads to ; and it is most desirable because the system it
row
immediately solved. In order to see this, we shall distin
guish between two cases by resolving the dotted ambiguity in the lower right
corner of (1.14).
represents
can
be
Let
Ax
be
a
=
b
system of linear
where A is
Let d be the number of
(1.14)).
Case 1
equations
.
d
=
In this
n.
xl+a21x2
x2
case
a32x3
the system of
+
+
x"-1
row-reduced matrix
nonzero rows
+--- +
+
a
+
an1x
a2x
cr-1xn
equations
=
bl
=
b2
=
bn-1
xn
=
b"
0
=
bn+1
0
=
Z>'
(of the
of A.
has this form :
form
10
Linear Functions
1.
solution if and only if b"+l
is found by successive substitutions: In this
Thus there is
=
In this
d<n.
Case 2.
xl
a
+
a^x2
+ +
x2
xd
case our
+
=
case
bm
=
0, and the solution
the solution is
unique.
system has the form :
=
b1
=
b2
=
bd
0
=
bd+1
0
=
bm
aV
+ a3 2x3 +
+
a2x"
add+1 'xd+l---
+
andxn
(We may have to reindex the variables in order to get all the leading l's in a
bm
0, and all the
line.) There is a solution if and only if bd+1
solutions are obtained by giving xd+1, ...,xn arbitrary values, and finding
the values of the remaining variables by substitutions.
We now summarize the factual (rather than the procedural) content of this
discussion in a theorem, the proof of which will appear in Section 1.3.
=
=
A matrix A is called
Definition 1.
a
=
row-reduced matrix if
(i) the first nonzero entry in any row is 1,
(ii) in any row this first 1 appears to the right of the first
1 in any
preceding
row.
The number of
of A is called its index.
nonzero rows
Theorem 1.1.
reduced form A'
Let A be
an
succession
m x n
matrix.
A
by
precisely the same solutions as
by the same sequence of row operations that led from
a
can
be
brought
=
=
EXERCISES
1
.
Find solutions for these systems
(a) 2x 3y 23
-
(b)
=
3x+ y=-4
10
hx + \y
=
-fx+8y= 0
(c)
x+ y+
z=
y+
z=
x-
2x-3y-5z
=
into
row-
b has
of
operations. The equation Ax
the equation A'x
b' ifV is obtained from b
row
15
3
-7
A to A'.
1.1
(d)
x+
y+ z+
x+ y+
y+
x
+
-x
(e)
w
=
w=2
z
h>=0
=
11
4
z
2y-3z + 4w
Simultaneous Equations
2
z=
-x+ y+
0
2x + 2y+10z
28
=
x+
y+
z
(f )
3x +
(g)
x+
y
=
x
y
=
\
=
0
x
=
3x
(h)
22
=
12
6y + 9z
+ 2y + 3z= 4
4y
x+ y
x + 2y
7
=
7
=
9
x+3y =11
(i)
jc
+ y + z+v
4
=
z w
6
+y
x + 2y+
z=
0
x
(j)
=
x
4
3y 6z
ll
4x+8y + 4z
2. A homogeneous system of linear equations is a system of the form
Ax
0; that is, the right-hand side is zero. Find nonzero solutions (if
possible) to these homogeneous systems.
(a) x+ y + z 0
=
=
=
=
y +
X-
z
0
+ 2y + z
0
x+y + z
x
(b)
=
=
x
(c)
z
y
=
0
y+ z+ w =0
2y + z 2w=0
x+
x
2x
3.
0
=
y + 2z
Suppose (x1,
...,
w=0
x") is
a
solution for
Show that for every real number /, (tx1,
...,
given homogeneous system.
tx") is also a solution.
a
x"), (y1, ...,y) are solutions for a given homogeneous
x" + y).
(xl + y1,
5. Find the row-reduced matrix which corresponds to the given matrix
according to Theorem 1.1.
4. If
(x\
system, then
.
.
so
,
is
.
0
7
1\
3
2
2
\-l
6
/
A=
B
.
.
.
,
4/
'l
0
0
6
5^
2
3
0
0
0
=
1
i0
0-1-1
0
0
2
1
0y
12
Linear Functions
1.
1\
2
6
-2
-4
0
2
0
0
8
8
3
6
9
12/
(1
i)
I
6. Let
^
i)
equations:
Ax=b, (b)
Solve these
(a)
(e)
Cx
=
(c)
Bx=a,
c, where A, B, C
are
Bx
given
=
(d)
c,
Cx
=
a,
in Exercise 5.
PROBLEMS
1. Show that the system
ax
+
by
=
cx
+ dy
=
has
a
a
fi
solution
no
matter what a,
jS are
if adbc^ 0, and there is only
one
such solution.
2. Can you suggest
by Example 3 ?
an
explanation of the ugly phenomenon illustrated
only one row-reduced matrix to which a given matrix may be
operations? If A' and A" are two such row-reduced
matrices, coming from a given matrix A show that they must have the same
3. Is there
reduced by
row
index.
4. Suppose we have a system of n equations in n unknowns, Ax
b.
row reduction the index of the row-reduced matrix is also n.
Show
=
After
that in this case the equation Ax
b always has one and only one solution
for every b.
5. Suppose that you have a system of m equations in n unknowns to
solve. What should you expect in the way of existence and uniqueness
of solutions in the cases m<n,m>nl
=
6.
Suppose we are given the n x n system Ax b, and all the rows of A
s" such that a/
multiples of the first row; that is, there are s1,
s'af
for all j and i
n.
Under what conditions will the given system
1,
=
are
.
=
have
a
.
.
.
.
.
=
,
,
solution ?
7. Suppose instead that the columns of A are multiples of the first column.
Can you make any assertions?
8. Verify that if the columns of a 3 x 3 matrix are multiples of the first
column, then the rows are multiples of one of the rows.
1.2
1.2
Numbers, Notation, and Geometry
13
Numbers, Notation, and Geometry
We
now
interrupt
our
discussion of simultaneous
equations
in order to
introduce certain facts and notational conventions which shall be in
use
throughout this text. We shall also describe the geometry of the plane from
the linear, or vector point of view as an alternative introduction to linear
algebra.
First of all, the collection {1, 2, 3,
.} of positive integers (the "counting
numbers ") will be denoted by P. Every integer n has an immediate successor,
If a fact is true for the integer 1 and also holds for the
denoted by n + 1
successor of every integer for which it is true, then it is true for all integers.
This is the Principle of Mathematical Induction, which we shall take as an
axiom, or defining property of the integers. We shall formulate it this way.
.
.
.
Principle of
properties:
(i)
(ii)
1 is
a
Mathematical Induction.
Let S be
a
subset of P with these
member of S,
whenever
a
particular integer
n
is in S,
so
also is its
successor n
+ 1
inS.
Then S must be the set P of all
This assertion is
intuitively
positive integers.
clear.
You
can
1 + 1
see, for
example
that 2 is in 5.
2 is also in 5.
Continuing,
(ii)
By applying (ii) another time 4 is in S.
Applying (ii) another 32 times we see that all the integers up to 36 are also
in S. No positive integer can escape : since 1 is in 5" we need only apply (ii) n
times to verify that the integer n is in S. In fact, the assertion of the principle
of mathematical induction is that there are no integers other than those that
can be captured in this way, and in this sense the principle is a defining
For by (i) 1 is in S, and thus by
3
2+1 is in S, again by (ii).
=
=
property of the integers.
of mathematical induction provides us with a tool for writing
for all positive integers which avoids the phrases:
assertions
proofs of
this
in
way," and so forth," "...,".... We shall find this a
continuing
in
helpful device verifying assertions concerning problems with an unspecified
number, n, of unknowns. Let us illustrate this method by proving a few
The
"
principle
"
propositions about integers.
14
Linear Functions
1.
The
1.
Proposition
of the first
sum
Let S be the set of
Proof.
n
(l/2)(
is
integers
integers for which Proposition
1).
+
1 is true.
Certainly
1 is in S:
1=*- 1(1 + 1)
Now, assuming the assertion of Proposition 1 for
holds for
+ 1
n
l +
+ 7j+l=l +
---
=
+
---
M
(n+ l)(in
Proposition 2.
The
l2
surely.
1
Proof.
=
+ 1
n
1 + 3 +
sum
1)
=
show that it also
we
n,
We
Thus
now assume
n
+ 1
K + \)(n + 2)
by
the
the
of mathematical in
principle
of the first n odd integers
is
n2.
proposition for
any n, and show that
:
+ 2(n + 1)
-
1
=
=
Proposition 3.
we can
integer
+ +l= in(n + 1) +
+
which is the appropriate conclusion.
duction, Proposition 1 is proven.
it follows for
any
.
Let K be
a
1 + 3 +
1 + In + 1
+ 2n
n2 + 2n+l=(n+l)2
-
given positive integer.
Then
for
integer
any
n=QK+R
with 0
<
R
<
K in
one
(U5)
and
only
one
way.
Proof. We may of course immediately discard the case K
1 for in that
(1 1 5) is just the trivial comment that n
n
1 for all n.
Thus take K > 1 and
proceed by mathematical induction. The proposition is true for n
1 :
=
case
=
.
,
=
\=0K+\
Now
we assume
n
for
n
write
some
n
=
that the proposition is true for
any
given integer
n.
Thus
QK+R
Q and R,0^R<K.
+ 1
=
QK+ (R + 1)
ThcnR+l^K.
If R + 1<
K,
we
have
now
1.2
with 0
< R
+ 1
n+l
as
=
<
K,
as
desired.
QK+K
=
Numbers, Notation, and Geometry
Otherwise,
R + 1
=
K, in which
case
(Q+ \)K+ 0
desired.
This
Thus, by mathematical induction, (1.15) is possible for
representation is unique, for if
=
n
is also
15
every
integer
n.
Q'K+R'
possible
with 0 <, R' < K, then
Q'K+R'
=
we
have
QK+R
or
(Q'-Q)K
R-R'
R' is between K and K.
and R
K is 0,
=
so
(Q!
-
Q)K
=
Now the
only multiple of
R-R'=0 from which
we
conclude R
K between
=
R', Q
=
K and
Q.
Set Notation
The set of
positive integers forms a subset of a larger number system, the
integers. Z consists of all positive integers, their negatives and 0.
The collection of all quotients of members of Z is the set of rational numbers,
denotedy by Q. Q is a very large subset of the set of all real numbers R.
For the purposes of geometric interpretation we will conceive of the real
number system R as being in one-to-one correspondence with the points
on a straight line.
That is, given a straight fine, we fix two points on it,
All
one is the origin O, and the other denotes the unit of measurement.
:
it
is
the
a
numerical
value
other points P on the line are given
displacement
from O as measured on the given scale (negative if O is between P and 1
and positive otherwise). (See Figure 1.1.)
set Z of all
There
are
certain ideas and notations in connection with sets which
Figure
1.1
we
16
Linear Functions
/.
shall standardize before proceeding. Customarily, in any given context
there is one largest set of objects under consideration, called the universe
(it may be the positive integers P, or the rabbits in Texas, or the people on
the
x
is
moon) and
an object
that
means
x
all sets
we
is not
The set with
no
specific
are
sets
actually subsets
are
shall write xelto
member of X.
a
is
x
a
set and
member of X.
a
Thus, for example,
-
7
e
x$X
Z, but -7 $P.
elements is called the empty set, and is designated 0. Most
defined by a property : the set in question is the set of all
We
elements of the universe that have that property.
hand form to represent that phrase
{xeU:
If X is
of this universe.
mean :
has that
x
use
the
following short
property}
example, the set of all positive real numbers is {x e R: x > 0}. The set
Englishmen who drink coffee is {x e England: x drinks coffee}. The
This is the same
set of all integers between 8 and 18 is {x e Z: 8 < x < 18}.
13 1 < 5}.
as {xeP: 8 <x< 18} and {xeZ: \x
For
of all
-
If X and Y are two sets, and every element of X is an element of Y we shall
Notice that 0 c X for every
say that X is contained in Y, written X <= Y.
set X.
We shall consider also these operations on sets :
X: the set of all
x
not in X
X
u
Y: the set of all
X
n
Y: the set of all
x
in both X and Y
Y: the set of all
x
in X, and not in Y
X
1.2 for
(Consult Figure
the
same
-X
u
Y
a
Xn-Y.
as
-(X
Y),
x
in either X
or
Y (or
both)
pictorial interpretation.)
There
X
are
(Y
many
Z)
(X
other
Notice that X
identities:
Y is
-
X
=
X,
(X n Z), and so on,
so don't be surprised when two different collections of
symbols identify the
same set.
A final operation is that of forming the Cartesian
product. If U
is a given universe, then U x U is the set of all ordered
of
elements in U.
pairs
U x U is often denoted by U2.
By extension we can define U3 as the set of
all ordered triples (x1, x2, x3) of elements of
U; and more generally U" is the
set of all ordered -tuples of elements of U.
-
X1,
If
=
X"
n
n
u
=
n
Y)
u
subsets of U, the set of all ordered n-tuples (x1,
...,x")
e Xn is denoted X1 x
x X".
Not every subset of
V is of the form X1 x--- xX", those which are of this form
are called
with x1
...
e
,
X1,
.
.
are
.
,
xn
rectangles.
Thus the space of n-tuples of real numbers is denoted R". If
I"
I1,
intervals in R, then I1 x--- xl" is indeed a
rectangle. A point
(x
x") in R" will be denoted, when specific reference to its elements is
...,
are
,
.
.
.
,
1.2
Numbers, Notation, and Geometry
17
^^^
X U Y
required, by a single boldfaced letter x (xl,
(ft1,
b") are two points in R" with b' > a',
notation [a, b] to denote the rectangle.
not
b
=
...,
.
.
.
,
{(x1,
If the
.
.
.
,
x"): a1
inequalities
(a, b)
=
A function from
a
x1
strict
are
{(x\
<
.
.
.
,
b\
<
we
.
.
.
,
a"
<
xh
shall denote the
x"): a1
<
x1
<
b1,
.
.
.
If a
x").
I <i<n
=
=
we
(a1,
shall
...
a"),
,
use
the
b"}
<
rectangle by (a, b) :
,
a"
<
x"
<
b"}
set X to another set Y is a rule which associates to each
point y in Y. It is customary to avoid the
by defining a function as a certain kind of subset
of X x Y.
Namely, a function is a set of ordered pairs (x, y) with xe X,
If
y e 7, with each x e X appearing precisely once as a first member.
(x, y) is such a pair we denote y by f(x) : y =f(x). We shall use the notation
X is called
Y to indicate that /is to mean a function from X to Y.
/: X
If every
the domain of/; the range of /is the set {f(x): x e X} of values off.
point of y appears as a value off we say that/maps X onto Y. If every point
More
of y is the value off at at most one x in X, we say that/is one-to-one.
precisely,/is one-to-one if x # x' implies /(x) #/(x')- Now, if/is a one-toone function from X onto Y, then for each y e 7, there is one and only one
point
use
of
x a
of the
uniquely
new
determined
word rule
-
xeX such
that/(x)
=y.
This defines
one-to-one and onto and has this
In this
case we
shall say that
/
a
function #: Y-+X which is also
x if and only \ff(x)
y.
property: g(y)
=
is invertible and g is its
=
inverse, denoted
18
1.
g=f~l.
Linear Functions
Y, f2 : Y-^Zare two functions, we can compose
Finally, iffx : X
a third function, denoted /2 of^.X^Z defined by
->
them to form
fi /iW =/2(/i(*))
Plane
Geometry
geometric study of the plane, as an alternative intro
duction to linear algebra. According to the notion of the Cartesian co
ordinate system we can make a correspondence between a plane, supposed
to be of infinite extent in all directions, and the collection R2 of ordered
pairs of real numbers. This is done in the following way: first a point on the
plane is chosen, to be called the origin and denoted O (Figure 1.3). Then
two distinct lines intersecting at O are drawn (it is ordinarily supposed that
These lines are
these lines are perpendicular, but it is hardly necessary).
called the coordinate axes; they are sometimes referred to more specifically
as the x and y axes, ordered counterclockwise (Figure 1.4).
Now a point
is chosen on each of these axes; we call these Et and E2 (Figure 1.5).
These
are the "unit lengths" in each of the directions of the coordinate axes.
Having chosen a unit on these lines, we can put each of them in one-to-one
correspondence with the real numbers. Now, letting P be any point in the
plane, we associate a pair of real numbers to P in this way. Draw the lines
through P which are parallel to the coordinate axes and let x be the inter
section with the line through EL and y the intersection with the line through
We
now
turn to the
o
Figure
1.3
Figure
1.4
1.2
Numbers, Notation, and Geometry
19
Figure 1.5
E2
.
Then
we
identify
P with the
pair of real numbers (x, y) (Figure 1.6.)
In this way to every point in the plane there corresponds a point in R2
(called its coordinates relative to the choice O, Elt E2). Clearly, for any pair
of real numbers
the
fourth vertex of the
co
ordinates
axes
(x, y) we have a point with those coordinates, namely
parallelogram of side lengths x and y along the
with one vertex at O (Figure 1.6).
P(x.y)
Figure 1.6
20
Linear Functions
1.
Figure 1.7
particular point in the plane is fixed as the origin, there can be
defined
operations on the points of the plane, and these operations form
of
linear
tools
the
algebra. Since they cannot be defined on the points until
an origin is chosen, we are forced to distinguish between the point set the
plane and the plane with chosen point. This distinction gives rise to the
notion of vector : a vector is a point in the plane-with-origin. The vector
can be physically realized as the directed line segment from the origin to the
given point; such a visualization is nothing more than a heuristic aid. It is
important to realize that as sets, the set of vectors in the plane is the same
The difference is that the set of vectors has
as the set of points in the plane.
additional structure : a particular point has been designated the origin. We
shall denote vectors by boldface letters; thus the point P becomes the vector
P, the origin O becomes the vector 0. We shall now describe these two
operations geometrically and then compute them in coordinates.
1.
Scalar Multiplication. Let P be a vector in the plane, and r a real
number. Consider the line through 0 and P.
Considering P now as a unit
length, we can put that line into one-to-one correspondence with R. Using
this scale, rP is the point corresponding to the real number r.
Said differently,
rP is one of the points on that line whose distance from 0 is \r\ times the
distance of P from 0 (Figure 1.7). Now, if P has the coordinates (x, y) we
shall see that rP has coordinates (rx, ry). First, suppose r > 0. Draw the
Once
a
two
1.2
Numbers, Notation, and Geometry
21
triangle formed by the fine through 0, P and rP, and the E^ axis and the lines
parallel to the E2 axis (Figure 1.8). Triangles I and II are similar. Thus,
referring to the lengths as denoted in Figure 1.8,
Jpl
=
|rP|
i
s
By definition |P|/|rP|
is
coordinates
2.
of
a
=
l/r,
thus the first coordinate of rP (here denoted
The second coordinate is
rx.
(rx, ry).
The
case r
by s),
similarly seen to be ry. Thus rP has the
< 0 is only slightly more complicated.
Let P, Q be vectors in R2. Then 0, P, Q are three vertices
determined parallelogram. We define P + Q to be the fourth
Addition.
uniquely
description of this operation in terms of coordinates is extremely
(x, y), and Q has coordinates (s, t), then P + Q
simple:
has (x + s, y + t) as coordinates. There is nothing profound to be learned
from the verification of this fact, so we shall not go through it in detail.
After all, it is not our purpose here to logically incorporate plane geometry
vertex.
The
if P has coordinates
Figure
1.8
22
1.
Linear Functions
Figure 1.9
to use it as an intuitive tool for comprehen
suspicious of our assurances we include the verifica
tion of a special case.
Consider Figure 1.9 and include the data relating
to the coordinates (Figure 1.10): To show that the length of the line segment
OB is s + x, we must verify that the length of AB is s.
Draw the line through
P and parallel to OE^ and let C be the intersection of that line with the line
through P + Q and B. The quadrilateral ABCP has parallel sides and thus
is a parallelogram.
Hence, AB and PC have the same length. Now triangles
and
+
O^Q
PC(P
Q) have pairwise parallel sides (as shown in Figure 1.10)
and further 0Q and P(P + Q) have the same length.
Thus these triangles
are congruent so the length of PC is the same as the
length of 0s, namely s.
Thus the length of AB is also s, and so OB has length s + x.
Notice that this
is a special case since it refers to an explicit picture which does not cover all
possibilities.
The operation inverse to addition is subtraction : P
Q is that vector
which must be added to Q in order to obtain P (Figure
1.11). The best way
to visualize P
Q is as the directed line segment running from the head of
Q to the head of P (denoted L in Figure 1.11). In actuality, P
Q is the
vector obtained by translating L to the origin; in
practice, it is customary not
to do this but to systematically confuse a vector with
any of its translates.
We shall do this only for purposes of pictorial
representation.
Notice that, having chosen the vectors E1 and
E2 we can express any
vector in the plane uniquely in terms of them and the
operations of addition
into
sion.
our
mathematics, but rather
For those who
are
-
-
-
,
1.2
Numbers, Notation, and Geometry
P +
Figure 1.11
Q
23
24
Linear Functions
/.
and scalar
multiplication:
(x, y)
=
(x, 0)
no
0, Ej, E2 do
not lie
Thus
Proposition
Then
we can
Q
=
(0, y)
matter how
This is true
collinear).
+
we
x(l, 0)
y(0, 1)
+
Ex and E2
same
have this
Let
4.
the
on
=
line
are
(we
=
xEt
chosen,
yE2
+
so
say the vectors
long as the points
E1 and E2 are not
fact.
important
EY and E2 be any two noncollinear
Q uniquely as
vectors in
the
plane.
write any vector
*%
+
x2E2
x1 and x2
are the coordinates of Q relative to the choice of origin 0 and the
Et and E2
If we state this geometric fact purely as a fact about R2, it turns out to be
a theoretical assertion about the solvability of a pair of linear equations.
Thus, let us suppose Ex
(ax1, a,2), E2 (a2x, a2) relative to some standard
coordinate system (for example, the usual rectangular coordinates).
First
of all, how do we express algebraically the assertion that Et and E2 do not
lie on the same line? We need an algebraic description of a straight line
through the origin.
vectors
.
=
=
Proposition 5.
(i) A set L is
a
straight
line
0
through
if and only if there
exists
(a, b)
e
R2
such that
L
=
{(x, y) :
ax
+
by
=
0}
(ii) The points (x, y), (x', y')
only if
-
=
x'
lie
on
the
same
line
through the origin if and
L
y'
You certainly recall these facts from
analytic geometry we leave the
verification to the exercises. Returning to the vectors
Ex and E2, these
geometric facts become the following algebraic fact.
Proposition 6.
Let
la,1 aA
be
a
2
x
2 matrix with
nonzero
columns.
Numbers, Notation, and Geometry
1.2
(i)
If at *a22
-
a2a2
for every be R2.
(ii) The equation Ax
0, then the equation Ax
=
0 has
a nonzero
=
b has
solution if and
a
unique
25
solution
only if a1 a2
=
afa1.
a^a2
al2b1.
If
(iii)
ifal1b2
=
=
a2a2
,
the
equation
Ax
=
b has
a
solution if and
only
Proof.
(i) This condition is according to Proposition 5 (ii) precisely the assertion that
the vectors (ai\ ai2), (a2l, a22) are not collinear.
Then, according to Proposition 4,
for any (b1, b2) there is a unique pair (x1, x2) such that
(b\ b2)
This is the
=
x'iai1, a,2) + x2(a2\ a22)
same as
the
pair of equations
b
=
Ax.
(ii) By the above, if aila22 ai2a2, then the only solution of Ax 0 is x 0.
ai2) or ( a2\ a22) is a non
On the other hand, if a?a22 =ai2a2\ then either (a22,
in which case everything
A
are
entries
of
all
the
Ax
or
of
zero,
zero solution
0,
=
=
=
=
solves Ax
=
0.
(iii) If a,la22 =ai2o21, then (a/, ai2), (a2l, a22) lie on the same line through the
origin by Proposition 5(ii). Any combination x1(a1\ ai2) + x2(a2\ a22) willhaveto
lie on that line, and conversely, any point on that line must be such a combination.
b has a solution if and only if (b\ b2) lies on the line through 0 deter
Thus Ax
mined by (at\ at2). The equation for this is, by Proposition 5(ii),
=
h1
h2
a,1
a,2
_=f_
or
bW=b1a12
Examples
the line
Lx be the line through (0, 0) and (3, 2) and L2
of
and
L2
Lv
through (1, 1) and (0, 6). Find the point of intersection
6.
the equation 5x + y
and
2x
0,
the
L2
has
3y
equation
Lt
8.
Let
.
-
point of intersection
(x, y) solving
The
3y
=
0
y
=
6
We find
x
=
2x
-
5x+
9.
18/17,
y
=
Find the line L
line L': 8x + 2y
=
17.
=
=
must lie
on
both lines, and thus is the
pair
12/17.
through
the
L will be
point (7, 3) that is parallel to the
given by an equation of the form
26
/.
Linear Functions
+
by
point
of
ax
8x +
ax
=
intersection,
=
have
can
so
the
L and L' must have
to
L',
we
must have
parallel
equations
no
17
2y=
by
+
In order to be
c.
c
no common
solution.
Thus
8_2
a~b
Furthermore, since (7, 3) is
la + 3b
This
6
c
4x + y
?
=
=
we
must also have
c
of
pair
1,
=
=
L,
on
equations
31.
in three unknowns has for
Thus Z, is
given by
the
a
solution
a
=
4,
equation
31
EXERCISES
7. Show that for every
l2 + 22 +
+ n2
=
9. Show that
=
n,
in(n + 1)(2 + 1)
8. Show that for every
2 + 22+--- + 2"
integer
integer
n,
2"+1-2
Xn(YvZ)
=
(Xn Y)u(XnZ)
and
Xu(YnZ)
=
(XuY)n(XvZ).
10. Give
an
example
of
a
subset of
a
Cartesian product which is not
a
rectangle.
11
.
Find the point of intersection of these
pairs of lines in R2 :
(a)
3x+
(b)
\
x~\ly
4
x-2y
0
2x+ y
y
=
l
(c)
=
2x+
2y=-l
at+ 12^
(d)
=
=
y=2x+
1
jc
3y + 18
through P which is parallel
(2, -l),Z.:3* + 7y 4
(8, \),L:x-y
-\
=
12. Find the line
(a)
(b)
(c)
P
=
P
=
V
=
14
=
to L-
=
=
(0,-l),L:y-2x
=
3
PROBLEMS
9. We
can
the vector X
-
define the line through P and
Q
P is parallel to the vector P
parallel if and only if one is
the set of all X such that
Show that two vectors are
of the other.
Conclude that the line
-
a
multiple
Q.
as
1.2
through
P and
Q is the
Numbers, Notation, and Geometry
27
set
{P + t(P~Q):teR}
10.
is, in
Using the definition in Exercise 9
coordinates, given as
show that
a
straight line in
the
plane
terms of
{(x,y)eR2:ax + by
+
c
=
0}
for suitable a, b, c.
1 1 Suppose Z, is
a line given by the
equation bx + ay + c 0
Show that the tangent of the angle this line makes with the
horizontal is b/a.
=
.
(a)
(b)
(c)
Show that the vector
(a, b) is perpendicular to L.
Find the point on L which is closest to the origin.
12. Find the line through the point P and perpendicular to L:
(a) P
(3, l),L:x-3y 2.
(b) P (-l, l),Z.:2x + 3y 0.
(c) P=(0, 2),L:5x 2y.
13. Suppose coordinates have been chosen in the plane. Let Ei E2 be
two vectors in the plane which are not collinear.
(That is, 0, Ei, E2 do not
lie on a straight line.) Then we can recoordinatize the plane relative to this
choice of principal directions. Give formulas which relate these two
coordinatizations in terms of the given coordinates of Ei E2 (see Figure 1.12).
=
=
=
=
=
,
,
p
/
<E, (x*
/
/
/
y
Uj)
(u.v)
)
"~
^-'
vE. r
/
uE,
Figure
1.12
E.U'.y1)
1.
28
Linear Functions
14. In the text, the equation 1 + 2 -\
hn
There is another way of doing this.
=
entries.
There
2 +
+
-
We
now
equations,
1) +
=
n
by
on the diagonal and 1+2+
diagonal. Thus
\-
n
1
n2
return to the
with
problem of analyzing systems of simultaneous linear
question in mind: given the m xn matrix A, for
solution of the equation Ax
b? In order to study this,
broader
a
which b is there
a
=
associate to A the function from Rm to R":
fA(x\ ...,*")
=
(V*1
+
Thus the set of b, such that Ax
a^x",
+
a,mx\
...,
...,
anmx")
b, is precisely the range offA
begin by introducing the two fundamental operations
the case n
2 studied in the previous section) :
Let
in
verified
matrix has n2
Linear Transformations
1.3
we
n
was
n x n
of these entries
are n
entries both above and below that
2(1 +
$n(n + 1)
An
induction.
(1.16)
=
.
us
on
R"
(just
as
=
1. Scalar
rx
=
(rx1,
...,
2. Addition: for
x
+
for
Multiplication:
x
R,
x
=
(x1,
.
.
.
,
xT),
+
(y1,
=
y
.
.
.
,
x")
R", define
e
.
.
.
,
y")
e
R", define
y")
A function
preserves these two
(x1,
=
rx")
y^x1 +y1,...,xn
Definition 2.
r e
/ from R"
operations, that is,
to Rm is a linear transformation if it
if
/(rx)
rf(x)
/(x + y) =/(*) +/(y)
=
The function fA defined above for the
fA(rx)
=
=
=
(a, 'rx1 +
(rfa1*1 +
rfA(x)
m x n
a1",
an V),
+
.
.
.
.
.
,
matrix A is linear:
afrx1 +
rfa"*1 +
.
,
a^rx")
anmxn))
+
+
1.3
/*(
+
y)
(xl
=
+
=
=
y1)
+
+
+<*."(*
(a11a:1
a^x1
//.*)
+
+
+
an\xr
y"),
+
Transformations
.
.
.
,
a^x1
+
29
yl)
/))
+
---+aV
+
Linear
+
a"V
+
a11y1 + ...+,,//,...,
+ anmy")
^V +
+ /,(y)
The
significance of the introduction of linear transformations, from the
of view of systems of equations, is that it provides a context in which to
consistently interpret the technique of row reduction. For, the application
point
of
a row
operation
to a
system of equations
amounts to composition of the
particular linear transformation
corresponding to the row operation. Once we have seen that we can analyze
the given system by studying these successive compositions.
Looking ahead,
it is even more important to recognize row reduction as a tool for analyzing
linear transformations. Let us now interpret the row operations as linear
associated linear transformation with
a
transformations.
Type
I.
of the rth
Multiply a row by a nonzero constant. Consider
row by c # 0.
Let Pt be the transformation on
Pl(b1,...,bm)
=
the
Rm
multiplication
:
(bl,...,cbr,...,bm)
(multiplication of the rth entry by c). The effect of this row operation is that
of composing the transformation/,: R"
Rm with Pu and changing the
Ax
b
into
the
equation PtAx Pib. These two equations have
equation
->
=
the
same
=
set of solutions since the transformation
Px
can
be reversed
its inverse is
invertible). Precisely,
Type II. Add one row
corresponds
this transformation
to
P2(b\
.
.
.
,
bT)
=
(it is
1/c.
given by multiplying the rth entry by
to another.
Adding the rth row to the 5th
(by
,
.
.
.
,
Rm
on
V,
.
.
.
,
row
:
b* + b',
.
.
.
,
bm)
Again, this step in the solution of the equations amounts to transforming the
equation Ax b to P2Ax P2b. Since P2 is invertible (what is its inverse?)
=
=
we
cannot have affected the solutions.
Type
III.
corresponds
Interchange
to
two
P3(b1,...,br,...,P,...,m
The
Interchanging
rows.
the rth and sth
rows
the transformation
importance
=
(bl,...,V,...,br,...,bm)
of these observations is this : the
to linear transformations which in turn
are
operations correspond
representable by matrices. The
row
30
1.
Linear Functions
solution of the system of equations Ax
b thus can be accomplished com
pletely in terms of manipulations with the matrix corresponding to the system.
=
It is
our
purpose now to study the representation of linear transformations
the representation of composition of transformations.
by matrices and
In Rn the
n
vectors
fundamental role.
Thus
(1,0,..., 0), (0, 1, 0,
We
Ef has all entries
.
.
.
zero, but the
a
.
as
rth, which is
in R" has
Proposition 7. Any vector
ofEu ...,En.
0),
E]
,
shall refer to them
1
.
.
,
(0,
E
.
,
.
.
.
,
.
.
,
,
0, 1) play a
respectively.
.
unique representation
as a
linear
combination
Proof.
Obviously,
b") =*! + +b'E
(b1
We shall refer to the set of vectors Ei,
, E as the standard basis for R".
Proposition 7 comes this more illuminating fact.
.
of
.
.
Proposition 8. Corresponding to any linear transformation
a unique m xn matrix
(a/) such that
L: R"
Out
->
Rm
there is
L(x'
x")
=
(
V/=l
Proof.
define
a
It is
clear, by
a/xi,
a
that, given the matrix (a,1), Equation (1.17) does
Now, given L, since it is linear, we can write
(E,)
Then
+
linear transformation is
standard basis.
=
(1 .17)
j=l
the way,
linear transformation.
L((x\ ...,x")) =L(x1E1
Thus
..., a/v)I
+
xE)
=
x'KEO + + x?L(En)
completely determined by
what it does to the
Let
(at\
...,
fll-),
.
.
.
,
Z,(E)
=
(an\
...,
a)
Equation (1.18) becomes
L((x\ ...,x"))= x\ai\
....
(xW,
...,
=
=
which is just Equation
(xW +
(1.17).
---
ai") + + x"(al
am)
x1*,"1) +
+ (x"a\
x"am)
+ x"a\..., x'ar + + x-V)
...,
(1.18)
1.3
31
Transformations
Multiplication
Matrix
Now
we
must discover how to
transformations by
make this
Rm ->RP
(b/),
Linear
an
operation
matrices.
There is
only
then that T: R"
discovery: compute.
are
represent the composition of
on
two linear
way to
Rm and S:
one
Suppose
represented by the matrices (a/) and
compute the composition ST as follows:
->
linear transformations
Then,
respectively.
we can
T(x1,...,xn)=(tcij1xj,...,
2>/v)
V/=i
j
ST(x\ ...,x")
=
=
Thus ST is
=
/
i
( Jx( ,f>,V)>
(]tl(jty*j)*i
the p
represented by
x n
-
*>*'( ,f>;V))
i(iy^
matrix
(jtw)
Definition 3.
matrix.
(U9>
Let A
=
(a/)
be
an
m x n
product BA is defined
Then the
as
matrix and B
the p
x n
=
(bf)
a
matrix whose
p
x m
(i,j)th
entry is given by (1 19).
.
The
preceding discussion thus provides
the verification of
Proposition 9. If T: R" -> Rm, S: Rm -? R" are represented by the
A, B, respectively, then ST is represented by the product BA.
The
product operation
may
seem a
bit obscure at first
sight;
matrices
but it is
easily
described in this way: the (i,j)th entry of BA is found by multiplying entry
by entry the rth row of B to they'th column of A, and adding.
Examples
10.
A=
/5
3
7\
6
5
1
\8
11
-4/
B=
/
\
0\
6
1
-3
2
5
4
4
4/
Linear Functions
/.
32
LetAB
c11
Cl2
Cl3
cj2
=
=
=
=
Then
(c/).
=
49
3(-3)+ 7.4
25
1.4
6.6+ 5(-3)+
+
8.6+ll(-3) (-4)4=-l
5.6+
=
=
1.4
5.5 +
6.0 +
=
29
'
AB
/ 49
39
43\
25
20
29
\-l
14
39/
=
11.
(2
5
(
0)(2
\
12.
0
1)
\2
0\/
/-I
0
1
7
lj\-l
4
1
+ 0-5 +
5-0
1
1
1
5
0\_/-l
-2) \-\
-
l-l)
=
1/
-7
0\
4
-2/
-2\
[l 0J\-1 4-2/^17
Of
17
(I 1\(
1
7
1
4
recapitulate the
0\/-l
0\
-v
0
11
-2\
l-i
4
-2)
/
(-l)
01
4
(0 1\(
[0 iA
us
l)=[2-0
5
1/
\
Now, let
=(-l-2
discussion of this section
so
far.
The
problem
of systems of m linear equations in n unknowns amounts to describing the
The technique of row reduc
range of a linear transformation T: R" -* Rm.
tion
on
we
be
corresponds to composing Tby a
Rm.
These transformations
shall call them
represented by
of the
solve
a
are
succession of invertible transformations
those which
elementary transformations.
provide
the
row
operations;
Linear transformations
can
of the standard basis
by matrices, and composition
transformations corresponds to matrix multiplication. Thus, we
system of linear equations as follows: Multiply the matrix on the left
means
succession of
elementary matrices in order to obtain a row-reduced
easily read off the solutions. Since multiplication
by an elementary matrix is the same as applying the corresponding row
operation to the matrix it is easy to keep track of this process.
by
a
matrix.
Then
we can
1.3
Linear
Transformations
33
Examples
1 3.
Let
consider the system of four equations in three unknowns
corresponding to the matrix
A
us
=
We shall record the process of row reduction in two columns. In
we shall list the succession of transformations which A under
the first
goes and in the second
shall accumulate the
we
products
of the
matrices.
corresponding elementary
(a) Multiply the third
first,
by
row
>\
0
-1\
/0
0
-1
3
2
2
1/0
1
0
0
4
0
1
II
1
0
0
0
\0
1
2/ \0
0
0
ly
1 and
interchange
0^
(b) Multiply the first row by 3 and subtract it from
multiply the first row by 4 and subtract it from the third.
'I
0
0
2
0
V0
0
-1\ /0
5 \ / 0
5 II 1
1
2/ \0
(c)
0
-1\
0
1
5/2 1
0
0
-1
1
3
0
0
4
0
0
0
1,
/o
0
1
1
2/ \o
(d)
row
0
1/5
row
by
0
-1
0'
1/2
3/2
4/5
0
0
0
0
1
Subtract the second
2 and the third
-1\
1
0
1
5/2
0
0
1
0
0
0/ \1/10
\
/
1
1/5
second;
row
0
-1
0'
3/2
4/5
-11/10
0
-1/2
by
5.
from and add one-half the third
1/2
0
row
0
to the fourth.
0
the
0\
Divide the second
1
it with the
0
1,
34
1.
Linear Functions
Let
us
denote the
is the last matrix
product of the elementary matrices by P; thus P
the right and the matrix on the left is PA. Now,
on
it is easy to see that if PAx
y has a solution, the fourth entry of y
must be zero.
Now our original problem
=
Ax
b
=
has
a solution if and only if PAx
Pb is solvable (since P is invertible).
Thus b is in the range of A if and only if the fourth entry of Pb is zero :
Ax
b can be solved if and only if
=
=
A*1-i*2-A*3
If b satisfies that
it
by solving
=
=
o
condition, there is
PAx
x1-x3= -b3
x2 + \x3
\b2
x3
ib1 + fb3
6*
+
=
an x
such that Ax
=
b;
we
find
Pb:
\ b3
+
=
14.
We
row
Consider
row
1
reduce
now
as
the system in three unknowns
above.
(a) Multiply row 1 by 4 and subtract
by 3 and subtract it from row 3.
'1
3
0
-10
,o
-5
~2\,/
1
3
0
1
\0
0
10
3/1
-3
0
-2 \ / 1
-3/5
0/\-l
+
2-b2
+
=
b3
of row 2 from
0
2; multiply
row
3 ; divide
row
2
by
-
10
0\
-1/10
1/2
b thus has
=
row
1,
0
2/5
it from
0"
-4
The system Ax
-b1
0
6
(b) Subtract 1/2
/I
given by
a
0
1/
solution if and
only if
(120)
1.3
In that
x1
3x2
x*
+
the solution is
case
2x3
3V3
-Lx*
TX
-
_
=
_
=
b1
21,
2^1
rO
T q-
ft2
of x3 will
Any arbitrarily chosen value
condition (1.20) is satisfied).
15.
A=
/2
0
0
2\
3
-1
1
0
\2
2
0
0/
(a) Divide
/I
0
3
-1
\2
2
(b)
from
row
0
0/
\
2.
0
0\
1
0
0
0
1/
Subtract 3 times
row
1/2
o
o\
1
-3-3
1
0
2
0
-2/\-2
0
1/
\0
provide
1 from
row
a
solution
(granted
2; subtract twice
the
row
1
3.
i\ /
oo
(c) Multiply
row
by
0
1W1/2
10
0
1
-1
n
(0
row
35
Transformations
given by
1
_
Linear
2
row
by -1, and subtract twice the result from
3.
condition for the equation PAx
y to be
solution
The
solvable.
is
also
b
Ax
thus
every problem
solvable,
Pb:
is found by writing the system PAx
Here there is
=
no
=
=
x1
+
x2- x3
+
x4
3xA
=
=
ii1
3b1
-
b2
2x3-8x*= -8Z>1+262
Clearly, the value of x* can
easily found by the equations.
be
+
&3
freely chosen,
and
x1, x2, x3
are
36
1.
Linear Functions
Validity of Row Reduction
The basic
point behind the present discussion
equations in n unknowns is the
taneous linear
study of m simul
study of linear
the study of m x n
is that the
same as
the
transformations of R" into Rm, which is the same as
matrices under multiplication by m x m matrices. The matrix version of
this story is the easiest to work, if only because it imposes the minimum
However, the linear transformation interpretation is
amount of notation.
the most
first, let
But
and in the next section
significant,
record
us
a
we
will follow that line of thought.
of the main result of Section 1 1 in terms of
proof
.
matrices.
Theorem 1.2.
E0
,
.
is in
Let A be
an
x n
m
matrix.
There is
a
finite collection
E0A
Es of elementary m xm matrices such that the product Es
row-reduced form. Let P
Es
E0 and let d be the index ofPA.
.
.
,
=
The system Ax
solution.
(i)
=
(ii) The system Ax
of Pb vanish.
(iii) n d unknowns
b has
=
Proof.
First of all,
matrix whose first
we
b has
can
a
solution
if
and
only if
by
sequence of
a
PAx
if and only if the last
be freely chosen in any solution
may,
nonzero
solution
a
row
=
Pb has
d entries
m
of Ax
a
=
operations, replace
b.
A with
a
column is
(!)
namely, supposing the y'th column is the first nonzero column.
in that column, say a/, is nonzero. Interchange the first and
Thus
yth
some
rows.
entry
This is
accomplished by multiplication on the left by an elementary matrix of Type III, call
Now, E0 A
(a/) with a/ #0. Multiply the first rowby (a/)"1; thismakes
the (l,y) entry 1 and is accomplished by means of an
elementary matrix, say Ei.
Now, let Et be the elementary matrix representing the operation of adding
-1
-<*/(/) times the first row to the y'th row (this makes the (/,;) entry zero). Then
Em
E0 A has its first nonzero column (1.21).
The proof now proceeds by induction on m. If m
1 the proof is concluded:
the 1 x n matrix (0,
0, 1, a}+1
al) is in row-reduced form. For m > 1,
itE0.
=
=
.
the matrix Em
.
.
,
,
E0A has the form
.
.
.
,
1.3
where B is
that Fs
Transformations
37
(m
1) x (nj) matrix. The induction assumption thus applies to
collection F0
F, of (m
1) x (m
1) elementary matrices such
F0B is in row-reduced form. Now let
an
There is
B.
Linear
Em+J
a
-
.
,
-\0
.
.
-
,
)
Fj
Then, it is easy to compute that further multiplication of (1.22) by these matrices
does not affect the first row, and in fact,
E0A:
/0
1
a}+1
\0
0
Fs
aA
F0B
which is in row-reduced form.
(i) Suppose there are x and b such that Ax
preserves the
equality,
such that PAx
so
Pb.
PAx
Let
/
=
Then
b.
multiplication by
P
On the other hand, supopse x, b are given
be the transformation on Rm corresponding to P.
=
Pb.
fP
fr
composition
operations which are invertible, thus/j. is invertible. In
particular, fP is one-to-one, so since fp(Ax) =fr(b) we must have Ax b.
(ii) If d is the index of the mx n matrix PA, its last m d rows vanish. Thus for
d rows of Pb must vanish. By (i)
PAx
Pb to hold for some x, the last m
is
=
of row
a
=
=
this is also the condition for Ax
(iii)
The solutions of Ax
=
b
=
b to have
are
the
a
solution.
same as
those of PAx
Pb.
=
This latter
system has the form
xl + a^x2 +
x2 + a32x3
+
+
x' +
Clearly, x1,
.
.
.
,
x"
are
ai+1xd+l +
are
+ andx"
uniquely determined once xd+1,
m
d equations of
restricted by the last
free to take any values.
The b's
an1x"=ZpJ1bJ
a2x"=2pj2bJ
=
.
.
.
Pb
,
2 p/bJ
b" are known.
x", b\
0, but x'*1, ...,x" are
.
.
.
,
=
EXERCISES
13.
Compute the products
AB:
3^
(b)
A=
j
3;
38
Linear Functions
1.
4\
2
(c)
A
B
(6, 6, 3, 2,1,)
=
-1
8
\ V
(d)
A
4
2
8
6
1
2
-3
0
0
1
1
0
1
0
-1
-1
B
=
14. Compute the products BA for the matrices A, B of Exercise 13.
1 5. Compute the matrix corresponding to the sequence of row operations
which
row
reduce the matrices of Exercise 5.
given m x n matrix A find conditions
equation Ax b has a solution.
as given in Exercise 13(a).
as given in Exercise 13(b).
16. For the
under which the
(a)
(b)
(c)
A
A
on
the vector b in Rm
=
/
2\
3
-1
0
2
-1
0
1
4
(d)
V
'2
6
0
4
2
0
2
1
0
1
^8
6
0
-1
0
17. Show that if A is
an m x n matrix with m>n, then there are always
equation Ax b has no solution.
18. Verify that the composition of two linear transformations is again
b for which the
=
linear.
19.
Suppose that
T: Rm
-*
Rm and has this property:
r(E1)=0,...,T(Em)=0.
Show that T(x)
0 for every x e Rm.
20. Show that there is only one linear function
=
/(E,)
=
E2 ,/(E2)
=
E3
,
.
.
.
,
/(E)
=
on
R" with this
property:
Ei
PROBLEMS
1 5. Let
is
a
/:
R"
-+
R be denned
linear function.
by f(x\
Is the function
g(x\...,x")= ]>7=i(*02
.
.
.
,
x)
=
2?=
i
x*
.
Show that
/
1.3
linear ?
Is the function
ftfpc1, x2)
=
Linear
jc1*2 linear ?
16. Suppose that S, Tare linear transformations of R" to Rm5 +T, defined by
(S+T)(x)
=
entry
Show that the matrix
of the matrices
A, B, C be
17. Let
Show that
S(x)+T(x)
is also linear.
sum
39
Transformations
representing S +
representing S, T, respectively.
n x n
matrices.
T is the entry
by
Show that
(AB)C A(BC)
(A + B)C=AC + BC
A(B+C)=AB + AC
=
Show that AB
=
BA need not be true.
18. Write down the
products of the elementary matrices which
row
reduce these matrices:
11
4
7
3
l\
0
-5
6
2
0
1
0
0
3
0
0
5
1
2
o
-2
-1
1
o
/ ~26
\
3
2
-5
4
3
2
1
3
6
3
3
0
possible to apply further operations to the matrices of Exercise 1 8
bring them to the identity? Notice that when this is possible
for a given matrix A, the product P of the elementary matrices corre
sponding to these operations has the property PA I. That is, P is an
inverse to A. Using this suggestion compute inverses to these matrices
19. Is it
in order to
=
also:
'3
2
8
6
0
0
0
1
A
0
0
1
0
,1
0
0
0
1
0
0
\2
0
-1
-1
!\
20. Find
Find
a
2
x
2 matrix A, different from the
I.
2 matrix such that A2
a
2
x
identity
such that A2
=
I.
=
21. Is the
equation (I + A)(I + B)=I
+A+B
possible (with
nonzero
AandB)?
22. An
n X n
Show that
diagonal
matrices, AB
=
matrix to have
(a/) is said to be diagonal if a/ 0 for i =j.
matrices commute; that is, if A and B are diagonal
Give necessary and sufficient conditions for a diagonal
matrix A
BA.
an
inverse.
=
=
Linear Functions
1.
40
Linear
1.4
In the last section
Ax
we saw
that the
equation
b
=
be solved
can
of R"
Subspaces
for b's restricted
just
by certain linear equations and that the
equation might have some degrees of freedom. In both
determined by some linear equations ; such sets are called
set of solutions of that
cases
these sets
are
subspaces of R". We will begin with an intrinsic definition of linear
subspace and the notion of its dimension. In the next section we shall find
a simple relation between the dimensions of the sets related to the equation
linear
Ax
=
b.
Definition 4.
(i)
A set V in R" is
addition and scalar
a
linear
subspace if it is closed under the operations of
That is, these conditions must be
multiplication.
satisfied :
(1)
(2)
v1; v2
(ii)
If S is
V implies v, + v2 e V.
rv 6 V.
e
reR,\eV implies
a
set of vectors in
R", the linear span of S, denoted [5] is the
set
of all vectors of the form
c'-Vi
-\
1-
ckvk
with v,
VjeS.
(iii) The dimension of
a
linear
subspace
V of R" is the minimum number
of vectors whose linear span is V.
Linear
Span
Having
we
now
had better
just
given the intuitively loaded word "dimension" a definition,
hope that it suits our preconception of that notion. It does
that in R3 :
line is
one dimensional since it is the linear
span of but one
is
two
dimensional because we need that many vectors
plane
to span it.
In fact, it is precisely those observations which have
motivated
the above definition. We should also ask that the above
definition makes
vector; and
a
a
1.4
this assertion true: R" has dimension
that this is not
Linear
You may need
n.
41
Subspaces ofR"
a
little
convincing
since you do know of n vectors (the
standard basis) whose linear span is R". But how can we be sure that we
cannot find less than n vectors with the same properties ?
Consider this
immediately obvious,
restatement of the notion of
If the vectors vl5
"spanning":
...,
yk span
R", then the system of n simultaneous linear equations
ZA
=
b
j=l
has
solution for every b
a
that this cannot be if k
We
n.
now
.
.
,
We
already know from the preceding section
gives us a proof that R" has dimension
and that
// the
set S
The
is
proof
R",
v* span
by induction
on
R", then S has
least
at
n
n.
n
of them must have
one
context.
in R" spans
of vectors
of R" is
Thus, the dimension
Proof.
.
R".
repeat the arguments in the present
Theorem 1.3.
members.
Vi,
e
< n,
Supposing that
Subtracting an
and goes like this.
first entry.
a nonzero
appropriate multiple of that from each of the others, we may suppose that the
1 vectors all have first entry equal to zero. Then they are the same
remaining k
v* spanned R" we can show that
as vectors in -R"-1, and since the original vx,
1 and we have it. (Notice
1 >n
these must span R"-1. Now, by induction k
Theorem
of
in
the
the
first step
1.2.) Here now is a
that this is the same as
proof
more precise argument.
0) could
If none of the Vi,
(1, 0,
vt has a nonzero first entry then Ei
hardly be in their linear span. Letting as be the first entry of v,, we may suppose
2,
Vi and Wj
v,
as ar v2 for /
(by reordering) that at = 0. Now let wx
v* (see
k. The vectors wi,
w* have the same linear span as the vectors Vi,
(ai bi),
Problem 1 8) ; the difference is that only Wi has a nonzero first entry. Let Wi
bt span
b are in Rn~\ Now, b2
wk
v/2
(0, b), where bi,
(0, b2),
v/k span R", there are
R"-1. For let c 6 R"-\ Then (0, c) e R", and since wi,
.
.
.
,
,
=
.
.
.
.
.
.
.
,
.
.
'
=
=
.
.
,
=
-
.
,
.
.
.
,
,
=
,
=
=
.
.
.
.
,
.
.
,
,
.
x1
.
.
.
.
.
,
,
xke R such that
i>'W,=(0,c)
1=1
Thus xlai +x2-0-\
equation implies xx
Thus, b2
dim R"
,
.
.
.
,
> n.
fact dim R"
=
h*"-0
=
0,
so
=
h xkb
0, *% H
the second
equation
=
1 >n
by induction k
bt span R"-\
On the other hand, the standard basis Ei,
so
n.
-
Since
c.
1 ; that
-
.
.
.
,
a,
# 0, the first
c.
h x*b*
is, k > n. Thus,
E clearly spans, so in
becomes x2b2 H
=
42
Linear Functions
1.
Examples
16.
Vi
=
v2
=
v3
=
Let
(0, 1, 0, 3)
(2, 2, 2, 2)
(3, 3, 3, 3)
R4, and let S be their linear span. Then clearly
But it is also clear that v3 is superfluous, since v3
Thus S is also the linear span of vx, v2 : if
be three vectors in
dim S
<
3.
3/2(v2).
v
a1v1
=
then
v
a2v2
+
+
(a2
Thus, dim S
were a
a3\3
+
also write
we can
a\
=
=
2.
<
vector
3/2(a3))v2
+
w
In
=
fact,
S has
(a1, a2, a3, a4)
precisely dim 2. For suppose there
which spanned S. Then we would
have numbers cu c2 such that vt
=
qw, v2
=
Explicitly
c2w.
this
becomes
0
=
1
=
0
=
3
=
CiO1
cta2
cta3
cta4
But this is
0,
ct /
so
2
=
2
=
2
=
2
=
clearly impossible. By the second equation
0.
But 2
by the first we must have a1
=
could not be.
17.
V=
c2a}
c2a2
c2a3
c2a4
Thus, dim 5
=
+
v3 -v4
=
c2
must have
a1,
which
2.
Let V be the subset of R4
{y.v1 +v2
we
=
given by
0}
V is certainly a linear subspace of R4.
We will shortly have the
theoretical tools to deduce that V has dimension
3; with a little
work we can show it now.
First of all, let Ax
(1, 0, 0, 1), A2
(0, 1, 0, 1), A3 (0, 0, 1, 1). Then A1; A2 A3 are all in V, and if
v
(v1, v2, v3, v4), since v4 vl + v2 + v3 we have
=
=
,
=
v
=
=
v1A1
+
v
2A2
+
v3A3
=
1.4
Linear
43
Subspaces ofR"
Thus V is the linear span of Al5 A2 A3 so dim V < 3. On the other
hand, if dim V < 3, then Al5 A2 A3 can all be expanded in terms of
,
,
,
of vectors
pair
some
Bl7 B2.
If
we
delete the fourth entry in all
these vectors this amounts to
saying
in R3
of vectors.
can
be
spanned by
a
Thus dim V
impossible.
pair
that the standard basis vectors
But dim
R3
=
3,
so
this is
3 also.
=
Independence
Repeating
the definition
once
again, dimension
is the minimum number of
linear space. There is another closely allied
intuitive concept: that of" degrees of freedom" or "independent directions."
In such phrases as "there is a four parameter family of curves," "two
vectors it takes to span
a
are involved," allusion is being made to a
dimension-like notion. Now, if we try to pin down this notion mathematic
ally and specify the concept of independence in the linear space context, it
turns out to be precisely the requirement for a spanning set of vectors to be
minimal. In other words, the dimension of a linear space is also the maxi
independently varying quantities
number of
mum
degrees of freedom,
Let S be
Definition 5.
vectors if the
independent
x1y1
x1,
with
0,
.
.
.
,
xk
.
.
.
=
,
x*vk
+
+
x*
e
indpendent
vectors in the space.
We say that S is
set of vectors in R".
a
set of
equation
=
and
R
a
or
0
\u...,vk distinct elements of S
implies x1
=
0.
The standard basis of R" is an
now verify that R" has in fact
We
independent set,
no more
than
n
as
is very easy to verify.
of freedom in this
degrees
sense.
Proposition
Proof.
since
n
> 1
The
Let \u...,\k be
10.
set in R".
Then k<n.
is
=
Now let
always.
the first entry of v,\
are zero, then bi,
.
independent
1 is automatically true,
by induction on k. The case k
Let as be
us proceed to the induction step (k > 1).
we can thus write v,
(at b,\ where b, e R-1. If all the a,
bk are an independent set in R"'\ By the induction assump
proof
.
an
.
=
,
,
We
tion then, k <n l,so k <n. Now suppose instead that some as is nonzero.
Then
for
i > 2.
Let
'v,
0.
Vi
a,
aT
that
w,
so
d
=
vectors
the
can reorder
given
P* are an
(0, (3,) with p, e R-1. p2
the first entry of w, is 0, so w,
=
=
-
44
/.
Linear Functions
set in R"'1.
independent
For if
|-ic'aiWIVi+
\
J
Since vi,
.
1
=
.
.
2
,
vk are an
1
2
=
2*=2 c'/J,
0, then also
2J=2
c'w(
0,
p2
0,
=
so
c'v,=0
2
c2
independent set,
Thus, by induction,
and the proposition is proved.
independent.
k <, n,
=
once
=
=
again
k
c*
=
1 <
n
so
1
.
are
also
Thus in every
case
,
.
.
.
,
pt
Examples
18.
vt
=
v2
=
v3
=
Let
(0, 3, 0, 2)
(5,l,l,2)
(1, 0, 2, 2)
In order to show that these vectors
are
independent
we
must show
that the system of equations
x1vl+x2\2+x3\3
=
0
(1.23)
has only the zero solution. But this system is the same as
corresponding to the matrix whose columns are vx, v2 v3 :
the system
,
V2
If
we row
y0
2
2)
reduce this matrix
0
x2 + *3
x2 + 2x3
x3
+
0
obtain
0y
Now the system PAx
x1
we
=
0
=
0
=
0
=
0
=
0
obviously
has
only the
zero
solution: if
O-24)
1.4
find, reading upward that x3
we
=
Linear
0, x2
=
Subspaces ofR"
0, x1
0.
=
45
Since P is
0 has only the zero solution. What is
invertible then the system Ax
x1
0. Thus
x2
the same, if (1.23) holds, so must (1.24), so x3
=
=
the vectors \t, v2
=
v2
=
v3
=
A
=
independent.
(3, 2, 1, 0)
(1, 2, 3, 1)
(2, 0,-2,-1)
Again,
A=
=
Now let
19.
Vl
v3
,
are
I
let A be the matrix whose columns
1
3
\0
1
row
are
vt, v2
,
v3:
-2]
-1/
reduces to
~4 :fi
rA
lo
\0
ol
0/
o
0
The system PAx
x1
x2
=
=
-3x2
x3
+
=
0 has the solutions
2x3
0 has the same solutions.
The system Ax
the particular solution (-1, 1, 1). Thus
=
Taking x3
=
-Vi + v2 + v3 =0
20.
Four vectors in R3 cannot be
?,=(2,1,2)
(0,3,0)
v2
(1,0,4)
v3
=
=
v4
=
(0,l,2)
independent.
Let
1
we
have
Linear Functions
/.
46
Find
a
linear relation which these vectors must
reduce the matrix whose columns
A=
/l
3
(0
1
\0
If
satisfy.
we row
obtain the matrix
we
1/3
-1/
-1/6
1
Now, corresponding
0, and thus of
Ax
the v's,
1 \
0
0
are
to any value of
=
xVj
x4
Take x4
0.
=
we
obtain
1
Then
=
.
a
solution of
1
x4
x3
x2
fr3-3-x4 \
x1= -3x2-x4=-i
=
=
=
=
Thus
Vl
2
~
iV2
+ V3 + V4
=
0
Now, the equivalent form of these
two propositions about R", that any
members, and any independent set has
spanning
set of vectors has at least
at most
members, holds for any linear subspace of R"
n
Proposition 11.
n
Let Vbea linear
well.
as
subspace ofR" of dimensions d.
(i) A spanning set has no less than d elements.
(ii) An independent set has no more than d elements.
Proof. Part (i) is of course just the definition, so we need only consider part (ii).
proof amounts to a reduction to the case where V is R", and an application of
The
Proposition
10.
Let Wi,...,wd span V; since V has dimension d there exists such vectors.
V. Then we can write
Suppose, as in (ii), that vt,
, vk are independent vectors in
.
each Vj
as a
vj
=
.
.
linear combination of Wi,
2 flj'w,
.
.
.
,
wd;
1 <, j <, k
j=i
for suitable numbers
a/.
corresponding to the
independent. For if,
R"
The vectors
vectors
io(fl/,...,a/)=0
j=i
{v,};
(/,
we
.
.
.,
shall
a/)
for j
1,
show that
now
=
.
.
.
,
k
are
they
vectors in
are
likewise
1.4
then also
25
cJ\j
i
=
=
ilk
i
%cJ\j= J=l
2cJ2/w,= 2
J=l
1
=
l=l\j=l
1
Vi,
independent in R",
a/)
Definition 6.
are
.
.
.
,
vt
hO-w
/
the k vectors
...,
,
we
must have c1
so
c*
0. Thus
0,
by Proposition 10 d> k.
v s
V
can
=
=
.
a linear space.
A basis of V is
be written in the form
Let Fbe
such that each
47
\
2cV|Wi=0-WiH
Thus, by the independence of
(a/,
Subspaces ofR"
by this computation:
0
k
k
Linear
a
.
.
,
set 5 of vectors
k
v
in
one
2! c%
;=i
=
and
only
cl
witn
one
R,\te S
e
way.
Another way of putting this is: a basis for
independent and spanning vectors in V.
Proposition
12.
S is
a
a
linear
basis for the linear space V
subspace
V is
a
if and only if both
set of
these
conditions hold:
S is
(i)
(ii)
an
independent set,
of S is V.
the linear span
Suppose that S is a basis of V. Since every vector in V can be written
linear combination of vectors in S, certainly (ii) is true: Kis the linear span of 5.
Since 0 can be written in only one way as a linear combination of vectors in V, any
Proof.
as a
time
we
have
c1vi +
with
0,
.
.
c1,
.
,
c*
.
.
.
=
---
+ c'vt
=
0
& in R and vi;
,
0
(since
0
=
0
Vi
.
.
.
,
distinct members of S, we must have c1
Thus (i) holds : S is an indepen
v* also).
=
vk
\- 0
-I
dent set.
Conversely,
any vector
with c'
In
v
v
=
e
R,
suppose now that
(i)
and
(ii)
are
true for the set S.
Then
(by (ii))
in V can be written
ciVl +
v, e
S.
...
(1.25)
+ c*Vt
This
can
be done in only
one
way because of the
independence.
fact, suppose (1.25) holds, and also
Y=aiVl +
is true, with (c1
..-
0-26)
+ akvk
a^vi -I
h
(ck
ak)yk
=
0,
so
c'
=
a' since the v,
are
independent.
48
Linear Functions
/.
Dimension and Basis
The
are
facts to know about dimension of linear
important
these : such
a
has
basis with
a
That number is the
greater than
We summarize this
n.
Theorem 1.4.
Let V be
a
linear
of R"
subspaces
finite number of elements.
space V always
same for all bases and is the dimension of
a
V, and is
not
follows :
as
subspace ofR".
(i) There is an integer d<n such that V has dimension
(ii) Any basis of V has precisely d elements.
(iii) d independent vectors in V form a basis.
(iv) d spanning vectors in V form a basis.
d.
The proof of this part of the theorem is by mathematical induction
1, either V= {0} or V has a nonzero vector, in which case V=R.
Now we proceed to the induction step.
Thus either dim V
0 or 1 so dim V < 1
(i)
Proof,
If
on n.
n
=
=
.
,
describe how it goes. We assume the assertion (i) for
R"'1 as the set of w-tuples in J?" with zero first entry. If Kis
1, and consider
of R", it
intersects this space in some subspace of R"~* which is, by induction spanned by
some S vectors, with 8<n 1.
Now, choosing any other vector in V with a
nonzero first entry, this together with the vectors referred to above will span V.
Let
us
Now
we
make this argument
a subspace of R".
=
coordinates
W
Wis
a
(a1,
...,
assume
=
If V
{0}, then dim V
=
0 ; if not, V has
=
a"). One of the entries is nonzero;
that a1 ^ 0. Let now
we may,
a nonzero
by reordering
the
{weR-1:(0,w)e V}
linear
subspace of R"-1.
c\0, wi) + c2(0, w2)
in V,
8 ^n
subspace
precise.
Let V be
vector v0
n
a
c'wi + c2w2
=
For if
Wi
,
W and
w2 e
c\ c2
e
R,
we
also have
(0 c'wi + c2w2)
,
W.
Now, by the induction hypothesis, W has dimension
span W.
By definition of W, Vi
\6
(0, Wi),
we need only show that v0
(0, wa) are
Let v e V, and
v span V.
Vi,
let c be its first entry. Then v
c(a*) 'v0 is also in V and its first entry is 0. Thus
this vector is of the form (0, w) with weI^. Then there are c1
c" such that
so
1.
-
e
Let wi,
in V. Now
.
.
.
,
ws
=
.
.
.
,
"
-
W
=
CJWi H
h cVa
Thus
v
-
c(a1)"1Vo
=
(0, w)
=
.
,
=
c'CO, w,) +
+
cs(0, w5)
.
.
,
1.4
Linear
49
Subspaces ofR"
or,
v
=
c(a1)~1v0
+ c'vi H
Thus, there are 8 + 1
h c'va
vectors which span
(n-l)+l=n.
(ii) This follows easily
from
V,
so
Khas dimension d with d < 8 + 1 <
Proposition 12. If 5 is a basis for V (dim V d),
elements, and since 5 is independent it has at
=
then since S spans, it has at least d
most d elements.
vd are independent vectors
(iii) Suppose that Vi
they span. Let y0eV. By Proposition 12(ii), since
c") # 0 such that
dependent, so there exist (c
cv0 H
h c-v,,
in
V;
we
dim V
=
must show that
d,
v0,
.
..,
vd are
0
=
cd
vd are independent we must also have c1
0, a
0, since Vi,
h cdvd) as desired.
contradiction. Thus c # 0, so v0
( c0)'1^! ^
If they are dependent, then the equation
vd span V.
(iv) Suppose that vi7
If c
=
=
.
.
.
=
=
,
=
.
1- Cvd
c'vi H
holds with at least
vr
=
one
.
=
.
,
0
c' = 0.
(-^"'(c'vi
If say c' # 0, then
+ C-'vr-i + c'+1vr+1 +
+
+
Cvd)
Hence, Khas dimension at most
Vi,...,vd, with vr excluded, also span V.
d 1, a contradiction, so we must have had Vi, ..., v independent and thus a
so
basis.
proposition, whose proof is
(theoretical) ease in finding bases.
This final
the
Proposition
13.
Let V be
a
linear
left
as an
exercise, is
an
indication of
subspace of R" of dimension
d.
(i) Any set of vectors whose span is V contains d vectors which form
(ii) Any set of independent vectors in Vis part of a basis for V.
Examples
21
.
Find
Vl
=
v2
=
v3
=
v4
=
(4,
(5,
(0,
(1,
a
3,
2,
1,
0,
basis for the linear span V of the vectors
2,
2,
0,
0,
1)
1)
1)
1)
and express V by
a
linear
equation.
a
basis.
50
Linear Functions
/.
We want to find all vectors b of the form
x\
and
we
(1-27)
b
=
want to find
system corresponding
Now
basis for such vectors.
a
(1.27)
is the
to the matrix A whose columns are the vectors
Trie sPan f these vectors is Just trie range of A. If
of
elementary matrices row reducing A, then any vector
product
Thus by
b is in the range of A if and only if Pb is in the range of PA.
row reduction we should easily be able to solve our problem.
v3
?i, v2
P is a
A
v4
.
=
The end result of
row
'11
1
PA=|
'
0
l
~!
0
yO
0
reduction
V
0
0
0
1
0
0
-4
1
|
1
1
0
0
0
Oy
,1
1
-1/2
-3/2
-4
rA
Thus, the
to
V
produces
P
r
=
range of A is obtained
by setting
the fourth entry of Pb
zero :
range of A
=
{(b1, b2, b3, b4) -.b1
=
+
b2
-
f b3
dim V
<
3.
22. Find
a
the
Thus, dim V
=
-
1/4).
4b4
-
V has dimension at least three since it contains the
(4, 0, 0, 1), (0, 4, 0, 1), (0, 0, 2/3,
so
1
+
8x2
x2
We
A=
+
-
independent vectors
hand, V # R4,
3 and these three vectors
basis for the linear subspace
are
(5
1
\0
3x3
x3
8
x5
x5
=
0
=
0
3
1
0
2
1\
-1
0/
=
are a
V of R5
=0
the solution space of Ax
0-10
1
+
2x4
+
seeking
x4
+
-
0}
On the other
equations
5*1
x1
=
0, where
basis.
given by
1.4
Linear
Subspaces of R"
51
Row reduction leads to
PA=
/l
0
-1
0
0
1
-1\
0
2
0
\0
0
-2
-15
6/
and V is the set of x such that PAx
0.
are to be freely chosen and
choice. Thus, dimF=2. Choosing
According to these equations
jc1, x2, x3 determined by this
(x4, x5) (1, 0), (0,1), re
=
x4 and x5
spectively,
we
obtain
as a
=
basis
(-. -2, -Y.1.0)
(4,0,3,0,1)
EXERCISES
21. What is the dimension of the linear span of these vectors?
(a)
=(-1,2, -1,0)
v2=(2,5,7,2)
?,=(0,2,1,1)
t4
(3, 5, 7,1)
Vl =(-1,0,2, 1)
Vl
=
(b)
?a
=(2, 2, -2,2)
=(1,1, 1,1)
(0,2,1, 1)
?a
(1,7, 3, 3)
v3=(0,0, 0,1)
v=(l,3, 1,2)
v5=(l,5,2,2)
Tl= (0,0, 1,1,1)
v2= (1,0, 0,1,1)
?a =(0,1, 0,1,0)
Ta
(C)
Vi=
=
(d)
22. What is the dimension of the space S given by these
(a) S {x Rs : x1 + x2 x3 x* 0, x1 + x3 0}
(b) 5 {x e Rs: x2 + x* + xs =0, x1 x3 + x* =0
=
-
-
=
=
equations:
=
-
x1
-
x2
-
x3
-
xs
=
0}
R*: x1 + x2 + x3
x3
x2
x1 + x*}
23. Determine the linear span of these vectors by a system of equations
(c)
S
(a)
^=(1,0,0,1)
?a=(0, 1,1,0)
?,= (0,1,0,1)
vx=(2,2,6,2)
v2
(l,2,3,0)
(b)
=
{x
e
=
=(0,1,0,-1)
?!=(!, 0,1)
?a
(-1,1,1)
vx= (1,0, 0,0,0)
v2=(2,0, 1,0, 1)
v3
(c)
=
(d)
=
-
-
52
/.
Linear Functions
independent?
given in Exercise 21(c).
Vi, v2 v3 as given in Exercise 23(a).
^=(0,2,0,2,0,6)
?a =0,1, -1,-1, 1,1)
v3=(2,4,6,8,10,12)
V4
(0,0, -2, -2,0,0)
v3= (0,1, 0,0, 1,0)
v6= (1,1, 1,1,1,1)
25. Find all linear relations involving these sets of vectors.
(a) Vl =(0,1,1)
v2
(5,3, 1)
v3= (0,2,0)
v4
(1,-1,1)
(b) v1=(0,2,0,2)
v2
(0,l,0,0)
v3= (0,1,0,1)
v*
(0, 0, 0, 1)
Ts=(l, 0,-1,0)
(c) ^=(0,0,0,0)
?a
(1,1, 1,1)
V3
(1,1,0,0)
V4
(0,0, -2, -2)
26. Find a basis for the linear subspace of Rs spanned by (0, 0, 0, 1, 1),
(0, 1, 0, 0, 0), (1, 0, 0, 0, 1), (1, 1, 0, 0, 1), (2, 1, 0, 1, 2)
24. Are these vectors
(a)
(b)
(c)
Vi,
.
.
.
v5 as
,
,
=
=
=
=
=
=
=
=
27. Find
(a)
a
basis for these linear spaces:
+ 2x2 + x3 =0, x1 + 2x* + x5 =0,
{(x1, ...,x5)eR5:x1
x' + x'^O}
(b)
{(x1, ...,xi)eR*:x1-x2 + x3-x*=0,x1-x3=0}
given vectors on Rs are independent, extend them to a basis:
(a) (0, 0, 0, 0, 1), (0, 0, 0, 1, 1), (0, 0, 1, 1, 1)
(b) (1, 5, 2, 0, -3), (6, 7, 0, 2, 1), (1, 0, -1, -2, 0), (1, 1, 1, 1, 1)
(c) (4, 4, 3, 2, 1), (3, 3, 3, 2, 1), (2, 2, 2, 2, 1)
28. If the
PROBLEMS
23.
Suppose
w2 =v2
given
we are
/J2Vi,...,
w*=v*
k vectors Vi, ...,v* in R". Let wi=Vi,
for some numbers /32, ...,/?*. Show
/3tvi
that the sets {vi,
vj and {wi,
24. The proof of Theorem 1.3
.
.
.
,
sists of the vectors Vi,
many elements?
25. Prove
.
.
.
,
vt.
w*} have the same linear span.
proceeds by assuming that the set S con
What of the case where S has infinitely
.
.
.
,
Proposition 13.
V, W are subspaces
26. Show that if
of R",
so
is Vn W.
1.5
Rank +
Nullity
=
Dimension
53
27. Show that if A is obtained from B
by a row operation, the linear span
of A is the linear span of the rows of B.
28. Show that if A is a row-reduced matrix the dimension of the linear
span of the rows of A is the same as its index.
of the
1.5
rows
Rank +
Now let
Nullity
us
spaces, and in
There
certain obvious linear spaces to be associated
are
given transformation.
Definition 7.
Let T: R"
Rm be
->
a
linear transformation.
The set
(i)
K(T)
a
Dimension
apply the propositions of the preceding section about linear
particular the notion of dimension, to the subject of linear
transformations.
to a
is
=
linear
=
{veR":T(\)
subspace
of T, denoted
(ii) The set
0}
=
of R", called the kernel of T.
Its dimension is the
nullity
v(T).
R(T)= {T(v):veRn}
subspace of R", called the
p(T).
is
a
of
linear
T, denoted
Let T: R"
Theorem 1.5.
n
v(T)
=
+
that is, dimension
For
Proof.
Let v+i,
R".
Let y/j
.
wv+i,
p(j)
(i)
.
=
.
.
.
,
-
a
linear
We have
transformation.
p(T)
=
nullity
+ rank.
short, write v(T)
v
Rm be
Its dimension is the rank
range of T.
be the rest of
as v.
a
Let Vi,
.
.
.
,
be
v,
basis for R": vi,
.
.
.
a
,
basis for the kernel of T.
v,
,
vv+i,
.
.
.
,
v
thus span
Now the crux of the matter is this:
n.
v+ 1,
of
T.
Once this is shown, we will have
the
range
w form a basis for
desired
the
is
which
equation.
n
v,
Then there is a v e R" such that w
T(v). Expand v in the
Let w e R(T).
.
=
T(\j) for j
...,
,
=
54
Linear Functions
1.
basis vi,
.
.
.
v: v
,
vi=T(v)
=
=
=
=
c'vi +
\- c"\
T(c1Yi +
c'^vi) +
+ c"\)
---
+
cT(v)
+ c"w
justified since T is linear
T(vv+i) wv+i,
The second line is
=
and the third follows since Vi,
vv are
Thus these last vectors
w.
, T(v)
.
.
.
,
=
R(T).
(ii)
wv+1,
w are
...,
cv+1wv+1 +
r(cv+1vv+1 +
cv+1vv+i -I
independent.
Suppose
+ c"w=0
-"
We must show that the
so
c"T(vv) + cv+1r(vv+1) +
+
cv+1wv+i +
in the kernel of T and
span
Then
.
{c1}
+
h c"v
e
are
c"v)
(1.28)
all
=
.(7").
In any event, from
zero.
cv+17(vv+1) +
vi,
.
.
.
,
+
vv span
cT(v)
^(J)
so
=
(1 .28)
we
have
0
there
are
c\
.
.
.
,
c" such
that
cv+1vv+1 H
h c"v
=
c'vi H
h cvvv
or
(-c^Vi +
Since vi,
.
.
.
,
+
v are
(-cv)vv + cv+1vv+1
independent,
+
all the cJ
are
+ c"v
zero,
=
as
0
required.
The theorem is
proven.
Examples
23. Let T: R4
A=
/l
3
2
(0
0
1
1
1
0
0/
\0
We
A
/l
0
\0
3
R3 be given by the matrix
7\
completely analyze
can
easily
->
row
2
this transformation
by
row
reduction.
reduces to
7\
10
0
0
1/
1
(1.29)
(1.30)
1.5
Rank +
merely by interchanging the last
transformation corresponding to
Nullity
two rows.
=
Dimension
Thus, letting
55
P be the
f. 2 !)
0/
1
\o
know that PT is the linear transformation
to (1.30).
R3, and the range of
1
3. The
T is P~ (range of PT), which is again all of R3, so p(T)
kernel of T is the same as the kernel of PT, which has the equations
given by (1.30):
we
Now, the range of PT is easily
seen
corresponding
to be all of
=
x1
+
3x2
+
2x3
x3
+
+
lx4
x2
x4
=
0
=
0
=
0
(1.31)
by letting x4 take on all real
remaining coordinates by (1.31). Thus
R}, which is one dimensional.
The set of all such solutions is found
values and
K(T)
=
{(
solving for the
St, 0, -t,t): te
24. Let T: R4
A
->
R4 be given by the matrix
=
Let
us row
reduce this matrix,
10
keeping
0
10
0
track of
our row
operations :
0^
0
-3010
-200L
P
0
0
0
=
Oy
Now, the kernel of T is easy
the transformation S
to
find; it is the
corresponding
to the last
same as
the kernel of
matrix PA (because
56
Linear Functions
/.
S
the
=
of T
composition
by
an
invertible transformation). Now the
corresponding to the matrix PA,
kernel of S, and thus also of T, has,
the form:
x1
or
x2
x2
+
x2
+
x3
+
+
x4, x1
=
K(T) ={(-(
so
2x4
x4
v(T)
=
+
0
=
0
0
=
0
0
=
0
=
v),
x3
x4.
-v, u,
Thus,
v): (u, v)
The range of T is
2.
transformation
matrices
=
corresponding
a
to
e
R2}
little harder to find.
If R is the
the
elementary
product
of the
the left, then S
R
T, so the range of T is R_1 of the
x4
0.
range of S, which has the equations x3
(That is, the vector
b4) is in the range of S if and only if there exist (x1,
(b1,
x4)
on
=
?
=
...
=
,
...,
such that
x1
+
x2
x3
x2
+
+
+
2x4
x4
=
=
0
=
0
=
b1
b2
b3
b4
The necessary and sufficient condition is b3
b4
0.) Thus the
necessary and sufficient condition for v to be in the range of T is that
Pv be in the range of S; that is, the third and fourth coordinates of
=
=
Pv must vanish :
-3*1 +4x2 +x3 =0
-2X1 -5x2 + x4
0
=
Thus, p(T)
25. Let
=
us
corresponds
/l2
0
1\
1
3
0
1
1
1
1
2
\4
3
7/
2.
do
one more
to the matrix
example briefly.
Suppose
that T: R3
->
Rs
1.5
This matrix
can
/I
0
1\
0
1
1
0
0
0
0
0
0
\0
0
0/
A
=
be
by multiplication
P
=
V
row
Nullity
=
Dimension
57
reduced to
the left
on
Rank +
by this matrix
1
0
0
0
0\
2
1
0
0
0
2
-1
1
0
0
1
-1
0
1
0
1
-1
-1
-1
1
The kernel of T can be found
by looking at the row-reduced form A;
0.
(x1, x2, x3)
Precisely, we
must have x1 + x3
x2
x3
0.
Thus
a
in K(T) if
vector
is
+
0,
its first and second coordinates are the negative of the third ; that is,
1.
The range of T is
K(T)= {(-t, -t,t):teR}. Thus v(T)
such
that
Px
is
in
the
the set of x
x5)
range of A (since
(x1,
of
A
are
A
in
the
The
precisely those with
range
PT).
5-tuples
zero.
Thus
the
third through fifth
fifth
coordinates
and
fourth,
third,
coordinates of Px must be zero for x to be in the range of T. Specifically
R(T) is the set of simultaneous solutions of
it is the set of
x
in R3 such that Ax
=
=
=
=
=
=
.
.
.
,
=
2X1 -x2 + x3
X1 -x2
+ x4
-x1 -x2 -x3 -x4
We
can
take
x3, x4, x5 ;
R(T)
so
These
purely
=
p(T)
x1, x2
+
as
x5
=
0
=
0
=
0
free variables and
these
equations
to define
thus
{(u, v-2u,v-u,3v- 2u) : (u, v)
=
use
e
R2}
2.
examples
illustrate the fact that Theorem 1.5
in terms of matrices.
We
now
do
just
that.
can
be formulated
58
/.
Linear Functions
Proposition 14. Let A be
formation T: R" ->Km. Then
p(T)
an m x n
matrix, representing the linear
of independent columns of A
of independent rows of A
of the row-reduced matrix to which
=
number
=
number
=
index
A
can
trans
be reduced.
Finally, we can also reformulate Theorem 1.5 as a conclusion for systems
equations, thus bringing us to the ultimate version of Theorems 1.1
of linear
and 1.2.
Suppose given a system of m linear equations in n unknowns,
or rank, of the corresponding matrix A.
Then
Theorem 1.6.
and suppose d is the index,
(i) d <m,d < n.
(ii) {x: Ax 0} is a vector space of dimension n d.
(iii) {b : there exists a solution of Ax b} is a vector
=
=
space
of dimension
d.
EXERCISES
29. Describe
by linear equations the
range and kernel of the linear trans
formations given by these matrices in terms of the standard basis :
(a)
(b)
(c)
(d)
/8
0
30. Find bases for
0
1
6\
K(T), R(T) for each
T
given by the matrices (a)-(d) of
Exercise 29.
31. Let
of /is
/:
R"->R be
a nonzero
linear function.
Show that the kernel
linear
1.
subspace of R" of dimension n
32. Let f(x\ ...,*")
*'
Find
a
i
spanning
2"=
kernel of/.
a
=
set of vectors for the
1.6
Invertible Matrices
59
PROBLEMS
29. Let T: R"
linear
->
Rm be
a
linear transformation.
of R", Rm, respectively.
30. Let T be the transformation represented
are
Show that R(T) is spanned by the columns of A.
K(T)={(x\...,x):
fCjxJ=0}
where d,
31. Let
the columns of A.
.
.
.
32. Let S
1.6
w e
C
,
are
Define
R".
w e
Show that for
all
Then K(T) and J?(D
subspaces
as
the set of
v
m x n
matrix A.
Show that
such that
2"=i
v'w'=0.
^ 0, J_(w) is a subspace of Rn of dimension n\.
R\ Define (S) as the set of v such that 2?=i v'w'
w
=
Show that
S.
_|_(w)
by the
(S) is a subspace of R",
and dim
(S)
+
=
dim[S]
0 for
=
.
Invertible Matrices
In this section
we
shall pay
particular attention
to the collection of linear
transformations of R" into R"
From the
the
case
or, what is the same, the n x n matrices.
of view of linear equations this is reasonable ; for it is usually
point
a given problem
that
will have
as
many
equations
as
unknowns.
First of all ; it is clear that there are certain operations which are defined
on the collection of all linear transformations of R", thus making of this set an
algebraic object of some
following definition.
Definition 8.
The
sort.
We collect
algebra of linear operators
collection of linear transformations
(i) if/is in ", and
(cf)(x)
(ii) if/
g
=
are
c
is
E",f+g
(f+g)(x)=f(x)
(iii) fgis
a
provided
real number,
cf(x)
in
defined
together all
by
(fff)(x)=f(9(x))
+
is defined
g(x)
by
on
these notions in the
R" denoted
with these
by E",
operations :
is the
60
Linear Functions
1.
important to think of the elements of E" as functions taking n-tuples
ra-tuples of numbers ; but in working with them it is convenient
Thus, we are
to represent them in terms of the standard basis by matrices.
with
the
led to consider also the algebra M" of real n xn matrices
operations
of scalar multiplication, addition and multiplication, the definitions of which
we now recapitulate.
It is
of numbers into
The
Definition 9.
with these
If A
(i)
cA
(ii)
operations
=
=
If A
algebra
M" is the collection of
n
xn
matrices
provided
:
(a/) is in M", and c is a real number,
(ca/)
=
A + B
(a/)
=
and B
=
(b/)
are
in M", then
(a/ + b/)
AB=(ak%k])
The two
algebras E", M" are completely interchangeable, for M" is just
explicit representation of E" relative to the standard basis.
Now the operations on M" obey certain laws, some of which we have
already observed in previous sections. Let us list some important ones.
the
Proposition
These equations hold for all
15.
n
xn
A, B, C and all
matrices
real numbers k.
(i)
(ii)
(iii)
(iv)
k(A + B) A:A + A:B
C(A + B)=CA + CB
(A + B)C AC + BC
A(BC) (AB)C
=
=
=
If A is
a given matrix, we shall let A2 denote A
A A A, and
A, A3
general A" is the -fold product of A with itself. Since we may also add
matrices, and multiply by real numbers, we may consider polynomials in a
given matrix. That is, A2 + 3A + A, A7 + 3ttA3 + A6,
In fact, if
we adopt the usual convention that A0
then
for
J,
any polynomial
PVQ X"=o ctX' in the indeterminate X, we may consider the matrix p(A)
Z"=o c;A'. A most remarkable observation can now be made, by noticing
=
in
.
.
.
.
=
=
=
that the collection M" of
2-tuples
of real numbers.
n
x n
matrices is the
same as
the collection R"2 of
1.6
p
Proposition 16. Given any n x n
of degree at most n2 such that p(A)
A, there
matrix
=
Invertible Matrices
is
a nonzero
61
polynomial
0.
Proof. An element of M" is a rectangular array of n2 real numbers, thus corre
We may make this correspondence explicit, by, say,
sponds to an element of R"
placing the rows one after another. That is, the matrix (a/) corresponds to the
vector (ai1,
al- 1, a") in R"2. In any event, the notions
a,1, ai2,
an2, d3,
of sum and scalar multiplication is the same in the two interpretations. Now con
A"2. These n2 + 1 vectors in R"2 cannot be inde
sider the matrices I, A, A2,
.
.
.
.
,
...
pendent
there
so
Thus the
+
c2
.
,
,
real numbers c0
are
c2A"2 +
.
.
...,
,
cu
.
.
.
,
c2
,
not
all zero, such that
A2 + CiA + coI=0
proposition is verified with p the polynomial
p(X)
=
c2
X"2 +
+
c2
X2 + ClX+
Co
rephrase this proposition in this way: Every matrix is a root of some
polynomial equation with real coefficients. From the purely
algebraic point of view this formulation is of some interest and raises the
converse speculation: given a polynomial with real coefficients, does it have
We shall verify this fact, and with n no greater
some n xn matrix as a root?
we
than two.
More precisely,
shall, in a later section, introduce the system
a
certain
collection of 2 x 2 matrices, and later verify
of complex numbers as
a
root in the system of complex numbers.
has
that every real polynomial
of algebra.
theorem
the
fundamental
This is known as
We may
nonzero
linear transformation in E" is invertible if it has an inverse as a
For this it must be one-to-one and onto; that is,
function from R" to R".
rank + nullity)
We have seen (n
it must have zero nullity and rank n.
Now,
a
=
that either of these assertions
assertions must be
The
Definition 10.
that BA
I
=
=
Proposition
expressible
AB.
17.
case
BA
I
=
BI
=
=
B is said to be
An invertible matrix has
This is clear : if B, C
Proof.
in terms of matrices;
AB
CA
=
=
B(AC)
=
a
an
I
AC
(BA)C
=
IC
=
C
a
matrix B such
unique inverse.
inverses to A, then all these
=
do that.
inverse for A.
are
Then
B
we now
matrix A is invertible if there is
nxn
In this
implies
Now it is clear that these
the other.
equations hold :
62
1.
Linear Functions
We shall denote the inverse of
ship
a
Let A be
Proposition 18.
(i)
(ii)
(iii)
(iv)
(v)
by A 1. The relation
gives us this propostion:
matrix A, if it exists,
between matrices and linear transformations
an n x n
matrix.
These assertions
are
equivalent:
A is invertible.
A represents an invertible transformation.
I.
There is a matrix B such that BA
=
There is
a
matrix B such that AB
A has index
=
I.
n.
Proof. We have already seen (in discussing systems of linear equations) that (ii)
(v) are equivalent. By definition (i) implies both (iii) and (iv). Thus we have
left to prove that (i) and (ii) are equivalent, (iii) implies (i), and (iv) implies (i).
and
(i) implies (ii).
it represents.
Since A A-1
Let A be the
given invertible matrix, and T the transformation
represented by the inverse, A-1, of A.
Let S be the transformation
I
A_1A, we have T- S I S T. Thus S is inverse to T, so
(ii) holds.
(ii) implies (i) by the same kind of reasoning with the roles of matrix and trans
formation interchanged.
(iii) implies (i). If T is the linear transformation represented by A, then by (iii),
there is a transformation S such that S T
I. Thus, if T(x)
0 we must also
have x
S(T(x)) S(0) 0, so T has nullity zero and is thus invertible. Thus
(iii) implies (ii), so also implies (i).
(iv) implies (i). If again, T is the transformation represented by A, by (iv) there
is a transformation S such that T S
I. This implies that T has rank n and thus
=
=
=
=
=
=
=
=
=
=
is invertible.
Computing
the Inverse
Now, it is clear that the question of invertibility for a given matrix is
important and that the problems arise of effectively deciding this question
and of effectively computing the inverse, if it exists. To ask that the rows
(or columns) be independent, or span R", while responsive to this question
hardly provides a procedure for determining invertibility. We shall now
introduce two such procedures: one is a continuation of row reduction and
the second is based on the notion of the determinant. The determinant is a
real-valued function defined on the algebra M" of n x n
its basic
matrices;
property is that it is
nonzero
only
on
the invertible matrices.
We shall
the determinant in the study of
eigenvectors (Section 1.7).
In Section 1.9 we shall explore the connection between the
determinant and
the notion of volume in R3.
depend heavily
on
1.6
In order to
Invertible Matrices
63
the critical
properties of the determinant function it is
elementary matrices, for they provide a technique
for decomposing an invertible matrix into a product of simple ones, and as
a result, a technique for computing inverses.
We recall these facts: the
are
matrices
the
matrices
which
elementary
represent the row operations.
Since the row operations are invertible, so are the elementary matrices
invertible. For any matrix A there is a sequence Ps
P0 of elementary
is
in
row-reduced
form.
The index
matrices such that B
PjPj-j
P0A
of A is the number of nonzero rows of B. We augment these facts by this
verify
necessary to return to the
.
,
.
.
,
=
further observation:
Proposition 19. Suppose that A is an invertible n x
P0o/ elementary n xn matrices such
sequence P,
matrix:
identity
P0A
P,
Proof.
The
Theorem 1.2.
n
matrix.
that
There is
a
P0 A is the
P,
I
=
proof
will be by induction
The first column of A is
As
independent (A is invertible).
exist elementary matrices P0
,
.
.
.
we
on n.
It is
a
slight modification of
since the columns of A must be
in the proof of Theorem 1.2, there
nonzero
have
seen
P* such that the first column of P*
,
P0A is Ei.
Thus,
1
n-1
/l
Al\
1
P*---P0A-|o A22j
Since P*
applies
to
P0A is invertible so is A22 (see Problem 37). Thus the proposition
1)
1) x (n
Q*+i of elementary (n
A22. There is a sequence Qs
-
,
matrices such that Qs
Pj
n-1
=
Q*+iA22
( q)
(J q'
P.-PA
*>r7
=
=
I.
.
.
.
-
,
Let
* + l,...,J
Qt+lA22) (o 1)
=
=
...
Now the matrix
(I -f)
is the product Ps+_,
Ps+i of elementary matrices corresponding to these
row
Linear Functions
1.
64
operations:
a/
subtract
times
they'th
from the first row, j
row
=
2,
...,n.
Finally,
p.+-i-poa=(j _f)(0 f)=(j ;)=i
as
required.
This
we
proposition provides
just continue
with
us
the process of
corresponding product
Then the
an
row
effective way for
reduction until
we
computing inverses;
obtain the identity.
of elementary matrices is the inverse.
Examples
26.
1
2
2\
-3
4
2
1
0
8/
/
A=
\
Product of elementary matrices
Row reduction
/I
2
2\
0
10
8)
2
6/
\o
-
/
1
0
3
l
0
\-l
0
1/
A
0
0
1
A\
*
/
\o
0
w
\-A
/i0
\o
-
*
/
1
\
0
13 7
190
1
190
0
1/
\
0
/
o\
-iV
iV
A
o
1/
21
190
1 5
190
2
76
Thus
A_1=Tfo
0\
/ 137
-21
65
15
\-10
5
-10\
-20
25/
27.
1\
(212
10
1
-1
-1
0
-1
4
1
2
1
-24/
76
-iV\
-A
w
1.6
T
0
1
0
0'
1
2
/
2\
1
1
2
I
0
0
0
0
-1
1
3
0
1
1
0
0
4
2
\o
1
0
1,
'10
-22/
1
2\
0
-31/
0
1
Oil
0
2
'10
0
1
0
^0
/
0
10
0^
1
-1
1
0
\-4
9
0
1;
l-\
0
2\
1
0
-31/
0
0
1
oil
0)
0
0
n
o
o
o\
o
i
o
oW
0
0
1
ON
^0
0
0
1/ \
65
0-200
-10/
0
Invertible Matrices
2
-1
1-2
0
1
-1
10
-10/\-6
11
/-u
if
ft
-I*
1-1
-4
10
10
0^
0
-2
1,
-u
a
1
10
Thus,
The Determinant Function
The determinant of a matrix is
into
of it and its
pretty complicated concept; before going
properties, we shall first see how to compute it.
a
study
Looking ahead, the method of computation comes from Equations (1.35)
and (1 .36), but we shall not use those equations to derive it. Instead we shall
simply describe the technique for finding determinants.
a
The determinant of
det I
,)
=
a
ad
The determinant of
a row.
2
x
2 matrix is defined
be
a
3x3 matrix is found
The determinant will be
a sum
of the
withy numbers, called their cofactors.
is (- 1)'+-' times the determinant of the 2 x
row and/th column are deleted.
row
by
follows.
as
products
First, select
of the elements of that
The cofactor of the
(i,j)th entry
2 matrix
when the rth
remaining
66
Linear Functions
1.
Examples
Compute the determinant of
28.
1
3
2\
-1
4
0
7
-2
/
A=
\
If
we
det A
select the first
=
1[4(1)
(-2)0]
the second
Selecting
=
-
=
7[3(0)
way.
=
(-2)2]
-
+
4[1(1)
-
2(7)]
-
4(2)]
-
(-2)[1(0)
2(- 1)]
-
For
a
example, selecting
sum
-3[(- 1)1
same
way.
of the
0(7)]
-
Select
products
+
4[1(1)
andy'th
-
7(0)]
(or column).
of the entries in that
(n
column
0\
(431
6
10
0-1
0
-
are
29. Let
2
a row
The cofactor of the
determinant of the
2
-
7(4)]
+
0[. .]
.
+
[1(4)
-
3(- 1)]
2[1(0)
-
-
2(- 1)]
-45
cofactors.
row
2[(- 1)(-2)
column first, and proceeded in the
the second column:
Now, in general, the determinant of the
the
+
row:
could also have selected
we
same
=
7(0)]
-
-45
Now,
det A
3[(- 1)0)
-45
Selecting the third
=
-
find
row:
-(- 1)[3(1)
=
det A
row we
-45
=
det A
1/
4l
11-1/
1)
x
(n
deleted.
-
x n
n
matrix is found in
The determinant is the
row
(or column)
with their
(i,j)ih entry is (-1)'+J' times the
1) matrix remaining when the rth
1.6
Select the first
det A
We
4 det
=
/6
0
j
0
0
\1
1
(2
12
0
1
0
4
\2
1
-l)
3 det
12
6
0\
0
Odet 1
0
0
\2
1
\2
1
1/
compute the determinants of the 3
3 matrices
by taking
since its factor is 0:
4(-l)[6(4) 0(-l)] 3(-l)[2(4) 1(-1)]
[2(4)
1(-1)]
+(-6)[l(-l) 4(2)]
-
-
-
-
-
-
-24.
=
30.
A
=
/6
2
1
0
4
3
8
1
~l\
0
0
0
2
0
0
8
1
4
0
1
\2
1
4
-1
Select the third
=
x
Select the second column in the
advantage
first three, and don't bother with the last
detA
-1\
6
of the location of the O's.
=
67
row
+ det 1
now
det A
Invertible Matrices
1/
row:
/6
(-l)3+3(2)det I48
2
0
-1
3
1
0
1
0
1
\2
1
-1
1
-
Select the third column:
/6
detA
=
2(-l)2+3det
8
\2
=
2
2
1
1
+
(-l)4+3(-l)det
-1\
4
3
0
\8
1
1/
96
theory of determinants. We begin with a definition
which is appropriate to the theoretical dis
function
of the determinant
the multiplicative property:
it
has
that
then
cussion and
verify
We turn
now
to the
det(AB)=
det A- det B
Linear Functions
1.
68
and
(1.36) below which form the basis for the preceding
a rewriting of the formula for the determinant.
The formulas
(1.35)
computations
will result from
The determinant of
the
sum
of all
matrix
an n x n
products
of
precisely
one
can
be described in this way: it is
row and column,
element from each
Our first business is to determine this appropriate
of
precisely one element from each row and column is
sign. A selection
In
the first row we select a certain element, say in the
follows:
as
described
with
appropriate signs.
In the second
7t(l) column.
column, say n(2).
ment
a\{i)
in the rth
row
...,
or
'
'
an
element, coming from
and
so
sure
These numbers then form
numbers 1,
.
.
.
,
n.
a
different
We select the ele
forth.
n(2) n(\),
7i(i)th column, making
^
that the numbers
a
rearrangement,
To form the determinant
then,
we
'
ait(n)
2(1)
as 7t
and
all distinct.
n(n)
permutation of the
consider all products
n(\),
are
select
row we
We have
ranges
over
all
n.
A
of the numbers 1,
,
of
two
successive
integers:
interchange
permutations
kind of permutation is
an
.
.
.
particular
1-1
2-2
+ 1
i-
i+1h
n
(We consider the integers as arranged in a circle, so that 1 is the successor to
n.) Now it is a fact about permutations, that any permutation consists of a
succession of such interchanges. There may be many ways to build up a
given permutation by these simple interchanges, but the parity of the number
involved is always the same. That is, if we can write a given permutation as a
succession of an even number of interchanges, then every way of writing that
permutation as a succession of interchanges will involve an even number.
For example, consider the permutation on four integers
1 2 3 4-3 1 4 2
This is obtained
12 3 4
2 13 4
2 3 14
3 2 14
3 12 4
3 14 2
by this succession of interchanges:
1.6
Here is
a
better way of
doing
Invertible Matrices
69
it:
12 3 4
13 2 4
3 12 4
3 14 2
Either way, there is an odd number of interchanges involved. We shall not
verify these facts about permutations; the verification would be tangential
to our present study.
However, we shall use these facts. We shall say that
a given permutation is even if it can be formed by an even numbered succession
interchanges ; the permutation is odd if an odd numbered
interchanges is required. For any permutation n, its sign,
of
succession of
denoted
e(n)
1 if n is odd. There is another way of defining
will be + 1 if n is even, and
the sign function on permutations which is described in Problem 36. This
does not involve the notion of
description
If A
Definition 11.
=
(a/)
detA=
all
a
an n x n
matrix its determinant is
e(X>fl<o
(1-32)
1=1
n
show that det A # 0 if and only if A is invertible, by
det A det B.
stronger statement : det (AB)
We shall
in fact
permutations
is
interchange.
now
showing
=
Lemma 1.
(i)
(ii)
det 1=1.
If A
has
a zero
row, det A
=
0.
Proof.
(i) Writing I (a/), we have ai<0 =0, unless ttQ) i. Thus, the sum (1.32)
has only one nonzero term, that corresponding to the identity permutation. Since
1
1.
1 1
each a,'
1, det I
(ii) If the /th row of A is zero, each term of the sum (1.32) has a factor a{u) 0,
=
=
=
=
=
=
so
is
zero.
Then det A
Lemma 2.
IfP
det(PA)
Proof.
Type
I.
=
Let A
If P
det PA
is
=
an
(a/),
PA
multiplies
=
elementary matrix, and A
any matrix,
(1 -33)
det A
det P
=
0.
=
(b/).
the rth
2 e(n) fl #<o
=
row
by
c, then
2 e(w)J<'>
'
caSw
aSoo
1=1
n
=
2 s(tt)c 1II1 alw
=
=
c
det A
Linear Functions
1.
70
In the
special
holds in this
Type II.
the
A
case
Suppose
det PA
=
det(PI)
P
now
c
=
det I
=
c.
Thus
(1.33)
interchanges the rth and sth rows. Let 77 represent
r and s.
Thus, b/
af'\ Now we compute:
which interchanges
2 W ii ,>
=
1
we
have det P
we
case.
permutation
Now
I,
=
=
1
=
=
2 eWll <l
change the index of summation.
(T V) 11 <%)
=2
det PA
Let
w
=
t
17, and
sum over t.
-2 to 11 <('.,)
=
1=1
sign changes since r; is an interchange; thus, if r is even, t 17 is odd. Now the
product FT" 1 a?(o> is tne same as tne product fl" 1 {(o (another change of index)
The
=
=
so
det PA
In
II {<o
2 *(t) j=i
=
particular, det
=
We
a/ + aaf.
bf
det(PI)
det A
-
=
1,
P adds
Suppose that
Type III.
and
P
=
now
a
(1.33) holds in this
so
times
row r
to row
s.
case.
Then
b/
=
a/
if i =
s
compute
det(PA)=2e("-)n#(o
(
=
1
n
=
n
2 EW II aid) +
2
1=1
(7r) 11
(1 .34)
ct*u> aUr) aUsy
<=i
(*r
The first term
the
on
right
s.
of the form
17, where
77
2
n even
=
2
n even
It is
11 aid)
important
-
n
is
The second term is
even.
a'n(r) a'Ms)
11 fl(D <&(D fl(s)
t^r,s
II i(i)(o(o a'Ms)
zero.
We
can see
that
by
Let 77 represent the inter
to note that the odd permutations are just those
l&r,s
Z
rt even
=
sum
is det A.
into odd and
splitting up the
change of r and
permutations.
even
Thus the last term in
2
nodd
EI i<o
'
Equation (1.34) is
a'*w *<>
'
t$r,s
2
n even
a;(s) a'Mr>)
11 flioKD>0ii(J|(r <(<)>
l<trts
=
0
l^r,s
det A. In particular, detP
Thus, det PA
Type III elementary matrices.
=
=
l,
so
(1.33) is verified also
for
1.6
Invertible Matrices
71
Now, lemma is a word denoting a logical particle of no particular intrinsic
interest, but of crucial importance in the verification of a theorem. Here
now
is the main theorem
Theorem 1 .7.
concerning determinants.
A matrix M is invertible
if and only if det M #
0. det AB
=
det B for any two matrices.
det A
Proof. Suppose M is an n x n matrix which is not invertible. Then there are
P0 such that P5
elementary matrices P,
P0M is row reduced and has zero
rows.
Thus, by the above lemma
.
,
0
=
det(P3
.
.
,
P0M)
det P3
=
det P^i
det P0
det M
Since the determinant of an
On the other
that I
=
elementary matrix is nonzero, we must have det M 0.
hand, if M is invertible, there are elementary matrices P,
P0 such
,
Ps
1
=
det I
=
det AB
so
det AB
is true.
=
=
det P,
det P5_,
det P
matrices.
n x n
0 and either det A
.
,
If
=
0
det M
one
or
of A
det B
=
or
B is not
0.
In any
invertible, neither
case
det B
det A
If A and B
.
Then
P0M.
Thus det M = 0.
Now let A, B be two
is AB,
=
.
invertible, there are elementary matrices P,
are
P0 Q
,
Qo
such that
Ps
PA
=
I
=
QB
Q
Then
Q
PoAB
QoP,
PoA)B
Q0(P5
0,
=
=
Q
Q0B
=
I
Thus
det Q
det Q
det Ps
Thus
det Q0
det Q0
det P0
again det(AB)
=
det P
det P.
det B
det A
det A
=
det(AB)
=
1
1
=
1
det B.
Notice that the formula det AB
the basis of the definition above.
=
det A
det B is far from transparent
on
fact, it is not at all derivable without
of n x n matrices. We have a
the
structure
some information regarding
matrix A; namely, the process
invertible
for
a
means of computing A-1
given
not
have
of row reduction. But we
given explicitly any formula for the
In
72
1.
Linear Functions
determinant.
This
formula is of theoretical interest, but not of any great
As far as computations are concerned, the surest and
computational
quickest route
value.
inverse is the process of row reduction.
Let A be an n x n matrix. The adjoint matrix of
an
Such is
inverse.
provided by
the cofactor
expansion
of
a
to the
entry of A is the
(n 1) x (n 1) matrix obtained by deleting the row and column of the
given entry (see Figure 1.13). Let A/ be the adjoint matrix of the entry a/.
Then the inverse to the matrix A (if it is invertible) is easily given by the
determinants of the adjoints: the (i,j)th entry of A-1 is
(-ir^
det A
More
A-1 A
precisely we have these formulas (the explicit version
I) known as Cramer's rule:
of AA_1
=
=
detA=
(-l)'+'a/detV
for alii
(1.35)
J(-l)l+'a/detA/
for
(1.36)
;=i
detA=
i
0
=
0
=
=
all;
l
f (- 1),+ V det V
for all i # k
I (- l)i+ V det V
for
all;
I
A;'
Figure 1.13
= k
(1.37)
(1.38)
1.6
The verifications of these formulas
be based
fix
directly
index i.
a row
formula
on
(1 .32).
are
Invertible Matrices
73
simpler than it may seem; they can
example, let us verify (1.35). First
the sum in (1.32) into n parts: those
For
We shall break up
permutations taking / 1, i 2, ...,/ n. Consider, for a fixed column
index the permutations taking i-*j. (That is, those n for which n(i) =j.)
These are precisely the same as all permutations on the indices of the
matrix adjoint to a/ (those permutations which take the integers 1, ...,,
Thus the terms appear
n, except for;').
except for i, into the integers 1,
ing in the sum (1.32) which have a/ as a factor, are the same as those in
(1.35): we must now verify that the signs agree. Let % be a permutation on
the indices of the adjoint to a/. The corresponding permutation n of
n) does the same as t and takes i into j. The number of interchanges
(1,
.
.
.
.
building
required
to send i to
i+j.
Thus, e(n)
(1 .32)
.
,
,
involved in
as
.
(1.35) also
and
that for t, with the
this
permutation is just
j.
The last number is j
interchanges
same parity
signs of corresponding terms in
( 1)'+J(r), so the
Thus (1.35) is true.
=
agree.
tions of the other formulas to the exercises.
require
a
small
i, which has the
We shall leave the verifica
(Equations (1.37)
and
(1.38)
trick.)
Cramer's rule allows for a simple description for solving the equation
Let A(i) be the matrix obtained
b when A is an invertible n xn matrix.
Ax
column
b. Then the equation
by replacing the rth column of A with the
turns
Ax
out, according to Cramer's rule
A_1b,
b, which is the same as x
=
=
=
to read
detA(i)
,
xl
.
t
1
=
< i < n
det A
This is checked out
by unraveling all the definitions and applying the formulas
of Cramer's rule: since
x
=
A_1b,
x-^A-^-^^-iy-detA^
But the summation is
just
the determinant of the matrix obtained by replacing
Thus we can solve by taking
the rth column of A with the column vector b!
quotients
of determinants.
Example
31. Solve the
equations
2x2- x3
**+ jt2 + 3x3
2x1 + 2x2+ x3
x1
+
=
2
=
0
=
l
74
1.
Linear Functions
The determinant of the matrix
is
easily found by cofactor expansion along the first
det A
=
1(1
-
6)
-
2(1
6)
-
-
1(2
-
2)
=
row :
5
By Cramer's rule
x1
=
|
(2
det 0
2
-1
-1\
1
=
|[2(-5)
+
1(6
+
1)]=-|
2
2
x2
x3
=
=
i
detj
1
idet(l
\2
-r
0
=
l[-2(-5)-(l+2)]=l
1
1
2
(the determinants
0)=i[2(0)+l(l-2)]=-i
1/
are
computed by
column cofactor
EXERCISES
33. Find the inverse of these matrices
(a)
(b)
(c)
(d)
expansion).
1.6
34. Solve the
Hi)
where A is
equation
the matrix in Exercise
the matrix in Exercise
(c)
A
'I
0
-1'
2
1
2
?
-1
ly
'4
6
1(
3
1
2
,0
4
0,
=
(d)
A
=
Suppose that the
if
=0
a/
75
given by
(a)
(b)
35.
Invertible Matrices
i
Show that A"
36. If A is
n x n
33(a)
33(b)
matrix A
=
(a/)
has this property:
<y
=
a
0.
matrix such that A"
=
0 show that I + A is invertible.
PROBLEMS
33. Show that if
Show that if there is
a
a
linear transformation T has rank n, it is invertible.
transformation S such that T S
I, T is invertible.
=
Equations (1.35)(1 .38) using the definition of the determinant.
35. Assume this fact about polynomials: A polynomial of degree d has
no more than a" roots.
Prove the following assertions:
(a) Let A be an n x n matrix. There are at most n numbers s such
34. Derive
that A + si is not invertible.
(b) The m x n matrix
/l
V=(l
\l
has
n2
r22
rV'X
r2
r
r2
rrl)
n
rrl)
determinant if and only if the r( are all distinct. (Hint:
0, there is a nonvanishing linear relation among the columns.)
a nonzero
If det V
=
36. Let
Rx1
*9
where
even
x\
.
.
=
.
,
n(*'-*0
x" axe,
distinct numbers.
if and only if
f(x'w
x")=f(x1,...,x")
Show that the
permutation
7t
is
76
1.
Linear Functions
and
similarly
f(x"w,
...,
w
**<">)
37. Let A be
x
is odd if and
-f(x\ ...,x")
=
an
only if
invertible
n x n
m) matrix formed from
(n
Show that B is also invertible.
A
matrix.
For m<n, let B be
by deleting any m rows and
(Hint: You need only take
an
(n
m)
m
columns.
m
=
\, and
proceed by induction.)
38. Let
H t)
Verify that
A2
-
(a + d)A + (ad-bc)I=0
that is, that A is a zero of a polynomial of degree 2.
39. The same fact is true for all n, that is an n x n matrix is the
zero of a
This is part of a famous theorem of algebra, which
goes like this: If A is any matrix, the polynomial
polynomial of degree
PA(x)
=
det(A
-
n.
xl)
is the characteristic polynomial of A. A is
PA(x) 0 (Cayley-Hamilton). That is,
a
root of the
polynomial equation
=
i^(A)=0
Verify the Cayley-Hamilton theorem for (i) a diagonal matrix, (ii) a triangular
matrix.
1.7
Eigenvectors
and
Change
of Basis
One fruitful way of studying linear transformations on R" is to find direc
along which they act merely by stretching the vector. For example,
if a transformation Tis represented by a diagonal matrix
tions
(1.39)
1.7
then
Eigenvectors and Change of Basis
T(Ej) ^E; where Ex
EB are the
a
factor
by stretching by
dt along the rth
standard basis vectors.
=
,
.
,
.
.
,
Tacts
T(x\
More
.
.
.
,
xn)
<*!*%
=
+
d2 x2E2
+
---
+
dx"En
generally, suppose we can find a basis \lt
by stretching along the direction of v;
Then, if
v
=
.
v of vectors in R"
for each i:
.
.,
d,yt
is any vector, the action of T is easily computed by referring v to
v
s'yt then T(v) Y,disiyi. T is represented
the basis ylt ...,y: if
by the diagonal matrix
a
Thus
direction:
such that T acts
T(yt)
77
basis of vectors
=
=
,
relative to this basis.
(1.39)
The process of
finding
which T acts
along
by stretching is called diagonalization.
not all transformations can be so diagonalized and this
Unfortunately,
a major difficulty in this line of investigation.
For example,
rotation in the plane clearly does not have any such directions in which
acts as a stretch.
More precisely, let T be represented by the matrix
presents
a
it
M-! i)
Then
If
v
=
T(x, y)
(a, b) is
(y, x). (T is
such that T(a, b)
=
da
=
b
Then d2a
=
db
db
=
(except 0).
Nevertheless, there
clockwise rotation
d(a, b),
we
through
a
right angle.)
must have
a
=
-a,
a
=
and there
are no
real numbers
d,
a
making this equation
true
this way, and it is
doing
many transformations which can be analyzed in
for
purpose in this section to study the techniques
are
our
so.
Definition 12.
R" be
Let T: R"
a
linear transformation.
number d for which there exists a nonzero vector
An eigenvector of T with eigenvalue d is a nonzero vector
of T is
a
An
eigenvalue
v
such that Tv
=
dy.
v
such that 7v
=
dy.
which there is
Proposition 20. // T: R" R" is a linear transformation for
with
a basis of eigenvectors y1,...,yn
eigenvalues du...,d, respectively,
then for any vector v
2! *S T(v) Z dis'y>
=
=
.
78
Linear Functions
1.
Compute T(y) using the fact that
Proof.
T is linear.
eigenvalues of a linear transformation T by making use
di is singular (not
an eigenvalue of T if and only if T
invertible). If A is the matrix representing T in terms of the standard basis,
[this condition is verified precisely when det(A dX) 0. Thus the eigen
Notice that when Tis rotation
values of T are just the roots of this equation.
by a right angle
Now
we
find the
of this remark: d is
-
u i)
det
which has
tion has
d2
=
+ l
explaining in
real roots, thus
no
no
d\
We shall
eigenvectors.
=
see
another way why this transforma
we extend the real number
that when
system in which every polynomial has
a root (the complex
represented in terms of (complex) eigenvectors.
This is one of the important reasons (particularly in the study of differential
equations, as we shall see) for so extending the number system. Let us now
to a
system
numbers),
then T
be
can
collect these observations.
Let T be
Proposition 21.
d is
A.
an
eigenvalue
det(A
-
fl)
=
a
transformation on R" represented by the
ifd is a root of the equation
matrix
ofT if and only
0
If d is an eigenvalue,
ofT-dl.
the
of eigenvectors corresponding
set
to
d is the kernel
Suppose d is an eigenvalue of T. Then there is a v ^ 0 such that
(T- dl)y 0. Thus the nullity of T di is positive, so T di is not
invertible. Thus, det(A
di) 0. On the other hand, if det(A dl) 0, then
Tdl is not invertible, so has a positive dimensional kernel. If v ^ 0 is in the
kernel, (T dl)(y)
0, or Ty
dy; thus d is an eigenvector of T.
7v
Proof.
dy,
=
or
=
-
=
=
=
Examples
32.
Let T be
-G D
represented by the matrix
-
=
1.7
Eigenvectors and Change of Basis
79
Then
*-H27' 1i,)
det(A fl) t2 3t + 2. The roots are r 2,
eigenvectors corresponding to t 2 is the kernel of
and
of
=
-
-
=
1.
The space
=
*--(! -?)
that is, the space of all vectors (x, y) such that x
0.
y
(1, 1) is an eigenvector with eigenvalue 2. The eigenvectors
=
-
sponding
to t
=
Thus
corre
1 lie in the kernel of
C o)
that
is, in the space of vectors (x, y) such that x 0. (0, 1) is such
eigenvector. Since (1, 1) and (0, 1) are a basis for R2, we have
diagonalized T. Relative to this basis T is represented by the matrix
=
an
33. Consider the transformation
basis
by
given,
relative to the standard
the matrix
GD
Then
fl)
det(A
eigenvalue of T.
-
=
r2
-
8r + 16
=
(t
-
4)2.
Thus
4
is
the
only
0}, which is one dimensional. Thus
there cannot be a basis of eigenvectors for the only eigenvectors he
on the line x
2y.
Notice that this example differs from that of a rotation, for there
is no problem with the roots; the difficulty lies with the transformation
has
as
kernel
{(x, y):
=
itself.
x
+ 2v
=
80
Linear Functions
1.
R3 be given by the matrix
34. Let T: R3
0
2
2
4
3
0
8/
\
Then
det(A
det(A
are
-18\
1-1
A=
tl)
-
fl)
-
=
=
1
-
3
+
3t2
-
0
2, -1.
2:
Eigenvalue
-18\
1-9
0
2
0
4
3
0
6/
A -21=
\
The kernel is the set of vectors
(x,
space is two dimensional,
with eigenvalue 2; for example, vx
so we can
1
Eigenvalue
A
+ 3z
=
1
4
3
0
9/
0
which is
or
=
one
0
or
[2
0
0
2
0
\0
0
-1/
a
(x,
x
=
3z
y
=
2z
such that
x
+ 2z
=
0.
This
=
independent eigenvectors
(0, 1, 0), v2 (2, 0, 1).
=
y,
z)
such that
( 3, 2, 1). These
eigenvector is v3
basis of eigenvectors, and T is represented by
An
dimensional.
v1; v2 v3 thus form
the matrix
,
z)
-18\
0
2
The kernel is the set of vectors
2x + y + 4z
y,
find two
:
1-6
-(-1)1=
\
x
The roots of
4.
0\
relative to the basis yu v2
=
(1.40)
,
v3
.
1.7
Eigenvectors and Change of Basis
81
Jordan Canonical Form
Notice that in general there are two difficulties with the procedure de
The polynomial det(A
fl) may not have many real roots,
scribed above.
and it may have multiple roots. As we shall see in the next section the first
difficulty can be overcome by transferring to the complex number system.
34 above demonstrates that the second
Example
possibility,
that of
multiple
roots, may not be severe, whereas Example 33 shows that it can seriously
handicap the diagonalization procedure. Continued study of this situation
becomes
the
quite difficult and we shall not enter into it. The conclusion is
typical matrix which cannot be diagonalized is of this form
0\
Id
1
0
0
0
0
d
1
0
0
0
0
0
d
1
0
0
0
0
0
d
0
0
[o
that
d)
0
the transformation
representing
T(x\
.
.
.
,
a")
=
(dx1
+
x2, dx2
+
x3,..., dx")
Given any matrix, we can find a basis of vectors (which includes all possible
eigenspaces) relative to which T decomposes into pieces, each of which has
the form (1.41). This is called the Jordan canonical form.
Change of Basis
Before
allow
x
us
in R"
leaving this subject, let
to change bases in Rn.
can
x
uniquely.
=
us
If
compute explicitly the formulas which
E} is a basis for R", then any
{E1;
.
.
.
,
be written
*%
+
+
x"En
w-tuple (x1,
E} denoted xE
We shall refer to the
relative to the basisE:
{Eu
.
.
.
,
.
.
.
,
x")
as
the coordinate
of x
.
of
F} be another basis for R". Let xF be the coordinates
{Fu
associate
we
can
To each set of E coordinates x
x relative to this new basis.
In this way we can
the F coordinates xF of the point corresponding to xE
The precise relation is this.
write xF as a function of xE
Let F:
.
.
.
,
.
.
82
Linear Functions
/.
E,.=
Proof.
=
x
E},
of the
^
=
yl
(x\
.
.
.
,
x"),
xF
=
(y\
...,
be two
F}
different
is denoted
AFE.
Then
y").
i>'F<
=
1
=
^
which is the
(1.42)
(AFE)~
that
=
*
=
AEF.
For, given any xf
AF AF XF
I.
relative to any basis E:
T(x)E
(1.42).
same as
Now, if T is any linear transformation
{Et
,
.
.
on
.
,
R", it
can
Let
E}.
be
us
represented by
a
denote that matrix
Tx
Proposition 23.
R" is
=
.,
2(1^/^^,
=2"=i aj'xJ>
AF XE
A/A/
Tf
.
(1.42)
Notice that it follows from
and T: R"
.
2ZxjVj=2Zx4Z"j'F<)
Thus for each i,
matrix,
byTE:
{F1;
A/x
1
a
F:
E's:
have this relation between its E andE coordinates:
we
=
Thus
.,
change of basis matrix, and
is called the
in R"
=
XF
.
aj%
Let xE
x=
.
;=i
For any point
xF
{E1;
Write the E's in terms
(a/)
The matrix
Let E:
22.
Proposition
bases for R".
IfE:{Elt..., E}, F: {Ft
,.F}
transformation, we have
,
a
linear
(A/)-1TA/
.
.
.
are
two
bases
of R",
1.7
Eigenvectors and Change of Basis
83
Proof.
T(x)F
On the other
r(x)F
Thus TF
=
=
AFET(x)E
=
A/Tx
=
AF*TA/xF
hand, by definition
=
TfxF
ArTEA/
=
(A/r^A/
Examples
35. Let T: R2
R2 be represented, relative
to the standard basis E
by
*-G J)
Let F:
{(1, 1), (2, -1)}
be another basis.
Find the matrix
Tf
Now,
*-G -?)
A/=(A/)-i=i(; _2)
Thus
*-G -96 K -?)
-i\
/7/3
\4/3
2/3,/
36. Let T be
given, relative
0
2
2
4
3
0
8/
\
and let F:
that F is
{(0, 1, 0), (-2, 0, 1), (-3, 2, 1)}.
a
respectively.
by
-18\
1-1
T=
to the standard basis E
basis of
Thus
we
We have
already
seen
for T, with eigenvalues 2,2, 1,
may conclude that Tf is given by (1 .40).
eigenvectors
.
84
Linear Functions
1.
EXERCISES
37. Find
represented
(a)
a
basis of
eigenvectors,
if
possible, for the transformation
by the matrix A:
in terms of the standard basis
1
eigenvalues: 2, 3,
(b)
eigenvalues: 1, 1, 0,
2
(c)
4
eigenvalues: 1,
(d)
/
1
1^
-1
0/
38. Show that for F: {Fi,
F} a basis for R", and E the standard basis,
the matrix AEF is just the matrix whose columns are Fi,
F
.
.
.
,
.
.
39. Find the matrix A/ for these
.
,
.
of bases in R".
pairs
(1, 0, 1), (0, 1, 1), (1, 0, 0)
E: (0,1, 2), (2, 0,1), (1,2,0).
(b) F: (1,0,0), (2, 0,1), (0,1,0)
E: (3, 1,5), (0,2, 3), (-1,-1,0).
(c) F: (1, 0, 1, 0), (0, 1, 1, 0), (0, 0, 2, 0), (0, 0, 1, 1)
E: (0, 2, 0, 2), (2, 0, 0, 0), (2, 0, 2, 0), (0, 2, 2, 0).
40. Let T: R3
R3 be a linear transformation represented by
(a)
(b)
/ 2 0 0\
/l 0 -1\
F:
(a)
TE=
\
-1
0
1
0
relative to the standard basis E.
(F
in Exercises
39(a)
3
TE=
1/
Find TE
,
0
1
\2
0
where F is
one
one
4
-1/
of these bases
and
39(b)).
independent eigenvectors with the
eigenvalue, then T is represented by a diagonal matrix in any basis.
as
of
41. If T:R2-rR2 has two
same
PROBLEMS
40. Prove
Proposition
41. If T is
which has
Pa(x)
and
=
n
a
=
on
R"
represented by the matrix A
distinct eigenvalues di,...,d, then
(- l)"(x
PA(A)
20.
linear transformation
0
-
dt)(x -d2)---(x- d)
(PA is defined in Problem 39).
1.8
Complex Numbers
42. Let T be a linear transformation on R". Let
E(r)
Show that E(r) is a linear subspace of R" (called the
Show that if r # s, then (r) n (5)
{0}.
=
r
{v
e
R" : Tv
eigenspace
85
=
of
rv}.
T).
=
43.
if ru
.
Pa(x)
Suppose
.
.
=
,
r are
A represents
the
(- l)"(x
-
a
linear transformation
eigenvalues of A, then
r,)
d'm
<ri>
(x
-
rk)
Verify the Cayley-Hamilton theorem for
44. Find
a
matrix with
no
n
d'm
=
2*=i
on
R" with this property :
dim
Then
(">
A.
nontrivial
eigenspaces.
expect to prove the Cayley-Hamilton theorem for such
1.8
E(rj).
How would you
a
matrix ?
Complex Numbers
Pythagoras' discovery, that yjl is not the quotient of two integers, was
day to be a geometric mystery. His conception of numbers
was limited to rational numbers and his desire to measure
lengths (to associate
numbers to line segments) led to this unhappy realization: there are some
lengths which are not measurable ! (as the hypotenuse of an isosceles right
triangle of leg length 1). It took a long time for mathematicians to realize
that the solution to this situation was to expand the notion of number. The
general liberation of thought that was the Renaissance led in mathematics to
the possibility of expressing the value of certain lengths by never-ending
decimals, or continued fractions, or other types of infinite expressions. It
was during those days that mathematicians formulated the view that such
expressions represented numbers and served to determine all lengths. Earlier,
Middle Eastern mathematicians were led from certain algebraic problems
considered in his
concept in another direction. As they
1 has no square root; some bold adventurer then
observed, quite clearly
suggested that we contemplate, in our minds, some purely imaginary quantity
to envision extension of the number
As
1 and treat it as if it were another number.
whose square would be
this supposition did not contradict any of the known facts concerning the
number system, it could do
(at least in our minds).
no
harm
and
might
do
a
great deal of good
Today we need not be so mysterious or cunning in our ways. We need
only recall that there is a 2 x 2 matrix (see Problem 20) whose square is the
negative of the identity. We can thus say quite factually that in the set of
1 does indeed have a square root. Well, there is also a
2x2 matrices,
5x5 matrix, and an n x n matrix for any n whose square is I, so we should
1 has a square root. The
ask for the smallest algebraic system in which
-
Linear Functions
1.
86
complex number system is this system and we shall later derive the remarkable
fact (the fundamental theorem of algebra) : Every polynomial has a root in
the complex number system.
Now,
to be
explicit,
the matrix
-G "D
(1.43)
-I. The complex number system is the collection
has the property that i2
the
form al + bi, where a, b are real numbers.
of
of all 2 x 2 matrices
=
Definition 13.
C,
the set of
complex
numbers is the collection of all 2
x
2
matrices of the form
Proposition 24.
(i) The operations of addition and multiplication are defined on
(ii) Every nonzero complex number has an inverse.
(iii) C is in one-to-one correspondence with R2.
Proof.
(i) (a
-b\
\b
(ii)
a)
+
(c -d\_(a +
c)
\d
la
-b\lc
\b
a)\d
\b
c
c)
-(b + d)\
+ d
-d\^/ac-bd
C.
a
+
c
J
-(ad+bc)\
ac-bdj
\ad+bc
If
r)
is nonzero, then one of a or A is nonzero,
an inverse.
By Cramer's rule
M
-,_
1
a2 +
/
a
so
det M
=
a2 + b2 = 0, and thus M has
b\
b2\-b a)
(iii) is obvious, since every complex number is given by
conversely every pair (a, b) of real numbers gives rise
and
a
pair of real numbers
a complex number.
to
1.8
Cartesian Form
We need
of a Complex
now a
Complex Numbers
87
Number
notation which is
more
convenient than the matrix notation,
from
(iii) above. The matrices I and i correspond to the
points (1, 0), (0, 1) of the plane and thus form a basis for C. More explicitly
and
we
get
our cue
eD-c.;K-i)
al + bi
=
(1.44)
identify the real number 1 with the identity matrix I; and more
generally the real number r with the complex number rl + Oi, then we can
In fact, the complex
say that every real number is also a complex number.
with
a
root
of
1 tacked on.
the
real
number
is
system
square
just
system
of
that
Arabian
circle
back
to
the
us
full
takes
original
conception
(This
adventurer. The difference here is that we now know what we mean by this
procedure and that it produces no inconsistencies.)
Thus, we can suppress the identity matrix in the expression and write a
complex number in the form a + bi. We now recapitulate the relevant facts.
If
we
C is the set of all 2
x
2 matrices
c
=
a
+ bi with a, b real numbers,
a
is
Re c, and b is the imaginary part, written
the real part of c, written a
lmc. And these following rules hold:
b
=
=
i2=-\
(a
+
bi)
(a
+
(c
+
di)
bi)(c
+
di)
(a
bi)-1^
+
=
=
(a
ac
+
Polar Form
(b
+
+
(bd
'
d)i
+
ac)i
whena2
b2
+
Z>2/0
of a Complex Number
Since C is in one-to-one
numbers
a2
+
bd +
-
U~
+
c)
by points
plex
numbers is the
seek
a
correspondence with R2, we can represent complex
plane (see Figure 1.14). Addition of com
in the
same
addition of vectors in the plane. We
of multiplication of complex numbers.
as
geometric description
this purpose it is convenient to
Let
Definition 14.
distance from the
\z\=(x2
+
z
origin:
y2y
=
x
move
+
yi.
to
now
For
polar coordinates.
The modulus of z, written
\z\, is
its
Linear Functions
1.
88
Z
=
x
+
iy
u
\argz
s'
i
x
0
Figure 1.J4
The
z, written arg z, is defined for
argument of
the ray
which
on
z
arg
=
z
tan
angle defining
x
write
can
=
z
= 0; it is the
-
complex numbers
r
coordinates (r, 9) then, since x
We
z
lies :
r(cos
=
9 +
sin
j
in
polar
cos
9,
y
form : If
=
r
sin 9,
a
=
we
x
+
yi has the polar
have
9)
have moved the i in front of sin 9 for the obvious notational con
venience which results.) The set of points of modulus 1 is the unit circle
(We
It is the set of all points of the form cos 9 + i sin 9.
this to cis 9. Precisely, cis 9 is the point of
abbreviate
We shall sometimes
on
the
unit
circle
the
Now, let z, w be two complex
ray of angle 9.
lying
centered at the
origin.
numbers,
z
=
r
cis 9
w
=
p cis (b
Then
zw
=
=
=
=
Thus
we
modulii and
r cis(9)p cis((j>)
(r cos 9 + ir sin 9)(p cos <b + ip sin <b)
rp(cos 9 cos (f> sin 9 sin <b) + irp(cos
rp cis(0 + <b)
form the
product
of two
complex
9 sin <b +
numbers
adding the arguments. (This does not make
numbers is zero, but that case is trivial anyway.)
cos
$
sin
9)
by multiplying
sense
if
one
the
of the
1.8
Notice
z"
then, if z
=
=
p cis
9, then z2
p2
=
Complex Numbers
cis 29 and
more
generally
p" cis n9
(1.45)
This observation leads to the fact that it is easy to extract roots.
converse of (1.45) is
zllk
p1* cis
=
k
and k
=
Write
rcisO
=
in
c
c
=
polar form:
zk=pk
Thus the modulus of
z
For the
j
Proposition 25. Let c be a complex number,
precisely k distinct solutions to the equation Xk
Proof.
89
cis
c
r
=
cis fl.
If
z
an
integer.
There
are
c.
=
p cis
is
</>
a
solution, then
<
is the kth root of the modulus of c, and the argument of a
an angle, but so is
angle such that k times it is fl. Well, (l/)fl is such
(l//fc)(fl + 2tt). In fact, each of the angles
is
an
1 (fl), i (fl + 2tt), 1 (fl + 4tt),
.
.
.
have the property that fc times it is fl.
has precisely these k roots :
r1/k
cis
fl, r1'*
cis
.
,
.
.
,
r1'*
,
i (fl + 2(A
All these
1)tt)
-
angles
are
distinct,
so c
=
r
cis fl
cis
k
k
Complex Eigenvalues
We shall work
with the
extensively
complex
number system in this text.
shall discover many situations besides the algebraic one above
where study within the system of complex numbers is beneficial. In par
In
fact,
we
ticular, let
us
return to the
eigenvalue problems
of the
preceding
section.
We can define
space of -tuples of complex numbers.
linear transformations on C just as we did on R". In fact, the entire theory
We consider
of linear
Ej
C, the
algebra through
=
(0, 0,
.
.
.
,
Section 1.7 holds
0, 1, 0,
.
.
.
,
0)
(1
over
C
as
in the rth
well
place)
as
R".
Let
90
/.
Linear Functions
be the standard basis vectors for C.
C" is
given by
a
matrix A
=
(a/)
Again,
any linear transformation
on
of complex numbers relative to the standard
basis :
T(zi,...,zn)=(i:Jv)...,i^v)
for all
(z1,
.
.
.
,
z")
C.
e
Examples
37. Consider the matrix
"(-? J)
as
representing
dard basis.
Its
det(A fl) f2
Eigenvalue i:
=
a
transformation T
eigenvalues
+
1,
so
are
the roots
on
C2 relative
the roots of
are
det(A
to
fl)
the stan
0.
But
fl is
given
=
i, i.
-(:: ')
The second
row
is
-
i times the
the
first,
so
the kernel of A
by
single equation ix + y 0. An eigenvector is (1, i).
The eigenvalue i has the eigenvector (1, 0- Now, F:
(1,
0}
=
are a
basis of
eigenvectors for C2,
relative to this basis :
Tf
"(i -)
if
-CM-,-)
then
CM-,)
so
T becomes
{(1,0,
diagonalized
1.8
Complex Numbers
91
38. Consider the matrix
H? i 1)
representing a transformation T
f + 1.
det(A- fl)= -f3 + f2
-
1 i,
Since the roots
i.
,
ing eigenvector, there is
a
are
on
to the standard basis.
has
polynomial
distinct and each must have
basis of
a
C3 relative
This
We
eigenvectors.
a
the
roots
correspond
now
find such
basis.
1
Eigenvalue
A-I=(\
:
i
-1
2
1
|)
0/
I is found
The kernel of A
Such
(recall Example 18).
C!
2C2
-
3C3
-
=
(1, -2, -3)
Eigenvalue /:
A-iI=U
\ 2
[o\0 i
a
row
with
eigenvalue
1.
//
-
relation among the columns
reduction is
we
must
row
reduce-
-i + iM
0
0
/
A solution of the
taking z3
(_3 + 4/,
we
linear relation among the columns
I
1
1
In order to find
eigenvector
an
-i-
The result of
as a
relation is
0
is
Thus
a
=
l
find the
corresponding homogeneous system
5, then
-
we
obtain z2
=
1
-
3i,
zt
=
is found
-3 + 4i.
by
Thus,
eigenvector with eigenvalue i. Similarly,
to the
eigenvector (-3 4i, 1 + 3i, 5) corresponding
3/, 5) is
an
-
92
/.
Linear Functions
i.
eigenvalue
Thus,
T is
represented by
relative to the basis
(1, -2, -3), (-3
+
4i, 1
3i, 5), (-3
-
-
4i, 1
+
3/, 5).
EXERCISES
42. Find the inverse of these
(a)
(b)
(c)
5-3/
(d)
(e)
(1-0/2
complex numbers:
4cis(2/3)
cis 7
3 + /
43. Show that z_1
44. Show that the
z if and only if z is on the unit circle.
complex number cis fl represents rotation in the plane
=
through the angle fl, when considered
45. Find all Mi roots of
as a
2
x
2 matrix.
z:
(a) k 2,z=-i.
(d) k 3,z i.
-l
(b) k 5,z
(e) k 2,z 3i-4
15 + 5/
(c) k=4,z l+i
(f) yt 3,z
46. Find, if possible, a basis of possibly complex eigenvectors for the
transformations represented by these matrices
(a)
=
=
=
=
=
=
=
=
=
=
G
O)
I
i
-i
0
0
PROBLEMS
45.
Compute that the matrix
(-!!)
G"J)
has square
so
equal
to I.
We have chosen i to be the 2
x
2 matrix
that the correspondence between complex numbers and
operations on
More precisely, we conceive a complex number in two
R2 will be correct.
ways: as a certain transformation on the plane, and as a vector on the plane.
Given two complex numbers z, w we may interpret their
product in two
1.9
Space Geometry
93
composition of the transformations, or the application of the transfor
corresponding to z to the vector vv. We would like these two
interpretations to have the same result. If z a + ib, w
c + id, show
ways :
mation
=
=
that zw,
c -% 1
and
all the
are
same
under that correspondence.
46. Show that the complex numbers z, z, when considered
R2 are independent (unless they are real or pure imaginary).
Why do the complex eigenvalues of
47.
real matrix
a
as
vectors in
in
conjugate
come
pairs?
1.9
Space Geometry
In this section
shall introduce the basic notions of three-dimensional
we
vector notation.
geometry, using
First of all,
as
in the
plane,
we
select
a
in space, called the origin and denoted 0. That being done
we may refer to the points of space as vectors and think heuristically of the
directed line segment from the origin to the point as a vector. The operations
particular point
and addition
be defined
of scalar
multiplication
expressed
in terms of coordinates in much the
can
same
as on
the
plane
and
way:
(i) If P is a vector and r a real number, rP is the vector lying on the line
through 0 and P and of distance from 0 equal to |r| times the length of the
If
segment OP.
r >
0, rP lies
on
the
same
opposite side.
(ii) if P, Q are two vectors in space,
in the plane determined by P and Q,
side of 0
as
P; if
r
<
0,
rP lies
on
the
We define P +
Q
there is
a
unique parallelogram lying
three of whose vertices
are
0, P, Q.
to be the fourth vertex.
the coordinatization of space.
Having chosen a point
new points with the property that 0, Eu E2
three
be
origin,
El5 E2 E3
E3 do not all lie on the same plane (we say the vectors E1; E2 E3 are not
coplanar). The three lines determined by the vectors E1; E2 E3 are called
Now,
we
turn to
let
as
,
,
,
,
the coordinate
Ej, E2 E3
,
axes.
enables
us
Just
to
in two dimensions the choices of the vectors
each line in one-to-one correspondence with the
as
put
real numbers.
In three dimensions two lines determine
a
plane.
We shall call the
planes
94
1.
Linear Functions
Figure
1.15
through 0 determined by Et and E2 the 1-2 plane, by Et and E3 the 1-3 plane,
and the plane determined by E2 and E3 is the 2-3 plane (Figure 1.15). These
three planes are called the coordinate planes. Each of these planes can be put
into one-to-one correspondence with R2 just as in the case of two dimensions.
Now, to each point in space we can associate a triple of numbers relative to
these choices in the following way. Let P be any such point. There is a
unique plane through P which is parallel to the 2-3 plane; and this plane
intersects the 1 axis in a unique point. This point has the coordinate x1
relative to the scale determined by E^ We shall call x1 the first coordinate
of P. The second, x2, is found in the same way : by intersecting the plane
through P and parallel to the 1-3 plane with the 2 axis. Finally we find
the third coordinate x3 similarly, and associate the triple (x1, x2, x3) to P.
In this way we put all of space into one-to-one correspondence with R3,
dependent upon the choice of vectors Els E2 E3 called a basis for space.
The expression in terms of coordinates of the operations of addition and
scalar multiplication are precisely the same as in R2 (no matter what basis
is chosen) :
,
r(x\ x2, x3)
(x1, x2, x3)
+
(y1, y2, y3)
=
=
,
(rx1, rx2, rx3)
(x1
+
y\ x2
+
v2, x3
+
y3)
1.9
Space Geometry
95
need to check that these formulas correspond to the geometric
descriptions given above; we need only refer to the computation in the plane.
When we are interested in the pictorial representation of problems of
There is
no
three-dimensional Eculidean geometry it is best if we consistently use a
particular coordinatization. For this purpose we select the "right-handed
coordinate system"; where the coordinate axes are mutually
perpendicular and the order 12 3 is that of a right-handed screw (see
rectangular
Figure 1.16). It is common in particular problems to refer to the coordinates
by the letters (x, y, z) rather than (x1, x2, x3). We shall use the numbered
coordinates when it is
more
convenient to do
so.
Inner Product
Now, the basic notions of Euclidean geometry
will be of
ordinates.
and the
importance
to us to derive
expressions
are
length
and
angle.
for these in terms of
It
co
Consider first the length of the line segment OP between the origin
P with coordinates (x, y, z). This can be easily computed
point
by use of the Pythagorean theorem (consult Figure 1.17).
point of intersection with the xz plane of the line through
Then OPP' is a right triangle, so
to the v axis.
|0P|2= |0P'|2+ |P'P|2
Figure
1.16
Let P' be the
P and
parallel
96
Linear Functions
/.
>
P{x,y,z)
.si,-
Figure 1.17
Letting P" be the point of intersection with
and parallel to the z axis, we obtain
|0P|2
But
now
|0P"|2
=
|0P"|2
+
|P"P'|2
x2, |P"P'|2
=
v2
+
Now, suppose P(x,
y,
|0P|
[x2
=
+
+
=
the
x
axis of the line
|P'P|2
z2, |P'P|2
=
y2,
so
z2]1/2
z), Q(a, b, c)
any two
are
definition of addition, P is the fourth vertex of the
whose vertices are 0, P
Q and Q. Thus, the side
-
the
same
length
IPQI
=
the side
as
l(P
-
Q)0|
through
=
[(*
-
Q and 0,
P
a)2
+
(y
b)2
-
Finally, we can compute the angle between
(consult Figure 1.18); if 9 is that angle, then
|PQ|2
In
=
|0P|2
+
through P'
|0Q|2
-
2|0P| |0Q|
+
so
(z- c)2]1/2
P and
cos
points in space. By
parallelogram three of
through P and Q has
Q by the law of cosines
9
coordinates,
(x
-
a)2
+
(y- b)2
+
(z- c)2
=
x2
(1.46)
+ y2 + z2 + a2 + b2 + c2
-2(x2 + y2+z2)112
x(a2 + b2 + c2)1/2cos9
1.9
97
Space Geometry
which reduces to
xa
COS0
+
=
(x2
y2
+
+
yb
+
z2)ll2(a2
cz
+
The form in the numerator thus has
with the notion of length determines
b2
+
(1.47)
c2)112
some special importance: it together
angles. It is called the inner product
of the two vectors P, Q.
Let P,
Definition 15.
Q be
two vectors in space.
Their Euclidean inner
product, denoted <P, Q>, is defined as |P| |Q| cos 9, where 9 is the angle
between P and Q. In coordinates, P
(xu yu Zj), Q (x2 y2 z2),
=
=
,
(P,Q}
Propositions
i/<p,Q> o.
=
,
x1x2 + y^2 + zxz2
26.
The
nonzero vectors
P and
Q
are
perpendicular if and only
=
Proof. P and Q are perpendicular if and only if the angle fl between them is a
right angle, fl is a right angle if and only if cos fl 0, and this holds precisely
0.
When <P, Q>
=
=
A
plane through
vector
the
perpendicular
origin
to such a
n=<x.N>=o
is the linear span of two vectors. If N is
plane n> then 11 is iven by the ecluat'on
a
98
1.
More
Linear Functions
generally, if p is a point on a plane (not necessarily through
orthogonal to FJ, F] is given by the equation
the
origin)
and N is
<x
-
N>
p,
0
=
through the origin is the linear span of a single vector, and can be
expressed by two linear equations (since a line is the intersection of two
planes).
A line
Examples
39. Find the
equation of the plane through (1, 2, 0) spanned by the
(1, 0, 1), (3, 1, 2). If N (n1, n2, n3) perpendicular to
two vectors
this
plane
=
must have
we
<N,(1,0, l)>
<N, (3, 0, 2)>
A
(1
=
of
1,
,
=
=
solution
N
n1+3
3k1 +n2
=
this
0
2n3=0
+
is (1,-1,-1), so
equation of the plane is
system
Then the
1).
<x- (1,2,0), (1, -1, -1)>
=
0
or
<N,
P
-
Letting
R
N
=
0
=
<N, Q
(m1, n2, n3),
=
take
equation of the plane through P (1, 0,
1), Q
(3, 1, 1). If N is perpendicular to the plane, we have
=
R>
may
x-y-z+l=0
40. Find the
(2, 2, 2),
we
-
we
R>
=
-
=
0
obtain this system of
equations:
-2n1-n2-2n3=0
n1
-
+
n2
which has
<x
-
Ix +
N, R>
4y
+
n3 =0
a
solution N
=
+ 3z
0,
=
41. Find the
or
l(x
=
-
(7, 4. 3).
3) + 4(y
Thus the
-
1)
+
3(z
-
equation we
1) 0 or
28
equations of the line L through (4, 0, 0) and perpen
Example 40. If x is on L we must have
dicular to the plane in
<x
-
(4, 0, 0),
P
seek is
=
-
R>
=
0
<x
-
(4, 0, 0), Q
-
R>
=
0
1.9
may take these
so we
2x + y + 2z
+ y +
-x
z
as
the
99
Space Geometry
equations :
8
=
-4
=
Vector Product
Given two noncollinear vectors vx, v2 in space, the set of vectors perpen
a line.
We shall now develop a useful formula for select
dicular to v1; v2 is
ing
a
If N is
vector on that
particular
on
that line, and
<x, N>
=
x
for all
x e
Vl
uniquely
a
R3.
=
vector
,
product
we
vt
have
0
On the other hand, since x, vx, v2
Now there is
line, called the
is in the linear span of vt and v2
This is
are
coplanar we
have
determined vector N such that
easily
(V, vt2, vt3),
seen
y2
=
using coordinates.
(v2\ v22, v23),
x
=
Write
(x1, x2, x3)
Then
det(
Vj
)
=
W
=
2
vt3
v^
.,2
.,3
vx2
22
detj V
1
V2/
x3
x2
Ix1
(x \
1
XKV.W V3V22) *W23 lW)
+ x3(v11v22-vl2v21)
(ix1, X2, X3), ((Vl2V23 1-! V), (V.W V^W),
-
~
-
-
-
=
(v.W-v.W))}
x
v2
.
100
1.
Linear Functions
Definition 16.
The vector
V x w
Proposition
(i)
<x,
(ii)
(iii)
(iv)
Let
product
=
v
=
v x w
(v2w3
(v1, v2, v3),
is defined
v3w2, v3w1
-
w
(w1, w2, w3)
=
be two vectors in R3.
by
-
v1w3, v1w2
-
v2wx)
27.
v x
v x w
w>
det
=
vl
forallxeR3.
w x v.
=
orthogonal to v and w.
The equation of the plane through
the equation <x, v x w>
0.
is
v x w
the origin
spanned by
v
and
w
has
=
The
proof
discussion.
for
example,
<u,
of this proposition is completely contained in the
preceding
The basic property of the vector product is the first; it follows,
that for any three vectors u, v, w
w>
v x
Notice that if v,
ordered basis u
proposition gives
of
=
<u
x
v,
w>
=
collinear,
<v,
w x
u>
=
<v
x
w,
u>
0. If they are not collinear, the
right handed (see Figure 1.19). The following
important geometric interpretation of the magnitude
w are
v x w
=
vxwis
v
an
v x w.
Proposition
(i)
The
28.
area
Let u, v,
w
be three noncollinear vectors.
of the parallelogram spanned by
wx v
Figure
1.19
u
and
v
is
||u
x
v||.
1.9
(ii)
The volume of the
|detQ
parallelepiped spanned by
J
v
This follows
u, v, and
w
101
is
'
Proof.
Step (a). The first step is to verify (ii) in case
perpendicular. In that case we must show that
det/
Space Geometry
||u||
=
the vectors u, v,
w are
mutually
||v||- ||w||
-
easily from the multiplicative property of the determinant.
First
we
note that
/u\
/<u,u>
(u,
v
v,
w)
=
\w/
\
0
0
0
<v, v>
0
0
0
\
<w,w>/
(i,j)th entry is the inner product of the rth
of the second (see Problem 49). Thus,
since the
;th
row
detjv j =detjv )(u,
Step (b). In particular,
perpendicular, so
v,
if u,
w)
=
v are
llu
so
||u
x
v||
=
||u||
x
of the first matrix with the
||u||2||v||2||w||2
perpendicular,
(u
=
row
x
v\
u
J
then u, v,
u x v are
mutually
||u||- ||v||
vll
||v||, when
is
perpendicular
part (i) in general.
u
to
v.
angle between
prove
the
of
by u and
area
the
spanned
Then
parallelogram
1.20).
(see Figure
is the product of the base and the height of the base:
Step (c).
and
Now
we
v
area
=
||u||a
=
||u||
||v||sinfl
Let fl be the the
u
v
102
/.
Linear Functions
Now the vector
by
u
and
u
<v,
=cosg-fl)
sin
Since
(u x v) is orthogonal to u and
orthogonal to u. We have
u x
and is
v
and
u x
HTll
l|U
(u
X
X
orthogonal, by Step (b)
u x v are
v,
so
lies in the plane spanned
v)>
x
(U
u x
T)||
we
have ||u
x
(u
x
v) ||
=
||u ||
||u
x
v||
Thus
area
=
<v,
||u
Nu
||r||
=
<U
(u
u x
X V, U X
v>
l|u|iVrrruxy ^-
v)>
x
(u
x
x
v)||
lluxv||
=
u
The volume of the parallele
Step (d). To prove part (ii) we refer to Figure 1 .21
piped spanned by u, v, w is the product of the area of the base and the altitude:
.
volume
Since
u X t
is
,
sin
=
||u
x
y\\b
orthogonal
<6=C0S
(tt
-
\2
=
||u
x
to the u,
\
4>\
7
||w||sin <j>
v||
v
|<w,
=--!
plane,
u>
xv>|
_
llwll-Hu xvll
u
X(u
Figure 1.20
X
v)
1.9
Space Geometry
x V
Figure
1.21
Thus
..
volume
A final
=
|lu
equality
x
v|
<w,uxv>
l|w||-n-T7]
w
uxt
-=
det
/w\
u
/
which will prove useful is this :
||uxv||2=||u||2||v||2-<u,v>2
This follows
easily
from the above arguments :
||uxv||2=||u||2||v||2sin24>
||u||2||v||2(l-cos20)
=
=
since the
angle
llu||2|
between
<u, v>2
u
and
v
is (b.
EXERCISES
47. Which
vx=(2, 1,2),
t5= (1,3,0),
pairs of the following
vectors
are
orthogonal?
t,=(3, -1,4), v3=(7,0,5),
v6= (0,0,1),
tt=(-15,5,21)
v*
=
(6, -2,5)
104
1.
Linear Functions
48. Find the vector
products
v, x V/
for all
pairs of
vectors
given in
Exercise 47.
49. Find
vector
a
<v,v,>=2
<v,v2>
where vi, v2 v3
50. Find the
such that
=
-l
<v,v3>=7
given in Exercise 47.
equation of the plane spanned by the
(c) t5, v6, (d) v2, t4, where the v,
,
(b)
v
v2, v5,
are
(a) Vi, y6
given in Exer
vectors
are
,
cise 47.
51. Find the
of the line
equation
spanned by the
vectors
given in
Exercise 47.
52. Find the equation of the plane
through (3, 2, 1) and orthogonal
(7, 1, 2).
53. Find the equation of the line through (0, 2, 0) and
orthogonal to the
plane spanned by (1, -1, 1) and (0, 3, 1).
54. Find the equation of the line through the origin and
perpendicular
to the plane through the points
(a) E,,E2,E3
(b) (1,1, 5), (0,0, 2), (-1,-1,0)
(c) (0, 0, 0), (0, 0, 1), (0, 1, 0)
55. Find the equation of the line of intersection of the two
planes,
(a) determined by (a), (b) of Exercise 54.
(b) determined by (a), (c) of Exercise 54.
(c) determined by (b), (c) of Exercise 54.
56. Find the plane of vectors perpendicular to each of the lines deter
to the vector
mined in Exercise 55.
57. Let A be
(a)
if the
tions of Ax
a
a
3
=
3 matrix.
x
rows
of A lie
b forms
a
line,
(b) if the rows of A lie
plane, or is empty.
58. Show that ||v x w||
the two vectors v and w.
59. Is the vector product
=
(u
x
v)
x w
=
u x
(v
x
Show that
on a
or
on a
||v||
plane (but
not on a line), the set of solu
is empty.
line, the set of solutions of Ax b forms
=
||w|| sin fl, where fl is the angle between
associative; that is, is
w)
always true?
60. If v,
v, v x w,
w are
v x
(v
X
two noncollinear vectors show that the three vectors
w)
are
pairwise orthogonal.
PROBLEMS
48. Prove the identities of Proposition 27.
49. Let vj, v2, v3 be three vectors in R3.
Let A be the matrix whose
1.10
rows are
Vi, v2
(a) the (i,/)th entry
(b) det
A
50. Let P be
=
of Linearity
and B the matrix with columns Vi,
of AB is <v, Vj>.
v3
,
Abstract Notions
v2
v3
,
.
105
Show that
,
det B.
given by
the coordinates
(x, y, z) relative to a choice Ei, E2
Show that the point of intersection of the line
through P parallel to the Ei axis with the 2-3 plane has coordinates (0, y, z).
,
E3 of basis for space.
Abstract Notions of
1.10
There
with
a
are
Linearity
many collections of mathematical objects which are endowed
algebraic structure which is very reminiscent of R". To be less
natural
vague, there is defined, within these collections, the operations of addition
and multiplication by real numbers. Furthermore, the problems that
naturally arise in these other contexts are reminiscent of the problems on R"
which we have been studying. The question to ask then, is this: does the
same theory hold, and will the same techniques work in this more general
We shall see in this section that for a large class of such objects
context ?
(the finite-dimensional vector spaces) the theory is the same. We shall see
later on that in many other cases, the techniques we have developed can be
modified to provide solutions to problems in the more general context.
First, let us consider some examples.
Examples
42.
[0, 1],
If/ and
then
g
are
we can
continuous real-valued functions on the interval
cf as follows:
define the functions /+ g,
(f+g)(x)=f(x) + g(x)
(cf)(x) cf(x)
=
Clearly,/ +
g and
c/are
of addition and scalar
C([0, 1])
also continuous.
multiplication
of all continuous functions
Thus
are
on
that
we see
defined
on
the interval
operations
the collection
[0, 1].
example, if/ and g are differentiable, so are/+ g
and cf. Thus the space ^([O, 1]) of functions on the interval [0, 1]
with continuous derivatives also has the operations of addition and
scalar multiplication. Notice that the operation of differentiation
takes functions in CHC0, 1]) into C([0, 1]): if /is in CH[0, 1]) it
has a continuous derivative, so /' is in C([0, 1]). Furthermore,
43. In the above
106
Linear Functions
/.
differentiation could be described
as a
linear transformation:
if+g)'=f'+3'
(cf)' cf
=
is, by the
So
way,
integration
a
linear transformation:
\if+g)=lf+U
l(cf)
ctf
=
The fundamental theorem of calculus says that differentiation is the
inverse operation for integration:
(J/)' =/
merely a curious way of describing the
implied point of view has led to a wide
The subject of functional analysis which
range of mathematical discoveries.
20th
in
the
was developed early
century came out of this geometric-algebraic
to
long standing problems of analysis.
approach
These remarks may strike you
well-known phenomena, but the
as
Examples
44. If S and T
are
linear transformations of R" to Rm, then
so
is
the function S + T defined by:
(S
+
We
T)(x)
also
can
(cS)(x)
=
=
S(x)
+
T(x)
multiply
a
linear transformation
by
a
scalar:
cS(x)
Thus the space L(R", Rm) of linear transformations of R" to Rm has
defined on it operations of addition and scalar multiplication.
45. We have
of
n x n
In
fact,
we
used, in
M" this way it
These
already observed (Section 1.6) that the collection M"
on it these two important operations.
matrices has defined
was
an
just
essential way, the fact that when
the
same as
examples, together with R", lead
we
viewed
R"2.
to the notion of
an
space : a set together with the operations of addition and scalar
We include in the definition the algebraic laws governing these
abstract vector
multiplication.
operations.
1.10
Definition 17.
Abstract Notions of Linearity
107
An abstract vector space is a set V with a
distinguished
on which are defined two
operations:
and vv are elements of V, then v + vv is a well-defined element
element, 0, called the origin,
Addition.
If v
of V.
Scalar
multiplication. If v is in V and c is a real number, cv is a well-defined
These operations must behave in accordance with these laws :
element of V.
(i)
(ii)
(iii)
(iv)
(v)
(vi)
The
of the
v
+
(vv
v
+
vv
t>
+ 0
+
=
=
c(v w)
ct(c2 vv)
+
\w
x)
(v
=
vv)
+
+ x,
+ v,
vv
v,
cv
=
=
+ cw,
(Ci c2)w,
vv.
=
preceding examples are all abstract
required laws are easily performed.
vector spaces; the verifications
We
now
want to
investigate
the
extent to which the ideas and facts discussed in the case of R" carry over to
abstract vector spaces. First of all, all the definitions carry over sensibly
to the abstract
case
vector space V.
case :
if
Thus
we
just replace
we
the word R"
take these notions
as
by
the words
an
abstract
defined also in the abstract
linear transformation, linear subspace, span, independent, basis, dimension.
one bit of amplification necessary in the case of dimension.
Now there is
We have until
now
encountered spaces of only finite dimension.
Example
46. Let R
an
be the collection of all sequences of real numbers.
an ordered oo-tuple,
Thus
element of I?00 is
(x1,*2,. ..,*",...)
R00 is
an
abstract vector space with these
operations:
(x1,x2,...,x,...) + (y1,y2,...,yn,...)
x" + y",
.)
(x1 + y1, x2 + y2,
c(x\ x2,...,xn,...) (ex1, ex2, ...,cx",...)
=
.
.
.
,
.
.
=
be the
has an infinite set of independent vectors. Let
This
are zero but for the nth, which is 1
entries
whose
all
of
sequence
entire collection {Eu ...,,.. .} is an independent set. For if
there is a relation among some finite subset of these, it must be of the
Now R
.
form
c1^
+
+
ckEk
=
0
108
Linear Functions
1.
(of course,
many of the c's may be
c1El
+
so
+
ckEk
if this vector is
indeed the set
We
=
must have
{Eu ...,,...}
make the
now
(c1, c2,..., ck, 0, 0,
zero we
following
shall
vector space; and
we
holds also in this
more
is
an
general
.
.
c1
.)
=
c2
infinite
=
=
ck
independent
=
0.
Thus
set on/?.
restriction to the so-called finite-dimensional
that all of the
see
But
zero).
preceding
information about R"
case.
Definition 18. A vector space V is finite dimensional if there is a finite set
of vectors vl,...,vk which span V. That Rx is not finite dimensional
follows from some of the observations to be made below. It can also be
verified in the terms of the above definition
(see Problem 53). The important
they are no different from
result about finite-dimensional vector spaces is that
the spaces R".
Proposition
Let V be
29.
a
finite-dimensional
basis for V, every vector in V
Ifvt,...,vdisa
combination of vu
v
=
x1v1
(x1, ...,xd)
The
.
.
.
,
+
xdvd
is called the coordinate of
correspondence
be
vd:
+
v
of dimension d.
expressed uniquely as a
vector space
can
(x1,
.
..,
xd)
is
v
a
relative to the basis vlt...,vd.
one-to-one linear transformation
of V onto R*.
Proof.
The definition of basis
(Definition 6) makes this proposition quite clear.
(Problem 54).
We leave the verifications to the reader
What is not
so clear is that every finite-dimensional vector
space has a
and
that
basis,
every basis has the same number of elements.
However,
once these facts are established the above
proposition serves to reduce the
general finite-dimensional space
1.3 through 1.6 carry over.
Proposition 30.
to one of the
R", and the results of Section
Every finite-dimensional vector space V has a finite basis,
same number of elements, the dimension
of V.
and every basis has the
Proof. Suppose V is finite dimensional. Then V has a finite spanning set. Let
{vi,
vd} be a spanning set with the minimal number of vectors; by definition V
has dimension d. We shall show that {vi,
vd} is a basis.
...,
...,
Abstract Notions of Linearity
1.10
Since {vu
vd}
span, every vector in V can be written as a linear combination
We have to show that there is only one way in which this can be
...,
of these vectors.
done.
Suppose for
some
x'Vi H
=
109
vector
h x*vd
=
have two different such ways :
v we
yhi
+ +
y"vd
(1 .48)
Then
(x>
y>i +
-
Vj-i, fJ+i,
equation
.
.
.
yd)vd
.
.
.
=
these elements
y' for
some
Thus
j.
Thus it must be
d.
That any two bases have the
same
Hence
{vL,
.
.
.
...,
But this contra
to span all of V also.
serve
in two different ways.
vd
,
0
must have xJ #
we
assumption about
dicts the minimal
in terms of Vi,
-
says that vj is in the linear span of the a- 1 elements vit
so
vd,
,
(xd
expressions differ
Since these two
Now this
+
impossible to express
vd} is a basis.
v
,
number of elements follows
easily
from
Proposition 28 (see also Problem 55). Let T: V^>Rd be the linear transfor
mation associating to each vector its coordinate relative to the above basis
{vu...,vd}. If {wu...,wd} is another basis, let S: K1R be the same
S T~ is a one-to-one
coordinate mapping relative to this basis. Then L
0. Thus (rank + nullity
linear mapping of Rd onto R3, so p(L)
5, v(L)
=
dimension) :
5
=
=
=
=
d.
PROBLEMS
a
{vi, ...,vk] in
51. Show that for any finite set of vectors S
vector weR which does not lie in their linear span [S].
v'
represent the first (k + l)-tuple of entries in
R", there is
=
span
there is
Rk+1,
a
.
.
Since v/,
.
vector w' in Rk+l which cannot be written
nation of v/,
vt'. Let
52. Are the vectors E,,
.
v.
,
.
w
.
.
=
,
(w', 0,
E
,
.
.
.
.
.
.
(Hint:
.
v/
,
Let
cannot
combin-
as a
.).)
in R described in
Example
43
a
basis
for Ra ?
53. Let Ro" be the collection of those sequences of real numbers
x",
.) such that x" 0 for all but finitely many n. Then R0^
(x1, x2,
are a
E
is a linear subspace of R". Show that the vectors Ei
=
.
.
.
.
,
.
,
.
.
.
,
,
.
.
.
basis for i?0.
54. Prove
of
a
Proposition
29.
that any two bases
by following the arguments in Section 1.4,
finite-dimensional vector space have the same number of elements.
55. Prove,
110
1.
Linear Functions
Show that the collection L(V,
56. Let V, W be two vector spaces.
linear transformations from V to W is
a
W) of
vector space under the two opera
tions :
(a)
(b)
if
e L( V, W), (cL)(x)
cL(x),
L(V, W), (L + L')(x) L(x) + L'(x).
R, L
c g
if L, L'
6
=
=
57. What is the dimension of L(R",
58. Show that
a
Rm)l
vector space V is finite dimensional if there is
a
one-to-one
linear transformation of F0 into J?" for some n.
59. Show that a vector space F is finite dimensional if there is
a
linear
transformation T of R" onto V for some n.
60. Verify that the collection P of polynomials is an abstract vector space.
For a positive integer n, let P be the collection of polynomials of degree not
Show that P is a linear subspace of P. Show that P is not
more than n.
finite dimensional, whereas P is. What is the dimension of P1
61 Let x0
c another collec
x be distinct real numbers and c0
.
,
.
.
.
,
,
Show that there is
tion of real numbers.
one
.
.
.
and only
,
one
polynomial p in
P such that
p(xt)
=
(Hint:
ct
0<Li<,n
Let L:P
that L has rank
62. Let g be
G(p)
n
J?"+1 be defined
(p(x0)
a
polynomial, and define the function
a
linear function.
63. Define Dk:P^-P:
of A?
64. Let x0
and
only
\
p(x0) =d0
(Hint:
e
Dk(p)
R, and let
one
=
p(x)).
G: P
Show
P:
Describe the range and kernel of G.
What are the range and kernel
dfy/dx!'.
c0
ck be given numbers.
polynomial p in P such that
,
dp
.
.
,
Use the
same
Show that there is
dkp
d~t(-X)=Cl'""dxk ^
idea
as
=
C"
in Exercise
65. Does Dk:P-+P have any
66. Show that C([0, 1]) is not
1.11
=
=pg
Show that G is
one
by L(p)
+ 1 .)
61.)
eigenvalues?
a
finite-dimensional vector space.
Inner Products
The notion of
length, or distance, is important in the geometric study of
spatial configurations. In Section 1.3 we studied these concepts
and related them to an algebraic concept, the inner product. From the
point of view of analysis also it is true that these concepts are significant:
planar
and
1.11
it is in terms of distance that
"convergence."
(vv1,
.
.
.
By analogy
express "closeness" and in particular
with R3 we define the inner product in R",
,
vv"),
<v, w>
product of
by <v, w> is defined
We shall say that v is orthogonal to
between v and w is defined by
\y\
=
v
than the
v
=
(v1,
...,
v"),
w
=
as
w
if <v,
w>
=
The distance
0.
d(v, w)
v
and 0,
d(y,0)=V(vt)2-l1<2
d(y, w)2
<v
two vectors
is the distance between
Distance in R" behaves much
Pythagorean theorem holds:
when
shall, in this section,
we
{X(vi-wi)2-\1'2
=
|v| of a vector
The modulus
here
v'w'
=
d(y,yv)
we are
The inner
denoted
111
we can
and in terms of it, distance. While
introduce some topological terms.
Definition 19.
Inner Products
w,
sum
=
d(y, x)2
+
as
it does in R2 and
d(y, x)
+
points
are no
further apart
(1 .50)
d(x,yv)
These facts will be verified in the
the
particular,
(1 .49)
=
<
in
d(x, w)2
w
x> 0. In any event, two
of the distances from a third,
d(v, w)
R3;
problems.
Topological Notions
The ball in R" of radius R > 0 and center c, denoted B(c,
points whose distance from c is less than R:
Definition 20.
is the set of all
R),
B(c,R)= {xeR":d(x,c)<R}
A set S is said to be
centered at
each of its
Thus,
a
c.
a
neighborhood
of
a
point
c
if it contains
A set V is said to be open if it contains
a
points.
set S is
a
neighborhood
of
c
if there is
some
R
some
ball
neighborhood
(presumably
of
very
112
Linear Functions
1.
small)
such that
d(x, c)
R
<
implies
x e
S
A set U is open if for every c e U, there is an R such that U => B(c, R).
For suppose xeB(c,R). Then d(x, c)
that any ball is open.
R
d(x, c) > 0. Now B(c, R) contains the ball of radius R
Notice
<
-
-
centered at
For if y is
x.
d(y, c)
Here is
a
d(y, x)
^
+
point in
a
that ball, then
d(x, c)<R- d(x, c)
collection of formal
properties
+
R,
so
d(x, c)
by (1 .50),
d(y, c)
=
R
of the collection of open sets.
Proposition 31.
(i) R" is open.
n U.
(ii) IfUu..., U are open, so is Ut n
(iii) // C is any collection of open sets, then the set of all points belonging
to any of the sets in C is open.
(This set is denoted [j U).
Proof.
(i) Clearly, R" contains a ball centered at every one of its points.
Then there are Ri
(ii) Suppose Ui,..., / are open, and x is in every U,
U => B(x, if). Let R
R such that Ui => B(x, Ri),
min[.Ri,
Rn]. Then if
In
is
in
each
Thus
is
in
n
n U
is
in
each
so
Ut
Ui
y
B(x,
Ri)
d(y, x)<R,y
n U => B(x, R).
Thus Ui n
n U is a neighborhood of
particular, Ui n
any one of its points x, and is thus open.
(iii) Suppose C is a collection of open sets. If x is in any one of them, say U,
then since U is open there is an R such that U => B(x, R). Thus, \JV ec U => B(x, R).
Thus [Jvee Uisa neighborhood of any one of its points, so is open.
.
=
.
.
.
.
,
.
.
,
.
.
Many of the concepts a mathematician studies are so-called local concepts:
They happen in a neighborhood of a point, or are determined by what goes
on near a point; far behavior being irrelevant.
Differentiation is thus local,
whereas
integration is
The
importance of open sets is that it is precisely
study these local concepts, since their definition
at a point depends on behavior in some neighborhood of the point.
If a set is open its complement, the set of all points not in the given set, is
said to be closed. Thus, S is a closed subset of R" if R"
S
{x e R" : x S}
is open. Corresponding to Proposition 31 we have this proposition about
on
such sets that
we
not.
should
-
closed sets.
Proposition 32.
(i) R" is closed.
(ii) If Su
Sn
.
.
.
,
are
closed,
so
is
S^v
u
S.
=
1.11
Inner Products
(iii) IfC is a collection of closed sets, then the set of all points
the sets of C is closed. (This set is denoted [)s c S).
113
common
to
all
6
Problem 67.
Proof.
Notice that there
are
sets which are both open and closed.
There
are
R" and 0 are the only ones. There are also sets which
many of them.
neither open nor closed, and there are many of them. For example,
not
are
an
interval is open in R1 if it contains neither end point, closed if it contains
both, and neither open nor closed if it contains only one end point.
We
are
acquainted
That
with the notion of
"dropping
a
perpendicular"
in the
line and p is a point not on the line, then we can drop
from p to / as in Figure 1 .22. The point p0 of intersection
is, if / is
plane.
a perpendicular
of the perpendicular with / is the point on / which is closest to p. A more
sophisticated way of describing this situation is to say that p0 is the orthogonal
projection of p on /. The concept of orthogonal projection generalizes to R"
and will prove quite useful there. In order to discuss this problem, we shall
generalize even further.
a
A Euclidean vector space is an abstract vector space V on
which is defined a real-valued function of pairs of vectors, called the inner
product, and denoted <, >. The inner product must obey these laws:
Definition 21.
(i)
(ii)
(iii)
(iv)
(v, v} > 0. If (v, v}
<v, w>
<w, v}.
a(v, vv>.
(av, vv>
=
0, then
v
=
0.
=
=
<>!
+ v2
,
w>
=
<!>!, w>
+
<v2 w}.
,
\
\
*p
\
\
\
\
\
\
\
\
\
\
\
^
\
\
Figure 1.22
114
1.
Linear Functions
Euclidean vector space when endowed with its inner
C[0, 1 ] of continuous functions on the unit interval is a
It is clear that R" is
The space
product.
a
Euclidean vector space with this inner
</, 9>
=
(f(t)g(t)
product:
dt
We leave it to the reader to
verify that the laws (i)-(iv) are obeyed. It is
(i)-(iv) are all that is essential to the notion of inner
interesting
that
such
function behaving in accordance with those laws
;
is,
any
product
will have all the properties of an inner product. Despite the inherent interest
in this
metamathematical
point, we shall not pursue it further, but take
it for granted that the above definition has indeed abstracted the essence of
that the laws
"
"
this notion.
In terms of an inner product
length and orthogonality:
llll
v
1
on a
vector space
we can
define the notions of
[<, t>>V'2
=
if and
vv
only
if
<y, vv>
=
0
The
important bases in a Euclidean vector space are those bases whose
are mutually orthogonal.
More specifically, we shall call a set
{Elt ...,} in a Euclidean vector space V an orthonormal set if
vectors
||,||
Et
1
for all
1
=
i
for all i # /
Ej
If the vectors
Eu
E span V we shall call them an orthonormal basis.
(Any orthonormal set of vectors is independent Problem 68.) The basic
geometric fact concerning orthonormal sets is the following:
Proposition
...,
33.
Let V be
orthonormal set in V.
is
orthogonal
to
a
Euclidean vector space and
v in V, the vector
v
v0
For any vector
the linear span
=
-
{l5 ...,}
?= <t>, jE;> Et
x
Sof{E1,...,En}.
n
Proof.
Let
w
=
i
<v, w>
=
<y
2 Ci E,
=
-
be in S.
Then
i
2
'=i
<v, ,> <, w>
,
=
iv, w>
-
J
i=i
an
<t>, E,> <, w>
,
1.11
Inner Products
115
Now
< w>
<,
=
,
2 CjEj>
,
=
,
=
c,
n
2 ctE,y
<y,
<, ,>
j=i
ft
<t>, w>
2 c./
=
j=i
2
=
c,
<y, (>
1=1
1=1
Thus
<t>, w>
=
1
2
Theorem 1.8.
be
an
c,
<v, E,y
-
1
=
1
Let V be
orthonormal set in V.
v0=
2
=
a
<y, -Ei)
c,
=
o
1
Euclidean vector space, and let
{Eu ...,}
For any vector v, let
<<;,,>,
i=l
Then
\\v\\2=\\v-v0\\2+\\vQ\\2;
(i)
(ii)
for any
vv
in the linear span of
{Eu ...,},
\\v-v0\\2<\\v-yv\\2
Proof.
(i)
II" II2
=
<(y v0) + v0,(v- v0) + v0>
\\v-v0\\2+ \M\2+ <.V0,V-Vo> + <V-V0,Vo>
<v v~>
The last two terms
span of
{Ei,
-
=
,
=
...,
(ii) ||y w||2
are zero
by the preceding proposition, since
=
=
<y
<,v
llw
vv,
vv>
v
wy
Vo + v0
w,v
foil2 + \\v0-w\\2+<v-v0,Vo-w) + <Vo-w,v-v0y
v0
+
v0
Again, the last two terms are zero for both
Thus
span of {,,...,}.
v0
,
w
and thus also
||t,_vv||2= ||_0||* + ||t;o-w||2> Hf-Uoll2
(ii) is
proven.
is in the linear
E}.
=
so
v0
v0
w
is in the linear
116
1.
Linear Functions
Gram-Schmidt Process
v' is orthogonal to the linear span S of {Eu ..., }.
v0
v0 is the vector in S which is closest to v; it is called the orthogonal projection
of v into S. It seems, by Theorem 1.8 that one needs an orthonormal basis
Notice that
v
=
orthogonal projections; the following proposition gives a
obtaining orthonormal basis for finite-dimensional vector
thus with it, orthogonal projections.
in order to find
for
procedure
spaces, and
Proposition 34. Let Fu ...,F be a basis for a Euclidean
can find an orthonormal basis Eu ...,Enso that the linear
Fjfor allj.
Ej is the same as the linear span ofF1,
We
.
.
The
Proof.
NPiir'pi.
Now in
proof is by induction
general, let Fi,
.
.
.
the linear span W of P\,
apply the proposition to W
.
.
orthonormal basis with the
F be
,
.
a
on
ofEu
...,
.,
n.
If
basis for
F-L is
vector space V.
span
a
n
=
1,
we
need
only take Ei
Euclidean vector space V.
=
Then
Euclidean vector space also, and we can
the inductive hypothesis. Let Eu ..., E-i be an
,
a
by
required properties.
Now,
we must
find
a
vector
En
such that
1
11^11
(E,E,)=0
=
all/^K
Fn is in the linear span of E,
If E is
Thus
a
we
.
.
.
,
E
vector that fulfills the last two
need
En
=
only find
F-
a
vector
conditions, then we can take E
||^ \\~1E
filling the last two conditions. That is easy ; take
=
.
2(Fn,E3)Ej
ji
Then, for
i < n,
( E,)
=
,
(F Ed
-
,
2 (Fn Ej)(Ej E,)
,
,
J<n
=
(Fn E,)
,
-
(F E,)(E, ,,)= 0
,
Furthermore,
F=+
so
>Z(F-,Ej)Ej
the last two conditions
The
are
fulfilled and the proposition is proven.
proof of this proposition provides a procedure for finding orthonormal
1.11
bases in
an
Euclidean vector space, known
as
Inner Products
the Gram-Schmidt process.
111
It
goes like this:
any basis
First, pick
1
=
Fu
=
.
,
F of
V.
Take
=
Fj+i
.
.,
Ej
be the vector of
J+1
are
by
the
length
to find
found, take
(Fj+i, i)i
~
and divide
(F2, E^)Elf
-
.
and let
.
11^111-^1
Then choose 2
F2
and so forth. If Eu
E+i
.
length
(FJ+l, E2)E2
-
one
-
collinear with
-
(FJ+l, ,);
+i.
Examples
47.
F1
F2
F3
=
=
=
Apply
the Gram-Schmidt process to this basis ofR3:
(1,0,1)
(3, -1,2)
(0, 0, 1)
Take
Fx
/
J_
_1_\
El"lFj'V2*0,>/2/
E2
a_1)2)_(3.^
=
-(3. -1.2)-
+
(-i).o
+
2.ii)(-L,o,ii)
(|,0.|) (|. -1.^)
-
Then
E2
=
E3
{i) [r~1'^r)
=
(0, 0, 1)
-
=
\Wr2'~\v
'(up)
-^ (-^,0, -^) -^3 (^75.
"(7) '(W71)
+
2,
118
1.
Linear Functions
and
finally
~2
I
-
~3
2
\
\(17)1/2'(17)1/2'(17)1/2j
3
48. Find
an
X(x\ x2, x3, x4)
orthonormal basis for the kernel of X: R4
=
of all,
First
(1, 0, 0,
x1
let
+
us
1/2), (0, 1, 0,
-
Schmidt process,
we
x2
-
+
x3
+
-*
R,
2x4.
pick a suitable basis for K(X); that is,
1/2), (0, 0, 1, 1/2). Applying the Gram-
obtain
2
E1
=
=
7^ (i, o,o,^i)
\(3V)m'\6) '^(mr^J
E:
(30)1
E3
=
(1094)1/2
49. Find the
T:
orthogonal projection of (3, 1, 2) into the kernel of
R3^R:
T(x,
y,
z)
=
x
+
2y
+
z
Now the kernel of T is
Applying
Ei
=
(l)1/2(2,
Thus the
spanned by Ft
the Gram-Schmidt process,
-
1, 0)
E2
=
orthogonal projection
(l)1/25(l)1/2(2, -1,0)
+
=
we
(2,
-
=
(0,
-
1, 2).
obtain the orthonormal basis
(fWj, -1,
of (3, 1,
1, 0), F2
1)
2) into this plane is
(f)i/2l(f)^(_i
_
|, 1}
=
m,
_az
f)
1.11
50. Find the
L:
x
+
y
3y
+
point
z
=
0
z
=
0
on
which is closest to
(
line
(the
^(7 1
'
'
closest
0)J'
(7, 1,0).
(-*. -L3)\
/
(26)1'2
L is
the
linear span of the vector
of (7, 1,0) on this
orthogonal projection
is
point)
(-4,-1,3)^27 l
(26)1'2
26
'
'
'
EXERCISES
61. Which of the
(a)
(b)
(c)
(d)
(e)
(f)
(g)
(h)
following sets are open; closed; or neither.
{xeR:2< |x-5|<13}.
{xeR:0<x<4}.
{xeR:x>32}.
{X6fi":<x,x>=4}.
{x6A3:<x,(0,2,l))=0}.
{xe#3:2<||x-(3,0,3)||<14}.
{xeR":x1>0,...,x">0}.
The set of integers (considered as a subset of R).
(i)
{xeR":
2 *''<)
1=1
x
|>'a'^l}.
(j)
{xeR-.
(k)
{xe/?":2(^)3<2^')2}.
62. Find the
+ 3y + 2z
on
the plane
4
=
closest to the
point
point (1, 0, 1).
point on the line
63. Find the
x
+ 7v +
x
z
=
2
z
=
0
closest to the
64. Find
(a)
(b)
(c)
(d)
point (7, 1, 0).
an
orthonormal basis for the linear span of
Vl
=
Vi
=
v,
=
vi
=
(0,
(0,
(0,
(1,
2,
1,
3,
2,
(1, 0, 2), t, (1, 2, 4).
(1, 1, 2, 3).
(1, 0, 1, 0), v3
0, 1), v2
(0, 0, 2,
(0, 6, 0, 3, 0), y3
0, 0, 0), y2
(2, 1, 4, 3).
(4, 3, 2, 1), v3
3, 4), y2
2),
119
the line
Thus the
4, -1,3).
Inner Products
y2
=
=
=
=
=
=
=
=
-
1, 1).
120
/.
Linear Functions
65. Find orthonormal bases for the linear span and kernel of these
transformations
(a)
R*:
on
0>)
'8
/8
6
1
0
0\
1
2
1
2
1
2
0
2
2
1
2
1
1
-1
1
1
0
1
0
3
3
0
1
7
4
1
-2
0
PROBLEMS
67. Prove
Proposition
69. Give
an
32.
orthogonal set of vectors is independent.
example of a sequence {U} of open sets such that f)"=i U
68. Show that
an
is not open.
70. Give an example of
is open.
71. Find
an
vector space
sequence
{C} of closed
sets such that
orthonormal basis for the linear span of 1,
C([0, 1])
In the next four
inner
a
with the inner
problems
product, denoted <
,
V
product </, gy
represents
a
=
C
x2, x3 in the
\fg.
vector space endowed with
an
>.
72. Let v, w, x be three points in V such that v
x.
Show that the Pythagorean theorem is valid :
w
x,
(J"=i
x
is orthogonal to
\\v-W\\2=\\v-x\\2+\\x-w\\2
of
73. Let v, w be two vectors in V.
w which is closest to v is
Show that the vector in the linear span
(1.51)
w
v0
(You can verify this by minimizing the function f(t)
calculus.)
74. Prove Schwarz's inequality:
\\v
tw\\2 by
\<v,wy\^\\v\\-\\w\\
for any two vectors in
V.
(Hint: [|f
o
II2
=> 0 where v0 is
(1.50).)
75. Prove the
triangle inequality:
ll-x||^||y-w||+||w-x|l
for any three vectors in V
(use Schwarz's inequality).
given by
1.11
76. Let V be
is
vector space with
a
an
Inner Products
inner product.
Suppose that
subspace of V. Let J_( W) {v: <u, w> 0 for all w e IF}.
called the orthogonal complement of W. Show that L(W) is
a
=
subspace of
V and
F is finite
(if
121
dimensional) that
W and
W
This is
=
linear
a
1_(W) together
span V.
11. Let T: R"
A.
-*
Show that the
78. Show that
/?m be
a
linear transformation represented by the matrix
of A span (K(T)).
linear transformation is one-to-one
rows
a
the
on
orthogonal
complement of its kernel.
FURTHER READING
R. E. Johnson, Linear Algebra, Prindle, Weber & Schmidt, Boston, 1968.
covers the same material and includes a derivation of the Jordan
This book
canonical form.
K. Hoffman and R. Kunze, Linear Algebra, Prentice-Hall, Englewood
This book is more thorough and abstract, and has a full
Cliffs, N.J., 1961.
discussion of canonical forms.
L. Fox, An Introduction to Numerical Linear Algebra, Oxford University
Press, 1965. This is a detailed treatment of computational problems in
matrix
theory.
Nickerson, D. C. Spencer, and N. Steenrod, Advanced Calculus,
Van Nostrand, Princeton, N.J., 1957. This set of notes has a full treatment
of all the abstract linear algebra required in modern analysis.
H. K.
MISCELLANEOUS PROBLEMS
79. Show that if A' is obtained from A
by
a
sequence of
row
operations
0.
0, A'x
80. Show that every nonempty set of positive integers has a least element.
81. Show that a set with n elements has precisely 2" subsets.
then these
equations have the
same
solutions: Ax
82. Show that the -fold Cartesian
product of
=
a
=
set with k elements has
k" elements.
83. Can you
interpret the
case
k
=
2in Problem 82
so as
to deduce the
assertion of Exercise 3 ?
84. Let A
r
the
as
=
(a/)
be
an n x n
Show that A"~r
> 0.
=
0.
i>r for
assumption j
matrix such that
Show that the
some r
> 0.
a/
same
Will the
=
0 if
i
j>
r
for
some
conclusion follows from
hypothesis \i
j\
>
r
do
well?
85. Let T: R"
there
->
Rm be
a
linear transformation of rank
r.
Show that
linear transformations S,: Rm-^Rm-', S2: R"-'^Rn such that
(a) Si has rank m r and b e R(T) if and only if Sib 0.
are
=
(b) S2 has rank
86. Suppose that T:
-
R"
r
->
and
x 6
K(T) if and only if
R" and Tk
=
/.
x e
Show that T is invertible.
Show that the linear span
intersection of all linear subspaces of R" containing S.
87. Let S be
a
subset of R".
R(S2).
[S] of
S is the
Linear Functions
1.
Show that
88. Let S, T be subsets of R".
dim([S
r])
u
<
dim([5]) + dim([r]),
equality holds if and only if [S] n [T]
89. Let V and W be subspaces of Rn.
=
and
{0}.
Let X be the set of all
sums
subspace of R". The
+
V,
V + W. If
relationship between X and V and W is indicated by writing X
the form
in
written
be
xeJcan
in addition V r\ W={0}, then every
we say
n
W
V
W
with
X
this
0,
In
V+
one
case,
v + w in only
way.
v
w
with
w 6
v e
W.
Show that X is
linear
a
=
=
=
of V and IF and write X=V@W.
90. Suppose X= V W. Then dim X= dim F + dim IF.
91. Show that if A: R" -> .R is a linear function, there exists a
that X is the direct
sum
w e
.R" such
that A(v)
<v, w> for all v e /?".
92. If S is a subset of R" define
=
1(5)
=
{v
e
R":
<v, s>
=
0 for all
s e
S}.
{0}.
(a) Show that (S) is a subspace of .R" and that S n _L(S)
that
Show
=(.(S)).
[S]
(b)
F J_(F).
(c) If V is a linear subspace of .R", R"
93. Suppose that T: V-> Wis a linear transformation and Fis not finite
dimensional. Show that either the rank or the nullity of T must be infinite.
=
=
94. Let V be
an
A bilinear function p
abstract vector space.
function of two variables in V with these
on
V is
a
properties:
cp(v, w)
p(v, cw)
p(cv, w) cp(v, w)
p(v,
p(vi + v2,w) =p(vu w) + p(v2 w)
=
=
,
Wi
+
w2) =p(v, Wi) + p(v, w2)
In fact, the space
sum of two bilinear functions is bilinear.
Bv of all bilinear functions is an abstract vector space. If V is finite dimen
sional, what is the dimension of Bv ? (Hint: See the next problem.)
Show that the
95. Let p be
a
bilinear function
on
R".
Let
a,;j=piE,,Ej)
Show
that/? is completely determined by
96. Let V be
(a)
the matrix
(at;j).
abstract vector space.
Show that the space V* of linear functions
an
under addition and scalar
on
Fis
a
vector space
multiplication.
d also.
(b) If dim V=d, show that dim V*
(c) Show that to every A e R"* there isaweff such that A(v)
<v, w> for all v e Rn. (Recall Problem 91.)
97. Suppose that V is a linear subspace of W. We define the annihilator
of V, denoted ann(F), to be the set of A e W* such that X(v)
0 if v e V.
=
=
=
1.11
Show that ann(F) is a linear subspace of IF*.
show that ann(F) has dimension n
d.
98. Let V be
linear
a
subspace
linear transformation.
R"
-=?
J?m defined
on
If dim W
=
123
dim V
n,
=
d,
of .R", and suppose that T: V->Rm is a
a linear transformation 7":
Show that there is
all of R" which extends T.
99. The closure of
every
Inner Products
a
neighborhood of
set
x
S, denoted 5, is the
contains
points of
S.
set of all
points
x
such that
Find the closure of all the
sets in Problem 61.
1 00. Show that the closure of a set S is the smallest closed set containing S.
101. The boundary of
a
set
S, denoted dS, is the
set of all
points x such
complement
that every neighborhood of x contains points of both S and the
of S. Find the boundary of all the sets in Problem 61.
102. Show that the
103. Show that the
boundary of a
boundary of a
set is
a
closed set.
set S is also the
S n (Rn
plement
fact, show that 8S
104. Let T: V-> W be a linear transformation of
R"
inner
S.
-
In
=
-
boundary of its
com
S).
vector space with
a
an
The adjoint of Tis the transformation T*: W-+ V defined
product.
in this way
<T*(w), vy
=
for all
<w, Tv>
(a) Show
(b) If T:
that T* is
a
v e
V
well-defined linear transformation.
represented by the matrix A (a/),
represented by the matrix A* (a*j), where a*j
(This matrix is called the adjoint or transpose of A.)
(c) Show that R(T*) is complementary to K(T).
v(T), v(T*) p(T).
(A) In fact, P(T*)
R"
--
Rm is
=
T* : Rm^ R" is
=
then
=
atJ.
=
=
105. A bilinear form pona vector space V is called symmetric if it obeys
the law: p(v, w) =/>(w, v) for all v and w. An inner product is a symmetric
manipulations with inner products
symmetric bilinear forms. For example, the GramSchmidt process (Proposition 32) gives rise to this fact (see if you can work
the proof of Proposition 32 to give it):
Proposition. Let p be a symmetric bilinear form on V. Suppose Fi,
We can find another basis, Eu...,EofV such that the
F is a basis for V.
Fj for all j, and
linear span of Eu...,Ej is the same as that of Fu
p(E,,Ej) 0ifi^j.
We shall call such a basis Ei,..., E p-orthogonal.
106. Let/? be a symmetric bilinear form on a vector space V, and suppose
E is a p-orthogonal basis.
Ei,
(a) Show that p(v, w) can be computed in terms of this basis as
bilinear form and much of the formal
remains valid for
...,
.
.
.
,
=
...,
follows : if
p(p, w)
=
v
=
2
v'Et
2 v'w'p(E,
1=1
,
,
w
E,)
=
2
W'E> then
>
(1.52)
/.
Linear Functions
(b) Show that
p is an inner
product
on
the linear span of the Et
such that p(E,, E,) >0.
(c) Similarly,
p is an
inner
product
on
the linear span of the Et
such that p(E, E,) < 0.
107. Prove this fact: Let p be a symmetric bilinear form on a finitedimensional vector space V. There is a basis Eu ...,E , integers r, s such
that r + s <; n and such that if v
2 v'-Ei then
,
=
,
Xm)=2M!fr
2 (f')2
0-53)
rlr+s
(Hint: Modify the basis {,} in Problem 106 so that (1.52) becomes (1.53).)
108. The integers r, s of Problem 107 are determined by p alone, and are
independent of the basis. Here is a sketch of how a proof would go.
Suppose Fi, ...,F is another p-orthogonal basis and p is the number of
r.
Let W be the linear
Ft's such that p(Ft Ft) > 0. We have to show p
E
span of these F's. Expressing points of W in terms of the basis Ei
we may consider the transformation T: W->R" given by
=
,
TQv'Et)=(v\...,tf)
T is one-to-one
on
W, for if
0<p(w,w)=2 ("')21ST
so we
w e
W, and
w
= 0,
2 (')2
rir+i
must have
2(')2>o
tsr
on
W.
Since T is one-to-one, it follows that r ;> p. The inequality p >; r
same argument with the roles of Eu
E and FU...,F
follows from the
.
.
.
,
interchanged.
109. Let A
=
(aci) be
Then A determines
PA(y, w)=
a
a symmetric n x n matrix, that is, a,;
j
symmetric bilinear form on R" as follows:
=
a,;
,
2 aiww'w-'
i.j
If P is the matrix
corresponding to the change of basis from the standard
basis to that described in Problem 105, then P*AP is diagonal. Verify that
assertion.
1.11
Inner Products
125
110. Find the p-orthogonal basis and the
107 for the symmetric bilinear forms
[4
(a)
3
p(y, y)
=
p(y, v) >0, =0,
(v1)2 + (i;2)2 + (v3)2
transformation
112. A
T:
v,
w e
Show that if T is
V).
self-
v, w are eigenvectors of a self-adjoint transformation T
eigenvalues. Show that <v, w> 0.
self-adjoint transformation on R", and v0 e Rn is such that
Suppose that
V with different
114. If T is
2(V)2
=
a
=
land
<Ty0, v0> =max{<rv, v>;
then v0 is
an
2 0>')2
eigenvector for
=
1}
T.
115. Use Problems 113 and 114 to prove the
adjoint operators
Theorem.
T
a
K(T)R(T)
113.
on
given by
self-adjoint if it is self-
V-> V is called
=
=
<0 in R* where p is
(v*)2
-
<v, 7w> for all
adjoint 7v, w>
on R", then
transformation
adjoint
R?
(b)
l\
0
111. Describe the sets
representation (1 .53) of Problem
given by these matrices:
can
be
on
There is
computed
r(2*'E,)
=
Spectral
theorem for self-
R":
an
orthonormal basis Ei,
in terms
.
.
.
,
E of eigenvectors of T.
of this basis by
2*'c.E,
1 1 6. Find a basis of eigenvectors in R* for the self-adjoint transformations
given by the matrices (a), (b) of Problem 110.
117. Orthonormalize these bases of R*:
(a)
(b)
(1, 0, 0, 0), (0, 1, 1, 1), (0, 0, 2, 2), (3, 0, 0, 3).
(-1, -1, -1, -D,(0, -1, -1, -1), (0,0, -1, -1),
(0,0,0,-1).
(c) (0, 1, 0, 1), (1, 0, 1, 0), (1, 0, 0, 1), (0, 1, 1, 0).
118. Find the orthogonal projection of R5 onto these spaces:
(a) The span of (0,1,0,0,1).
(b) The
(c) The
(d) The
(1, 1, 0, 0, 0), (1, 0, 1, 0, 0).
(1, 0, 0, 0, 1), (0, 1, 0, 0, 1), (0, 0, 1, 0, 1).
of the vectors given in (c) and the vector (0, 0, 0, 1, 1).
span of
span of
span
Download