Uploaded by 나총

Complete Solutions for Manual-4th linear algebra

advertisement
Complete Solutions Manual
Linear Algebra
A Modern Introduction
FOURTH EDITION
David Poole
Trent University
Prepared by
Roger Lipsett
Australia • Brazil • Japan • Korea • Mexico • Singapore • Spain • United Kingdom • United States
ISBN-13: 978-128586960-5
ISBN-10: 1-28586960-5
© 2015 Cengage Learning
ALL RIGHTS RESERVED. No part of this work covered by the
copyright herein may be reproduced, transmitted, stored, or
used in any form or by any means graphic, electronic, or
mechanical, including but not limited to photocopying,
recording, scanning, digitizing, taping, Web distribution,
information networks, or information storage and retrieval
systems, except as permitted under Section 107 or 108 of the
1976 United States Copyright Act, without the prior written
permission of the publisher except as may be permitted by the
license terms below.
For product information and technology assistance, contact us at
Cengage Learning Customer & Sales Support,
1-800-354-9706.
For permission to use material from this text or product, submit
all requests online at www.cengage.com/permissions
Further permissions questions can be emailed to
permissionrequest@cengage.com.
Cengage Learning
200 First Stamford Place, 4th Floor
Stamford, CT 06902
USA
Cengage Learning is a leading provider of customized
learning solutions with office locations around the globe,
including Singapore, the United Kingdom, Australia,
Mexico, Brazil, and Japan. Locate your local office at:
www.cengage.com/global.
Cengage Learning products are represented in
Canada by Nelson Education, Ltd.
To learn more about Cengage Learning Solutions,
visit www.cengage.com.
Purchase any of our products at your local college
store or at our preferred online store
www.cengagebrain.com.
NOTE: UNDER NO CIRCUMSTANCES MAY THIS MATERIAL OR ANY PORTION THEREOF BE SOLD, LICENSED, AUCTIONED,
OR OTHERWISE REDISTRIBUTED EXCEPT AS MAY BE PERMITTED BY THE LICENSE TERMS HEREIN.
READ IMPORTANT LICENSE INFORMATION
Dear Professor or Other Supplement Recipient:
Cengage Learning has provided you with this product (the
“Supplement”) for your review and, to the extent that you adopt
the associated textbook for use in connection with your course
(the “Course”), you and your students who purchase the
textbook may use the Supplement as described below. Cengage
Learning has established these use limitations in response to
concerns raised by authors, professors, and other users
regarding the pedagogical problems stemming from unlimited
distribution of Supplements.
Cengage Learning hereby grants you a nontransferable license
to use the Supplement in connection with the Course, subject to
the following conditions. The Supplement is for your personal,
noncommercial use only and may not be reproduced, posted
electronically or distributed, except that portions of the
Supplement may be provided to your students IN PRINT FORM
ONLY in connection with your instruction of the Course, so long
as such students are advised that they
Printed in the United States of America
1 2 3 4 5 6 7 17 16 15 14 13
may not copy or distribute any portion of the Supplement to any
third party. You may not sell, license, auction, or otherwise
redistribute the Supplement in any form. We ask that you take
reasonable steps to protect the Supplement from unauthorized
use, reproduction, or distribution. Your use of the Supplement
indicates your acceptance of the conditions set forth in this
Agreement. If you do not accept these conditions, you must
return the Supplement unused within 30 days of receipt.
All rights (including without limitation, copyrights, patents, and
trade secrets) in the Supplement are and will remain the sole and
exclusive property of Cengage Learning and/or its licensors. The
Supplement is furnished by Cengage Learning on an “as is” basis
without any warranties, express or implied. This Agreement will
be governed by and construed pursuant to the laws of the State
of New York, without regard to such State’s conflict of law rules.
Thank you for your assistance in helping to safeguard the integrity
of the content contained in this Supplement. We trust you find the
Supplement a useful teaching tool.
Contents
1 Vectors
1.1 The Geometry and Algebra of Vectors
1.2 Length and Angle: The Dot Product .
Exploration: Vectors and Geometry . . . .
1.3 Lines and Planes . . . . . . . . . . . .
Exploration: The Cross Product . . . . . .
1.4 Applications . . . . . . . . . . . . . . .
Chapter Review . . . . . . . . . . . . . . . .
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
3
3
10
25
27
41
44
48
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
2 Systems of Linear Equations
2.1 Introduction to Systems of Linear Equations . . . . .
2.2 Direct Methods for Solving Linear Systems . . . . . .
Exploration: Lies My Computer Told Me . . . . . . . . . .
Exploration: Partial Pivoting . . . . . . . . . . . . . . . . .
Exploration: An Introduction to the Analysis of Algorithms
2.3 Spanning Sets and Linear Independence . . . . . . . .
2.4 Applications . . . . . . . . . . . . . . . . . . . . . . . .
2.5 Iterative Methods for Solving Linear Systems . . . . .
Chapter Review . . . . . . . . . . . . . . . . . . . . . . . . .
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
53
. 53
. 58
. 75
. 75
. 77
. 79
. 93
. 112
. 123
3 Matrices
3.1 Matrix Operations . . . . . . . . . . . .
3.2 Matrix Algebra . . . . . . . . . . . . . .
3.3 The Inverse of a Matrix . . . . . . . . .
3.4 The LU Factorization . . . . . . . . . .
3.5 Subspaces, Basis, Dimension, and Rank
3.6 Introduction to Linear Transformations
3.7 Applications . . . . . . . . . . . . . . . .
Chapter Review . . . . . . . . . . . . . . . . .
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
129
129
138
150
164
176
192
209
230
4 Eigenvalues and Eigenvectors
4.1 Introduction to Eigenvalues and Eigenvectors . .
4.2 Determinants . . . . . . . . . . . . . . . . . . . .
Exploration: Geometric Applications of Determinants
4.3 Eigenvalues and Eigenvectors of n ⇥ n Matrices .
4.4 Similarity and Diagonalization . . . . . . . . . .
4.5 Iterative Methods for Computing Eigenvalues . .
4.6 Applications and the Perron-Frobenius Theorem
Chapter Review . . . . . . . . . . . . . . . . . . . . . .
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
235
235
250
263
270
291
308
326
365
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
1
.
.
.
.
.
.
.
.
2
CONTENTS
5 Orthogonality
5.1 Orthogonality in Rn . . . . . . . . . . . . . . . . . . . . . .
5.2 Orthogonal Complements and Orthogonal Projections . . .
5.3 The Gram-Schmidt Process and the QR Factorization . . .
Exploration: The Modified QR Process . . . . . . . . . . . . . .
Exploration: Approximating Eigenvalues with the QR Algorithm
5.4 Orthogonal Diagonalization of Symmetric Matrices . . . . .
5.5 Applications . . . . . . . . . . . . . . . . . . . . . . . . . . .
Chapter Review . . . . . . . . . . . . . . . . . . . . . . . . . . . .
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
371
371
379
388
398
402
405
417
442
6 Vector Spaces
6.1 Vector Spaces and Subspaces . . . . . . . . . . . . . . . . .
6.2 Linear Independence, Basis, and Dimension . . . . . . . . .
Exploration: Magic Squares . . . . . . . . . . . . . . . . . . . . .
6.3 Change of Basis . . . . . . . . . . . . . . . . . . . . . . . . .
6.4 Linear Transformations . . . . . . . . . . . . . . . . . . . .
6.5 The Kernel and Range of a Linear Transformation . . . . .
6.6 The Matrix of a Linear Transformation . . . . . . . . . . .
Exploration: Tiles, Lattices, and the Crystallographic Restriction
6.7 Applications . . . . . . . . . . . . . . . . . . . . . . . . . . .
Chapter Review . . . . . . . . . . . . . . . . . . . . . . . . . . . .
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
451
451
463
477
480
491
498
507
525
527
531
7 Distance and Approximation
7.1 Inner Product Spaces . . . . . . . . . . . . . . . . . . . . . .
Exploration: Vectors and Matrices with Complex Entries . . . .
Exploration: Geometric Inequalities and Optimization Problems
7.2 Norms and Distance Functions . . . . . . . . . . . . . . . .
7.3 Least Squares Approximation . . . . . . . . . . . . . . . . .
7.4 The Singular Value Decomposition . . . . . . . . . . . . . .
7.5 Applications . . . . . . . . . . . . . . . . . . . . . . . . . . .
Chapter Review . . . . . . . . . . . . . . . . . . . . . . . . . . . .
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
537
537
546
553
556
568
590
614
625
8 Codes
8.1 Code Vectors . . . . . . .
8.2 Error-Correcting Codes .
8.3 Dual Codes . . . . . . . .
8.4 Linear Codes . . . . . . .
8.5 The Minimum Distance of
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
633
633
637
641
647
650
.
.
.
.
a
. . . .
. . . .
. . . .
. . . .
Code
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
Chapter 1
Vectors
1.1
The Geometry and Algebra of Vectors
!!2, 3"
1.
!2, 3"
3
2
1
!3, 0"
!2
1
!1
2
3
!1
!3, !2"
!2
2. Since



2
3
5
+
=
,
3
0
3



2
2
4
+
=
,
3
3
0

2
+
3


2
0
=
,
3
0
plotting those vectors gives
1
2
3
4
!1
c
b
!2
a
!3
d
!4
!5
3
5

2
+
3

3
=
2

5
,
5
4
CHAPTER 1. VECTORS
3.
2
c
z
b
1
!2
0
!1
0
1
y2
1
0
3
2
!1
a
x
!1
d
!2
4. Since the heads are all at (3, 2, 1), the tails are at
2 3
3
425
1
2 3 2 3
0
3
425 = 405 ,
0
1
2 3
3
425
1
2 3 2 3
0
3
425 = 405 ,
1
0
2 3
3
4 25
1
2
3 2 3
1
2
4 25 = 445 ,
1
0
# »
5. The four vectors AB are
3
c
2
1
d
a
1
2
3
4
!1
b
!2
In standard position, the vectors are
# »
(a) AB = [4
# »
(b) AB = [2
# » ⇥
(c) AB = 12
# » ⇥
(d) AB = 16
1, 2
0, 1
2, 3
1 1
3, 2
( 1)] = [3, 3].
( 2)] = [2, 1]
⇤ ⇥ 3 3⇤
3
2 =
2, 2
⇤
⇥
⇤
1
1 1
3 =
6, 6 .
3
2
a
1
c
b
d
!1
1
2
3
2 3
3
425
1
2
3 2 3
1
4
4 15 = 435 .
2
3
1.1. THE GEOMETRY AND ALGEBRA OF VECTORS
5
6. Recall the notation that [a, b] denotes a move of a units horizontally and b units vertically. Then during
the first part of the walk, the hiker walks 4 km north, so a = [0, 4]. During the second part of the
walk, the hiker walks a distance of 5 km northeast. From the components, we get
" p
p #
5 2 5 2
b = [5 cos 45 , 5 sin 45 ] =
,
.
2
2
Thus the net displacement vector is
" p
p #
5 2
5 2
c=a+b=
, 4+
.
2
2
7. a + b =




3
2
3+2
5
+
=
=
.
0
3
0+3
3
3
2
a!b
b
1
a
1
8. b c =


2
3


2
2 ( 2)
4
=
=
.
3
3 3
0
2
3
4
5
3
2
b
!c
1
b!c
1
9. d
c=


3
2
2
=
3

5
.
5
2
1
2
3
3
4
4
5
d
!1
!2
d!c
!3
!c
!4
!5
10. a + d =

6
.
2

3
+
0

3
2
=

3+3
0 + ( 2)
a
=
1
2
3
4
5
6
d
!1
a!d
!2
11. 2a + 3c = 2[0, 2, 0] + 3[1, 2, 1] = [2 · 0, 2 · 2, 2 · 0] + [3 · 1, 3 · ( 2), 3 · 1] = [3, 2, 3].
12.
3b
2c + d = 3[3, 2, 1] 2[1, 2, 1] + [ 1, 1, 2]
= [3 · 3, 3 · 2, 3 · 1] + [ 2 · 1, 2 · ( 2), 2 · 1] + [ 1, 1, 2]
= [6, 9, 1].
6
CHAPTER 1. VECTORS
13. u = [cos 60 , sin 60 ] =
h
p
3
2
1
2,
"
i
, and v = [cos 210 , sin 210 ] =
1
u+v =
2
#
1
,
2
p
p
3
3
,
2
2
"
h
p
3
2 ,
1
2
i
, so that
#
p p
1
3
3 1
v=
+
,
+
.
2
2
2
2
u
# »
14. (a) AB = b a.
# » # »
# » # »
(b) Since OC = AB, we have BC = OC b = (b a) b = a.
# »
(c) AD = 2a.
# »
# »
# »
(d) CF = 2OC = 2AB = 2(b a) = 2(a b).
# » # » # »
(e) AC = AB + BC = (b a) + ( a) = b 2a.
# »
# »
# »
# »
(f ) Note that F A and OB are equal, and that DE = AB. Then
# » # » # »
BC + DE + F A =
15. 2(a
3b) + 3(2b + a)
property e.
distributivity
=
(2a
# » # »
AB + OB =
a
6b) + (6b + 3a)
a
(b
property b.
associativity
=
a) + b = 0.
(2a + 3a) + ( 6b + 6b) = 5a.
16.
3(a
c) + 2(a + 2b) + 3(c
property e.
distributivity
b)
=
property b.
associativity
=
=
17. x
a = 2(x
2a) = 2x
4a ) x
( 3a + 2a) + (4b
3b)
3b) + (3c + 3c)
a + b + 6c.
4a )
2x = a
( 3a + 3c) + (2a + 4b) + (3c
3a ) x = 3a.
x=
18.
x + 2a
b = 3(x + a)
x
2(2a
3x =
a
b) = 3x + 3a
4a + 2b
)
2a + 2b + b
3a + 3b )
3
3
b.
x= a
2
2
2x =
)
19. We have 2u + 3v = 2[1, 1] + 3[1, 1] = [2 · 1 + 3 · 1, 2 · ( 1) + 3 · 1] = [5, 1]. Plots of all three vectors are
2
w
1
v
v
1
!1
2
3
4
v
u
!1
v
u
!2
5
6
1.1. THE GEOMETRY AND ALGEBRA OF VECTORS
20. We have u
vectors are
2v =
[ 2, 1]
7
2 · 2, 1
2[2, 2] = [ ( 2)
2 · ( 2)] = [ 2, 3]. Plots of all three
!u
!v
3
!u
2
w
!v
1
u
!2
1
!1
2
v
!1
!2
21. From the diagram, we see that w =
2u + 4v.
6
!u
5
!u
w
4
v
3
v
2
v
1
v
1
!1
2
3
4
6
5
u
!1
22. From the diagram, we see that w = 2u +
3v.
9
8
u
7
6
w
5
u
4
3
v
2
u
v
1
v
!2
!1
1
2
3
4
5
6
23. Property (d) states that u + ( u) = 0. The first diagram below shows u along with
diagonal of the parallelogram, the resultant vector is 0.
Property (e) states that c(u + v) = cu + cv. The second figure illustrates this.
7
u. Then, as the
8
CHAPTER 1. VECTORS
cv
c!u ! v"
cu
u
v
cu
u
u!v
u
cv
v
!u
24. Let u = [u1 , u2 , . . . , un ] and v = [v1 , v2 , . . . , vn ], and let c and d be scalars in R.
Property (d):
u + ( u) = [u1 , u2 , . . . , un ] + ( 1[u1 , u2 , . . . , un ])
= [u1 , u2 , . . . , un ] + [ u1 , u2 , . . . , un ]
= [u1
u 1 , u2
u2 , . . . , un
un ]
= [0, 0, . . . , 0] = 0.
Property (e):
c(u + v) = c ([u1 , u2 , . . . , un ] + [v1 , v2 , . . . , vn ])
= c ([u1 + v1 , u2 + v2 , . . . , un + vn ])
= [c(u1 + v1 ), c(u2 + v2 ), . . . , c(un + vn )]
= [cu1 + cv1 , cu2 + cv2 , . . . , cun + cvn ]
= [cu1 , cu2 , . . . , cun ] + [cv1 , cv2 , . . . , cvn ]
= c[u1 , u2 , . . . , un ] + c[v1 , v2 , . . . , vn ]
= cu + cv.
Property (f):
(c + d)u = (c + d)[u1 , u2 , . . . , un ]
= [(c + d)u1 , (c + d)u2 , . . . , (c + d)un ]
= [cu1 + du1 , cu2 + du2 , . . . , cun + dun ]
= [cu1 , cu2 , . . . , cun ] + [du1 , du2 , . . . , dun ]
= c[u1 , u2 , . . . , un ] + d[u1 , u2 , . . . , un ]
= cu + du.
Property (g):
c(du) = c(d[u1 , u2 , . . . , un ])
= c[du1 , du2 , . . . , dun ]
= [cdu1 , cdu2 , . . . , cdun ]
= [(cd)u1 , (cd)u2 , . . . , (cd)un ]
= (cd)[u1 , u2 , . . . , un ]
= (cd)u.
25. u + v = [0, 1] + [1, 1] = [1, 0].
26. u + v = [1, 1, 0] + [1, 1, 1] = [0, 0, 1].
1.1. THE GEOMETRY AND ALGEBRA OF VECTORS
9
27. u + v = [1, 0, 1, 1] + [1, 1, 1, 1] = [0, 1, 0, 0].
28. u + v = [1, 1, 0, 1, 0] + [0, 1, 1, 1, 0] = [1, 0, 1, 0, 0].
29.
30.
+
0
1
2
3
4
+
0
1
2
3
0
0
1
2
3
1
1
2
3
0
2
2
3
0
1
3
3
0
1
2
·
0
1
2
3
0
0
0
0
0
1
0
1
2
3
2
0
2
0
2
3
0
3
2
1
0
0
1
2
3
4
1
1
2
3
4
0
2
2
3
4
0
1
3
3
4
0
1
2
4
4
0
1
2
3
·
0
1
2
3
4
0
0
0
0
0
0
1
0
1
2
3
4
2
0
2
4
1
3
3
0
3
1
4
2
4
0
4
3
2
1
31. 2 + 2 + 2 = 6 = 0 in Z3 .
32. 2 · 2 · 2 = 3 · 2 = 0 in Z3 .
33. 2(2 + 1 + 2) = 2 · 2 = 3 · 1 + 1 = 1 in Z3 .
34. 3 + 1 + 2 + 3 = 4 · 2 + 1 = 1 in Z4 .
35. 2 · 3 · 2 = 4 · 3 + 0 = 0 in Z4 .
36. 3(3 + 3 + 2) = 4 · 6 + 0 = 0 in Z4 .
37. 2 + 1 + 2 + 2 + 1 = 2 in Z3 , 2 + 1 + 2 + 2 + 1 = 0 in Z4 , 2 + 1 + 2 + 2 + 1 = 3 in Z5 .
38. (3 + 4)(3 + 2 + 4 + 2) = 2 · 1 = 2 in Z5 .
39. 8(6 + 4 + 3) = 8 · 4 = 5 in Z9 .
40. 2100 = 210
10
= (1024)10 = 110 = 1 in Z11 .
41. [2, 1, 2] + [2, 0, 1] = [1, 1, 0] in Z33 .
42. 2[2, 2, 1] = [2 · 2, 2 · 2, 2 · 1] = [1, 1, 2] in Z33 .
43. 2([3, 1, 1, 2] + [3, 3, 2, 1]) = 2[2, 0, 3, 3] = [2 · 2, 2 · 0, 2 · 3, 2 · 3] = [0, 0, 2, 2] in Z44 .
2([3, 1, 1, 2] + [3, 3, 2, 1]) = 2[1, 4, 3, 3] = [2 · 1, 2 · 4, 2 · 3, 2 · 3] = [2, 3, 1, 1] in Z45 .
44. x = 2 + ( 3) = 2 + 2 = 4 in Z5 .
45. x = 1 + ( 5) = 1 + 1 = 2 in Z6
46. x = 2
1
= 2 in Z3 .
47. No solution. 2 times anything is always even, so cannot leave a remainder of 1 when divided by 4.
48. x = 2
1
49. x = 3
1
= 3 in Z5 .
4 = 2 · 4 = 3 in Z5 .
50. No solution. 3 times anything is always a multiple of 3, so it cannot leave a remainder of 4 when
divided by 6 (which is also a multiple of 3).
51. No solution. 6 times anything is always even, so it cannot leave an odd number as a remainder when
divided by 8.
10
CHAPTER 1. VECTORS
52. x = 8
1
9 = 7 · 9 = 8 in Z11
53. x = 2
1
(2 + ( 3)) = 3(2 + 2) = 2 in Z5 .
54. No solution. This equation is the same as 4x = 2 5 = 3 = 3 in Z6 . But 4 times anything is even,
so it cannot leave a remainder of 3 when divided by 6 (which is also even).
55. Add 5 to both sides to get 6x = 6, so that x = 1 or x = 5 (since 6 · 1 = 6 and 6 · 5 = 30 = 6 in Z8 ).
56. (a)
All values.
(b)
All values.
(c)
All values.
57. (a) All a 6= 0 in Z5 have a solution because 5 is a prime number.
(b) a = 1 and a = 5 because they have no common factors with 6 other than 1.
(c) a and m can have no common factors other than 1; that is, the greatest common divisor, gcd, of
a and m is 1.
1.2
Length and Angle: The Dot Product

1
3
·
= ( 1) · 3 + 2 · 1 = 3 + 2 = 1.
2
1


3
4
2. Following Example 1.15, u · v =
·
= 3 · 4 + ( 2) · 6 = 12 12 = 0.
2
6
2 3 2 3
1
2
3. u · v = 425 · 435 = 1 · 2 + 2 · 3 + 3 · 1 = 2 + 6 + 3 = 11.
3
1
1. Following Example 1.15, u · v =

4. u · v = 3.2 · 1.5 + ( 0.6) · 4.1 + ( 1.4) · ( 0.2) = 4.8 2.46 + 0.28 = 2.62.
2 3 2
3
4
p1
p
6 27 6
p
p
p
27
7
p 7 6
5. u · v = 6
2) + 3 · 0 + 0 · ( 5) = 4 2 = 2.
4 35 · 4 0 5 = 1 · 4 + 2 · (
5
0
2
3 2
3
1.12
2.29
6 3.257 6 1.727
7 6
7
6. u · v = 6
4 2.075 · 4 4.335 = 1.12 · 2.29 3.25 · 1.72 + 2.07 · 4.33 1.83 · ( 1.54) = 3.6265.
1.83
1.54
7. Finding a unit vector v in the same direction as a given vector u is called normalizing the vector u.
Proceed as in Example 1.19:
p
p
kuk = ( 1)2 + 22 = 5,
so a unit vector v in the same direction as u is
1
1
v=
u= p
kuk
5
"
#
1
2
=
"
p1
5
p2
5
#
.
8. Proceed as in Example 1.19:
kuk =
p
p
p
32 + ( 2)2 = 9 + 4 = 13,
so a unit vector v in the direction of u is
1
1
v=
u= p
kuk
13
"
#
3
2
=
"
p3
13
p2
13
#
.
1.2. LENGTH AND ANGLE: THE DOT PRODUCT
9. Proceed as in Example 1.19:
kuk =
so a unit vector v in the direction of u is
p
11
12 + 22 + 32 =
p
14,
2 3 2
3
p1
1
14
6 2 7
1
1 6 7
7
v=
u= p 6
=6
27
4
5
4 p14 5 .
kuk
14
3
p
3
14
10. Proceed as in Example 1.19:
p
p
p
kuk = 3.22 + ( 0.6)2 + ( 1.4)2 = 10.24 + 0.36 + 1.96 = 12.56 ⇡ 3.544,
so a unit vector v in the direction of u is
11. Proceed as in Example 1.19:
3 2
3
1.5
0.903
1
1 4
0.45 ⇡ 4 0.1695 .
v=
u=
kuk
3.544
2.1
0.395
kuk =
r
2
12 +
so a unit vector v in the direction of u is
⇣p ⌘2 ⇣p ⌘2
p
2 +
3 + 02 = 6,
2
3 2 1 3 2 1 3 2p 3
6
p
p
1
6
6
6
p
p
p
6 7 6 27 6 1 7 6 37
6p 7 6p 7 6 3 7
1
1 6 27
p6 7 = 6 3 7 = 6 p 7
v=
u= p 6
=6
p 7
6
7
6 1 7 6 27
37
kuk
6 4 35 6
4 p6 5 4 p2 5 4 2 5
0
0
0
0
12. Proceed as in Example 1.19:
p
p
kuk = 1.122 + ( 3.25)2 + 2.072 + ( 1.83)2 = 1.2544 + 10.5625 + 4.2849 + 3.3489
p
= 19.4507 ⇡ 4.410,
so a unit vector v in the direction of u is
1
1 ⇥
1.12
3.25 2.07
v
u=
kuk
4.410
13. Following Example 1.20, we compute: u
d(u, v) = ku
v=

⇤ ⇥
1.83 ⇡ 0.254
vk =

1
2
q

3
=
1
2

0.737
4
, so
1
( 4) + 12 =
p
17.


3
4
1
14. Following Example 1.20, we compute: u v =
=
, so
2
6
8
q
p
2
2
d(u, v) = ku vk = ( 1) + ( 8) = 65.
2 3 2 3 2 3
1
2
1
15. Following Example 1.20, we compute: u v = 425 435 = 4 15, so
3
1
2
q
p
2
2
d(u, v) = ku vk = ( 1) + ( 1) + 22 = 6.
0.469
⇤
0.415 .
12
CHAPTER 1. VECTORS
2
3
3.2
v = 4 0.65
1.4
16. Following Example 1.20, we compute: u
d(u, v) = ku
vk =
p
2
3 2
3
1.5
1.7
4 4.15 = 4 4.75, so
0.2
1.2
1.72 + ( 4.7)2 + ( 1.2)2 =
p
26.42 ⇡ 5.14.
17. (a) u · v is a real number, so ku · vk is the norm of a number, which is not defined.
(b) u · v is a scalar, while w is a vector. Thus u · v + w adds a scalar to a vector, which is not a
defined operation.
(c) u is a vector, while v · w is a scalar. Thus u · (v · w) is the dot product of a vector and a scalar,
which is not defined.
(d) c · (u + v) is the dot product of a scalar and a vector, which is not defined.
18. Let ✓ be the angle between u and v. Then
u·v
3 · ( 1) + 0 · 1
p
cos ✓ =
=p
=
2
kuk kvk
3 + 02 ( 1)2 + 12
Thus cos ✓ < 0 (in fact, ✓ =
3⇡
4 ),
3
p =
3 2
p
2
.
2
so the angle between u and v is obtuse.
19. Let ✓ be the angle between u and v. Then
cos ✓ =
Thus cos ✓ > 0 (in fact, ✓ =
u·v
2 · 1 + ( 1) · ( 2) + 1 · ( 1)
1
p
=p
= .
2
2
2
2
2
2
kuk kvk
2
2 + ( 1) + 1 1 + ( 2) + 1
⇡
3 ),
so the angle between u and v is acute.
20. Let ✓ be the angle between u and v. Then
cos ✓ =
u·v
4 · 1 + 3 · ( 1) + ( 1) · 1
0
p
=p
= p p = 0.
2
2
2
2
2
2
kuk kvk
26 3
4 + 3 + ( 1) 1 + ( 1) + 1
Thus the angle between u and v is a right angle.
21. Let ✓ be the angle between u and v. Note that we can determine whether ✓ is acute, right, or obtuse
u·v
by examining the sign of kukkvk
, which is determined by the sign of u · v. Since
u · v = 0.9 · ( 4.5) + 2.1 · 2.6 + 1.2 · ( 0.8) = 0.45 > 0,
we have cos ✓ > 0 so that ✓ is acute.
22. Let ✓ be the angle between u and v. Note that we can determine whether ✓ is acute, right, or obtuse
u·v
by examining the sign of kukkvk
, which is determined by the sign of u · v. Since
u · v = 1 · ( 3) + 2 · 1 + 3 · 2 + 4 · ( 2) =
3,
we have cos ✓ < 0 so that ✓ is obtuse.
23. Since the components of both u and v are positive, it is clear that u · v > 0, so the angle between
them is acute since it has a positive cosine.
⇣ p ⌘
p
2
2
1
24. From Exercise 18, cos ✓ =
= 3⇡
2 , so that ✓ = cos
2
4 = 135 .
25. From Exercise 19, cos ✓ = 12 , so that ✓ = cos
26. From Exercise 20, cos ✓ = 0, so that ✓ =
⇡
2
1 1
2
=
⇡
3
= 60 .
= 90 is a right angle.
1.2. LENGTH AND ANGLE: THE DOT PRODUCT
13
27. As in Example 1.21, we begin by calculating u · v and the norms of the two vectors:
u · v = 0.9 · ( 4.5) + 2.1 · 2.6 + 1.2 · ( 0.8) = 0.45,
p
p
kuk = 0.92 + 2.12 + 1.22 = 6.66,
p
p
kvk = ( 4.5)2 + 2.62 + ( 0.8)2 = 27.65.
So if ✓ is the angle between u and v, then
cos ✓ =
u·v
0.45
0.45
p
=p
⇡p
,
kuk kvk
6.66 27.65
182.817
so that
✓ = cos
1
✓
0.45
p
182.817
◆
⇡ 1.5375 ⇡ 88.09 .
Note that it is important to maintain as much precision as possible until the last step, or roundo↵
errors may build up.
28. As in Example 1.21, we begin by calculating u · v and the norms of the two vectors:
u · v = 1 · ( 3) + 2 · 1 + 3 · 2 + 4 · ( 2) = 3,
p
p
kuk = 12 + 22 + 32 + 42 = 30,
p
p
kvk = ( 3)2 + 12 + 22 + ( 2)2 = 18.
So if ✓ is the angle between u and v, then
cos ✓ =
u·v
=
kuk kvk
p
3
p =
30 18
1
p
2 15
so that
✓ = cos
1
✓
1
p
2 15
◆
⇡ 1.7 ⇡ 97.42 .
29. As in Example 1.21, we begin by calculating u · v and the norms of the two vectors:
u · v = 1 · 5 + 2 · 6 + 3 · 7 + 4 · 8 = 70,
p
p
kuk = 12 + 22 + 32 + 42 = 30,
p
p
kvk = 52 + 62 + 72 + 82 = 174.
So if ✓ is the angle between u and v, then
u·v
70
35
cos ✓ =
=p p
= p
kuk kvk
30 174
3 145
so that
✓ = cos
1
✓
35
p
3 145
◆
⇡ 0.2502 ⇡ 14.34 .
30. To show that 4ABC is right, we need only show that one pair of its sides meets at a right angle. So
# »
# »
# »
let u = AB, v = BC, and w = AC. Then we must show that one of u · v, u · w or v · w is zero in
order to show that one of these pairs is orthogonal. Then
# »
u = AB = [1
# »
2] = [4, 2] , v = BC = [4 1, 6
# »
w = AC = [4 ( 3), 6 2] = [7, 4] ,
( 3), 0
0] = [3, 6] ,
and
u · v = 4 · 3 + ( 2) · 6 = 12
12 = 0.
# » # »
Since this dot product is zero, these two vectors are orthogonal, so that AB ? BC and thus 4ABC is
a right triangle. It is unnecessary to test the remaining pairs of sides.
14
CHAPTER 1. VECTORS
31. To show that 4ABC is right, we need only show that one pair of its sides meets at a right angle. So
# »
# »
# »
let u = AB, v = BC, and w = AC. Then we must show that one of u · v, u · w or v · w is zero in
order to show that one of these pairs is orthogonal. Then
# »
u = AB = [ 3 1, 2 1, ( 2) ( 1)] = [ 4, 1, 1] ,
# »
v = BC = [2 ( 3), 2 2, 4 ( 2)] = [5, 0, 2],
# »
w = AC = [2 1, 2 1, 4 ( 1)] = [1, 1, 3],
and
u·v =
u·w =
4·5+1·0
4·1+1·1
1 · ( 2) =
18
1 · ( 3) = 0.
# » # »
Since this dot product is zero, these two vectors are orthogonal, so that AB ? AC and thus 4ABC is
a right triangle. It is unnecessary to test the remaining pair of sides.
32. As in Example 1.22, the dimensions of the cube do not matter, so we work with a cube with side length
1. Since the cube is symmetric, we need only consider one diagonal and adjacent edge. Orient the cube
as shown in Figure 1.34; take the diagonal to be [1, 1, 1] and the adjacent edge to be [1, 0, 0]. Then the
angle ✓ between these two vectors satisfies
✓
◆
1·1+1·0+1·0
1
1
p p
cos ✓ =
=p ,
so
✓ = cos 1 p
⇡ 54.74 .
3 1
3
3
Thus the diagonal and an adjacent edge meet at an angle of 54.74 .
33. As in Example 1.22, the dimensions of the cube do not matter, so we work with a cube with side length
1. Since the cube is symmetric, we need only consider one pair of diagonals. Orient the cube as shown
in Figure 1.34; take the diagonals to be u = [1, 1, 1] and v = [1, 1, 0] [0, 0, 1] = [1, 1, 1]. Then the
dot product is
u · v = 1 · 1 + 1 · 1 + 1 · ( 1) = 1 + 1 1 = 1 6= 0.
Since the dot product is nonzero, the diagonals are not orthogonal.
34. To show a parallelogram is a rhombus, it suffices to show that its diagonals are perpendicular (Euclid).
But
2 3 2 3
2
1
d1 · d2 = 425 · 4 15 = 2 · 1 + 2 · ( 1) + 0 · 3 = 0.
0
3
To determine its side length, note that since the diagonals are perpendicular, one half of each diagonal
are the legs of a right triangle whose hypotenuse is one side of the rhombus. So we can use the
Pythagorean Theorem. Since
2
2
kd1 k = 22 + 22 + 02 = 8,
kd2 k = 12 + ( 1)2 + 32 = 11,
we have for the side length
s2 =
so that s =
p
19
2
✓
kd1 k
2
◆2
+
✓
kd2 k
2
◆2
=
8 11
19
+
=
,
4
4
4
⇡ 2.18.
35. Since ABCD is a rectangle, opposite sides BA and CD are parallel and congruent. So we can use
# »
the method of Example 1.1 in Section 1.1 to find the coordinates of vertex D: we compute BA =
# »
# »
[1 3, 2 6, 3 ( 2)] = [ 2, 4, 5]. If BA is then translated to CD, where C = (0, 5, 4), then
D = (0 + ( 2), 5 + ( 4),
4 + 5) = ( 2, 1, 1).
1.2. LENGTH AND ANGLE: THE DOT PRODUCT
15
36. The resultant velocity of the airplane is the sum of the velocity of the airplane and the velocity of the
wind:



200
0
200
r=p+w =
+
=
.
0
40
40
37. Let the x direction be east, in the direction of the current, and the y direction be north, across the
river. The speed of the boat is 4 mph north, and the current is 3 mph east, so the velocity of the boat
is



0
3
3
v=
+
=
.
4
0
4
38. Let the x direction be the direction across the river, and the y direction be downstream. Since vt = d,
use the given information to find v, then solve for t and compute
d. Since the speed of the boat is 20

20
km/h and the speed of the current is 5 km/h, we have v =
. The width of the river is 2 km, and
 5
2
the distance downstream is unknown; call it y. Then d =
. Thus
y


20
2
vt =
t=
.
5
y
Thus 20t = 2 so that t = 0.1, and then y = 5 · 0.1 = 0.5. Therefore
(a) Ann lands 0.5 km, or half a kilometer, downstream;
(b) It takes Ann 0.1 hours, or six minutes, to cross the river.
Note that the river flow does not increase the time required to cross the river, since its velocity is
perpendicular to the direction of travel.
39. We want to find the angle between Bert’s resultant vector, r, and his velocity vector upstream, v. Let
the first coordinate of the vector be the direction across the river, and the second
be the direction

x
upstream. Bert’s velocity vector directly across the river is unknown, say u =
. His velocity vector
0

0
upstream compensates for the downstream flow, so v =
. So the resultant vector is r = u + v =
1



x
0
x
+
=
. Since Bert’s speed is 2 mph, we have krk = 2. Thus
0
1
1
2
x2 + 1 = krk = 4,
so that
x=
p
3.
If ✓ is the angle between r and v, then
p
r·v
3
cos ✓ =
=
,
krk kvk
2
so that
✓ = cos
40. We have
u·v
( 1) · ( 2) + 1 · 4
proju v =
u=
u·u
( 1) · ( 1) + 1 · 1

1
1
=3
1
p !
3
= 60 .
2

1
=
1

3
.
3
A graph of the situation is (with proju v in gray, and the perpendicular from v to the projection also
drawn)
16
CHAPTER 1. VECTORS
4
v
3
proju !v"
2
1
u
!3
41. We have
u·v
proju v =
u=
u·u
!2
!1
3
5 ·1+
3 3
·
5 5 +
4
5
4
5
·
·2
4
5
"
3
5
4
5
#
=
1
u=
1
u.
A graph of the situation is (with proju v in gray, and the perpendicular from v to the projection also
drawn)
2
v
1
proju !v"
1
u
42. We have
proju v =
u·v
u=
u·u
43. We have
proju v =
1
2
1
2
·
1
2
+
·2
1
4
1
4
·2
1
4
1
2
+
· ( 2)
1
2
1
2
2
6
4
3
1
2
17
45
1
2
2
u·v
1 · 2 + ( 1) · ( 3) + 1 · ( 1) + ( 1) · ( 2) 6
6
u=
6
u·u
1 · 1 + ( 1) · ( 1) + 1 · 1 + ( 1) · ( 1) 4
44. We have
proju v =
=
2
86
4
3
3
1
2
17
45
1
2
2
6
=4
3
4
3
27
.
35
4
3
3
2 3 2
1
1
7
6
6
17 6 6 17
7 6
7= 6 7=6
15 4 4 15 4
1
1



u·v
0.5 · 2.1 + 1.5 · 1.2 0.5
2.85 0.5
0.57
u=
=
=
= 1.14u.
1.71
u·u
0.5 · 0.5 + 1.5 · 1.5 1.5
2.5 1.5
3
3
2
37
27
37
5
2
3
2
=
3
u.
2
1.2. LENGTH AND ANGLE: THE DOT PRODUCT
17
45. We have
2
3
3.01
0.33 · 4.25 + 2.52 · ( 1.66) 4
0.335
0.33 · ( 0.33) + 2.52 · 2.52
2.52
2
3 2
3
3.01
0.301
1.5523 4
0.335 ⇡ 4 0.0335 ⇡
=
15.5194
2.52
0.252
u·v
3.01 · 1.34
proju v =
u=
u·u
3.01 · 3.01
# »
46. Let u = AB =

2
2
1
u.
10



# »
1
1
4 1
3
=
and v = AC =
=
.
( 1)
3
0 ( 1)
1
(a) To compute the projection, we need


1
3
u·v =
·
= 6,
3
1
Thus
u·u=


1
1
·
= 10.
3
3
" # " #
3
u·v
3 1
proju v =
u=
= 59 ,
u·u
5 3
5
so that
v
" #
3
proju v =
1
" #
3
5
9
5
Then
kuk =
so that finally
p
12
+
32
=
p
kv
10,
proju vk =
=
"
s✓
12
5
4
5
12
5
#
.
◆2
p
✓ ◆2
4
4 10
+
=
,
5
5
p
1
1p
4 10
A = kuk kv proju vk =
10 ·
= 4.
2
2
5
p
p
p
(b) We already know u · v = 6 and kuk = 10 from part (a). Also, kvk = 32 + 12 = 10. So
u·v
6
3
=p p = ,
kuk kvk
5
10 10
cos ✓ =
so that
sin ✓ =
Thus
p
1
cos2 ✓ =
s
1
✓ ◆2
3
4
= .
5
5
1
1p p
4
A = kuk kvk sin ✓ =
10 10 · = 4.
2
2
5
2
3 2 3
2
3 2 3
4 3
1
5 3
2
# »
# »
47. Let u = AB = 4 2 ( 1)5 = 4 15 and v = AC = 40 ( 1)5 = 4 15.
6 4
2
2 4
2
(a) To compute the projection, we need
2 3 2 3
1
2
u · v = 4 15 · 4 15 =
2
2
3,
Thus
proju v =
u·v
u=
u·u
2
3 2 3
1
1
u · u = 4 15 · 4 15 = 6.
2
2
2
3 2 3
1
1
3 6 7 6 21 7
4 15 = 4 2 5 ,
6
2
1
18
CHAPTER 1. VECTORS
so that
v
Then
kuk =
p
2
3
2
6 7
proju v = 4 15
2
p
12 + ( 1)2 + 22 =
so that finally
so that
kv
6
4
3
1
2
1 7
2 5
1
2
6
=4
3
5
2
17
25
1
s✓ ◆
p
✓ ◆2
2
5
1
30
2
proju vk =
+
+ ( 1) =
,
2
2
2
p
p
1p
30
3 5
proju vk =
6·
=
.
2
2
2
p
p
3 and kuk = 6 from part (a). Also, kvk = 22 + 12 + ( 2)2 = 3.
p
u·v
3
6
cos ✓ =
= p =
,
kuk kvk
6
3 6
A=
(b) We already know u · v =
So
6,
2
1
kuk kv
2
sin ✓ =
Thus
p
1
v
u
u
2
cos ✓ = t1
p !2
p
6
30
=
.
6
6
p
p
1
1p
30
3 5
kuk kvk sin ✓ =
6·3·
=
.
2
2
6
2
48. Two vectors u and v are orthogonal if and only if their dot product u · v = 0. So we set u · v = 0 and
solve for k:


1
2
k+1
u·v =
·
= 0 ) 2(k + 1) + 3(k 1) = 0 ) 5k 1 = 0 ) k = .
3
k 1
5
A=
Substituting into the formula for v gives
v=
As a check, we compute
"
1
5
1
5
# "
+1
=
1
" # "
2
u·v =
·
3
6
5
4
5
#
=
12
5
6
5
4
5
#
.
12
= 0,
5
and the vectors are indeed orthogonal.
49. Two vectors u and v are orthogonal if and only if their dot product u · v = 0. So we set u · v = 0 and
solve for k:
2 3 2 23
1
k
u · v = 4 15 · 4 k 5 = 0 ) k 2 k 6 = 0 ) (k + 2)(k 3) = 0 ) k = 2, 3.
2
3
Substituting into the formula for v gives
2
3 2 3
( 2)2
4
k = 2 : v 1 = 4 2 5 = 4 25 ,
3
3
As a check, we compute
2 3 2 3
1
4
u · v1 = 4 15 · 4 25 = 1 · 4
2
3
k=
2
3 2 3
32
9
3 : v2 = 4 3 5 = 4 35 .
3
3
3 2 3
1
9
1 · ( 2) + 2 · ( 3) = 0, u · v2 = 4 15 · 4 35 = 1 · 9
2
3
and the vectors are indeed orthogonal.
2
1 · 3 + 2 · ( 3) = 0
1.2. LENGTH AND ANGLE: THE DOT PRODUCT
19
50. Two vectors u and v are orthogonal if and only if their dot product u · v = 0. So we set u · v = 0 and
solve for y in terms of x:


3
x
u·v =
·
= 0 ) 3x + y = 0 ) y = 3x.
1
y
Substituting y =
3x back into the formula for v gives


x
1
v=
=x
.
3x
3
Thus any vector orthogonal to
u·v =

3
is a multiple of
1

3
·
1

x
= 3x
3x

1
. As a check,
3
3x = 0 for any value of x,
so that the vectors are indeed orthogonal.
51. As noted in
0 is orthogonal to all vectors in
 the remarks just prior
 to Example 1.16, the zero vector

a
x
a
2
R . So if
= 0, any vector
will do. Now assume that
6= 0; that is, that either a or b is
b
y
b
nonzero. Two vectors u and v are orthogonal if and only if their dot product u · v = 0. So we set
u · v = 0 and solve for y in terms of x:


a
x
u·v =
·
= 0 ) ax + by = 0.
b
y
First assume b 6= 0. Then y =
a
b x,
so substituting back into the expression for v we get
"
#
" #
" #
x
1
b
x
v=
=x
=
.
a
a
b
x
a
b
b
Next, if b = 0, then a 6= 0, so that x =
"
v=
and substituting back into the expression for v gives
" #
" #
b
b
y
b
y
a
a =
=y
.
a
y
1
a
So in either case, a vector orthogonal to
check, note that

a
·
b

b
a y,
#

a
, if it is not the zero vector, is a multiple of
b
rb
= rab
ra

b
. As a
a
rab = 0 for all values of r.
52. (a) The geometry of the vectors in Figure 1.26 suggests that if ku + vk = kuk + kvk, then u and v
point in the same direction. This means that the angle between them must be 0. So we first prove
Lemma 1. For all vectors u and v in R2 or R3 , u · v = kuk kvk if and only if the vectors point
in the same direction.
Proof. Let ✓ be the angle between u and v. Then
cos ✓ =
u·v
,
kuk kvk
so that cos ✓ = 1 if and only if u · v = kuk kvk. But cos ✓ = 1 if and only if ✓ = 0, which means
that u and v point in the same direction.
20
CHAPTER 1. VECTORS
We can now show
Theorem 2. For all vectors u and v in R2 or R3 , ku + vk = kuk + kvk if and only if u and v
point in the same direction.
Proof. First assume that u and v point in the same direction. Then u · v = kuk kvk, and thus
2
ku + vk =u · u + 2u · v + v · v
2
By Example 1.9
2
= kuk + 2u · v + kvk
2
2
Since w · w = kwk for any vector w
2
= kuk + 2 kuk kvk + kvk
By the lemma
2
= (kuk + kvk) .
Since ku + vk and kuk+kvk are both nonnegative, taking square roots gives ku + vk = kuk+kvk.
For the other direction, if ku + vk = kuk + kvk, then their squares are equal, so that
2
2
2
(kuk + kvk) = kuk + 2 kuk kvk + kvk and
2
ku + vk = u · u + 2u · v + v · v
2
are equal. But kuk = u · u and similarly for v, so that canceling those terms gives 2u · v =
2 kuk kvk and thus u · v = kuk kvk. Using the lemma again shows that u and v point in the same
direction.
(b) The geometry of the vectors in Figure 1.26 suggests that if ku + vk = kuk kvk, then u and v
point in opposite directions. In addition, since ku + vk
0, we must also have kuk
kvk. If
they point in opposite directions, the angle between them must be ⇡. This entire proof is exactly
analogous to the proof in part (a). We first prove
Lemma 3. For all vectors u and v in R2 or R3 , u · v =
in opposite directions.
kuk kvk if and only if the vectors point
Proof. Let ✓ be the angle between u and v. Then
cos ✓ =
u·v
,
kuk kvk
so that cos ✓ = 1 if and only if u · v = kuk kvk. But cos ✓ =
means that u and v point in opposite directions.
1 if and only if ✓ = ⇡, which
We can now show
Theorem 4. For all vectors u and v in R2 or R3 , ku + vk = kuk
point in opposite directions and kuk kvk.
kvk if and only if u and v
Proof. First assume that u and v point in opposite directions and kuk
kuk kvk, and thus
2
ku + vk =u · u + 2u · v + v · v
2
2
= kuk
= (kuk
By Example 1.9
2
= kuk + 2u · v + kvk
kvk. Then u · v =
2
2
2 kuk kvk + kvk
Since w · w = kwk for any vector w
By the lemma
2
kvk) .
Now, since kuk kvk by assumption, we see that both ku + vk and kuk kvk are nonnegative,
so that taking square roots gives ku + vk = kuk kvk. For the other direction, if ku + vk =
1.2. LENGTH AND ANGLE: THE DOT PRODUCT
21
kuk kvk, then first of all, since the left-hand side is nonnegative, the right-hand side must be
as well, so that kuk kvk. Next, we can square both sides of the equality, so that
(kuk
2
2
kvk) = kuk
2
2
2 kuk kvk + kvk and
ku + vk = u · u + 2u · v + v · v
2
are equal. But kuk = u · u and similarly for v, so that canceling those terms gives 2u · v =
2 kuk kvk and thus u · v = kuk kvk. Using the lemma again shows that u and v point in
opposite directions.
53. Prove Theorem 1.2(b) by applying the definition of the dot product:
u · v = u1 (v1 + w1 ) + u2 (v2 + w2 ) + · · · + un (vn + wn )
= u1 v 1 + u1 w 1 + u2 v 2 + u2 w 2 + · · · + un v n + un w n
= (u1 v1 + u2 v2 + · · · + un vn ) + (u1 w1 + u2 w2 + · · · + un wn )
= u · v + u · w.
54. Prove the three parts of Theorem 1.2(d) by applying the definition of the dot product and various
properties of real numbers:
Part 1: For any vector u, we must show u · u
0. But
u · u = u1 u1 + u2 u2 + · · · + un un = u21 + u22 + · · · + u2n .
Since for any real number x we know that x2 0, it follows that this sum is also nonnegative, so
that u · u 0.
Part 2: We must show that if u = 0 then u · u = 0. But u = 0 means that ui = 0 for all i, so that
u · u = 0 · 0 + 0 · 0 + · · · + 0 · 0 = 0.
Part 3: We must show that if u · u = 0, then u = 0. From part 1, we know that
u · u = u21 + u22 + · · · + u2n ,
and that u2i
0 for all i. So if the dot product is to be zero, each u2i must be zero, which means
that ui = 0 for all i and thus u = 0.
55. We must show d(u, v) = ku vk = kv uk = d(v, u). By definition, d(u, v) = ku vk. Then by
Theorem 1.3(b) with c = 1, we have k wk = kwk for any vector w; applying this to the vector u v
gives
ku vk = k (u v)k = kv uk ,
which is by definition equal to d(v, u).
56. We must show that for any vectors u, v and w that d(u, w)  d(u, v) + d(v, w). This is equivalent to
showing that ku wk  ku vk + kv wk. Now substitute u v for x and v w for y in Theorem
1.5, giving
ku wk = k(u v) + (v w)k  ku vk + kv wk .
57. We must show that d(u, v) = ku vk = 0 if and only if u = v. This follows immediately from
Theorem 1.3(a), kwk = 0 if and only if w = 0, upon setting w = u v.
58. Apply the definitions:
u · cv = [u1 , u2 , . . . , un ] · [cv1 , cv2 , . . . , cvn ]
= u1 cv1 + u2 cv2 + · · · + un cvn
= cu1 v1 + cu2 v2 + · · · + cun vn
= c(u1 v1 + u2 v2 + · · · + un vn )
= c(u · v).
22
CHAPTER 1. VECTORS
59. We want to show that ku vk kuk kvk. This is equivalent to showing that kuk  ku
This follows immediately upon setting x = u v and y = v in Theorem 1.5.
vk + kvk.
60. If u · v = u · w, it does not follow that v = w. For example, since 0 · v = 0 for every vector v in Rn ,
the zero vector is orthogonal to every vector v. So if u = 0 in the above equality, we know nothing
about v and w. (as an example, 0 · [1, 2] = 0 · [ 17, 12]). Note, however, that u · v = u · w implies
that u · v u · w = u(v w) = 0, so that u is orthogonal to v w.
2
2
61. We must show that (u + v)(u v) = kuk
kvk for all vectors in Rn . Recall that for any w in Rn
2
that w · w = kwk , and also that u · v = v · u. Then
(u + v)(u
v) = u · u + v · u
u·v
2
v · v = kuk + u · v
2
u·v
kvk = kuk
2
62. (a) Let u, v 2 Rn . Then
2
ku + vk + ku
2
vk = (u + v) · (u + v) + (u
v) · (u
v)
= (u · u + v · v + 2u · v) + (u · u + v · v 2u · v)
⇣
⌘
⇣
⌘
2
2
2
2
= kuk + kvk + 2u · v + kuk + kvk
2u · v
2
2
= 2 kuk + 2 kvk .
(b) Part (a) tells us that the sum of the
squares of the lengths of the diagonals
of a parallelogram is equal to the sum
of the squares of the lengths of its four
sides.
u"v
u!v
u
v
63. Let u, v 2 Rn . Then
1
2
ku + vk
4
1
ku
4
2
1
[(u + v) · (u + v) ((u v) · (u v)]
4
1
= [(u · u + v · v + 2u · v) (u · u + v · v 2u · v)]
4
⌘ ⇣
⌘
i
1 h⇣
2
2
2
2
=
kuk
kuk + kvk
kvk + 4u · v
4
= u · v.
vk =
64. (a) Let u, v 2 Rn . Then using the previous exercise,
ku + vk = ku
2
vk , ku + vk = ku
2
vk
2
2
, ku + vk
ku vk = 0
1
1
2
2
, ku + vk
ku vk = 0
4
4
,u·v =0
, u and v are orthogonal.
2
kvk .
1.2. LENGTH AND ANGLE: THE DOT PRODUCT
23
(b) Part (a) tells us that a parallelogram is
a rectangle if and only if the lengths of
its diagonals are equal.
u"v
u!v
u
v
2
2
2
2
65. (a) By Exercise 55, (u+v)·(u v) = kuk kvk . Thus (u+v)·(u v) = 0 if and only if kuk = kvk .
It follows immediately that u + v and u v are orthogonal if and only if kuk = kvk.
(b) Part (a) tells us that the diagonals of a
parallelogram are perpendicular if and
u"v
only if the lengths of its sides are equal,
i.e., if and only if it is a rhombus.
u!v
u
v
2
2
2
2
66. From Example 1.9 and the fact that w · w = kwk
q, we have ku + vk = kuk + 2u · v + kvk . Taking
2
2
the square root of both sides yields ku + vk = kuk + 2u · v + kvk . Now substitute in the given
p
values kuk = 2, kvk = 3, and u · v = 1, giving
r
⇣p ⌘2 p
p
ku + vk = 22 + 2 · 1 +
3 = 4 + 2 + 3 = 9 = 3.
67. From Theorem 1.4 (the Cauchy-Schwarcz inequality), we have |u · v|  kuk kvk. If kuk = 1 and
kvk = 2, then |u · v|  2, so we cannot have u · v = 3.
68. (a) If u is orthogonal to both v and w, then u · v = u · w = 0. Then
u · (v + w) = u · v + u · w = 0 + 0 = 0,
so that u is orthogonal to v + w.
(b) If u is orthogonal to both v and w, then u · v = u · w = 0. Then
u · (sv + tw) = u · (sv) + u · (tw) = s(u · v) + t(u · w) = s · 0 + t · 0 = 0,
so that u is orthogonal to sv + tw.
69. We have
u · (v
⇣
proju v) = u · v
=u·v
=u·v
=u·v
= 0.
⇣u · v⌘ ⌘
u
u
·u ⌘
⇣u
·v
u·
u
u
·
⇣ u · v ⌘u
(u · u)
u·u
u·v
24
CHAPTER 1. VECTORS
u·v
u·u u
70. (a) proju (proju v) = proju
=
u·v
u·u
proju u =
u·v
u·u u
= proju v.
(b) Using part (a),
proju (v
⇣
proju v) = proju v
u · v ⌘ ⇣u · v⌘
u =
u
u·u
u·u
⇣u · v⌘
proj u
u·u ⇣ u⌘
u·v
=
u
u·u
(c) From the diagram, we see that
proju vku, so that proju (proju v) =
proju v. Also, (v proju v) ? u, so
that proju (v proju v) = 0.
⇣u · v⌘
u·u
u = 0.
u
proju !v"
v
v ! proju !v"
!proju !v"
71. (a) We have
(u21 + u22 )(v12 + v22 )
(u1 v1 + u2 v2 )2 = u21 v12 + u21 v22 + u22 v12 + u22 v22
= u21 v22 + u22 v12
u21 v12
2u1 v1 u2 v2
u22 v22
2u1 u2 v1 v2
2
= (u1 v2
u2 v 1 ) .
But the final expression is nonnegative since it is a square. Thus the original expression is as well,
showing that (u21 + u22 )(v12 + v22 ) (u1 v1 + u2 v2 )2 0.
(b) We have
(u21 + u22 + u23 )(v12 + v22 + v32 )
(u1 v1 + u2 v2 + u3 v3 )2
= u21 v12 + u21 v22 + u21 v32 + u22 v12 + u22 v22 + u22 v32 + u23 v12 + u23 v22 + u23 v32
u21 v12
=
u21 v22
+
u21 v32
2u1 v1 u2 v2
+
u22 v12
2u1 u2 v1 v2
= (u1 v2
+
u22 v32
+
u22 v22
2u1 v1 u3 v3
u23 v12
u23 v22
2u1 v1 u3 v3
2
u2 v1 ) + (u1 v3
2
+
u23 v32
2u2 v2 u3 v3
2u2 v2 u3 v3
u3 v1 ) + (u3 v2
u2 v 3 ) 2 .
But the final expression is nonnegative since it is the sum of three squares. Thus the original
expression is as well, showing that (u21 + u22 + u23 )(v12 + v22 + v32 ) (u1 v1 + u2 v2 + u3 v3 )2 0.
72. (a) Since proju v =
u·v
u·u u,
we have
proju v · (v
⇣
u·v
u·v ⌘
u· v
u
u·u⇣
u·u
⌘
u·v
u·v
=
u·v
u·u
u·u
u·u
u·v
=
(u · v u · v)
u·u
= 0,
proju v) =
Exploration: Vectors and Geometry
25
so that proju v is orthogonal to v proju v. Since their vector sum is v, those three vectors form
a right triangle with hypotenuse v, so by Pythagoras’ Theorem,
2
2
2
kproju vk  kproju vk + kv
(b)
2
proju vk = kvk .
Since norms are always nonnegative, taking square roots gives kproju vk  kvk.
kproju vk  kvk ()
()
()
⇣u · v⌘
u  kvk
u·u !
u·v
u  kvk
2
kuk
u·v
kuk
2
kuk  kvk
|u · v|
 kvk
kuk
() |u · v|  kuk kvk ,
()
which is the Cauchy-Schwarcz inequality.
73. Suppose proju v = cu. From the figure, we see that cos ✓ =
two expressions are equal, i.e.,
ckuk
kvk .
But also cos ✓ =
u·v
kukkvk .
Thus these
c kuk
u·v
u·v
u·v
u·v
=
) c kuk =
)c=
=
.
kvk
kuk kvk
kuk
kuk kuk
u·u
74. The basis for induction is the cases n = 1 and n = 2. The n = 1 case is the assertion that kv1 k  kv2 k,
which is obviously true. The n = 2 case is the Triangle Inequality, which is also true.
Now assume the statement holds for n = k
2; that is, for any v1 , v2 , . . . , vk ,
kv1 + v2 + · · · + vk k  kv1 k + kv2 k + · · · + kvk k .
Let v1 , v2 , . . . , vk , vk+1 be any vectors. Then
kv1 + v2 + · · · + vk + vk+1 k = kv1 + v2 + · · · + (vk + vk+1 )k
 kv1 k + kv2 k + · · · + kvk + vk+1 k
using the inductive hypothesis. But then using the Triangle Inequality (or the case n = 2 in this
theorem), kvk + vk+1 k  kvk k + kvk+1 k. Substituting into the above gives
kv1 + v2 + · · · + vk + vk+1 k  kv1 k + kv2 k + · · · + kvk + vk+1 k
 kv1 k + kv2 k + · · · + kvk k + kvk+1 k ,
which is what we were trying to prove.
Exploration: Vectors and Geometry
# »
# »
# »
1. As in Example 1.25, let p = OP . Then p a = AP = 13 AB = 13 (b a), so that p = a + 13 (b a) =
# »
# »
1
1
a = AP =
3 (2a + b). More generally, if P is the point n of the way from A to B along AB, then p
1# »
1
a), so that p = a + n1 (b a) = n1 ((n 1)a + b).
n AB = n (b
# »
2. Use the notation that the vector OX is written x. Then from exercise 1, we have p = 12 (a + c) and
q = 12 (b + c), so that
# »
PQ = q
p=
1
(b + c)
2
1
1
(a + c) = (b
2
2
a) =
1# »
AB.
2
26
CHAPTER 1. VECTORS
# »
# »
# »
# »
# »
3. Draw AC. Then from exercise 2, we have P Q = 12 AB = SR. Also draw BD. Again from exercise 2,
# »
# » # »
we have P S = 12 BD = QR. Thus opposite sides of the quadrilateral P QRS are equal. They are also
parallel: indeed, 4BP Q and 4BAC are similar, since they share an angle and BP : BA = BQ : BC.
Thus \BP Q = \BAC; since these angles are equal, P QkAC. Similarly, SRkAC so that P QkSR. In
a like manner, we see that P SkRQ. Thus P QRS is a parallelogram.
4. Following the hint, we find m, the point that is two-thirds of the distance from A to P . From exercise
1, we have
✓
◆
1
1
1
1
1
p = (b + c), so that m = (2p + a) =
2 · (b + c) + a = (a + b + c).
2
3
3
2
3
Next we find m0 , the point that is two-thirds of the distance from B to Q. Again from exercise 1, we
have
✓
◆
1
1
1
1
1
0
q = (a + c), so that m = (2q + b) =
2 · (a + c) + b = (a + b + c).
2
3
3
2
3
Finally we find m00 , the point that is two-thirds of the distance from C to R. Again from exercise 1,
we have
✓
◆
1
1
1
1
1
00
r = (a + b), so that m = (2r + c) =
2 · (a + b) + c = (a + b + c).
2
3
3
2
3
Since m = m0 = m00 , all three medians intersect at the centroid, G.
# »
# »
# » # »
5. With notation as in the figure, we know that AH is orthogonal to BC; that is, AH · BC = 0. Also
# »
# »
# » # »
# » # »
BH is orthogonal to AC; that is, BH · AC = 0. We must show that CH · AB = 0. But
# » # »
AH · BC = 0 ) (h a) · (b c) = 0 ) h · b h · c a · b + a · c = 0
# » # »
BH · AC = 0 ) (h b) · (c a) = 0 ) h · c h · a b · c + b · a = 0.
Adding these two equations together and canceling like terms gives
0=h·b
h·a
c · b + a · c = (h
c) · (b
# » # »
a) = CH · AB,
so that these two are orthogonal. Thus all the altitudes intersect at the orthocenter H.
# »
# »
# »
# »
6. We are given that QK is orthogonal to AC and that P K is orthogonal to CB, and must show that
# »
# »
RK is orthogonal to AB. By exercise 1, we have q = 12 (a + c), p = 12 (b + c), and r = 12 (a + b). Thus
✓
◆
1
# » # »
QK · AC = 0 ) (k q) · (c a) = 0 ) k
(a + c) · (c a) = 0
2
✓
◆
1
# » # »
P K · CB = 0 ) (k p) · (b c) = 0 ) k
(b + c) · (b c) = 0.
2
Expanding the two dot products gives
k·c
k·a
k·b
k·c
1
1
a·c+ a·a
2
2
1
1
b·b+ b·c
2
2
1
1
c·c+ a·c=0
2
2
1
1
c · b + c · c = 0.
2
2
Add these two together and cancel like terms to get
✓
◆
1
1
1
0=k·b k·a
b·b+ a·a= k
(b + a) · (b
2
2
2
a) = (k
r) · (b
# » # »
a) = RK · AB.
# »
# »
Thus RK and AB are indeed orthogonal, so all the perpendicular bisectors intersect at the circumcenter.
1.3. LINES AND PLANES
27
2
2
7. Let O, the center of the circle, be the origin. Then b = a and kak = kck = r2 where r is the radius
# »
# »
of the circle. We want to show that AC is orthogonal to BC. But
# » # »
AC · BC = (c a) · (c b)
= (c
2
a) · (c + a)
= kck + c · a
= (a · c
kak
c · a) + (r
2
a·c
2
r2 ) = 0.
Thus the two are orthogonal, so that \ACB is a right angle.
8. As in exercise 5, we first find m, the point that is halfway from P to R. We have p = 12 (a + b) and
r = 12 (c + d), so that
✓
◆
1
1 1
1
1
m = (p + r) =
(a + b) + (c + d) = (a + b + c + d).
2
2 2
2
4
Similarly, we find m0 , the point that is halfway from Q to S. We have q = 12 (b + c) and s = 12 (a + d),
so that
✓
◆
1
1 1
1
1
0
m = (q + s) =
(b + c) + (a + d) = (a + b + c + d).
2
2 2
2
4
#
»
#
»
Thus m = m0 , so that P R and QS intersect at their mutual midpoints; thus, they bisect each other.
1.3
Lines and Planes
1. (a) The normal form is n · (x

x
(b) Letting x =
, we get
y


✓
3
p) = 0, or
· x
2
3
·
2
✓
x
y

0
0
◆

0
0
◆
= 0.


3
x
=
·
= 3x + 2y = 0.
2
y
The general form is 3x + 2y = 0.

✓
3
· x
4
2. (a) The normal form is n · (x p) = 0, or

x
(b) Letting x =
, we get
y

✓
 ◆ 

3
x
1
3
x
·
=
·
4
y
2
4
y

(b)
4. (a)
(b)
◆
= 0.
1
= 3(x
2
Expanding and simplifying gives the general form 3x
3. (a)
1
2
4y =
1)
4(y
2) = 0.
5.


1
1
In vector form, the equation of the line is x = p + td, or x =
+t
.
0
3



x
x
1 t
Letting x =
and expanding the vector form from part (a) gives
=
, which yields
y
y
3t
the parametric form x = 1 t, y = 3t.


4
1
In vector form, the equation of the line is x = p + td, or x =
+t
.
4
1

x
Letting x =
and expanding the vector form from part (a) gives the parametric form x = 4+t,
y
y = 4 + t.
28
CHAPTER 1. VECTORS
5. (a)
(b)
6. (a)
(b)
7. (a)
(b)
2 3
2 3
0
1
In vector form, the equation of the line is x = p + td, or x = 405 + t 4 15.
0
4
2 3
x
Letting x = 4y 5 and expanding the vector form from part (a) gives the parametric form x = t,
z
y = t, z = 4t.
2 3
2 3
3
2
In vector form, the equation of the line is x = p + td, or x = 4 05 + t 455.
2
0
2 3
2 3
2
3
x
x
3 + 2t
Letting x = 4y 5 and expanding the vector form from part (a) gives 4y 5 = 4 5t 5, which
z
z
2
yields the parametric form x = 3 + 2t, y = 5t, z = 2.
2 3 0
2 31
3
0
The normal form is n · (x p) = 0, or 425 · @x 415A = 0.
1
0
2 3
x
Letting x = 4y 5, we get
z
2 3 02 3 2 31 2 3 2
3
3
x
0
3
x
425 · @4y 5 415A = 425 · 4y 15 = 3x + 2(y 1) + z = 0.
1
z
0
1
z
Expanding and simplifying gives the general form 3x + 2y + z = 2.
2 3 0
2 31
2
3
8. (a) The normal form is n · (x p) = 0, or 455 · @x 4 05A = 0.
0
2
2 3
x
(b) Letting x = 4y 5, we get
z
2 3 02 3 2 31 2 3 2
3
2
x
3
2
x 3
455 · @4y 5 4 05A = 455 · 4 y 5 = 2(x 3) + 5y = 0.
0
z
2
0
z+2
Expanding and simplifying gives the general form 2x + 5y = 6.
9. (a) In vector form, the equation of the line is x = p + su + tv, or
2 3
2 3
2 3
0
2
3
x = 405 + s 415 + t 4 25 .
0
2
1
2 3
x
(b) Letting x = 4y 5 and expanding the vector form from part (a) gives
z
2 3 2
3
x
2s 3t
4y 5 = 4 s + 2t 5
z
2s + t
which yields the parametric form the parametric form x = 2s
3t, y = s + 2t, z = 2s + t.
1.3. LINES AND PLANES
29
10. (a) In vector form, the equation of the line is x = p + su + tv, or
2 3
2 3
2 3
6
0
1
x = 4 45 + s 415 + t 4 15 .
3
1
1
2 3
x
(b) Letting x = 4y 5 and expanding the vector form from part (a) gives the parametric form x = 6 t,
z
y = 4 + s + t, z = 3 + s + t.
11. Any pair of points on ` determine a direction vector, so we use P and Q. We choose P to represent
# »
the point on the line. Then a direction vector forthe line is
 d = P Q = (3, 0) (1, 2) = (2, 2). The
1
2
vector equation for the line is x = p + td, or x =
+t
.
2
2
12. Any pair of points on ` determine a direction vector, so we use P and Q. We choose P to represent the
# »
point on the line. Then a direction vector for the line is
2d=
3 P Q2= (3 2, 1, 3) (0, 1, 1) = ( 2, 0, 4).
0
2
The vector equation for the line is x = p + td, or x = 4 15 + t 4 05.
1
4
13. We must find two direction vectors, u and v. Since P , Q, and R lie in a plane, we compute We get
two direction vectors
# »
u = P Q = q p = (4, 0, 2) (1, 1, 1) = (3, 1, 1)
# »
v = P R = r p = (0, 1, 1) (1, 1, 1) = ( 1, 0, 2).
Since u and v are not scalar multiples of each other, they will serve as direction vectors (if they were
parallel to each other, we2would
have
3
2 3not a2plane
3 but a line). So the vector equation for the plane is
1
3
1
x = p + su + tv, or x = 415 + s 4 15 + t 4 05.
1
1
2
14. We must find two direction vectors, u and v. Since P , Q, and R lie in a plane, we compute We get
two direction vectors
# »
u = P Q = q p = (1, 0, 1) (1, 1, 0) = (0, 1, 1)
# »
v = P R = r p = (0, 1, 1) (1, 1, 0) = ( 1, 0, 1).
Since u and v are not scalar multiples of each other, they will serve as direction vectors (if they were
parallel to each other, we2would
have
3
2 3not a2plane
3 but a line). So the vector equation for the plane is
1
0
1
x = p + su + tv, or x = 415 + s 4 15 + t 4 05.
0
1
1
15. The parametric and associated vector forms x = p + td found below are not unique.
(a) As in the remarks prior to Example 1.20, we start by letting x = t. Substituting x = t into
y = 3x 1 gives
So we get parametric equations x = t, y = 3t 1, and corresponding
 y=
 3t 1. 
x
0
1
vector form
=
+t
.
y
1
3
(b) In this case since the coefficient of y is 2, we start by letting x = 2t. Substituting x = 2t into
3x + 2y = 5 gives 3 · 2t + 2y = 5, which gives y = 3t + 52 . So we get parametric equations x = 2t,
y = 52 3t, with corresponding vector equation
" # " #
" #
x
0
2
= 5 +t
.
y
3
2
30
CHAPTER 1. VECTORS
Note that theequation was of the form ax + by = c with a = 3, b = 2, and that a direction vector
b
was given by
. This is true in general.
a
16. Note that x = p + t(q p) is the line that passes through p (when t = 0) and q (when t = 1). We
write d = q p; this is a direction vector for the line through p and q.
(a) As noted above, the line p + td passes through P at t = 0 and through Q at t = 1. So as t varies
from 0 to 1, the line describes the line segment P Q.
(b) As shown in Exploration: Vectors and Geometry, to find the midpoint of P Q, we start at P
# »
and travel half the length of P Q in the direction of the vector P Q = q p. That is, the midpoint
of P Q is the head of the vector p + 12 (q p). Since x = p + t(q p), we see that this line passes
through the midpoint at t = 12 , and that the midpoint is in fact p + 12 (q p) = 12 (p + q).
(c) From part (b), the midpoint is
(d) From part (b), the midpoint is
1
2
1
2
([2, 3] + [0, 1]) = 12 [2, 2] = [1, 1].
⇥
([1, 0, 1] + [4, 1, 2]) = 12 [5, 1, 1] = 52 , 12 ,
1
2
⇤
.
(e) Again from Exploration: Vectors and Geometry, the vector whose head is 13 of the way from
P to Q along P Q is x1 = 13 (2p + q). Similarly, the vector whose head is 23 of the way from P to
Q along P Q is also the vector one third of the way from Q to P along QP ; applying the same
formula gives for this point x2 = 13 (2q + p). When p = [2, 3] and q = [0, 1], we get
1
x1 = (2[2, 3] + [0, 1]) =
3
1
x2 = (2[0, 1] + [2, 3]) =
3

1
4
[4, 5] =
,
3
3

1
2
[2, 1] =
,
3
3
5
3
1
.
3
(f ) Using the formulas from part (e) with p = [1, 0, 1] and q = [4, 1, 2] gives

1
1
(2[1, 0, 1] + [4, 1, 2]) = [6, 1, 4] = 2,
3
3

1
1
x2 = (2[4, 1, 2] + [1, 0, 1]) = [9, 2, 5] = 3,
3
3
x1 =
1
,
3
2
,
3
4
3
5
.
3
17. A line `1 with slope m1 has equation y = m1 x + b1 , or
m1 x + y = b1 . Similarly, a line `2 with slope
m1
m2 has equation y = m2 x + b2 , or m2 x + y = b2 . Thus the normal vector for `1 is n1 =
, and
1

m2
the normal vector for `2 is n2 =
. Now, `1 and `2 are perpendicular if and only if their normal
1
vectors are perpendicular, i.e., if and only if n1 · n2 = 0. But


m1
m2
n1 · n2 =
·
= m1 m2 + 1,
1
1
so that the normal vectors are perpendicular if and only if m1 m2 +1 = 0, i.e., if and only if m1 m2 =
1.
18. Suppose the line ` has direction vector d, and the plane P has normal vector n. Then if d · n = 0 (d
and n are orthogonal), then the line ` is parallel to the plane P. If on the other hand d and n are
parallel, so that d = n, then ` is perpendicular to P.
2 3
2
(a) Since the general form of P is 2x + 3y z = 1, its normal vector is n = 4 35. Since d = 1n, we
1
see that ` is perpendicular to P.
1.3. LINES AND PLANES
31
2
3
4
(b) Since the general form of P is 4x y + 5z = 0, its normal vector is n = 4 15. Since
5
2 3 2 3
2
4
d · n = 4 35 · 4 15 = 2 · 4 + 3 · ( 1) 1 · 5 = 0,
1
5
` is parallel to P.
2
3
1
(c) Since the general form of P is x y z = 3, its normal vector is n = 4 15. Since
1
2 3 2 3
2
1
d · n = 4 35 · 4 15 = 2 · 1 + 3 · ( 1) 1 · ( 1) = 0,
1
1
` is parallel to P.
2
3
4
(d) Since the general form of P is 4x + 6y 2z = 0, its normal vector is n = 4 65. Since
2
2 3
2 3
2
4
1
1
d = 4 35 = 4 65 = n,
2
2
1
2
` is perpendicular to P.
19. Suppose the plane P1 has normal vector n1 , and the plane P has normal vector n. Then if n1 · n = 0
(n1 and n are orthogonal), then P1 is perpendicular to P. If on the other hand n1 and n are
parallel, so that n1 = cn, then2P13is parallel to P. Note that in this exercise, P1 has the equation
4
4x y + 5z = 2, so that n1 = 4 15.
5
2 3
2
(a) Since the general form of P is 2x + 3y z = 1, its normal vector is n = 4 35. Since
1
2 3 2 3
4
2
n1 · n = 4 15 · 4 35 = 4 · 2 1 · 3 + 5 · ( 1) = 0,
5
1
the normal vectors are perpendicular, and thus P1 is perpendicular to P.
(b) Since the general form of P is 4x
P1 is parallel to P.
3
4
y + 5z = 0, its normal vector is n = 4 15. Since n1 = n,
5
2
2
3
1
(c) Since the general form of P is x y z = 3, its normal vector is n = 4 15. Since
1
2 3 2 3
4
1
n1 · n = 4 15 · 4 15 = 4 · 1 1 · ( 1) + 5 · ( 1) = 0,
5
1
the normal vectors are perpendicular, and thus P1 is perpendicular to P.
32
CHAPTER 1. VECTORS
(d) Since the general form of P is 4x + 6y
2
3
4
2z = 0, its normal vector is n = 4 65. Since
2
3 2 3
4
4
n1 · n = 4 15 · 4 65 = 4 · 4
5
2
2
1 · 6 + 5 · ( 2) = 0,
the normal vectors are perpendicular, and thus P1 is perpendicular to P.
20. Since the vector form is x = p + td, we use the given information to determine
p and d. The general

2
equation of the given line is 2x 3y = 1, so its normal vector is n =
. Our line is perpendicular
3

2
to that line, so it has direction vector d = n =
. Furthermore, since our line passes through the
3

2
point P = (2, 1), we have p =
. Thus the vector form of the line perpendicular to 2x 3y = 1
1
through the point P = (2, 1) is



x
2
2
=
+t
.
y
1
3
21. Since the vector form is x = p + td, we use the given information todetermine p and d. The general
2
equation of the given line is 2x 3y = 1, so its normal vector is n =
. Our line is parallel to that
3

3
(note that d · n = 0). Since our line passes through the point
line, so it has direction vector d =
2

2
P = (2, 1), we have p =
, so that the vector equation of the line parallel to 2x 3y = 1 through
1
the point P = (2, 1) is



x
2
3
=
+t
.
y
1
2
22. Since the vector form is x = p + td, we use the given information to determine p and d. A line is
perpendicular to a plane if its direction vector d is the normal vector n of
2 the
3 plane. The general
1
equation of the given plane is x 3y + 2z = 5, so its normal vector is n = 4 35. Thus the direction
2
2 3
1
vector of our line is d = 4 35. Furthermore, since our line passes through the point P = ( 1, 0, 3), we
2
2 3
1
have p = 4 05. So the vector form of the line perpendicular to x 3y + 2z = 5 through P = ( 1, 0, 3)
3
is
2 3 2 3
2 3
x
1
1
4y 5 = 4 05 + t 4 35 .
z
3
2
23. Since the vector form is x = p + td, we use the given information to determine p and
given line has parametric equations
2 3 2 3
2
x
1
x = 1 t, y = 2 + 3t, z = 2 t, it has vector form 4y 5 = 4 25 + t 4
z
2
d. Since the
3
1
35 .
1
1.3. LINES AND PLANES
33
2
3
1
So its direction vector is 4 35, and this must be the direction vector d of the line we want, which is
1
2 3
1
parallel to the given line. Since our line passes through the point P = ( 1, 0, 3), we have p = 4 05.
3
So the vector form of the line parallel to the given line through P = ( 1, 0, 3) is
2 3 2 3
2 3
1
1
x
4y 5 = 4 05 + t 4 35 .
z
3
1
24. Since the normal form is n · x = n · p, we use the given information to determine n and p. Note that
a plane is parallel to a given plane if their normal vectors
2 are
3 equal. Since the general form of the
6
given plane is 6x y + 2z = 3, its normal vector is n = 4 15, so this must be a normal vector of the
2
desired
plane
as
well.
Furthermore,
since
our
plane
passes
through the point P = (0, 2, 5), we have
2 3
0
p = 4 25. So the normal form of the plane parallel to 6x y + 2z = 3 through (0, 2, 5) is
5
2 3 2 3 2 3 2 3
2 3 2 3
6
x
6
0
6
x
4 15 · 4y 5 = 4 15 · 4 25 or 4 15 · 4y 5 = 12.
2
z
2
5
2
z
25. Using Figure 1.34 in Section 1.2 for reference, we will find a normal vector n and a point vector p for
each of the sides, then substitute into n · x = n · p to get an equation for each plane.
(a) Start with 2P3
1 determined by the face of the cube in the xy-plane. Clearly a normal vector for
1
P1 is n = 405, or any vector parallel to the x-axis. Also, the plane passes through P = (0, 0, 0),
02 3
0
so we set p = 405. Then substituting gives
0
2 3 2 3 2 3 2 3
1
x
1
0
405 · 4y 5 = 405 · 405 or x = 0.
0
z
0
0
So the general equation for P1 is x = 0. Applying the same argument above to the plane P2
determined by the face in the xz-plane gives a general equation of y = 0, and similarly the plane
P3 determined by the face in the xy-plane gives a general equation of z = 0.
Now consider P4 , the plane containing the face parallel to the face 2
in 3
the yz-plane but passing
1
through (1, 1, 1). Since P4 is parallel to P1 , its normal vector is also 405; since it passes through
0
2 3
1
(1, 1, 1), we set p = 415. Then substituting gives
1
2 3 2 3 2 3 2 3
1
x
1
1
405 · 4y 5 = 405 · 415 or x = 1.
0
z
0
1
So the general equation for P4 is x = 1. Similarly, the general equations for P5 and P6 are
y = 1 and z = 1.
34
CHAPTER 1. VECTORS
2 3
x
(b) Let n = 4y 5 be a normal vector for the desired plane P. Since P is perpendicular to the
z
xy-plane, their normal vectors must be orthogonal. Thus
2 3 2 3
x
0
4y 5 · 405 = x · 0 + y · 0 + z · 1 = z = 0.
z
1
2 3
x
Thus z = 0, so the normal vector is of the form n = 4y 5. But the normal vector is also
0
perpendicular to the plane in question, by definition. Since that plane contains both the origin
and (1, 1, 1), the normal vector is orthogonal to (1, 1, 1) (0, 0, 0):
2 3 2 3
x
1
4y 5 · 415 = x · 1 + y · 1 + z · 0 = x + y = 0.
0
1
2 3
x
Thus x + y = 0, so that y = x. So finally, a normal vector to P is given by n = 4 x5 for
0
2 3
1
any nonzero x. We may as well choose x = 1, giving n = 4 15. Since the plane passes through
0
(0, 0, 0), we let p = 0. Then substituting in n · x = n · p gives
2 3 2 3 2 3 2 3
1
x
1
0
4 15 · 4y 5 = 4 15 · 405 , or x y = 0.
0
z
0
0
Thus the general equation for the plane perpendicular to the xy-plane and containing the diagonal
from the origin to (1, 1, 1) is x y = 0.
(c) As in Example 1.22 (Figure 1.34)
2 3 in Section 1.2, use u = [0, 1, 1] and v = [1, 0, 1] as two vectors
x
in the required plane. If n = 4y 5 is a normal vector to the plane, then n · u = 0 = n · v:
z
2 3 2 3
2 3 2 3
x
0
x
1
n · u = 4y 5 · 415 = y + z = 0 ) y = z,
n · v = 4y 5 · 405 = x + z = 0 ) x = z.
z
1
z
1
2 3
2 3
z
1
Thus the normal vector is of the form n = 4 z 5 for any z. Taking z = 1 gives n = 4 15.
z
1
Now, the side diagonals pass through (0, 0, 0), so set p = 0. Then n · x = n · p yields
2 3 2 3 2 3 2 3
1
x
1
0
4 15 · 4y 5 = 4 15 · 405 , or x + y z = 0.
1
z
1
0
The general equation for the plane containing the side diagonals is x + y
z = 0.
26. Finding the distance between points A and B is equivalent to finding d(a, b), where a is the vector
from the origin to A, and similarly for b. Given x = [x, y, z], p = [1, 0, 2], and q = [5, 2, 4], we want
to solve d(x, p) = d(x, q); that is,
p
p
d(x, p) = (x 1)2 + (y 0)2 + (z + 2)2 = (x 5)2 + (y 2)2 + (z 4)2 = d(x, q).
1.3. LINES AND PLANES
35
Squaring both sides gives
(x
x
2
1)2 + (y
2
0)2 + (z + 2)2 = (x
2
2x + 1 + y + z + 4z + 4 = x
2
5)2 + (y
10x + 25 + y
8x + 4y + 12z = 40
2x + y + 3z = 10.
2)2 + (z
2
4y + 4 + z
4)2
2
)
8z + 16
)
)
Thus all such points (x, y, z) lie on the plane 2x + y + 3z = 10.
27. To calculate d(Q, `) =
|ax0 +by0 c|
p
,
a2 +b2
we first put ` into general form. With d =
since then n · d = 0. Then we have

n·x=n·p )


1
1
, we get n =
1
1

 
1
x
1
1
·
=
= 1.
1
y
1
2
Thus x + y = 1 and thus a = b = c = 1. Since Q = (2, 2) = (x0 , y0 ), we have
p
|1 · 2 + 1 · 2 1|
3
3 2
p
d(Q, `) =
=p =
.
2
2
12 + 12
2
3
2
28. Comparing the given equation to x = p + td, we get P = (1, 1, 1) and d = 4 05. As suggested by
3
# »
Figure 1.63, we need to calculate the length of RQ, where R is the point on the line at the foot of the
# »
perpendicular from Q. So if v = P Q, then
# »
Now, v = P Q = q
# »
# »
P R = projd v,
RQ = v
2 3 2 3 2 3
0
1
1
p = 415 415 = 4 05, so that
0
1
1
projd v =
✓
d·v
d·d
Thus
v
◆
d=
2 · ( 1) + 3 · ( 1)
2 · ( 2) + 3 · 3
3
1
6 7
projd v = 4 05
1
Then the distance d(Q, `) from Q to ` is
kv
✓
projd v.
2
2
6
4
2
13
3
◆
2
7 6
05 = 4
3
13
2
3 2
2
6 7 6
4 05 = 4
3
15
13
2
13
3
7
05 .
3
13
3
7
05 .
10
13
2 3
p
3
5 4 5
5p 2
5 13
0 =
projd vk =
3 + 22 =
.
13
13
13
2
0 +by0 +cz0 d|
29. To calculate d(Q, P) = |axp
, we first note that the plane has equation x + y
a2 +b2 +c2
that a = b = 1, c = 1, and d = 0. Also, Q = (2, 2, 2), so that x0 = y0 = z0 = 2. Hence
p
|1 · 2 + 1 · 2 1 · 2 0|
2
2 3
d(Q, P) = p
=p =
.
3
3
12 + 12 + ( 1)2
z = 0, so
36
CHAPTER 1. VECTORS
0 +by0 +cz0 d|
30. To calculate d(Q, P) = |axp
, we first note that the plane has equation x 2y + 2z = 1, so
a2 +b2 +c2
that a = 1, b = 2, c = 2, and d = 1. Also, Q = (0, 0, 0), so that x0 = y0 = z0 = 0. Hence
d(Q, P) =
|1 · 0
p
2·0+2·0
1|
=
12 + ( 2)2 + 22
1
.
3
# »
# »
31. Figure 1.66 suggests that we let v = P Q;
line ` to
 then w = P R = projd v. Comparing

 the given

# »
1
2
1
3
x = p + td, we get P = ( 1, 2) and d =
. Then v = P Q = q p =
=
. Next,
1
2
2
0
w = projd v =
✓
d·v
d·d
◆
d=
✓
1 · 3 + ( 1) · 0
1 · 1 + ( 1) · ( 1)
So
# »
r = p + P R = p + projd v = p + w =
1 1
2, 2
So the point R on ` that is closest to Q is
"
◆
# "
1
+
2
3
1
=
1
2
3
2
3
2
#
=

1
.
1
" #
1
2
1
2
.
.
# »
# »
32. Figure 1.66 suggests that we let
= P Q; then P R = projd v. Comparing
given3line ` to x = p+td,
2 v3
2 3 2 3the 2
2
0
1
1
# »
we get P = (1, 1, 1) and d = 4 05. Then v = P Q = q p = 415 415 = 4 05. Next,
3
0
1
1
projd v =
So
✓
d·v
d·d
◆
d=
✓
2 · ( 1) + 3 · ( 1)
( 2)2 + 32
2 3 2
1
# »
6 7 6
r = p + P R = p + projd v = 415 + 4
1
15
10
13 , 1, 13
So the point R on ` that is closest to Q is
◆
2
3 2
2
6 7 6
4 05 = 4
3
2
13
3
2
15
13
2
13
3
7
05 .
3
13
3
7 6 7
05 = 4 1 5 .
3
13
.
10
13
# »
# »
33. Figure 1.67 suggests we let v = P Q, where P2is some
point on the plane; then QR = projn v. The
3
1
equation of the plane is x + y z = 0, so n = 4 15. Setting y = 0 shows that P = (1, 0, 1) is a point
1
on the plane. Then
2 3 2 3 2 3
2
1
1
# »
v = P Q = q p = 425 405 = 425 ,
2
1
1
so that
projn v =
⇣n · v⌘
n·n
n=
✓
Finally,
# » # »
r = p + PQ + PR = p + v
1·1+1·1 1·1
12 + 12 + ( 1)2
◆
2
3 2
1
6 7 6
4 15 = 4
1
2 3 2 3
1
1
6 7 6 7
projn v = 405 + 425
1
1
Therefore, the point R in P that is closest to Q is
4 4 8
3, 3, 3
.
2
6
4
3
2
3
27
35
2
3
3
2
3
27
.
35
2
3
=
2 3
4
3
647
435 .
8
3
1.3. LINES AND PLANES
37
# »
34. Figure 1.67 suggests we let v = P Q, where P is2 some
3
1
equation of the plane is x 2y + 2z = 1, so n = 4 25.
2
point on the plane. Then
2 3
0
# »
v = P Q = q p = 405
0
so that
projn v =
⇣n · v⌘
n·n
n=
Finally,
# » # »
r = p + PQ + PR = p + v
✓
12
# »
point on the plane; then QR = projn v. The
Setting y = z = 0 shows that P = (1, 0, 0) is a
2 3 2 3
1
1
405 = 4 05 ,
0
0
1 · ( 1)
+ ( 2)2 + 22
◆
2
3 2
1
6 7 6
4 25 = 4
2
2 3 2 3
1
1
6 7 6 7
projn v = 405 + 4 05
0
0
Therefore, the point R in P that is closest to Q is
1 2
9, 9,
2
9
.
2
6
4
3
3
1
9
27
95
2
9
1
9
27
.
95
2
9
2
6
=4
3
1
9
27
.
95
2
9
35. Since the given lines `1 and `2 are parallel, choose arbitrary points Q on `1 and P on `2 , say Q = (1, 1)
and P = (5, 4). The direction vector of `2 is d = [ 2, 3]. Then



# »
1
5
4
v = PQ = q p =
=
,
1
4
3
so that
✓
◆
✓
◆

d·v
2 · ( 4) + 3 · ( 3)
1
2
=
projd v =
d=
3
d·d
( 2)2 + 32
13
Then the distance between the lines is given by




1
1
18 3
4
2
54
kv projd vk =
+
=
=
=
3
3
36
13
13
13 2
2
.
3
18 p
13.
13
36. Since the given lines `1 and `2 are parallel, choose arbitrary points Q on `1 and P on `2 , say Q =
(1, 0, 1) and P = (0, 1, 1). The direction vector of `2 is d = [1, 1, 1]. Then
2 3 2 3 2 3
1
0
1
# »
v = P Q = q p = 4 05 415 = 4 15 ,
1
1
2
so that
projd v =
✓
d·v
d·d
◆
d=
✓
1 · 1 + 1 · ( 1) + 1 · ( 2)
12 + 12 + 12
Then the distance between the lines is given by
2 3
2 3
2
1
1
6 7 26 7
6
kv projd vk = 4 15 + 415 = 4
3
2
1
3
5
3
17
35
4
3
=
2 3
◆ 1
4 15 =
1
2 3
1
24 5
1 .
3
1
1p 2
5 + ( 1)2 + ( 4)2 =
3
p
42
.
3
37. Since P1 and P2 are parallel, we choose an arbitrary point on P1 , say Q = (0, 0, 0), and compute
d(Q, P2 ). Since the equation of P2 is 2x + y 2z = 5, we have a = 2, b = 1, c = 2, and d = 5; since
Q = (0, 0, 0), we have x0 = y0 = z0 = 0. Thus the distance is
d(P1 , P2 ) = d(Q, P2 ) =
|ax0 + by0 + cz0 d|
|2 · 0 + 1 · 0 2 · 0
p
p
=
2
2
2
a +b +c
22 + 12 + 22
5|
=
5
.
3
38
CHAPTER 1. VECTORS
38. Since P1 and P2 are parallel, we choose an arbitrary point on P1 , say Q = (1, 0, 0), and compute
d(Q, P2 ). Since the equation of P2 is x + y + z = 3, we have a = b = c = 1 and d = 3; since
Q = (1, 0, 0), we have x0 = 1, y0 = 0, and z0 = 0. Thus the distance is
p
|ax0 + by0 + cz0 d|
|1 · 1 + 1 · 0 + 1 · 0 3|
2
2 3
p
p
d(P1 , P2 ) = d(Q, P2 ) =
=
=p =
.
3
3
a 2 + b2 + c 2
12 + 1 2 + 1 2

# »
a
0 c|
39. We wish to show that d(B, `) = |axp0a+by
,
where
n
=
, n · a = c, and B = (x0 , y0 ). If v = AB =
2 +b2
b
b a, then


a
x
n · v = n · (b a) = n · b n · a =
· 0
c = ax0 + by0 c.
b
y0
Then from Figure 1.65, we see that
⇣n · v⌘
|n · v|
|ax0 + by0 c|
p
=
.
n·n
knk
a 2 + b2
2 3
a
0 +by0 +cz0 d|
4
b 5, n · a = d, and B = (x0 , y0 , z0 ). If
40. We wish to show that d(B, `) = |axp
,
where
n
=
a2 +b2 +c2
c
# »
v = AB = b a, then
2 3 2 3
a
x0
n · v = n · (b a) = n · b n · a = 4 b 5 · 4 y0 5 d = ax0 + by0 + cz0 d.
c
z0
d(B, `) = kprojn vk =
n =
Then from Figure 1.65, we see that
d(B, `) = kprojn vk =
⇣n · v⌘
n·n
n =
|n · v|
|ax0 + by0 + cz0 d|
p
=
.
knk
a 2 + b2 + c 2
41. Choose B = (x0 , y0 ) on `1 ; since `1 and `2 are parallel, the distance between them is d(B, `2 ). Then
a
x
since B lies on `1 , we have n · b =
· 0 = ax0 + by0 = c1 . Choose A on `2 , so that n · a = c2 . Set
b
y0
v = b a. Then using the formula in Exercise 39, the distance is
d(`1 , `2 ) = d(B, `2 ) =
|n · v|
|n · (b a)|
|n · b n · a|
|c1 c2 |
=
=
=
.
knk
knk
knk
knk
42. Choose B = (x0 , y0 , z0 ) on P1 ; since P1 and
parallel, the distance between them is d(B, P2 ).
2P
3 2 2are 3
a
x0
Then since B lies on P1 , we have n · b = 4 b 5 · 4 y0 5 = ax0 + by0 + cz0 = d1 . Choose A on P2 , so
c
z0
that n · a = d2 . Set v = b a. Then using the formula in Exercise 40, the distance is
|n · v|
|n · (b a)|
|n · b n · a|
=
=
knk
knk
knk
2 3
2
1
43. Since P1 has normal vector n1 = 415 and P2 has normal vector n2 = 4
1
the normal vectors satisfies
d(P1 , P2 ) = d(B, P2 ) =
cos ✓ =
Thus
=
|d1 d2 |
.
knk
3
2
15, the angle ✓ between
2
n1 · n2
1 · 2 + 1 · 1 + 1 · ( 2)
1
p
=p
= p .
2
2
2
2
2
2
kn1 k kn2 k
3 3
1 + 1 + 1 2 + 1 + ( 2)
✓ = cos
1
✓
1
p
3 3
◆
⇡ 78.9 .
1.3. LINES AND PLANES
39
2
3
2 3
3
1
44. Since P1 has normal vector n1 = 4 15 and P2 has normal vector n2 = 4 45, the angle ✓ between
2
1
the normal vectors satisfies
cos ✓ =
n1 · n2
3 · 1 1 · 4 + 2 · ( 1)
p
=p
=
2
kn1 k kn2 k
3 + ( 1)2 + 22 12 + 42 + ( 1)2
p
3
p =
14 18
1
p .
28
This is an obtuse angle, so the acute angle is
⇡
✓=⇡
cos
1
✓
1
p
28
◆
⇡ 79.1 .
45. First, to see that P and ` intersect, substitute the parametric equations for ` into the equation for P,
giving
x + y + 2z = (2 + t) + (1 2t) + 2(3 + t) = 9 + t = 0,
so that t =
9 represents the point of intersection,
which is thus (2 + ( 9), 1 2( 29), 33+ ( 9)) =
2 3
1
1
( 7, 19, 6). Now, the normal to P is n = 415, and a direction vector for ` is d = 4 25. So if ✓ is
2
1
the angle between n and d, then ✓ satisfies
cos ✓ =
so that
n·d
1 · 1 + 1 · ( 2) + 2 · 1
1
p
=p
= ,
2
2
2
2
2
2
knk kdk
6
1 + 1 + 1 1 + ( 2) + 1
✓ = cos
1
✓ ◆
1
⇡ 80.4 .
6
Thus the angle between the line and the plane is 90
80.4 ⇡ 9.6 .
46. First, to see that P and ` intersect, substitute the parametric equations for ` into the equation for P,
giving
4x y z = 4 · t (1 + 2t) (2 + 3t) = t 3 = 6,
9 represents the point of intersection,
· ( 9)) =
2 3 which is thus ( 9, 1 + 2 · ( 9), 2 +23 3
4
1
( 9, 17, 25). Now, the normal to P is n = 4 15, and a direction vector for ` is d = 425. So if ✓
1
3
is the angle between n and d, then ✓ satisfies
so that t =
cos ✓ =
n·d
4·1 1·2 1·3
p
=p
=
knk kdk
42 + 12 + 12 12 + 22 + 32
p
1
p .
18 14
This corresponds to an obtuse angle, so the acute angle between the two is
✓
◆
1
1
p p
✓ = ⇡ cos
⇡ 86.4 .
18 14
Thus the angle between the line and the plane is 90
47. We have p = v
c n, so that c n = v
86.4 ⇡ 3.6 .
p. Take the dot product of both sides with n, giving
(c n) · n = (v
c(n · n) = v · n
p) · n
p·n
)
)
c(n · n) = v · n (since p and n are orthogonal)
n·v
c=
.
n·n
)
40
CHAPTER 1. VECTORS
Note that another interpretation of the figure is that c n = projn v =
n·v
c = n·n
.
n·v
n·n
n, which also implies that
Now substitute this value of c into the original equation, giving
⇣n · v⌘
p = v cn = v
n.
n·n
2 3
1
48. (a) A normal vector to the plane x + y + z = 0 is n = 415. Then
1
so that c =
1
3.
2 3 2 3
1
1
n · v = 415 · 4 05 = 1 · 1 + 1 · 0 + 1 · ( 2) =
1
2
2 3 2 3
1
1
n · n = 415 · 415 = 1 · 1 + 1 · 1 + 1 · 1 = 3,
1
1
Then
3
2 3 2
1
1
6 7 16 7 6
n = 4 05 + 415 = 4
n·n
3
2
1
(b) A normal vector to the plane 3x
2
1
11 .
2
⇣n · v⌘
p=v
so that c =
1
2
3
3
y + z = 0 is n = 4 15. Then
1
3 2
3
n · v = 4 15 · 4
1
2 3 2
3
n · n = 4 15 · 4
1
3
1
05 = 3 · 1
2
3
3
15 = 3 · 3
1
3
4
3
17
.
35
5
3
1 · 0 + 1 · ( 2) = 1
1 · ( 1) + 1 · 1 = 11,
Then
3
1
6 7
n = 4 05
n·n
2
⇣n · v⌘
p=v
(c) A normal vector to the plane x
2
2
3 2
3
1 6 7 6
4 15 = 4
11
1
3
1
2z = 0 is n = 4 05. Then
2
3 2
1
n · v = 4 05 · 4
2
2 3 2
1
n · n = 4 05 · 4
2
2
2
3
1
05 = 1 · 1 + 0 · 0
2
3
1
05 = 1 · 1 + 0 · 0
2
3
8
11
1 7
.
11 5
23
11
2 · ( 2) = 5
2 · ( 2) = 5,
Exploration: The Cross Product
41
so that c = 1. Then
p=v
3
1
6 7
n = 4 05
n·n
2
⇣n · v⌘
2
2
3
1
6 7
4 05 = 0.
2
Note that the projection is 0 because the vector is normal to the plane, so its projection onto the
plane is a single point.
2 3
2
(d) A normal vector to the plane 2x 3y + z = 0 is n = 4 35. Then
1
2
3 2
2
n · v = 4 35 · 4
1
2 3 2
2
n · n = 4 35 · 4
1
2
3
1
05 = 2 · 1
2
3
2
35 = 2 · 2
1
3 · 0 + 1 · ( 2) = 0
3 · ( 3) + 1 · 1 = 14,
3
1
so that c = 0. Thus p = v = 4 05. Note that the projection is the vector itself because the
2
vector is parallel to the plane, so it is orthogonal to the normal vector.
Exploration: The Cross Product
1. (a)
(b)
(c)
(d)
2
u2 v 3
u ⇥ v = 4 u3 v 1
u1 v 2
2
u2 v 3
u ⇥ v = 4 u3 v 1
u1 v 2
2
u2 v 3
u ⇥ v = 4 u3 v 1
u1 v 2
2
u2 v 3
u ⇥ v = 4 u3 v 1
u1 v 2
3 2
3 2 3
u3 v 2
1 · 3 0 · ( 1)
3
u1 v3 5 = 4 1 · 3 0 · 2 5 = 4 35.
u2 v 1
0 · ( 1) 1 · 3
3
3 2
3 2 3
u3 v 2
1·1 2·1
3
u1 v3 5 = 4 2 · 0 3 · 1 5 = 4 35.
u2 v 1
3 · 1 ( 1) · 0
3
3 2
3
u3 v 2
2 · ( 6) 3 · ( 4)
u1 v3 5 = 43 · 2 ( 1) · ( 6)5 = 0.
u2 v 1
1 · ( 4) 2 · 2
3 2
3 2 3
u3 v 2
1·3 1·2
1
u1 v3 5 = 41 · 1 1 · 35 = 4 25.
u2 v 1
1·2 1·1
1
2. We have
2 3 2 3 2
1
0
0·0
e1 ⇥ e2 = 405 ⇥ 415 = 40 · 0
0
0
1·1
2 3 2 3 2
0
0
1·1
e2 ⇥ e3 = 415 ⇥ 405 = 40 · 0
0
1
0·0
2 3 2 3 2
0
1
0·0
e3 ⇥ e1 = 405 ⇥ 405 = 41 · 1
1
0
0·0
3 2 3
0·1
0
1 · 05 = 405 = e3
0·0
1
3 2 3
0·0
1
0 · 15 = 405 = e1
1·0
0
3 2 3
1·0
0
0 · 05 = 415 = e2 .
0·1
0
42
CHAPTER 1. VECTORS
3. Two vectors are orthogonal if their dot product equals zero. But
2
3 2 3
u2 v 3 u3 v 2
u1
(u ⇥ v) · u = 4u3 v1 u1 v3 5 · 4u2 5
u1 v 2 u2 v 1
u3
= (u2 v3
u3 v2 )u1 + (u3 v1
u1 v3 )u2 + (u1 v2
= (u2 v3 u1 u1 v3 u2 ) + (u3 v1 u2
2
3 2 3
u2 v 3 u3 v 2
v1
(u ⇥ v) · v = 4u3 v1 u1 v3 5 · 4v2 5
u1 v 2 u2 v 1
v3
= (u2 v3
= (u2 v3 v1
u3 v2 )v1 + (u3 v1
u2 v1 u3 ) + (u1 v2 u3
u1 v3 )v2 + (u1 v2
u2 v1 v3 ) + (u3 v1 v2
u2 v1 )u3
u3 v 2 u1 ) = 0
u2 v1 )v3
u3 v2 v1 ) + (u1 v2 v3
u1 v3 v2 ) = 0.
4. (a) By Exercise 1, a vector normal to the plane is
2 3 2 3 2
3 2 3
0
3
1 · 2 1 · ( 1)
3
n = u ⇥ v = 415 ⇥ 4 15 = 4 1 · 3 0 · 2 5 = 4 35 .
1
2
0 · ( 1) 1 · 3
3
So the normal form for the equation of this plane is n · x = n · p, or
2 3 2 3 2 3 2 3
3
x
3
1
4 35 · 4y 5 = 4 35 · 4 05 = 9.
3
z
3
2
This simplifies to 3x + 3y
3z = 9, or x + y z = 3.
2 3
2 3
2
1
# »
# »
(b) Two vectors in the plane are u = P Q = 415 and v = P R = 4 35. So by Exercise 1, a vector
1
2
normal to the plane is
2 3 2 3 2
3 2 3
2
1
1 · ( 2) 1 · 3
5
n = u ⇥ v = 415 ⇥ 4 35 = 41 · 1 2 · ( 2)5 = 4 55 .
1
2
2·3 1·1
5
So the normal form for the equation of this plane is n · x = n · p, or
2 3 2 3 2 3 2 3
5
x
5
0
4 55 · 4y 5 = 4 55 · 4 15 = 0.
5
z
5
1
5. (a)
(b)
(c)
(d)
This simplifies to 5x + 5y + 5z = 0, or x y z = 0.
2
3
2
3
v 2 u3 v 3 u2
u2 v 3 u3 v 2
v ⇥ u = 4v3 u1 u3 v1 5 = 4u3 v1 u1 v3 5 = (u ⇥ v).
v 1 u2 v 2 u1
u1 v 2 u2 v 1
2 3 2 3 2
3 2 3
u1
0
u2 · 0 u3 · 0
0
u ⇥ 0 = 4u2 5 ⇥ 405 = 4u3 · 0 u1 · 05 = 405 = 0.
u3
0
u1 · 0 u2 · 0
0
2
3 2 3
u2 u3 u3 u2
0
u ⇥ u = 4u3 u1 u1 u3 5 = 405 = 0.
u1 u2 u2 u1
0
2
3
2
3
u2 kv3 u3 kv2
u2 v 3 u3 v 2
u ⇥ kv = 4u3 kv1 u1 kv3 5 = k 4u3 v1 u1 v3 5 = k(u ⇥ v).
u1 kv2 u2 kv1
u1 v 2 u2 v 1
Exploration: The Cross Product
43
(e) u ⇥ ku = k(u ⇥ u) = k(0) = 0 by parts (d) and (c).
(f ) Compute the cross-product:
2
3
u2 (v3 + w3 ) u3 (v2 + w2 )
u ⇥ (v + w) = 4u3 (v1 + w1 ) u1 (v3 + w3 )5
u1 (v2 + w2 ) u2 (v1 + w1 )
2
3
(u2 v3 u3 v2 ) + (u2 w3 u3 w2 )
= 4(u3 v1 u1 v3 ) + (u3 w1 u1 w3 )5
(u1 v2 u2 v1 ) + (u1 w2 u2 w1 )
2
3 2
3
u2 v 3 u3 v 2
u2 w 3 u3 w 2
= 4 u3 v 1 u1 v 3 5 + 4 u3 w 1 u1 w 3 5
u1 v 2 u2 v 1
u1 w 2 u2 w 1
= u ⇥ v + u ⇥ w.
6. In each case, simply compute:
(a)
2
3 2
u1
v2 w 3
u · (v ⇥ w) = 4u2 5 · 4v3 w1
u3
v1 w 2
= u1 v 2 w 3
= (u2 v3
3
v3 w2
v1 w3 5
v2 w1
u1 v 3 w 2 + u2 v 3 w 1
u3 v2 )w1 + (u3 v1
u2 v 1 w 3 + u3 v 1 w 2
u1 v3 )w2 + (u1 v2
u3 v 2 w 1
u2 v1 )w3
= (u ⇥ v) · w.
(b)
2
3 2
3
u1
v2 w3 v3 w2
u ⇥ (v ⇥ w) = 4u2 5 ⇥ 4v3 w1 v1 w3 5
u3
v1 w2 v2 w1
2
3
u2 (v1 w2 v2 w1 ) u3 (v3 w1 v1 w3 )
= 4u3 (v2 w3 v3 w2 ) u1 (v1 w2 v2 w1 )5
u1 (v3 w1 v1 w3 ) u2 (v2 w3 v3 w2 )
2
3
(u1 w1 + u2 w2 + u3 w3 )v1 (u1 v1 + u2 v2 + u3 v3 )w1
= 4(u1 w1 + u2 w2 + u3 w3 )v2 (u1 v1 + u2 v2 + u3 v3 )w2 5
(u1 w1 + u2 w2 + u3 w3 )v3 (u1 v1 + u2 v2 + u3 v3 )w3
2 3
2 3
v1
w1
= (u1 w1 + u2 w2 + u3 w3 ) 4v2 5 (u1 v1 + u2 v2 + u3 v3 ) 4w2 5
v3
w3
= (u · w)v
(u · v)w.
(c)
2
u2 v 3
2
ku ⇥ vk = 4u3 v1
u1 v 2
= (u2 v3
3
u3 v 2
u1 v 3 5
u2 v 1
2
u3 v2 )2 + (u3 v1
u1 v3 )2 + (u1 v2
= (u21 + u22 + u23 )2 (v12 + v22 + v32 )2
2
2
= kuk kvk
(u · v)2 .
u2 v 1 ) 2
(u1 v1 + u2 v2 + u3 v3 )2
44
CHAPTER 1. VECTORS
7. For problem 2, use the computation in the solution above to show that e1 ⇥ e2 = e3 . Then
e2 ⇥ e3 = e2 ⇥ (e1 ⇥ e2 )
= (e2 · e2 )e1
= 1 · e1
= e1 .
0 · e2
(e2 · e1 )e2
6(b)
ei · ej = 0 if i 6= j, ei · ei = 1
Similarly,
e3 ⇥ e1 = e3 ⇥ (e2 ⇥ e3 )
= (e3 · e3 )e2
= 1 · e2
= e2 .
0 · e3
(e3 · e2 )e3
6(b)
ei · ej = 0 if i 6= j, ei · ei = 1
For problem 3, we have u · (u ⇥ v) = (u ⇥ u) · v by 6(a), and then since u ⇥ u = 0 by Exercise 5(c), this
reduces to 0 · v = 0. Thus u is orthogonal to u ⇥ v. Similarly, v · (u ⇥ v) = v · ( v ⇥ u) = v · (v ⇥ u)
by Exercise 5(a), and then as before, v · (v ⇥ u) = (v ⇥ v) · u = 0 · u = 0, so that v is also
orthogonal to u ⇥ v.
8. (a) We have
2
2
2
2
2
2
2
2
2
(u · v)2
ku ⇥ vk = kuk kvk
2
= kuk kvk
2
kuk kvk cos2 ✓
= kuk kvk
1
cos2 ✓
= kuk kvk sin2 ✓.
Since the angle between u and v is always between 0 and ⇡, it always has a nonnegative sine, so
taking square roots gives ku ⇥ vk = kuk kvk sin ✓.
(b) Recall that the area of a triangle is A = 21 (base)(height). If the angle between u and v is ✓, then
the length of the perpendicular from the head of v to the line determined by u is an altitude of
the triangle; the corresponding base is u. Thus the area is
1
1
1
kuk · (kvk sin ✓) = kuk kvk sin ✓ = ku ⇥ vk .
2
2
2
2 3
2 3
1
4
# »
# »
(c) Let u = AB = 4 15 and v = AC = 4 35. Then from part (b), we see that the area is
1
2
2 3 2 3
2 3
1
4
5
1 4 5 4 5
1 4 5
1p
1 ⇥
3
6 =
A=
=
62.
2
2
2
1
2
1
A=
1.4
Applications
1. Use the method of example 1.34. The magnitude of the resultant force r is
q
p
2
2
krk = kf1 k + kf2 k = 122 + 52 = 13 N,
while the angle between r and east (the direction of f2 ) is
✓ = tan
1
12
⇡ 67.4 .
5
Note that the resultant is closer to north than east; the larger force to the north pulls the object more
strongly to the north.
1.4. APPLICATIONS
45
2. Use the method of example 1.34. The magnitude of the resultant force r is
q
p
2
2
krk = kf1 k + kf2 k = 152 + 202 = 25 N,
while the angle between r and west (the direction of f1 ) is
✓ = tan
1
20
⇡ 53.1 .
15
Note that the resultant is closer to south than west; the larger force to the south pulls the object more
strongly to the south.



4
8
8 cos 60
3. Use the method of Example 1.34. If we let f1 =
, then f2 =
= p . So the resultant
0
8 sin 60
4 3
force is



4
12
8
r = f1 + f2 =
+ p = p .
0
4 3
4 3
q
p 2 p
p
The magnitude of r is krk = 122 + 4 3 = 192 = 8 3 N, and the angle formed by r and f1 is
✓ = tan
1
p
4 3
= tan
12
1
1
p = 30 .
3
Note that the resultant also forms a 30 degree angle with f2 ; since the magnitudes of the two forces
are the same, the resultant points equally between them.


 p
4
6 cos 135
3p 2
4. Use the method of Example 1.34. If we let f1 =
, then f2 =
=
. So the
0
6 sin 135
3 2
resultant force is
p

 p

4
3p 2
4 p3 2
r = f1 + f2 =
+
=
.
0
3 2
3 2
The magnitude of r is
q
krk = (4
p
p
3 2)2 + (3 2)2 =
and the angle formed by r and f1 is
✓ = tan
1
q
52
p
24 2 ⇡ 4.24 N,
p
3 2
p ⇡ 93.3
4 3 2



2
6
4 cos 60
5. Use the method of Example 1.34. If we let f1 =
, then f2 =
, and f3 =
0
0
4 sin 60
So the resultant force is




2
2
2
6
r = f1 + f2 + f3 =
+
+ p = p .
0
0
2 3
2 3
The magnitude of r is
krk =
and the angle formed by r and f1 is
=

2
p
.
2 3
q
p
p
( 2)2 + (2 3)2 = 16 = 4 N,
✓ = tan
1
p
2 3
= tan
2
1
(
p
3) = 120 .
p
(Note that many CAS’s will return 60 for tan 1 (
3); by convention we require an angle between
0 and 180 , so we add 180 to that answer, since tan ✓ = tan(180 + ✓)).
46
CHAPTER 1. VECTORS
6. Use the method of Example 1.34. If we let f1 =
the resultant force is




10
0
5
0
, then f2 =
, f3 =
, and f4 =
. So
0
13
0
8





10
0
5
0
5
r = f 1 + f 2 + f3 + f 4 =
+
+
+
=
.
0
13
0
8
5
The magnitude of r is
krk =
and the angle formed by r and f1 is
p
p
p
52 + 52 = 50 = 5 2 N,
✓ = tan
1
5
= tan
5
1
1 = 45 .
7. Following Example 1.35, we have the following diagram:
f
fy
60°
fx
Here f makes a 60 angle with fy , so that
Thus kfx k = 10 ·
p
3
2
kfy k = kf k cos 60 ,
p
= 5 3 ⇡ 8.66 N and kfy k = 10 ·
fx =
 p
5 3
,
0
kfx k = kf k sin 60 .
1
2
= 5 N. Finally, this gives
fy =

0
.
5
8. The force that must be applied parallel to the ramp is the force needed to counteract the component
of the force due to gravity that acts parallel to the ramp. Let f be the force due to gravity; this
acts downwards. We can decompose it into components fp , acting parallel to the ramp, and fr , acting
orthogonally to the ramp. Since the ramp makes an angle of 30 with the horizontal, it makes an angle
of 60 with f . Thus
1
kfp k = kf k cos 60 = kf k = 5 N.
2
9. The vertical force is the vertical component of the force vector f . Since this acts at an angle of 45 to
the horizontal, the magnitude of the component of this force in the vertical direction is
p
p
2
kfy k = kf k cos 45 = 1500 ·
= 750 2 ⇡ 1060.66 N.
2
10. The vertical force is the vertical component of the force vector f , which has magnitude 100 N and acts
at an angle of 45 to the horizontal. Since it acts at an angle of 45 to the horizontal, the magnitude
of the component of this force in the vertical direction is
p
p
2
kfy k = kf k cos 45 = 100 ·
= 50 2 ⇡ 70.7 N.
2
Note that the mass of the lawnmower itself is irrelevant; we are not considering the gravitational force
in this exercise, only the force imparted by the person mowing the lawn.
1.4. APPLICATIONS
47
11. Use the method of Example 1.36. Let t be the force vector on the cable; then the tension on the cable
is ktk. The force imparted by the hanging sign is the y component of t; call it ty . Since t and ty form
an angle of 60 , we have
1
kty k = ktk cos 60 = ktk .
2
The gravitational force on the sign is its mass times the acceleration due to gravity, which is 50·9.8 = 490
N. Thus ktk = 2 kty k = 2 · 490 = 980 N.
12. Use the method of Example 1.36. Let w be the force vector created by the sign. Then kwk = 1·9.8 = 9.8
N, since the sign weighs 1 kg. By symmetry, each string carries half the weight of the sign, since the
angles each string forms with the vertical are the same. Let s be the tension in the left-hand string.
Since the angle between s and w is 45 , we have
p
1
2
kwk = ksk cos 45 =
ksk .
2
2
p
9.8
Thus ksk = p12 kwk = p
= 4.9 2 ⇡ 6.9 N.
2
13. A diagram of the situation is
25
15
20
15 kg
The triangle is a right triangle, since 152 + 202 = 625 = 252 . Thus if ✓1 is the angle that the left-hand
4
wire makes with the ceiling, then sin ✓1 = 20
25 = 5 ; likewise, if ✓2 is the angle that the right-hand wire
15
3
makes with the ceiling, then sin ✓2 = 25 = 5 . Let f1 be the force on the left-hand wire and f2 the
force on the right-hand wire. Let r be the force due to gravity acting on the painting. Then following
Example 1.36, we have, using the Law of Sines,
kf1 k
kf2 k
krk
15 · 9.8
=
=
=
= 147 N.
sin ✓1
sin ✓2
sin 90
1
Then
kf1 k = 147 sin ✓1 = 147 ·
4
= 117.6 N,
5
kf2 k = 147 sin ✓2 = 147 ·
3
= 88.2 N.
5
14. Let r be the force due to gravity acting on the painting, f1 be the tension on the wire opposite the
30 angle, and f2 be the tension on the wire opposite the 45 angle (don’t these people know to hang
paintings straight?). Then krk = 20 · 9.8 = 196 N. Note that the remaining angle in the triangle is
105 . Then using the method of Example 1.36, we have, using the Law of Sines,
kf1 k
kf2 k
krk
=
=
.
sin 30
sin 45
sin 105
Thus
196 · 12
krk · sin 30
⇡
⇡ 101.46 N
sin 105
0.9659p
196 · 22
krk · sin 45
kf2 k =
⇡
⇡ 143.48 N.
sin 105
0.9659
kf1 k =
48
CHAPTER 1. VECTORS
Chapter Review
1. (a) True. This follows from the properties of Rn listed in Theorem 1.1 in Section 1.1:
u=u+0
Zero Property, Property (c)
= u + (w + ( w))
Additive Inverse Property, Property (d)
= (u + w) + ( w)
Distributive Property, Property (b)
= (v + w) + ( w)
By the given condition u + w = v + w
= v + (w + ( w))
Distributive Property, Property (b)
=v+0
Additive Inverse Property, Property (d)
=v
Zero Property, Property (c).
(b) False. See Exercise 60 in Section 1.2. For one counterexample, let w = 0 and u and v be arbitrary
vectors. Then u · w = v · w = 0 since the dot product of any vector with the zero vector is zero.
But certainly u and v need not be equal. As a second counterexample, suppose that both u and
v are orthogonal to w; clearly they need not be equal, but u · w = v · w = 0.
(c) False. For example, let u be any nonzero vector, and v any vector orthogonal to u. Let w = u.
Then u is orthogonal to v and v is orthogonal to w = u, but certainly u and w = u are not
orthogonal.
(d) False. When a line is parallel to a plane, then d · n = 0; that is, d is orthogonal to the normal
vector of the plane.
(e) True. Since a normal vector n for P and the line ` are both perpendicular to P, they must be
parallel. See Figure 1.62.
(f ) True. See the remarks following Example 1.31 in Section 1.3.
(g) False. They can be skew lines, which 2
are3nonintersecting lines with nonparallel 2
direction
3
2vectors.
3
1
0
0
For example, let `1 be the line x = t 405 (the x-axis), and `2 be the line x = 405 + t 415 (the
0
1
0
line through (0, 0, 1) that is parallel to the y-axis). These two lines do not intersect, yet they are
not parallel.
2 3 2 3
1
1
(h) False. For example, 405 · 405 = 1 + 0 + 1 = 0 in Z2 .
1
1
(i) True. If ab = 0 in Z/5, then ab must be a multiple of 5. But 5 is prime, so either a or b must be
divisible by 5, so that either a = 0 or b = 0 in Z5 .
(j) False. For example, 2 · 3 = 6 = 0 in Z6 , but neither 2 nor 3 is zero in Z6 .

10
2. Let w =
. Then the head of the resulting vector is
10




1
3
10
9
4u + v + w = 4
+
+
=
.
5
2
10
12
So the coordinates of the point at the head of 4u + v are (9, 12).
3. Since 2x + u = 3(x
v) = 3x
3v, simplifying gives u + 3v = x. Thus



1
3
8
x = u + 3v =
+3
=
.
5
2
11
# »
# »
4. Since ABCD is a square, OC = OA, so that
# » # » # »
BC = OC OB =
# »
OA
# »
OB =
a
b.
Chapter Review
49
5. We have
2
3 2 3
1
2
u · v = 4 15 · 4 15 = 1 · 2 + 1 · 1 + 2 · ( 1) =
2
1
p
p
kuk = ( 1)2 + 12 + 22 = 6
p
p
kvk = 22 + 12 + ( 1)2 = 6.
3
Then if ✓ is the angle between u and v, it satisfies
cos ✓ =
u·v
=
kuk kvk
Thus
✓ = cos
6. We have
1
✓
3
p p =
6 6
1
2
◆
1
.
2
= 120 .
2
3
2 3 2
1
1
1·1 2·1+2·1 6 7 16 7 6
proju v =
u=
4 25 = 4 25 = 4
u·u
1 · 1 2 · ( 2) + 2 · 2
9
2
2
⇣u · v⌘
3
1
9
27
.
95
2
9
7. We are looking for a vector
2 3 in the xy-plane; any such vector has a z-coordinate of 0. So the vector we
a
are looking for is u = 4 b 5 for some a, b. Then we want
0
so that a =
2 3 2 3 2 3
1
a
1
u · 425 = 4 b 5 · 425 = a + 2b = 0,
3
0
3
2b. So for example choose b = 1; then a = 2, and the vector
2 3
2
u = 4 15
0
2 3
1
is orthogonal to 425. Finally, to get a unit vector, we must normalize u. We have
3
kuk =
to get
p
p
( 2)2 + 12 + 02 = 5
w=
2
6
1
u=6
4
kuk
3
p2
57
p1 7
55
0
that is orthogonal to the given vector. We could have chosen any value for b, but we would have gotten
either w or w for the normalized vector.
8. The vector form of the given line is
2
3
2 3
2
1
x = 4 35 + t 4 25 ,
1
1
50
CHAPTER 1. VECTORS
2
3
1
so the line has a direction vector d = 4 25. Since the plane is perpendicular to this line, d is a normal
1
vector n to the plane. Then the plane passes through P = (1, 1, 1), so equation n · x = n · p becomes
2
3 2 3 2 3 2 3
1
x
1
1
4 25 · 4y 5 = 4 25 · 415 = 2.
1
z
1
1
Expanding gives the general equation
x + 2y + z = 2.
2
3
2
9. Parallel planes have parallel normals, so the vector n = 4 35, which is a normal to the given plane,
1
is also a normal to the desired plane. The plane we want must pass through P = (3, 2, 5), so equation
n · x = n · p becomes
2 3 2 3 2 3 2 3
2
x
2
3
4 35 · 4y 5 = 4 35 · 425 = 7.
1
z
1
5
Expanding gives the general equation 2x + 3y
z = 7.
10. The three points give us two vectors that lie in the plane:
2 3
1
# » 4 5
d1 = AB = 0
1
2 3
0
# »
d2 = BC = 415
2
2 3 2
1
4 15 = 4
0
2 3 2
1
4 05 = 4
1
3
0
15 ,
1
3
1
15 .
1
2 3
a
A normal vector n = 4 b 5 must be orthogonal to both of these vectors, so
c
2 3 2
a
n · d1 = 4 b 5 · 4
c
2 3 2
a
n · d2 = 4 b 5 · 4
c
3
0
15 =
1
3
1
15 =
1
b+c=0
)
a+b+c=0
b=c
)
a = b + c.
2 3
2c
Since b = c and a = b + c, we get a = 2c, so n = 4 c 5 for any value of c. Choosing c = 1 gives the
c
2 3
2
vector n = 415. Let P be the point A = (1, 1, 0) (we could equally well choose B or C), and compute
1
n · x = n · p:
2 3 2 3 2 3 2 3
2
x
2
1
415 · 4y 5 = 415 · 415 = 3.
1
z
1
0
Expanding gives the general equation 2x + y + z = 3.
Chapter Review
51
11. Let
2
1
# »
u = AB = 40
1
2
0
# »
v = AC = 41
2
3 2
1
15 = 4
0
3 2
1
15 = 4
0
3
0
15
1
3
1
05 .
2
Using the first method from Figure 1.39 (Exercises 46 and 47) in Section 1.2, we have
2 3
2 3 2 3
0
0
0
⇣u · v⌘
0 · ( 1) 1 · 0 + 1 · 2 4 5 2 4 5 4 5
1
1
1
proju v =
u=
=
=
u·u
02 + ( 1)2 + 12
2
1
1
1
2 3 2 3 2 3
1
0
1
v proju v = 4 05 4 15 = 4 15 .
2
1
1
Then the area of the triangle is
1
kuk kv
2
proju vk =
p
1p 2
1p p
1p
0 + ( 1)2 + 12 ( 1)2 + 12 + 12 =
2 3=
6.
2
2
2
12. From the first example in Exploration: Vectors and Geometry, we have a formula for the midpoint:
02 3 2 3 1
2 3 2 3
5
3
8
4
1
1 @4 5 4 5 A 1 4 5 4 5
7
1 +
6 =
3 .
m = (a + b) =
=
2
2
2
2
0
2
1
13. Suppose that kuk = 2 and kvk = 3. Then from the Cauchy-Schwarcz inequality, we have
|u · v|  kuk kvk = 2 · 3 = 6.
Thus
6  u · v  6, so the dot product cannot equal
7.
14. We will apply the formula
|ax0 + by0 + cz0 d|
p
a 2 + b2 + c 2
where A = (x0 , y0 , z0 ) is a point, and a general equation for the plane is ax + by + cz = d. Here, we
have a = 2, b = 3, c = 1, and d = 0; since the point is (3, 2, 5), we have x0 = 3, y0 = 2, and z0 = 5.
So the distance from the point to the plane is
p
7
|2 · 3 + 3 · 2 1 · 5 0|
14
d(A, P) = p
=p =
.
2
14
22 + 32 + ( 1)2
d(A, P) =
15. As in example 1.32 in Section 1.3, we have B = (3, 2, 5), and the line ` has vector form
2 3
2 3
0
1
x = 415 + t 415 ,
2
1
2 3
1
so that A = (0, 1, 2) lies on the line, and a direction vector for the line is d = 415. Then
1
2 3 2 3 2 3
3
0
3
# »
v = AB = 425 415 = 415 ,
5
2
3
52
CHAPTER 1. VECTORS
and then
projd v =
✓
d·v
d·d
◆
d=
✓
2 3 2 3
7
◆ 1
1 · 3 + 1 · 1 + 1 · 3 6 7 6 37 7
415 = 4 3 5 .
12 + 12 + 12
7
1
3
Then the vector from B that is perpendicular to the
2 3
3
6 7
v projd v = 415
3
So the distance from B to ` is
kv
line is the vector
2 3 2 3
7
3
677 6
435 = 4
7
3
2
3
47
.
35
2
3
s✓ ◆
✓
◆2 ✓ ◆ 2
2
2
4
2
1p
2p
projd vk =
+
+
=
4 + 16 + 4 =
6.
3
3
3
3
3
16. We have
3
(2 + 4)3 (4 + 3)2 = 3
13 · 22 = 3
in Z5 . Note that 2 + 4 = 6 = 1 in Z5 , 4 + 3 = 7 = 2 in Z5 , and
1·4=
1=4
1 = 4 in Z5 .
17. 3(x + 2) = 5 implies that 5 · 3(x + 2) = 5 · 5 = 25 = 4. But 5 · 3 = 15 = 1 in Z7 , so this is the same as
x + 2 = 4, so that x = 2. To check the answer, we have
3(2 + 2) = 3 · 4 = 12 = 5 in Z7 .
18. This has no solutions. For any value of x, the left-hand side is a multiple of 3, so it cannot leave a
remainder of 5 when divided by 9 (which is also a multiple of 3).
19. Compute the dot product in Z45 :
[2, 1, 3, 3] · [3, 4, 4, 2] = 2 · 3 + 1 · 4 + 3 · 4 + 3 · 2 = 6 + 4 + 12 + 6 = 1 + 4 + 2 + 1 = 8 = 3.
20. Suppose that
[1, 1, 1, 0] · [d1 , d2 , d3 , d4 ] = d1 + d2 + d3 = 0 in Z2 .
Then an even number (either zero or two) of d1 , d2 , and d3 must be 1 and the others must be zero. d4
is arbitrary (either 0 or 1). So the eight possible vectors are
[0, 0, 0, 0], [0, 0, 0, 1], [1, 1, 0, 0], [1, 1, 0, 1], [1, 0, 1, 0], [1, 0, 1, 1], [0, 1, 1, 0], [0, 1, 1, 1].
Chapter 2
Systems of Linear Equations
2.1
Introduction to Systems of Linear Equations
1. This is equation
is linear since each of x, y, and z appear to the first power with constant coefficients.
p
1, ⇡, and 3 5 are constants.
2. This is not linear since each of x, y, and z appears to a power other than 1.
3. This is not linear since x appears to a power other than 1.
4. This is not linear since the equation contains the term xy, which is a product of two of the variables.
5. This is not linear since cos x is a function of x other than a constant times a multiple of x.
6. This is linear, since here cos 3 is a constant, and x, y, and
p z all appear to the first power with constant
coefficients cos 3, 4, and 1, and the remaining term, 3, is also a constant.
7. Put the equation into the general form ax + by + c by adding y to both sides to get 2x + 4y = 7. There
are no restrictions on x and y since there were none in the original equation.
8. In the given equation, the denominator must not be zero, so we must have x 6= y. With this restriction,
we have
x2 y 2
(x + y)(x y)
= 1 ()
= 1 () x + y = 1, x 6= y.
x y
x y
The corresponding linear equation is x + y = 1, for x 6= y.
9. Since x and y each appear in the denominator, neither one can be zero. With this restriction, we have
1
1
4
y
x
4
+ =
()
+
=
() y + x = 4 () x + y = 4.
x y
xy
xy xy
xy
The corresponding linear equation is x + y = 4, for x 6= 0 and y 6= 0.
10. Since log10 x and log10 y are defined only for x > 0 and y > 0, these are the restrictions on the variables.
With these restrictions, we have
log10 x
log10 y = 2 () log10
x
x
= 2 ()
= 102 = 100 () x = 100y () x
y
y
The corresponding linear equation is x
100y = 0.
100y = 0 for x > 0 and y > 0.
11. As in Example 2.2(a), we set x = t and solve for y. Setting x = t in 3x 6y =⇥0 gives
⇤ 3t 6y = 0, so
that 6y = 3t and thus y = 12 t. So the set of solutions has the parametric form t, 12 t . Note that if we
had set x = 2t to start with, we would have gotten the simpler (but equivalent) parametric form [2t, t].
53
54
CHAPTER 2. SYSTEMS OF LINEAR EQUATIONS
12. As in Example 2.2(a), we set x1 = t and solve for x2 . Setting x1 = t in 2x1 +3x2 = 5 gives 2t+3x
2 = 5,
⇥
⇤
so that 3x2 = 5 2t and thus x2 = 53 23 t. So the set of solutions has the parametric form t, 53 23 t .
Note that if we had set x = 3t to start with, we would have gotten the simpler (but equivalent)
parametric form [3t, 5 2t].
13. As in Example 2.2(b), we set y = s and z = t and solve for x. (We choose y and z since the coefficient
of the remaining variable, x, is one, so we will not have to do any division). Then x + 2s + 3t = 4, so
that x = 4 2s 3t. Thus a complete set of solutions written in parametric form is [4 2s 3t, s, t].
14. As in Example 2.2(b), we set x1 = s and x2 = t and solve for x3 . This substitution yields 4s+3t+2x3 =
1
3
1, so that
⇥ 2x31 = 1 4s 3 ⇤3t and x3 = 2 2s 2 t. Thus a complete set of solutions written in parametric
form is s, t, 2 2s 2 t .
3
15. Subtract the first equation from the second to get x = 3; substituting x = 3
back into the first equation gives y =
3. Thus the unique intersection point
is (3, 3).
2x!y!3
2
1
1
2
3
4
5
"1
"2
!3, "3"
"3
x!y!0
"4
"5
"6
"7
16. Add twice the second equation to the first
equation to get 7x = 21 and thus x =
3. Substitute x = 3 back into the first
equation to get y = 2. So the unique
intersection point is (3, 2).
6
3x!y!7
4
2
1
2
3
4
5
"2
!3, "3"
x"2y!7
"4
"6
"8
3
17. Add three times the second equation to
the first equation. This gives 0 = 6,
which is impossible. So this system has
no solutions — the two lines are parallel.
3x!6y!3
2
1
!x " 2 y ! 1
!2
1
!1
2
3
4
5
4
5
!1
18. Add 35 times the first equation to the second, giving 0 = 0. Thus the system has
an infinite number of solutions — the two
lines are the same line (they are coincident).
6
4
0.1 x ! 0.05 y ! 0.2
2
!2
1
!1
!2
2
3
!0.06 x " 0.03 y ! !0.12
!4
!6
!8
19. Starting from the second equation, we find y = 3. Substitute y = 3 into x
so that x = 7. The unique solution is [x, y] = [7, 3].
2y = 1 to get x
20. Starting from the second equation, we find 2v = 6 so that v = 3. Substitute v = 3 into 2u
get 2u 3 · 3 = 5, or 2u = 14, so that u = 7. The unique solution is [u, v] = [7, 3].
2 · 3 = 1,
3v = 5 to
21. Starting from the third equation, we have 3z = 1, so that z = 13 . Now substitute this value into
the second equation:
✓
◆
1
1
2
1
2y z = 1 ) 2y
= 1 ) 2y + = 1 ) 2y = ) y = .
3
3
3
3
2.1. INTRODUCTION TO SYSTEMS OF LINEAR EQUATIONS
55
Finally, substitute these values for y and z into the first equation:
✓
◆
1
1
2
2
x y+z =0)x
+
=0)x
=0)x= .
3
3
3
3
⇥2 1
⇤
So the unique solution is [x, y, z] = 3 , 3 , 13 .
22. The third equation gives x3 = 0; substituting into the second equation gives 5x2 = 0, so that
x2 = 0. Substituting these values into the first equation gives x1 = 0. So the unique solution is
[x1 , x2 , x3 ] = [0, 0, 0].
23. Substitute x4 = 1 from the fourth equation into the third equation, giving x3 1 = 0, so that x3 = 1.
Substitute these values into the second equation, giving x2 + 1 + 1 = 0 and thus x2 = 2. Finally,
substitute into the first equation:
x1 + x2
x3
x4 = 1 ) x1 + ( 2)
1
1 = 1 ) x1
4 = 1 ) x1 = 5.
The unique solution is [x1 , x2 , x3 , x4 ] = [5, 2, 1, 1].
24. Let z = t, so that y = 2t
x
1. Substitute for y and z in the first equation, giving
3y + z = 5 ) x
3(2t
1) + t = 5 ) x
5t + 3 = 5 ) x = 5t + 2.
This system has an infinite number of solutions: [x, y, z] = [5t + 2, 2t
1, t].
25. Working forward, we start with x = 2. Substitute that into the second equation to get 2 · 2 + y =
or 4 + y = 3, so that y = 7. Finally, substitute those two values into the third equation:
3x
4y + z =
10 )
3·2
4 · ( 7) + z =
10 ) 22 + z =
10 ) z =
3,
32.
The unique solution is [x, y, z] = [2, 7, 32].
26. Working forward, we start with x1 = 1. Substituting into the second equation gives 21 ·( 1)+x2 = 5,
or 12 + x2 = 5. Thus x2 = 92 . Finally, substitute those two values into the third equation:
3
3
9
15
x1 + 2x2 + x3 = 7 ) · ( 1) + 2 · + x3 = 7 )
+ x3 = 7 ) x3 =
2
2
2
2
⇥
⇤
The unique solution is [x1 , x2 , x3 ] = 1, 92 , 12 .
27. As in Example 2.6, we create the augmented matrix from the coefficients:

1
1 0
2
1 3
28. As in Example 2.6, we create the augmented matrix
2
2 3
4 1 0
1 2
from the coefficients:
3
1 1
1 05
2 0
29. As in Example 2.6, we create the augmented matrix from the coefficients:
2
3
1 5
1
4 1 1
55
2 4
4
30. As in Example 2.6, we create the augmented matrix from the coefficients:

1
2
0
1 2
1
1
1
3 1
1
.
2
56
CHAPTER 2. SYSTEMS OF LINEAR EQUATIONS
31. There are three variables, since there are three columns not counting the last one. Use x, y, and z.
Then the system becomes
y+z=1
x y
=1
2x y + z = 1 .
32. There are five variables, since there are five columns not counting the last one. Use x1 , x2 , x3 , x4 , and
x5 . Then the system becomes
x1 x2
+ 3x4 + x5 = 2
x1 + x2 + 2x3 + x4
x5 = 4
x2
+ 2x4 3x5 = 0 .
33. Add the first equation to the second, giving 3x = 3, so that x = 1. Substituting into the first equation
gives 1 y = 0, so that y = 1. The unique solution is [x, y] = [1, 1].
34. As in Example 2.6, we perform row operations on the augmented matrix of this system (from Exercise
28) to get it into a form where we can read o↵ the solution.
2
2
4 1
1
3
0
2
3
1
05
0
1
1
2
2
3
2
1 0
1 0
1 0
R2 2R1
R1 $R2 4
5
4
2
3
1
1
!
0 3
!
1
1 2
2 0 R3 +R!
0 2
2
3
2
1 0
1 0
1 0
6
7
6
. . . 40 1
1 13 5
40 1
R3 2R2
0 2
1 0
! 0 0
From this we get the unique solution [x1 , x2 , x3 ] =
⇥2
1
3,
3,
2
3
⇤
1
3
1
1
1
1
3
0 1
2
15 3 R!
...
0
3 R1 R3 2
0
! 1
1 7 R2 +R3 6
! 40
35
2
0
3
0
1
0
0
0
1
3
2
3
17
.
35
2
3
.
35. Add the first two equations to give 6y = 6, so that y = 1. Substituting back into the second
equation gives x 1 = 5, so that x = 4. Further, 2 · 4 + 4 · ( 1) = 4, so this solution satisfies the
third equation as well. Thus the unique solution is [x, y] = [4, 1].
36. As in Example 2.6, we perform row operations on the augmented matrix of this system (from Exercise
30) to get it into a simpler form.

1
1
2
1
0
1
1
3
2
1

1
! 0
R2 +R1
2
1
0
1
1
2

1
...
0
2
3
R2
!
2 0 1
1 1 2
...
2
3

! 1
0
R1 +2R2
0
1
2
1
5
2
4
.
3
This augmented matrix corresponds to the linear system
a
+ 2c + 5d =
b + c + 2d =
so that, letting c = s and d = t, we get a =
solutions is parametrized by [a, b, c, d] = [ 4
4
3
4 2s 5t and b = 3 s
2s 5t, 3 s 2t, s, t].
2t. So a complete set of
37. Starting with the linear system given in Exercise 31, add the first equation to the second and to the
third, giving
x+ z=2
2x + 2z = 2 .
Subtracting twice the first of these from the second gives 0 =
is inconsistent and has no solution.
2, which is impossible. So this system
2.1. INTRODUCTION TO SYSTEMS OF LINEAR EQUATIONS
57
38. As in Example 2.6, we perform row operations on the augmented matrix of this system (from Exercise
32) to get it into a simpler form.
2
1
41
0
1 0 3
1 2 1
1 0 2
2
1
. . . 40
0
1
1
3
3
2
45
0
1 0
1 0
2 2
R2
2
1
R1 4
! 0
0
3
2
2
1
3
2
2
1 0
2 2
1 0
3
2
05
2 R3
1
. . . 40
0
3
2
2
1
2
3
2
1
40
2R2
! 0
1 0
1 0
0 1
3
2
3
3
2
25
0
+ 5s + 4t = 2
x2
+ 2s + 3t = 0
x3 3s 4t = 1 ,
!...
3
1 2
3 05 1
...
3
8 2 2 R!
3
2
2
2 R1 +R!
1 0 0
5
40 1 0
05
2
1
0 0 1
3
1 0
1 0
0 2
1
3
4
Then setting x4 = s and x5 = t, we have
x1
R2 $R3
3
2
6
x1 = 2
so that
x2 =
5s
2s
4
3
4
3
2
05
1
4t
3t
x3 = 1 + 3s + 4t.
So the general solution to this system has parametric form
[x1 , x2 , x3 , x4 , x5 ] = [2
39. (a) Since t = x and y = 3
2x + y = 3.
5s
4t, 2s
3t, 1 + 3s + 4t, s, t] .
2t, we can substitute x for t in the equation for y to get y = 3
2x, or
(b) If y = s, then substituting s for y in the general equation ⇥above gives
⇤ 2x + s = 3, or 2x = 3
so that x = 32 12 s. So the parametric solution is [x, y] = 32 12 s, s .
s,
40. (a) Since x1 = t, substitute x1 for t in the second and third equations, giving
x2 = 1 + x1 and x3 = 2
x1 .
This gives the linear system
x1 + x2
=1
x1
+ x3 = 2 .
(b) Substitute s for x3 into the second equation, giving x1 + s = 2, so that x1 = 2 s. Substitute that
value for x1 into the first equation, giving (2 s) + x2 = 1, or x2 = 3 s. So the parametric
form given in this way is
[x1 , x2 , x3 ] = [2 s, 3 s, s].
41. Let u = x1 and v = y1 . Then the system of equations becomes 2u + 3v = 0 and 3u + 4v = 1. Solving
the second equation for v gives v = 14 34 u. Substitute that value into the first equation, giving
✓
◆
1 3
1
3
2u + 3
u = 0, or
u=
, so u = 3.
4 4
4
4
⇥
⇤ ⇥
⇤
Then v = 14 34 · 3 = 2, so the solution is [u, v] = [3, 2], or [x, y] = u1 , v1 = 13 , 12 .
42. Let u = x2 and v = y 2 . Then the system of equations becomes u + 2v = 6 and u v = 3. Subtracting
the second of these from the first gives 3v = 3, so that v = 1. Substituting that into the second equation
gives u 1 = 3, so that u = 4. The solution in terms of u and v is thus [u, v] = [x2 , y 2 ] = [4, 1]. Since
1 and 4 have both a positive and a negative square root, the original system has the four solutions
[x, y] = [2, 1],
[ 2, 1],
[2, 1],
or
[ 2, 1].
58
CHAPTER 2. SYSTEMS OF LINEAR EQUATIONS
43. Let u = tan x, v = sin y, and w = cos z. Then the system becomes
u
u
2v
=
v+w=
v w=
2
2
1.
Subtract the first equation from the second, giving v + w = 0. Add that to the third equation, so that
1
1
2v = 1 and v = 12 . Thus w = 12 . Substituting
these
2 + 2 = 2,
⇥
⇤ into the second equation gives u
1 1
so that u = 1. The solution is [u, v, w] = 1, 2 , 2 , so that
⇡
1 = + k⇡,
✓
◆4
1
7⇡
y = sin 1 v = sin 1
=
+ 2`⇡ or
2
6
1
⇡
z = cos 1 w = cos 1 = ± + 2n⇡,
2
3
x = tan
1
u = tan
1
⇡
+ 2`⇡,
6
where k, `, and n are integers.
44. Let r = 2a and s = 3b . Then the system becomes r + 2s = 1 and 3r 4s = 1. Adding three times the
first equation to the second gives 2s = 4, so that s = 2. Substituting that back into the first equation
gives r + 4 = 1, so that r = 3. So the solution is [r, s] = [2a , 3b ] = [3, 2]. In terms of a and b, this
gives [a, b] = [log2 3, log3 2].
2.2
Direct Methods for Solving Linear Systems
1. This matrix is not in row echelon form, since the leading entry in row 2 appears to the right of the
leading entry in row 3, violating condition 2 of the definition.
2. This matrix is in row echelon form, since the row of zeros is at the bottom and the leading entry in
each nonzero row is to the left of the ones in all succeeding rows. It is not in reduced row echelon form
since the leading entry in the first row is not 1.
3. This matrix is in reduced row echelon form: the leading entry in each nonzero row is 1 and is to the left
of the ones in all succeeding rows, and any column containing a leading 1 contains zeros everywhere
else.
4. This matrix is in reduced row echelon form: all zero rows are at the bottom, and all other conditions
are vacuously satisfied (that is, they are true since no rows or columns satisfy the conditions).
5. This matrix is not in row echelon form since it has a zero row, but that row is not at the bottom.
6. This matrix is not in row echelon form since the leading entries in rows 1 and 2 appear to the right of
leading entries in succeeding rows.
7. This matrix is not in row echelon form since the leading entry in row 1 does not appear to the left of
the leading entry in row 2. It appears in the same column, but that is not sufficient for the definition.
8. This matrix is in row echelon form, since the zero row is at the bottom and the leading entry in each
row appears in a column to the left of the leading entries in succeeding rows. However, it is not in
reduced row echelon form since the leading entries in rows 1 and 3 are not 1. (Also, it is not in reduced
row echelon form since the third column contains a leading 1 but also contains other nonzero entries).
9. (a)
2
0
40
1
0
1
1
3
1
15
1
2
1
R1 $R2 4
! 0
0
1
1
0
3
1
15 · · ·
1
2.2. DIRECT METHODS FOR SOLVING LINEAR SYSTEMS
(b)
2
! 1
R3 4 0
!
0
R1 R3
R2
···
10. (a)
"
4
2
#
3
1
"
#
1
3
2
!
4
R2 $R1
3
0
05
1
1
1
0
"
! 1
4
R1
2
3
5
7
25
4
3
6
45
2
(b)
2

2
3
4
1
(b)
2
! 1
6
45
3
3
2
7
15
1

! 1
3
"
1
· · · 17 R2
! 0
2
1
2
1
3
9
7
3
7
#
R2
3
0
05
1
0
1
0
"
1
2
1
4R1
! 0
1
#
···
#
3
2
7
25
5
2
R2
R3
2
! 1
6
40
R3 +R2
! 0
3
0
7
15
0

1
! 0
R2 3R1
3
2
7
125 · · ·
1
1
5R1 6
! 40
3R1
! 0
R1 2R2
1 3
6 6
1
#
0
1
1
2 R1
1
1
6
12 R2
! 40
0
1
2 R1
2 6
6 6
3
"
! 1
0
2
···
12. (a)
3
4
7
25
5
2
R1 $R3 6
! 45
3
1
2
1
2 R2
···
11. (a)
2
! 1
40
0
R1 R2
1
2 R1
(b)
59
"
! 1
0
R1 +2R2
0
1
2
7
1
9
11
7
9
7
15
7
3
7
3
···
3
#
13. (a)
2
3
42
4
3
1
15
1
2
1
3
2
3
3R2
! 4 6
3R3
12
!
(b)
3
1
35
3
2
3
9
2
! 3
40
0
R1 +2R2
···
14. (a)
(b)
2
2
4 3
1
4
6
2
3
7
105
3
2
1
R1 $R3 4
3
!
2
2
6
4
2
3
R2 2R1
! 40
R3 +4R1
! 0
3
3
15
0
0
1
0
3
3
105
7
···
2
! 1
40
0
1
3 R1
2
1
R2 +3R1
! 40
R3 +2R1
! 0
2
! 1
40
0
R1 3R2
2
0
0
3
1
15
1
2
1
1
3
0
15
0
0
1
0
2
0
0
R3
2
3
40
R2
! 0
2
1
0
3
1
15 · · ·
0
3
1
15
0
3
3
15
1
R3
2
1
40
R2
! 0
2
0
0
3
3
15 · · ·
0
60
15.
CHAPTER 2. SYSTEMS OF LINEAR EQUATIONS
2
1
60
6
40
0
2
1
0
0
4
10
1
0
4
9
1
0
3
2
5
1
60
57
7
6
40
15
3
24 R4 +29R!
0
2
1
60
6
40
R4 3R2
! 0
2
1
R2 +2R1
!6 2
6
R3 +2R1 4
2
!
R4 R1
1
!
2
1
0
0
4
10
1
29
4
9
1
29
2
1
0
3
4
10
8
1
4
9
8
2
2
4
3
1
4
0
2
3
4
0
1
6
3
2
5
1
60
57
7
6
3 4
15 8R!
0
5
0
3
2
5
1
60
57
R
$R
2 6
7 3 !
40
85
10
0
3
5
27
7
55
5
2
1
0
0
4
10
8
29
2
0
1
3
4
9
8
29
4
8
10
1
4
8
9
2
3
5
57
7
85
5
3
5
87
7
55
10
16. Ri $ Rj undoes itself; k1 Ri undoes kR1 ; and Ri kRj undoes Ri + kRJ .


 1

1
7
2
2 R1
1 2
1 2
3
1
1 R1 + 2 R!
!
2
17. A =
= B. So A and B are row equivalent,
1
3 4 R2 2R!
1 0
1
0
1
0
and we can convert A to B by the sequence R2 2R1 , 12 R1 , and R1 + 72 R2 .
18.
2
2
A=4 1
1
0
1
1
3
1
05
1
R1 +R2
!
2
3
4 1
1
1
1
1
2
3
4 1
R3 +R1
2
!
3
1
05
1
1
3
2
2
3
!4 1
1
1
3
1
3
2
1
3
1 4
25 R2 +R!
2
0
2
R2 +2R3
1
4
2
3
1
25
1
Therefore the matrices A and B are row equivalent.
3
1
15
0
2
3
! 43
2
R2 + 12 R3
1
5
2
3
1
15 = B.
0
19. Performing R2 + R1 gives a new second row R20 = R2 + R1 . Then performing R1 + R2 gives a new first
row with value R1 + R20 = 2R1 + R2 . Only one row operation can be performed at a time (although in
a sequence of row operations, we may perform more than one in one step as long as any rows a↵ected
do not appear in other transformations in that step, just for efficiency).





R1 R2
R1
x1
x1
x2
x2
x2
!
!
20.
. The net e↵ect is to interchange the
1
1
x2 R2 +R!
x1 + x2
x1 + x2 R2 +R!
x1
x1
first and second rows. So this sequence of elementary row operations is the same as R1 $ R2 .
21. Note that 3R2 2R1 is not an elementary row operation since it
Ri + kRj (it is not of the third form since the coefficient of R2 is



3
1
3 1
3
2
3R2
1
2 4 R2 3 R!
0 10
! 0
3
is not of the form Ri $ Rj , kRi , or
3, not 1). However,
1
,
10
so this is in fact a composition of elementary row operations.
22. We can create a 1 in the top left corner as follows:




1
1
3 2 R1 $R2 1 4
3 2 3 R!
1
! 3 2 ,
1 4
1 4
1
2
3
4
,

3
1
2
4

! 1
1
R1 2R2
6
.
4
The first method, simply interchanging the two rows, requires the least computation and seems simplest.
Our next choice would probably be the third method, since at least it produces integer results.
2.2. DIRECT METHODS FOR SOLVING LINEAR SYSTEMS
61
23. The rank of a matrix is the number of nonzero rows in its row echelon form.
(1) Since A is not in row echelon form,
2
1
40
0
(2)
(3)
(4)
(5)
we first put it in row echelon form:
3
2
3
0 1
1 0 1
3 4
0 35 R2 $R!
0 1 05 .
1 0
0 0 3
Since the row echelon form has three nonzero rows, rank A = 3.
A is already in row echelon form, and it has two nonzero rows, so rank A = 2.
A is already in row echelon form, and it has two nonzero rows, so rank A = 2.
A is already in row echelon form, and it has no nonzero rows, so rank A = 0.
Since A is not in row echelon form, we first put it in row echelon form:
2
3
2
3
1 0 3
4 0
1 0 3
4 0
3 4
40 0 0
0 05 R2 $R!
0 1 5
0 15 .
0 1 5
0 1
0 0 0
0 0
Since the row echelon form has two nonzero rows, rank A = 2.
(6) Since A is not in row echelon form, we first put it in row echelon form:
2
3
2
3
0 0 1
1 0 0
3 4
40 1 05 R1 $R!
0 1 05
1 0 0
0 0 1
Since the row echelon form has three nonzero rows, rank A = 3.
(7) Since A is not in row echelon form, we first put it in row echelon form:
2
3
2
3
2
3
1 2 3
1
2
3
1
2
3
61 0 07
6
60
2
37
2
37
6
7 R2 R1 6 0
7
6
7
6
7
7 R3 + 12 R2 6
!6
17
40 1 15
40
5
1
15
0
! 40
2
0
0
1
0
0
1
0
0
1
2
1
60
6
6
40
R4 +2R3
! 0
Since the row echelon form has three nonzero rows, rank A = 3.
(8) A is already in row echelon form, and it has three nonzero rows, so rank A = 3.
24. Such a matrix has 0, 1, 2, or 3 nonzero rows.
2
2
0
0
3
3
37
7
.
17
5
2
0
2
3
0 0 0
• If it has no nonzero rows, then all rows are zero, so the only such matrix is 40 0 05.
0 0 0
• If it has one nonzero row, that must be the first row, and any entries in that row following the 1
are arbitrary, so the possibilities are
2
3
2
3
2
3
1 ⇤ ⇤
0 1 ⇤
0 0 1
4 0 0 05 ,
40 0 05 ,
40 0 05 ,
0 0 0
0 0 0
0 0 0
where ⇤ denotes that any entry is allowed.
• If it has two nonzero rows, then the third row is zero. There are several possibilities for the
arrangements of 1s in the first two rows: they could be in columns 1 and 2, 1 and 3, or 2 and
3. In each case, the entries following each 1 are arbitrary, except for the column in the first row
above the 1 in the second row. This gives the possibilities
2
3
2
3
2
3
1 0 ⇤
1 ⇤ 0
0 1 0
4 0 1 ⇤5 ,
40 0 15 ,
40 0 15 .
0 0 0
0 0 0
0 0 0
62
CHAPTER 2. SYSTEMS OF LINEAR EQUATIONS
2
1
• If the matrix has all nonzero rows, then the only possibility is 40
0
entries must be zero since each column has a one in it.
0
1
0
3
0
05. Note that all other
1
25. First row-reduce the augmented matrix of the system:
2
1
6
42
4
2
1
1
3
1
1
3
9
7R
05 2
4 R3
2
3
2
1
2
3
9
1
7 1 R2 6
2R1 6
! 40
5
7
185 5 !
40
4R1
9 13
32
0
! 0
2
3
2
1 2
3
9
1 2
6
6
7
18 7
40 1
40 1
5
5 5 5
R3 +9R2
2
2
2 R3
! 0 0
! 0 0
5
5
2
3
2
R1 +3R3
2
! 1 2 0 12 R1 2R!
1 0
7
6
R2 + 75 R3 6
55
40 1
! 40 1 0
0 0 1
1
0 0
2
1
9
3
9
18 7
5 5
32
3
7
5
13
3
3
9
7
18 7
5
5 5
1
1
3
0 2
7
0 55
0 1
2
3 2 3
x1
2
Thus the solution is 4x2 5 = 455.
x3
1
26. First row-reduce the augmented matrix of the system:
2
1
6
4 1
3
1 1
3 1
1 7
3
0
7
55
2
2
1
R2 +R1 6
! 40
R3 3R1
! 0
1 1
2 2
4 4
3
0
7
55
2
2
1
1
6
2 R2
! 40
0
1 1
1 1
4 4
3
0
57
25
2
R3
2
1
6
40
4R2
! 0
3
0
57
25
8
1 1
1 1
0 0
This is an inconsistent system since the bottom row of the matrix is zero but the constant is nonzero.
Thus the system has no solution.
27. First row-reduce the augmented matrix of the system:
2
1
4 1
2
3
2
4
2
1
6
3
0
05
0
3
1
3
2 0
! 40
1
1 05
2R1
! 0 10 10 0
2
3
1
3
2 0
40
1
1 05
0 10 10 0 R3
R2 +R1
R3
2
R2
!
2
1
40
10R2
! 0
3
1
0
2
1
0
3
0
05
0
2
! 1
40
0
R1 +3R2
0
1
0
1
1
0
3
0
05
0
Since there is a row of zeros, and no leading 1 in the third column, we know that x3 is a free variable,
say x3 = t. Back substituting gives x2 + t = 0 and x1 + t = 0, so that x1 = x2 = t. So the solution is
2
3 2 3
2 3
x1
t
1
4x2 5 = 4 t5 = t 4 15 .
x3
t
1
2.2. DIRECT METHODS FOR SOLVING LINEAR SYSTEMS
63
28. First row-reduce the augmented matrix of the system:
2
3 1 R1 2
3
3
1
2
3
1
4 1 2 !
1
2 12
2
2
6
7
6
7 R
1
0
1 15
1
0
1 15 2
43
43
3
4
1
1 2
3
4
1
1 2 R3
2
3
1
1
2
2
2
2
6
R
2
3
10
11
0
1
4
!
11
11
17
5
7
0
2
2
2
3
2
11
2
17
2
1
3R1 6
! 40
3R1
! 0
3
1
2
3
2
5
2
2
1
1
2
6
1 7
0
5
4
11
R3 + 17
1
2 R2
! 0
2
3
2
1
2
3
11
2
11
1
0
This matrix is in row echelon form, but we can continue, putting it in reduced row
1
2
3 R1 3 R2 2
3 R1 + 11
R3 2
1
1
1
7
4
2
1 32
2
! 1
! 1 0
2
2
11
11
11
6
6
3
3
10
1 7
3
10
1 7 R2 + 11
R3 6 0
···
0
1
40 1
5
4
5
11
11
11
11
11
11
!4
11
2 R3
1
4
7
0 0
1
4
7
0
! 0 0
Using z = t as the parameter gives w = 1 t, x = 2 2t, y = 7
2 3 2
3 2 3
2
w
1 t
1
6 x 7 62 2t7 627
6
6 7=6
7 6 7
6
4 y 5 47 4t5 = 475 + t 4
z
t
0
29. First row-reduce the augmented matrix of the system:
2
3 1 R1 2
3
2
3
1
3
3
2 1
3 2 !
1 12
1
2
2
2
6
7
6
7 R 4R1 6
7
75
75 2 !
1
15
44 1
44 1
40
1
2 5
1
2 5
1 R3 2R!
0
4
4


x
2
The solution is 1 =
.
x2
1
3
2
5
7
1
2
17
25
1
2
2
10
11
8
11
3
1
2
1 7
11 5
14
11
echelon form:
3
0 0 1 1
7
1 0 2 25
0 1 4 7
4t, z = t for the solution, or
3
1
27
7.
45
1
2
1
R2 6
! 40
0
1
2
1
4
3
2
3
7
15
4
R1
R3
2
! 1
6
40
4R2
! 0
1
2 R2
0
1
0
3
2
7
15
0
0
2
0
3
2
15
0
30. First row-reduce the augmented matrix of the system:
2
1
4 2
1
3
6
3
2
1
4
3
0
35
2
4
2
8
2
!
2
1
R2 +2R1 4
! 0
R3 +R1
! 0
1
4
! 0
0
1
3 R2
R1
3 2
0 1
0 2
3
0
0
2
3
2
3
2
15
2
4
2
4
R3
4
6
4
2
3
0
35
2
1
40
2R2
! 0
The system has been reduced to
x1
3x2 =
2
x3
2x4 =
1.
3 2
0 1
0 0
4
2
0
2
! 1
40
0
R1 2R2
3
2
15
0
3 0
0 1
0 0
x2 and x4 are free variables; setting x2 = s and x4 = t and back-substituting gives for the solution
2 3 2
3 2 3
2 3
2 3
x1
3s 2
2
3
0
6x2 7 6 s 7 6 07
617
607
6 7=6
7 6 7
6 7
6 7
4x3 5 4 2t + 1 5 = 4 15 + s 405 + t 425 .
x4
t
0
0
1
64
CHAPTER 2. SYSTEMS OF LINEAR EQUATIONS
31. First row-reduce the augmented matrix of the system:
2
1
2
61
46
1
3
1
1
2
0
1
0
2
6
3
0
3 2R1 2
2
! 1
7 6R2 4
15 ! 1
3
1
8 3R!
0
1
4
2
3
0
2
0
6
2
1
2
! 40
1
R3 R1
2
! 0
2
1 2
40 1
R3 +2R2
! 0 0
2
R1 2R2
! 1 0
40 1
0 0
R2 R1
12
18
0
3
4
65
24
0
6
12
2
2
4
12
6
12
0
6
12
2
2
0
12 0
6 6
0 0
6
2
0
0
6
0
3
4
105
0
3
4
105
20
3
24
105
0
12
6
0
Since the matrix has rank 2 but there are 5 variables, there are three free variables, say x3 = r,
x4 = s, and x5 = t. Back-substituting gives x1 6r 12t = 24 and x2 + 2r 6s + 6t = 10, so that
x1 = 24 + 6r + 12t and x2 = 10 2r + 6s 6t. Thus the solution is
2 3 2
3
2 3
2 3
2 3
x1
24
6
0
12
6x2 7 6 107
6 27
667
6 67
6 7 6
7
6 7
6 7
6 7
6x3 7 = 6 07 + r 6 17 + s 607 + t 6 07 .
6 7 6
7
6 7
6 7
6 7
4x4 5 4 05
4 05
415
4 05
x5
0
0
0
1
32. First row-reduce the augmented matrix of the system:
2p
2
6
4 0
0
1
p
2
1
2
3
p
2
3
1
p 7
25
1
2
! 1
1
p
R 6
2 2 40
!
0
1
p
R
2 1
2
1
6
0
4
p
2 R3
! 0
p
p
2
2
2
p
3 2
1
1
p2
2
p
0
1
0
3
2+
p 2
3 2
2
1
2
! 1
7
6
15
40
R
+R
3
2
1
! 0
p
2
2
3
p 3
2
7
15
0
R1
p
2
2 R2
0
1
0
p
R1 ( 2+ 32 )R3
2
! 1
6
R2 + 3 2 2 R3
0
! 4
0
p
p
3
2
3 2
p2
2
2
2+
p
0
1
0
p 3
2
7
15
0
p 3
2
7
15
0
0
0
1
33. First row-reduce the augmented matrix of the system:
2
1
61
6
40
1
1
1
1
1
2 1
1 1
1 0
0 1
3
1
R2
07
7
15
2 R4
2
1
!6
60
40
R1
! 0
2
1
60
6
R3 +2R2 4
! 0
0
R1
1
2
1
0
1
1
0
0
2 1
3 0
1 0
2 0
2 1
1 0
1 0
2 0
3
2
1
1
60
17
3 6
7 R2 $R!
40
15
1
0
3
2
1
1
60
17
7
6
40
35
R4 2R3
1
! 0
1
1
2
0
1
1
0
0
2
1
3
2
2 1
1 0
1 0
0 0
1
0
0
0
3
1
17
7
15
1
3
1
17
7
35
7
We need not proceed any further — since the last row says 0 = 7, we know the system is inconsistent
and has no solution.
2.2. DIRECT METHODS FOR SOLVING LINEAR SYSTEMS
65
34. First row-reduce the augmented matrix of the system:
2
1
61
6
41
1
1
2
3
4
1
3
6
10
1
4
10
20
3
4
107
7
205
35
2
R4
1
60
6
40
3R3
! 0
2
1
! 60
6
R1 4
! 0
R1
! 0
R2 R1
R3
R4
1
1
0
0
1
2
1
0
1
3
3
1
1
1
2
3
3
4
67
7
45
1
1
2
5
9
1
3
9
19
R1 R4
R2
R3
!
2
3
4
67
7
165
31
1
3R4 6
! 60
3R4 40
!
0
R3
R4
1
1
0
0
1
2
1
0
2
1
60
2R2 6
! 40
3R2
! 0
3
3
37
7
15
1
0
0
0
1
1
1
0
0
1
2
1
3
R1 R3
R2
!
1
3
3
10
2
1
2R3 60
!6
40
0
3
4
67
7
45
13
1
1
0
0
R1 R2
2 3 2 3
a
1
6 b 7 6 17
7 6 7
so that 6
4 c 5 = 415 is the unique solution.
d
1
0
0
1
0
2
0
0
0
1
! 1
60
6
40
0
0
1
0
0
3
2
17
7
15
1
0
0
1
0
0
0
0
1
3
1
17
7
15
1
35. If we reverse the first and third rows, we get a system in row echelon form with a leading 1 in every
row, so rank A = 3. Thus this is a system of three equations in three variables with 3 3 = 0 free
variables, so it has a unique solution.
36. Note that R3 2R2 results in the row 0 0 0 0 | 2, so that the system is inconsistent and has no solutions.
37. This system is a homogeneous system with four variables and only three equations, so the rank of the
matrix is at most 3 and thus there is at least one free variable. So this system has infinitely many
solutions.
38. Note that R3
R3 R1 R2 gives a zero row, and then R2
R2 6R1 gives a matrix in row echelon
form. So there are three free variables, and this system has infinitely many solutions.
39. It suffices to show that the row-reduced matrix has no zero rows, since it then has rank 2, so there are
no free variables and no possibility of an inconsistent system.
First, if a = 0, then since ad
the augmented matrix:
bc 6= 0, we see that both b and c are nonzero. Then row-reduce
bc =

0
c
b
d
r
s

c
! 0
R1 $R2
d
b
s
.
r
Since b and c are both nonzero, this is a consistent system with a unique solution.
Second, if c = 0, then since ad
augmented matrix is
bc = ad 6= 0, we see that both a and d are nonzero. Then the

a
0
b
d
r
.
s
This is row-reduced since both a and d are nonzero, and it clearly has rank 2, so this is a consistent
system with a unique solution.
Finally, if a 6= 0 and c 6= 0, then row-reducing gives

a
c
Since ac 6= 0 and also ad
b
d
r
s

! ac
aR2
! ac
cR1
bc
ad
rc
as

ac
! 0
R2 R1
bc
ad
bc
rc
.
as
bc 6= 0, this is a rank 2 matrix, so it has a unique solution.
66
CHAPTER 2. SYSTEMS OF LINEAR EQUATIONS
40. First reduce the augmented matrix of the given system to row echelon form:

k
2
2
4
3
6

2
! k
R1 $R2
6
3
4
2
R2

2
! 0
1
2 kR1
4
2 + 2k
6
3 + 3k
(a) The system has no solution when this matrix has a zero row with a corresponding nonzero constant.
Now, 2 + 2k = 0 for k = 1, in which case 3 + 3k = 0 as well. Thus there is no value of k for
which the system has no solutions.
(b) The system has a unique solution when rank A = 2, which happens when the bottom row is
nonzero, i.e., for k =
6
1.
(c) The system has an infinite number of solutions when A has a zero row with a zero constant. From
part (a), this happens when k = 1.
41. First row-reduce the augmented matrix of the system:

1
k
k
1
1
1

1
! 0
R2 kR1
k
1
1
k2
1
k
(a) If 1 k 2 = 0 but 1 k 6= 0, this system will be inconsistent and have no solution. 1 k 2 = 0 for
k = 1 and k = 1; of these, only k = 1 produces 1 k 6= 0. Thus the system has no solution
for k = 1.
k 2 6= 0, so for k 6= ±1.
(b) The system has a unique solution when 1
k2 = 1
(c) The system has infinitely many solutions when the last row is zero, so when 1
This happens for k = 1.
k = 0.
42. First reduce the augmented matrix of the given system to row echelon form:
2
1
41
2
2 3
1 1
1 4
3
2
k5
k2
R2
R3
2
1
R1
! 40
2R1
! 0
2
3
3
3
2
2
k
k2
3
2
25
4
R3
2
1
40
R2
! 0
2
3
0
3
2
0
k
k
k2
3
2
25
2
(a) The system has no solution when the matrix has a zero row with a corresponding nonzero constant.
This occurs for k 2 k 2 6= 0, i.e. for k 6= 2, 1.
(b) Since the augmented matrix has a zero row, the system does not have a unique solution for any
value of k; z is a free variable.
(c) The system has an infinite number of solutions when the augmented matrix has a zero row with
a zero constant. This occurs for k 2 k 2 = 0, i.e. for k = 2, 1.
43. First reduce the augmented matrix of the given system to row echelon form:
2
1
41
k
1
k
1
k
1
1
3
1
15
2
R2
R3
2
1
R1
! 40
kR1
! 0
1
k
1
3
1
0 5
2 k
k
1
k
1 k
1 k2
2
1
40
R3 +R2
! 0
1
k
k
1
0
1
2
k
k
k2
3
1
0 5
2 k
(a) The system has no solution when the matrix has a zero row with a corresponding nonzero constant.
This occurs when 2 k k 2 = (1 k)(2 + k) = 0 but 2 k 6= 0. For k = 1, 2 k = 3 6= 0,
but for k = 2, 2 k = 0. Thus the system has no solution only when k = 1.
(b) The system has a unique solution when 2
k
k 2 6= 0; that is, when k 6= 1 and k 6=
2.
(c) The system has an infinite number of solutions when the augmented matrix has a zero row with
a zero constant. This occurs when 2 k k 2 = 0 and 2 k = 0. From part (a), we see that
this is exactly when k = 2.
2.2. DIRECT METHODS FOR SOLVING LINEAR SYSTEMS
67
44. There are many examples. For instance,
(a) The following system of n equations in n variables has infinitely many solutions (for example,
choose k and set x1 = k and x2 = k, with the other xi = 0):
x1 + x2 + · · · + xn = 0
2x1 + 2x2 + · · · + 2xn = 0
..
.
nx1 + nx2 + · · · + nxn = 0 .
Let m be any integer greater than n. Then the following system of m equations in n variables
has infinitely many solutions (for example, choose k and set x1 = k and x2 = k, with the other
xi = 0):
x1 + x2 + · · · + xn = 0
2x1 + 2x2 + · · · + 2xn = 0
..
.
mx1 + mx2 + · · · + mxn = 0 .
(b) The following system of n equations in n variables has a unique solution:
x1 = 0
x2 = 0
..
.
xn = 0 .
Let m be any integer greater than n. Then a system whose first n equations are as above, and
which continues with equations x1 + x2 = 0, x1 + 2x2 = 0, and so forth until there are m equations
clearly also has a unique solution.
45. Observe that the normal vectors to the planes, which are [3, 2, 1] and [2, 1, 4], are not parallel, so the
two planes will indeed intersect in a line. The line of intersection is the points in the solution set of
the system
3x + 2y + z =
1
2x
5
y + 4z =
Reduce the corresponding augmented matrix to row echelon form:
"
3
2
2 1
1 4
1
5
#
"
! 1
2
1
3 R1
2
3
1
3
1
4
1
3
#
"
2
3
7
3
1
R2 2R1
5
! 0
"
1 23
3
7 R2
! 0 1
1
3
10
3
1
3
10
7
1
3
17
7
#
1
3
17
3
#
R1
Thus z is a free variable; we set z = 7t to eliminate some fractions. Then
x=
so that the line of intersection is
9
9t + ,
7
2 3 2
x
6 7 6
4y 5 = 4
z
y = 10t
3
3
9
6 7
+ t 4 105 .
0
7
9
7
17 7
7 5
2
17
,
7
"
! 1
0
2
3 R2
0
1
9
7
10
7
9
7
17
7
#
68
CHAPTER 2. SYSTEMS OF LINEAR EQUATIONS
46. Observe that the normal vectors to the planes, which are [4, 1, 1] and [2, 1, 3], are not parallel, so the
two planes will indeed intersect in a line. The line of intersection is the points in the solution set of
the system
4x + y + z = 0
2x
y + 3z = 2
Reduce the corresponding augmented matrix to row echelon form:
"
4
2
1 1
1 3
0
2
#
"
! 1
2
1
4 R1
1
4
1
4
1
3
#
0
2
R2
"
1
2R1
! 0
"
1
2
3 R2
! 0
1
4
3
2
1
4
5
2
1
4
1
#
0
2
1
4
5
3
#
0
R1
4
3
"
! 1
0
1
4 R2
0
1
2
3
5
3
1
3
4
3
#
Letting the parameter be z = t, we get the parametric form of the solution:
x=
1
3
2
t,
3
y=
4 5
+ t,
3 3
z = t.
47. We are looking for examples, so we should keep in mind simple planes that we know, such as x = 0,
y = 0, and z = 0.
(a) Starting with x = 0 and y = 0, these planes intersect when x = y = 0, so they intersect on the
z-axis. So we must find another plane that contains the z-axis. Since the line y = x in R2 passes
through the origin, the plane y = x in R3 will have z as a free variable, and it contains (0, 0, 0),
so it contains the z-axis. Thus the three planes x = 0, y = 0, and x = y intersect in a line, the
z-axis. A sketch of these three planes is:
(b) Again start with x = 0 and y = 0; we are looking for a plane that intersects both of these, but not
on the z-axis. Looking down from the top, it seems that any plane whose projection onto the xy
plane is a line in the first quadrant will suffice. Since x + y = 1 is such a line, the plane x + y = 0
should intersect each of the other planes, but there will be no common points of intersection.
To find the line of intersection of x = 0 and x + y = 1, solve that pair of equations. Using
2.2. DIRECT METHODS FOR SOLVING LINEAR SYSTEMS
69
forward substitution we get y = 1, so the line of intersection is the line x = 0, y = 1, which is
the line [0, 1, t]. To find the line of intersection of y = 0 and x + y = 1, again solve using forward
substitution, giving the line x = 1, y = 0, which is the line [1, 0, t]. Clearly these two lines have
no common points. A sketch of these three planes is:
(c) Clearly x = 0 and x = 1 are parallel, while y = 0 intersects each of them in a line. A sketch of
these three planes is:
70
CHAPTER 2. SYSTEMS OF LINEAR EQUATIONS
(d) The three planes x = 0, y = 0, and z = 0 intersect only at the origin:
48. The two lines have parametric forms p + su and q + tv, where s and t are the parameters. These two
are equal when
p + su = q + tv, or su tv = q p.
Substitute the given values of the four vectors to get
2 3
2 3 2 3 2 3 2 3
1
1
2
1
3
s+t=
s 4 25 t 4 15 = 425 4 25 = 4 05 ) 2s t =
1
0
0
1
1
s
=
3
0
1.
From the third equation, s = 1; substituting into the first equation gives t = 2. Since s = 1, t = 2
satisfies the second equation, this is the solution. Thus the point of intersection is
2 3
2 3
x
0
4y 5 = p + 1u = 445 .
z
0
49. The two lines have parametric forms p + su and q + tv, where s and t are the parameters. These two
are equal when
p + su = q + tv, or su tv = q p.
Substitute the given values of the four vectors to get
2 3
2 3 2 3 2 3 2 3
1
2
1
3
4
s
s 405 t 435 = 4 15 415 = 4 05 )
1
1
1
0
1
s
2t =
3t =
t=
4
0
1.
From the second equation, t = 0; setting t = 0 in the other two equations gives s =
This is impossible, so the system has no solution and the lines do not intersect.
4 and s =
1.
50. Let Q = (a, b, c); then we want to find values of Q such that the line q + sv intersects the line p + tu.
The two lines intersect when
p + su = q + tv, or su
tv = q
p.
2.2. DIRECT METHODS FOR SOLVING LINEAR SYSTEMS
Substitute the values of the vectors to get
2 3
2 3 2 3
1
2
a
s 4 15 t 415 = 4 b 5
1
0
c
Solving those equations for a, b, and c gives
2
3 2
3
1
a 1
4 25 = 4 b 2 5 )
3
c+3
Q = (a, b, c) = (s
2t + 1, s
71
s
s
s
t + 2, s
2t = a 1
t= b 2
= c+3.
3).
Thus any pair of numbers s and t determine a point Q at which the two given lines intersect.
51. If x = [x1 , x2 , x3 ] satisfies u · x = v · x = 0, then the components of x satisfy the system
u1 x1 + u2 x2 + u3 x3 = 0
v1 x1 + v2 x2 + v3 x3 = 0.
There are several cases. First, if either vector is zero, then the cross product is zero, and the statement
does not hold: any vector that is orthogonal to the other vector is orthogonal to both, but is not a
multiple of the zero vector. Otherwise, both vectors are nonzero. Assume that u is the vector with a
lowest-numbered nonzero component (that is, if u1 = 0 but u2 6= 0, while v1 6= 0, then reverse u and
v). The augmented matrix of the system is

u1 u2 u3 0
.
v1 v2 v3 0
If u1 = u2 = 0 but u3 6= 0, then v1 = v2 = 0 and v3 6= 0 as well, so that u and v are parallel and thus
u ⇥ v = 0. However, any vector of the form
2 3
s
4t5
0
is orthogonal to both u and v, so the result does not hold in this case either.
Next, if u1 = 0 but u2 6= 0, then v1 = 0 as well, so the augmented matrix has the form



0 u2 u3 0
0 u2
u3
0
0 u2
u3
0
.
2
1
0 v2 v3 0 u2 R!
0 u2 v2 u2 v3 0 R2 v2 R!
0 0 u2 v 3 u3 v 2 0
2 3
2 3
0
0
Then x1 is a free variable. If u2 v3 u3 v2 = 0, then u = 4u2 5 and v = 4v2 5 are again parallel, so
u3
v3
their dot product is zero and again the result does not hold. But if u2 v3 u3 v2 6= 0, then x3 = 0,
which implies that x2 = 0 as well. So any vector that is orthogonal to both u and v in this case is of
the form
2 3 2 3
x1
s
4x2 5 = 405 ,
x3
0
and
and since u2 v3
2
3 2 3 2
0
0
u2 v 3
4u2 5 ⇥ 4v 2 5 = 4u3 · 0
u3
v3
0 · v2
3 2
3
u3 v 2
u2 v 3 u3 v 2
5,
0 · v3 5 = 4
0
0 · u2
0
u3 v2 6= 0, the orthogonal vectors are multiples of the cross product.
Finally, assume that u1 6= 0. Then the augmented matrix reduces as follows:



u1 u2 u3 0
u1
u2
u3
0
u1
u2
2
1
v1 v2 v3 0 u1 R!
u1 v1 u1 v2 u1 v3 0 R2 v1 R!
0 u1 v 2 u2 v 1
u3
u1 v 3
u3 v 1
0
.
0
72
CHAPTER 2. SYSTEMS OF LINEAR EQUATIONS
Now consider the last row of this matrix. If it consists of all zeros, then u1 v2 = u2 v1 and u1 v3 = u3 v1 .
If v1 6= 0, then uu21 = vv21 and also uu31 = vv31 , so that again u and v are parallel and the result does not
hold. Otherwise, if v1 = 0, then u1 v2 = u1 v3 = 0, and u1 was assumed nonzero, so that v2 = v3 = 0
and thus v is the zero vector, which it was assumed not to be. So at least one of u1 v2 u2 v1 and
u1 v3 u3 v1 is nonzero.
If they are both nonzero, then x3 = t is free, and
(u1 v2
u2 v1 )x2 = (u3 v1
u1 v3 )t,
so thatx2 =
u3 v 1
u1 v 2
u1 v 3
t.
u2 v 1
But then u1 x1 + u2 x2 + u3 x3 = 0, so that
✓
◆
u3 v 1 u1 v 3
u1 (u3 v2 u2 v3 )
u1 x1 + u2
t + u3 t = u1 x1 +
t = 0,
u1 v 2 u2 v 1
u1 v 2 u2 v 1
and thus
Thus the orthogonal vectors are
or, clearing fractions,
2
x1 =
u2 v 3
u1 v 2
3
2
x1
6 7
6x2 7 =
4 5
x3
2
x1
3
u3 v 2
t.
u2 v 1
u2 v3
6 u1 v2
u3 v1
t6
4 u1 v2
2
u2 v 3
6 7
6
6 x2 7 = t 6 u3 v 1
4 5
4
x3
u1 v 2
3
1
u3 v2
u2 v1 7
u1 v3 7 ,
u2 v1 5
u3 v 2
3
7
u1 v 3 7
5,
u2 v 1
so that orthogonal vectors are multiples of the cross product.
If u1 v2 u2 v1 6= 0 but u1 v3 u3 v1 = 0, then again v1 = 0 forces v = 0, which is impossible. So v1 =
6 0
and thus uu31 = vv31 . Now, x3 = t is free and x2 = 0, so that u1 x1 + u3 t = 0 and then x1 = uu31 t. Let
t = (u1 v2 u2 v1 )s; then
x1 =
u3 (u1 v2 u2 v1 )
s=
u1
u3 u1 v 2
u3
s + (u2 v1 )s =
u1
u1
u3 v 2 s +
v3
(u2 v1 )s = (u2 v3
v1
u3 v2 )s.
Thus orthogonal vectors are of the form
2 3
2
3
x1
u2 v 3 u3 v 2
4x2 5 = s 4
5,
0
x3
u1 v 2 u2 v 1
and since u1 v3
u3 v1 = 0, this column vector is u ⇥ v, and again the result holds.
The final case is that u1 v3 u3 v1 6= 0 but u1 v2 u2 v1 = 0. Again v1 = 0 forces v = 0, which is
impossible, so that v1 6= 0 and thus uu21 = vv21 . Now, x2 = t is free and x3 = 0, so that u1 x1 + u2 t = 0
and then x1 = uu21 t. Let t = (u1 v3 u3 v1 )s; then
x1 =
u2 (u1 v3 u3 v1 )
s=
u1
u2 u1 v 3
u2
s + u3 v 1 s =
u1
u1
u2 v 3 s +
Thus orthogonal vectors are of the form
2 3
2
3
x1
u2 v 3 u3 v 2
4 x2 5 = s 4 u1 v 3 u3 v 1 5 ,
x3
0
and since u1 v2
Whew.
v2
u3 v1 s = (u3 v2
v1
u2 v1 = 0, this column vector is u ⇥ v, and again the result holds.
u2 v3 )s.
2.2. DIRECT METHODS FOR SOLVING LINEAR SYSTEMS
73
52. To show that the lines are skew, note first that they have nonparallel direction vectors, so that they
are not parallel. Suppose there is some x such that
x = p + su = q + tv.
Then su
tv = q
p. Substituting the given values into this equation gives
2 3
2 3 2 3 2 3 2 3
1
1
2s
= 1
2
0
0
s 4 35 t 4 65 = 4 15 415 = 4 05 ) 3s 6t = 0
1
1
1
0
1
s+ t= 1.
But the first equation gives s = 12 , so that from the second equation t = 14 . But this pair of values
does not satisfy the third equation. So the system has no solution and thus the lines do not intersect.
Since they are not parallel, they are skew.
Next, we wish to find a pair of parallel planes, one containing each line. Since u and v are direction
vectors for the lines, those direction vectors will lie in both planes since the planes are required to be
parallel. So a normal vector is given by
2
3 2 3
3 · ( 1) 1 · 6
3
u ⇥ v = 4 1 · 0 2 · ( 1) 5 = 4 25 .
2 · 6 ( 3) · 0
12
So the planes we are looking for will have equations of the form 3x + 2y + 12z = d. The plane that
passes through P = (1, 1, 0) must satisfy 3 · 1 + 2 · 1 + 12 · 0 = d, so that d = 1. The plane that
passes through Q = (0, 1, 1) must satisfy 3 · 0 + 2 · 1 + 12 · ( 1) = d, so that d = 10. Hence the
two planes are
3x + 2y + 12z =
53. Over Z3 ,

1
1
2
1
1 containing p + su,
1
2

! 1
0
R2 +2R1

x
0
Thus the solution is
=
.
y
2
54. Over Z2 ,
2
1
40
1
1
1
0
0
1
1

3
1
05
1
2
1
40
R3 +R1
! 0
1
1
1
2
2
0
1
1
3x + 2y + 12z =

1
1
1
2R2
! 0
3
1
05
0
2
2
1
1
40
R3 +R2
! 0

! 1
0
1
2
1
1
0
10 containing q + tv.
0
1
0
R1 +R2
0
1
3
1
05
0
2
! 1
40
0
Thus z = t is a free variable, and we have x + t = 1, so that x = 1
that y t = t. Thus the solution is
2 3 2
3 8 2 3 2 39
x
1+t
0 =
< 1
4y 5 = 4 t 5 = 405 , 415 ,
:
;
z
t
0
1
R1 +R2
0
.
2
0
1
0
1
1
0
3
1
05
0
t = 1 + t and also y + t = 0, so
since the only possible values for t in Z2 are 0 and 1.
55. Over Z3 ,
2
1 1 0
40 1 1
1 0 1
3
1
05
1
2
1
40
R3 +2R1
! 0
1
1
2
0
1
1
3
1
05
0
2
! 1
40
R3 +R2
! 0
R1 +2R2
0
1
0
3
1
05
···
3
0 2R!
2
3
1 0 2 1
· · · 4 0 1 1 05
0 0 1 0
2
1
2
R1 +R3
!
2
1
R2 +2R3 40
!
0
0
1
0
0
0
1
3
1
05
0
74
CHAPTER 2. SYSTEMS OF LINEAR EQUATIONS
Thus the unique solution is
56. Over Z5 ,

2 3 2 3
x
1
4 y 5 = 4 05 .
z
0
3
1
2
4
1
1

! 1
1
2R1
4
4
2
1

1
! 0
R2 +4R1
4
0
1
.
4
Since the bottom row says 0 = 4, this system is inconsistent and has no solutions.
57. Over Z7 ,


1
2 1 5R!
1
4 1
1


x
3
.
Thus the solution is
=
y
3
3
1
3
4
5
1
R2 +6R1
0
4
0
3
4
1
3
1
3
1
27
7
45
1

1
! 0
3
1
R3 +4R2
···

! 1
0
R1 +4R2
5
3
0
1
3
.
3
58. Over Z5 ,
2
1
61
6
42
1
0
2
2
0
0
4
0
3
4
0
1
0
3
1
37
7
15
2
2
1
! 60
6
R3 +3R1 4
! 0
R4 +4R1
! 0
R2 +4R1
2
1
60
···6
40
0
0
2
0
0
0
2
2
0
0
4
1
3
Thus x4 = t is a free variable, and
4
1
2
1
3
1
27
7
25
1
!
2
1
! 6
60
40
R4 +2R3
! 0
3R2
0
1
0
0
0
2
1
0
4
3
2
0
3
1
17
7
25
0
2
1
R2 +3R3 6
! 60
40
0
0
1
0
0
0
0
1
0
4
4
2
0
3
1
27
7
25
0
2
3 2
3 2
3
x1
1 4t
1+t
6x2 7 62 4t7 6 2 + t 7
6 7=6
7 6
7
4x3 5 42 2t5 = 42 + 3t5 .
x4
t
t
Since t = 0, 1, 2, 3, or 4, there are five solutions:
2 3
2 3
2 3
1
2
3
627
6 37
6 47
6 7, 6 7, 6 7,
425
4 05
4 35
0
1
2
2 3
4
607
6 7,
415
3
2 3
0
617
7
and 6
445 .
4
59. The Rank Theorem, which holds over Zp as well as over Rn , says that if A is the matrix of a consistent
system, then the number of free variables is n rank(A), where n is the number of variables. If there
is one free variable, then it can take on any of the p values in Zp ; since 1 = n rank(A), the system
has exactly p = p1 = pn rank(A) solutions. More generally, suppose there are k free variables. Then
k = n rank(A). Each of the free variables can take any of p possible values, so altogether there are
pk choices for the values of the free variables. Each of these yields a di↵erent solution, so the total
number of solutions is pk = pn rank(A) .
60. The first step in row-reducing the augmented matrix of this system would be to multiply by the
appropriate number to get a 1 in the first column. However, neither 2 nor 3 has a multiplicative
inverse in Z6 , so this is impossible. So the best we can do in terms of row-reduction is


2 3 4
2 3 4
1
4 3 2 R2 +R!
0 0 0
Exploration: Lies My Computer Told Me
75
Thus y is a free variable, and 2x + 3y = 4. To solve this equation, choose each possible value for y in
turn:
• When y = 0, y = 2, or y = 4 we get 2x = 4, so that x = 2 or x = 5.
• When y = 1 or y = 3, we get 2x + 3 = 4, or 2x = 1, which has no solution.
So altogether there are six solutions:


2
5
,
,
0
0

2
,
2

5
,
2

2
,
4

5
.
4
Exploration: Lies My Computer Told Me
1.
2.
x+
x+
y=0
=0
800x 800y = 0
800x + 801y = 800
)
801
800 y
x+
y=0
x + 1.0012y = 0
)
)
x=
y=
800
800 .
x
y=0
x + 1.0012y = 1
)
x=
y=
833.33
833.33 .
x
y=0
x + 1.001y = 1
)
x=
y=
1000
1000 ,
3. At four significant digits, we get
x+
y=0
x + 1.001y = 0
)
while at three significant digits, we get
x+
y=0
x + 1.00y = 0
)
x
y=0
x + 1.00y = 1
)
Inconsistent system; no solution.
4. For the equations we just looked at, the slopes of the two lines were very close (the angle between the
lines was very small). That means that changing one of the slopes even a little bit will result in a
large change in their intersection point. That is the cause of the wild changes in the computed output
values above — rounding to five significant digits is indeed the same as a small change in the slope of
the second line, and it caused a large change in the computed solution.
For the second example, note that the two slopes are
7.083
⇡ 1.55602,
4.552
2.693
⇡ 1.55575,
1.731
so that again the slopes are close, so we would expect the same kind of behavior.
Exploration: Partial Pivoting
1. (a) Solving 0.00021x = 1 to five significant digits gives x =
1
0.00021
⇡ 4761.9.
1
(b) If only four significant digits are kept, we have 0.0002x = 1, so that x = 0.0002
= 5000. Thus the
e↵ect of an error of 0.00001 in the input is an error of 238.1 in the result.
1
2. (a) Start by pivoting on 0.4. Multiply the first row by 0.4
= 250 to get [1, 249, 250]. Now we want
to subtract 75.3 times that row from the second row. Multiplying by 75.3 and keeping three
significant digits gives
75.3[1, 249, 250] = [75.3, 18749.7, 18825.0] ⇡ [75.3, 18700, 18800].
Then the second row becomes (again keeping only three significant digits)
[75.3, 45.3, 30]
[75.3, 18700, 18800] = [0, 18654.7, 18770] ⇡ [0, 18700, 18800].
76
CHAPTER 2. SYSTEMS OF LINEAR EQUATIONS
This gives y = 18800
18700 ⇡ 1.00535 ⇡ 1.01. Substituting back into the modified first row, which says
x + 249y = 250 gives x = 250 251.49 ⇡ 250 251 = 1.
Substituting x = y = 1 into the original system, however, yields two true equations, so this is in
fact the solution. The problem was that the large numbers in the second row, combined with the
fact that we were keeping only three significant digits, overwhelmed the information from adding
in the first row, so all that information was lost.
(b) First interchanging the rows gives the matrix

75.3
45.3
0.4
99.6
30
100
1
Multiply the first row by 75.3
to get (after reducing to three significant digits) [1, 0.602, 0.398].
Then subtract 0.4 times that row from the second row:
[0.4, 99.6, 100]
0.4[1, 0.602, 0.398] ⇡ [0, 99.8, 99.8].
This gives 99.8y = 99.8, so that y = 1; from the first equation, we have x
that x 0.602 · 1 = 0.398 and thus x = 1. We get the correct answer.
0.602y = 0.398, so
3. (a) Without partial pivoting, we pivot on 0.001 to three significant digits, so we divide the first row
by 0.001, giving [1, 995, 1000], then multiply it by 10.2 and add the result to row 2, giving the
matrix

1
995
1000
)
y ⇡ 1.01.
0
10100
10200
Back-substituting gives x = 1.00 2 1.00 = 0 (again remembering to keep only three significant
digits). This error was introduced because 1.00 and 50.0 were ignored when reducing to three
significant digits in the original row-reduction, being overwhelmed by numbers in the range of
10000.
Using partial pivoting, we pivot on 10.2 to three significant digits: the second row becomes
[1, 0.098, 4.9]. Then multiply it by 0.001 and subtract it from the first row, giving [0, 0.995, 1.00].
Thus 0.995y = 1.00, so that to three significant digits y = 1.00. Back-substituting gives x =
50.0 1.00
= 51.0
10.2
10.2 = 5. Pivoting on the largest absolute value reduced the error in the solution.
(b) Start by pivoting on the 10.0 at upper left:
2
3 2
10.0
7.00 0.00 7.00
10.0
4 3.0 2.09 6.00 3.915 ! 4 0.0
5.0
1.00 5.00 6.00
0.0
7.00
0.01
2.50
0.00
6.00
5.00
3
7.00
6.015 .
2.50
If we continue, pivoting on 0.01 to three significant digits, the second row then becomes
[0, 1, 600, 601]. Multiply it by 2.5 and add it to the third row:
2
3 2
3
10.0
7.00 0.00 7.00
10.0
7.00 0.00
7.00
4 0.0
0.01 6.00 6.015 ! 4 0.0
0.01 6.00
6.01 5 .
0.0
2.50 5.00 2.50
0.0
0
1500
1500
This gives z = 1.00, y = 1.00, and x = 0.00.
If instead we reversed the second and third rows and pivoted on the 2.5, we would first divide
that row by 2.5, giving [0, 1, 2, 1], then multiply it by 0.01 and add it to the third row:
2
3 2
3 2
3
10.0
7.00 0.00 7.00
10.0
7.00 0.00 7.00
10.0
7.00 0.00 7.00
4 0.0
2.50 5.00 2.505 ! 4 0.0
1
2
1 5 ! 4 0.0
1
2
1 5.
0.0
0.01 6.00 6.01
0.0
0.01 6.00 6.01
0.0
0
6.02 6.02
This also gives z = 1.00, y =
1.00, and x = 0.00.
Exploration: An Introduction to the Analysis of Algorithms
77
Exploration: An Introduction to the Analysis of Algorithms
1. We count the number of operations, one step at a time.
2
3 1 1 2
2 4
6
8 2 R!
1
4 3 9
4 3
6 125
1 1
1
1
1
There were three operations here: 12 · 4 = 2, 12 · 6 = 3, and
don’t actually perform that operation — once we chose 2,
2
3
2
1 2
3
4
R
3R
1 4
4 3 9
6 125 2 !
1 1
1
1
2
9
1
3
6
1
1
2
3
4
125
1
· 8 = 4. We don’t count 12 · 2 = 1 since we
we knew we were going to turn it into 1.
3
1 2
3 4
0 3
3 05
1 1
1 1
There were three operations here: 3 · 2 = 6, 3 · 3 = 9, and 3 · 4 = 12, fr a total of six operations so far.
Note that we are counting only multiplications and divisions, and that again we don’t count 3 · 1 = 3
since we didn’t actually have to perform it — we just replace the 3 in the second row by a 0.
2
3
2
3
1 2
3 4
1 2
3 4
4 0 3
40 3
3 05
3 05
1
1 1
1 1 R3 +R!
0 3
2 5
There were 3 operations here: 1 · 2 = 2,
2
1 2
40 3
0 3
1 · 3 = 3, and 1 · 4 = 4,
3
2
3 4 1
1 2
2 4
3 05 3 R!
0 1
2 5
0 3
In this step, there were 2 operations:
· ( 3) and
1
3
3
1
2
1
40
3R2
! 0
2
1
40
0
1
3
2
1
3
1
40
0
2
1
0
In this step, there was one operation,
R3
· 0 = 0, for a total of eleven so far.
2
3 · ( 1) = 3 and
In this step, there were 2 operations:
2
3
4
05
5
3
3
1
5
1
5
4
05
5
for a total of nine operations so far.
3
3 4
1 05
2 5
2
1
40
1
5 R3
! 0
2
1
0
3
1
5
3
4
05
5
3 · 0 = 0, for a total of 13 so far.
2
1
0
3
4
05
1
3
1
1
· 5 = 1, for a total of 14.
Finally, to complete the back-substitution, we need three more operations:
and 3 · x3 to find x1 . So in total 17 operations were required.
2. To compute the reduced row echelon form, we first compute the
14 operations. Then:
2
3
2
2
1 2
3 4 R1 2R!
1 0
40 1
40 1
1 05
0 0
1 1
0 0
This required two operations,
2
2 · ( 1) and
1
40
0
0
1
0
5
1
1
3
4
05
1
1 · x3 to find x2 , and 2 · x2
row echelon form above, which took
5
1
1
3
4
05
1
2 · 0, for a total of 16. Finally,
2
! 1
R2 +R3 40
!
0
R1 5R3
0
1
0
0
0
1
3
1
15
1
78
CHAPTER 2. SYSTEMS OF LINEAR EQUATIONS
This required two operations, both in the constant column: 1 · 1 and 5 · 1, for a total of 18. So for this
example, Gauss-Jordan required one more operation than Gaussian elimination. We would probably
need more data to conclude that it is generally less efficient.
3. (a) There are n operations required to create the first leading 1 because we have to divide every entry
in the first row, except a11 , by a11 (we don’t have to divide a11 by itself because the 1 results
from the choice of pivot. Next, there are n operations required to create the first zero in column
1, because we have to multiply every entry in the first row, except the 1 in column one, by a21 .
Again, we don’t have to multiply the 1 by itself, since we know we are simply replacing a21 by
zero. Similarly, there are n operations required to create each zero in column 1. Since there are
n 1 rows excluding the first row, introducing the 1 in row one and the zeros below it requires
n + (n 1)n = n2 operations.
(b) We can think of the resulting matrix, below and to the right of a22 , as an (n 1) ⇥ n augmented
matrix. Applying the process in part (a) to this matrix, we see that turning a22 into a 1 and
creating zeros below it requires (n 1) + (n 2)(n 1) = (n 1)2 operations. Continuing to the
matrix starting at row 3, column 3; this requires (n 2)2 operations. So altogether, to reduce the
matrix to row echelon form requires n2 + (n 1)2 + · · · + 22 + 12 operations.
(c) To find xn 1 requires one operation: we must multiply xn by the entry an 1,n . Then there are
two operations required to find xn 2 , and so forth, until we require n 1 operations to find x1 .
So the total number required for the back-substitution is 1 + 2 + · · · + (n 1).
(d) From Appendix B, we see that
k(k + 1)(2k + 1)
6
k(k + 1)
1 + 2 + ··· + k =
.
2
12 + 22 + · · · + k 2 =
Thus the total number of operations required is
12 + 22 + · · · + n2 + 1 + 2 + · · · + (n
n(n + 1)(2n + 1) n(n 1)
+
6
2
1 3 1 2 1
1 2 1
= n + n + n+ n
n
3
2
6
2
2
1
1
= n3 + n2
n.
3
3
1) =
For large values of n, the cubic term dominates, so the total number of operations required is
⇡ 13 n3 .
4. As we saw in Exercise 2 in this section, we get the same n2 + (n 1)2 + · · · + 22 + 12 operations to get
the matrix into row echelon form. Then to create the 1 zero above row 2 requires 1(n 1) operations,
since each of the remaining n 1 entries in row 2 must be multiplied. To create the 2 zeros above row
3 requires 2(n 2) operations, since each of the remaining entries in row 3 must be multiplied twice.
Continuing, we see that the total number of operations is
(n
1) + 2(n
2) + · · · +(n
1)(n
(n
1))
= (1 · n + 2 · n + · · · + (n
= n(1 + 2 + · · · + (n
1))
1) · n)
(12 + 22 + · · · + (n
(12 + 22 + · · · + (n
1)2 )
1)2 ).
Adding that to the operations required to get the matrix into row echelon form, and using the formulas
2.3. SPANNING SETS AND LINEAR INDEPENDENCE
79
from Appendix B, we get for the total number of operations
n2 + (n
1)2 + · · · + 22 + 12 + n(1 + 2 + · · · + (n
1)) (12 + 22 + · · · + (n
= n2 + n ·
n(n
1))2
1)
2
1 3 1 2
=n + n
n
2
2
1
1
= n3 + n2 .
2
2
Again, for large values of n, the cubic term dominates, so the total number of operations required is
⇡ 12 n3 .
2
2.3
Spanning Sets and Linear Independence
1. Yes, it is. We want to write
x

1
+y
1

so we want to solve the system of linear equations

2
1
,
=
1
2
x + 2y = 1
x
y = 2.
Row reducing the augmented matrix for the system gives


1
2 1
1 2 1 R1
R2 +R1
1
1 2
! 0 1 3
So the solution is x =

! 1
0
2R2
0
1
5 and y = 3, and the linear combination is v =
2. No, it is not. We want to write
x

4
+y
2

so we want to solve the system of linear equations
4x
5
.
3
5u1 + 3u2 .

2
2
=
,
1
1
2y = 2
2x + y = 1
Row-reducing the augmented matrix for the system gives
"
# 1
"
#
R1
1
1
4
2 2 4 !
1
2
2
2
1 1
2
1 1
"
1
R2 +2R1
! 0
1
2
1
2
0
2
#
So the system is inconsistent, and the answer follows from Theorem 2.4.
3. No, it is not. We want to write
2 3
2 3 2 3
1
0
1
x 415 + y 415 = 425 ,
0
1
3
so we want to solve the system of linear equations
x
=1
x+y =2
y = 3.
Rather than constructing the matrix and reducing it, note that the first and third equations force x = 1
and y = 3, and that these values of x and y do not satisfy the second equation. Thus the system is
inconsistent, so that v is not a linear combination of u1 and u2 .
80
CHAPTER 2. SYSTEMS OF LINEAR EQUATIONS
4. Yes, it is. We want to write
2 3
2 3 2 3
1
0
3
x 415 + y 415 = 4 25 ,
0
1
1
so we want to solve the system of linear equations
x
=
3
x+y =
2
y=
1
Rather than constructing the matrix and reducing it, note that the first and third equations force x = 3
and y = 1, and that these values of x and y also satisfy the second equation. Thus v = 3u1 u2 .
5. Yes, it is. We want to write
2 3
2 3
2 3 2 3
1
0
1
1
x 415 + y 415 + z 405 = 425 ,
0
1
1
3
so we want to solve the system of linear equations
x
+z = 1
x+y
=2
y+z = 3.
Row-reducing the associated augmented matrix gives
2
1 0 1
41 1 0
0 1 1
3
1
25
3
R2
2
1
R1 4
! 0
0
0
1
1
1
1
1
3
1
15
3
2
1
40
R2
! 0
R3
0
1
0
3
1 1
1 15
2 2
2
1 0
· · · 40 1
0 0
1
2 R3
!
1
1
1
This gives the solution x = 0, y = 2, and z = 1, so that v = 2u2 + u3 .
···
3
1
15
1
2
! 1
R2 +R3 40
!
0
R1 R3
0
1
0
0
0
1
3
0
25
1
6. Yes, it is. The associated linear system is
1.0x + 3.4y
1.2z =
3.2
0.4x + 1.4y + 0.2z =
2.0
4.8x
2.6
6.4y
1.0z =
Solving this system with a CAS gives x = 1.0, y = 1.0, z = 1.0, so that v = u1 + u2 + u3 .
7. Yes, it is. b is in the span of the columns of A if and only if the system Ax = b has a solution, which
is to say if and only if the linear system
x + 2y = 5
3x + 4y = 6
has a solution. Row-reducing the augmented
"
#
"
1 2 5
1
2
R2 3R1
3 4 6
2
! 0
Thus x =
4 and y = 92 , so that
A
matrix for the system gives
#
"
#
"
R1 2R2
5
1 2 5
! 1
1
9
2 R2
9
0
! 0 1 2

4
9
2
=

5
.
6
0
1
4
9
2
#
2.3. SPANNING SETS AND LINEAR INDEPENDENCE
81
8. Yes, it is. b is in the span of the columns of A if and only if the system Ax = b has a solution, which
is to say if and only if the linear system
x + 2y + 3z = 4
5x + 6y + 7z = 8
9x + 10y + 11z = 12
Row-reducing the augmented matrix for the system gives
2
1
45
9
2
6
10
3
7
11
3
4
85
12
R2
R3
2
1
5R1
! 40
9R1
! 0
2
4
8
3
8
16
3
4
125
24
2
1
1
4 R2 4 0
!
0
2
1
8
3
4
35
24
3
2
16
2
1
40
R3 +8R2
! 0
2
1
0
3
2
0
3
4
35
0
This system has an infinite number of solutions, so that b is indeed in the span of the columns of A.

a
9. We must show that for any vector
, we can write
b

1
x
+y
1


1
a
=
1
b
for some x, y. Row-reduce the associated augmented matrix:
#
"
#
"
"
1
1 a
1 1
a
1 1
1
R2 R1
2 R2
1
1 b
2 b a
! 0
! 0 1
a
a b
2
#
"
! 1
0
R1 R2
0
1
a+b
2
a b
2
#
So given a and b, we have



a+b 1
a b
1
a
·
+
·
=
.
1
1
b
2
2

a
10. We want to show that for any vector
, we can write
b
x



3
0
a
+y
=
2
1
b
for some x, y. Row-reduce the associated augmented matrix:
"
# 1
"
#
"
R1
3 0 a 3 !
1 0 a3
1
R2 +2R1
2 1 b
2 1 b
! 0
0
1
a
3
b+
2a
3
#
which shows that the vector can be written


✓
◆ 
a
2a
a
3
0
= ·
+ b+
·
.
b
2
1
3
3
11. We want to show that any vector can be written as a linear combination of the three given vectors, i.e.
that
2 3
2 3
2 3
2 3
a
1
1
0
4 b 5 = x 405 + y 415 + z 415
c
1
0
1
82
CHAPTER 2. SYSTEMS OF LINEAR EQUATIONS
for some x, y, z. Row-reduce the associated augmented matrix:
2
1
40
1
Thus
1
1
0
3
a
b5
c R3
0
1
1
2
1
40
R1
! 0
2
1 1
6
40 1
1
2 R3
! 0 0
2
R1 R2
! 1
6
40
0
2 3
a
4 b 5 = 1 (a
2
c
1
1
1
0
1
1
0
1
1
0
1
0
a
b
3
a
b 5
c a
3
b+c a
2
0
0
1
7
5
R2
3
a b+c
2
a+b c 7
5
2
b+c a
2
2 3
1
1
b + c) · 405 + (a + b
2
1
2
1 1 0
a
40 1 1
b
R3 +R2
! 0 0 2 b+c
2
3
1 1 0
a
7
R3 6
! 40 1 0 a+b2 c 5
0 0 1 b+c2 a
a
3
5
2 3
2 3
1
0
1
c) · 415 + ( a + b + c) · 415 .
2
0
1
12. We want to show that any vector can be written as a linear combination of the three given vectors, i.e.
that
2 3
2 3
2 3
2 3
a
1
1
2
4 b 5 = x 415 + y 425 + z 4 15
c
0
3
1
for some x, y, z. Row-reduce the associated augmented matrix:
2
1 1
41 2
0 3
2
1
1
3
a
b5
c
2
1
6
40
1
2 R3
! 0
Thus
2
1
1
3
1 1 2
a
40 1
b a5
1
b a 5
R3 3R2
c
0
0
2
c
+
3a 3b
!
3
2
3
a
1 1 0
3b 2a c
R1 2R3
!6
7
7
1
b a
b) 5
5 R2 +R3 40 1 0
2 (c + a
!
1
3b)
0 0 1 12 (c + 3a 3b)
2 (c + 3a
2
3
R1 R2
! 1 0 0 12 (7b 5a 3c)
6
7
1
b) 5
40 1 0
2 (c + a
0 0 1 12 (c + 3a 3b)
1
1
0
2 3
a
4 b 5 = 1 (7b
2
c
5a
2 3
1
1
3c) 415 + (c + a
2
0
1
1
3
2
1
1
a
3
2
2 3
1
1
b) 425 + (c + 3a
2
3
2
3
2
3b) 4 15 .
1

1
2
, these vectors are parallel, so their span is the line through the origin with
2

1
direction vector
.
2


x
1
(b) The vector equation of this line is
=t
. Note that this is just another way of saying
y
2
that the line is the span of the given vector. To derive the general equation of the line, expand
the above, giving the parametric equations x = t and y = 2t. Substitute x for t in the second
equation, giving y = 2x, or 2x + y = 0.
13. (a) Since

2
1
! 40
0
R2 R1
2
=
4
2.3. SPANNING SETS AND LINEAR INDEPENDENCE
83


0
3
is just the origin, and the origin is in the span of
, the span of these two
0
4


3
3
vectors is just the span of
, which is the line through the origin with direction vector
.
4
4


x
3
(b) The vector equation of this line is
=t
. Note that this is just another way of saying that
y
4
the line is the span of the given vector. To derive the general equation of the line, expand the
above, giving the parametric equations x = 3t and y = 4t. Substitute 13 x for t in the second
equation, giving y = 43 x. Clearing fractions gives 3y = 4x, or 4x 3y = 0.
14. (a) Since the span of
15. (a) Since the given vectors are not parallel, their span is a plane through the origin containing these
two vectors as direction vectors.
(b) The vector equation of the plane is
2 3
2 3
2 3
x
1
3
4 y 5 = s 4 25 + t 4 25 .
z
0
1
One way of deriving the general equation of the plane is to solve the system of equations arising
from the above vector equation. Row-reducing the corresponding augmented system gives
2
3
2
3
2
3
1 3
x
1 3 x
1 3
x
R
2R
2 4
42 2 y 5 2 !
40
5
4
y 4x
0
4 y 2x5
R3 14 R2
1
0
1 z
0
1
z
0
0
z
+
(2x
y)
!
4
Since we know the system is consistent, the bottom row must consist entirely of zeros. Thus
the points on the plane must satisfy z + 14 (2x y) = 0, or 4z + 2x y = 0. This is the plane
2x y + 4z = 0.
An alternative method is to find a normal vector to the plane by taking the cross-product of the
two given vectors:
2 3 2 3 2
3 2 3
1
3
0 · 2 2 · ( 1)
2
425 ⇥ 4 25 = 41 · ( 1) 0 · 35 = 4 15 .
0
1
2·3 1·2
4
So the equation of the plane is 2x
2x y + 4z = 0.
16. (a) First, note that
y + 4z = d; since it passes through the origin, we again get
2
3
0
4 15 =
1
2
3
1
4 05
1
2
3
1
4 15 ,
0
so that the third vector is in the span of the first two. So we need only2consider
the
3
2
1
vectors. Their span is the plane through the origin with direction vectors 4 05 and 4
1
(b) The vector equation of this plane is
2 3
2 3 2 3
1
1
x
s 4 05 + t 4 15 = 4y 5 .
1
0
z
first
3 two
1
15 .
0
To find the general equation of the plane, we can solve the corresponding system of equations for
s and t. Row-reducing the augmented matrix gives
2
3
2
3
2
3
1
1 x
1
1
x
1
1
x
6
7
6
7
6
7
1 y5
y 5
y
40
40 1
40 1
5
R3 +R1
R3 +R2
1 0 z
0
1
x
+
z
0
0
x
+
y
+
z
!
!
84
CHAPTER 2. SYSTEMS OF LINEAR EQUATIONS
Since we know the system is consistent, the bottom row must consist of all zeros, so that points
on the plane satisfy the equation x + y + z = 0.
Alternatively, we could find a normal vector to the plane by taking the cross product of the
direction vectors:
2 3 2 3 2
3 2 3
1
1
0 · 0 ( 1) · 1
1
4 05 ⇥ 4 15 = 4 1 · 1 1 · 05 = 415 .
1
0
1 · 1 0 · ( 1)
1
Thus the equation of the plane is x + y + z = d; since it passes through the origin, we again get
x + y + z = 0.
17. Substituting the two points (1, 0, 3) and ( 1, 1, 3) into the equation for the plane gives a linear system
in a, b, and c:
a
+ 3c = 0
a+b
3c = 0.
Row-reducing the associated augmented matrix gives

1
1
0
1
3
3
0
0

1
! 0
R2 +R1
0
1
3
0
0
,
0
so that b = 0 and a = 3c. Thus the equation of the plane is 3cx + cz = 0. Clearly we must have
c 6= 0, since otherwise we do not get a plane. So dividing through by c gives 3x + z = 0, or 3x z = 0,
for the general equation of the plane.
18. Note that
u = 1u + 0v + 0w,
v = 0u + 1v + 0w,
w = 0u + 0v + 1w.
Or, if one adopts the usual convention of not writing terms with coefficients of zero, u is in the span
because u = u.
19. Note that
u = u,
v = (u + v)
u,
w = (u + v + w)
(u + v),
and the terms in parentheses on the right, together with u, are the elements of the given set.
20. (a) To prove the statement, we must show that anything that can be written as a linear combination
of the elements of S can also be written as a linear combination of the elements of T . But if x is
a linear combination of elements of S, then
x = a1 u 1 + a2 u 2 + · · · + ak u k
= a1 u1 + a2 u2 + · · · + ak uk + 0uk+1 + 0uk+1 + · · · + 0um .
This shows that x is also a linear combination of elements of T , proving the result.
(b) We know that span(S) ✓ span(T ) from part (a). Since T is a set of vectors in Rn , we know that
span(T ) ✓ Rn . But then
Rn = span(S) ✓ span(T ) ✓ Rn ,
so that span(T ) = Rn .
2.3. SPANNING SETS AND LINEAR INDEPENDENCE
85
21. (a) Suppose that
ui = ui1 v1 + ui2 v2 + · · · + uim vm ,
i = 1, 2, . . . , k.
Suppose that w is a linear combination of the ui . Then
w = w1 u1 + w2 u2 + · · · + wk uk
= w1 (u11 v1 + u12 v2 + · · · + u1m vm ) + w2 (u21 v1 + u22 v2 + · · · + u2m vm ) + · · ·
wk (uk1 v1 + uk2 v2 + · · · + ukm vm )
= (w1 u11 + w2 u21 + · · · + wk uk1 )v1 + (w1 u12 + w2 u22 + · · · + wk uk2 )v2 + · · ·
=
w10 v1
(w1 u1m + w2 u2m + · · · + wk ukm )vm
0
+ w20 v2 + · · · + wm
vm .
This shows that w is a linear combination of the vj . Thus span(u1 , . . . , uk ) ✓ span(v1 , . . . , vk ).
(b) Reversing the roles of the u’s and the v’s in part (a) gives span(v1 , . . . , vk ) ✓ span(u1 , . . . , uk ).
Combining that with part (a) shows that each of the spans is contained in the other, so they are
equal.
(c) Let
2 3
1
u1 = 405 ,
0
2 3
1
u 2 = 4 15 ,
0
2 3
1
u3 = 415 .
1
We want to show that each ui is a linear combination of e1 , e2 , and e3 , and the reverse. The
first of these is obvious: since e1 , e2 , and e3 span R3 , any vector is R3 is a linear combination of
those, so in particular u1 , u2 , and u3 are. Next, note that
e1 = u 1
e2 = u 2
u1
e3 = u 3
u2 ,
so that the ej are linear combinations of the ui . Part (b) then implies that span(u1 , u2 , u3 ) =
span(e1 , e2 , e3 ) = R3 .
2 3
2 3
2
1
22. The vectors 4 15 and 4 25 are not scalar multiples of one another, so are linearly independent.
3
3
23. There is no obvious dependence relation, so follow Example 2.23. We want scalars c1 , c2 , c3 so that
2 3
2 3
2 3 2 3
1
1
1
0
c1 415 + c2 425 + c3 4 15 = 405 .
1
3
2
0
Row-reduce the associated augmented matrix:
2
1
41
1
1
2
3
1
1
2
3
0
05
0
R3
2
1
! 40
R1
! 0
R2 R1
1
1
2
3
1 0
2 05
1 0
2
1 1
· · · 40 1
0 0
2
3
1 1
1 0
40 1
2 05 1
···
R3 2R2
3
5 0 5 R!
! 0 0
3 R1 R3 2
3
1 0
! 1 1 0 0 R1
3 40
2 05 R2 +2R!
1 0 05
1 0
0 0 1 0
2
! 1
40
0
R3
0
1
0
0
0
1
Thus the system has only the solution c1 = c2 = c3 = 0, so the vectors are linearly independent.
3
0
05
0
86
CHAPTER 2. SYSTEMS OF LINEAR EQUATIONS
24. There is no obvious dependence relation, so follow Example 2.23. We want scalars c1 , c2 , c3 so that
2 3
2 3
2 3 2 3
2
3
1
0
c1 425 + c2 415 + c3 4 55 = 405 .
1
2
2
0
Row-reduce the associated augmented matrix:
2
2 3
42 1
1 2
1
5
2
3 R3 2
0 ! 1
1 4
05 R!
2
2
0 R!
2
2
3
1
2
1
5
2
1
R2 4
! 0
0
3
0
05
0
2
1
3
R2
R3
2
1
2
2
2R1
! 40
1
3
2R1
3
9
! 0
3
2
2 0
1
40
3 05
2
9 0 R3 +3R!
0
3
0
05
0
2
1
0
2
3
0
3
0
05
0
2
! 1
40
0
R1 2R2
0
1
0
4
3
0
3
0
05
0
Thus the system has a nontrivial solution, and the vectors are linearly dependent. From the reduced
matrix, the general solution is c1 = 4t, c2 = 3t, c3 = t. Setting t = 1, for example, we get the relation
2 3
2 3 2 3 2 3
2
3
1
0
4 425 3 415 + 4 55 = 405 .
1
2
2
0
25. Since by inspection
v1
the vectors are linearly dependent.
2 3
0
v2 + v3 = 415
2
2 3 2 3
2
2
415 + 405 = 0,
3
1
26. Since we are given 4 vectors in R3 , we know they are linearly dependent vectors by Theorem 2.8. To
find a dependence relation, we want to find scalars c1 , c2 , c3 , c4 such that
2 3
2 3
2 3
2 3 2 3
2
4
3
5
0
c1 4 35 + c2 4 15 + c3 415 + c4 405 = 405 .
7
5
3
2
0
Row-reduce the associated augmented matrix:
2
3 1
2
3
2 R1
2
4 3 5 0
1
2
!
2
6
7
6
6 3
7
63
1
1
0
0
1
1
4
5
4
7
5 3 2 0
7
5
3
2
3
1
2
2
6
1
R
5 2 6
1 11
! 40
10
0 19 27
2
2
3
1
2
2
6
11
60
1
4
10
5
37 R3
0
1
! 0
2
R1 +2R2
! 1 0 0
6
60 1 0
4
0 0 1
3
2
0
1
7 R2 3R1 6
60
0 07
!
5
4
R3 7R1
2 0
! 0
3
2
5
0
1
2
7
6
3
7
6
05
40
2
R3 19R2
39
0
! 0
2
3 R1 + 3 R3 2
2
5
0
! 1
2
6
7
11
R2 10 R3 6
3
7
0
!
5
40
2
45
0
0
37
3
13
0
37
7
6
07
5
37
45
0
37
5
2
2
5
19
2
1
0
2
3
2
11
2
27
2
5
2
15
2
39
2
3
2
11
10
37
5
5
2
3
2
0
25
37
6
37
45
37
1 0
0 1
9
0
3
7
07
5
0
3
0
7
07
5
0
3
0
7
07
5
0
2.3. SPANNING SETS AND LINEAR INDEPENDENCE
87
Thus the system has a nontrivial solution, and the vectors are linearly dependent. From the reduced
6
45
matrix, the general solution is c1 = 13
37 t, c2 =
37 t, c3 =
37 t, c4 = t. Setting t = 37, for example, we
get the relation
2 3
2 3
2 3
2 3 2 3
2
4
3
5
0
13 4 35 6 4 15 45 415 + 37 405 = 405 .
7
5
3
2
0
2 3
0
27. Since v3 = 405 is the zero vector, these vectors are linearly dependent. For example,
0
2 3
2 3
2 3
3
6
0
0 · 445 + 0 · 475 + 137 405 = 0.
5
8
0
Any set of vectors containing the zero vector is linearly dependent.
28. There is no obvious dependence relation, so follow Example 2.23. We want scalars c1 , c2 , c3 so that
2
3
2 3
2
1
3
6 17
6 27
6
7
6 7
6
c1 6
4 25 + c2 425 + c3 4
1
4
3 2 3
2
0
607
37
7=6 7
15 405
1
0
Row-reduce the associated augmented matrix:
2
1
6 1
6
4 2
1
3
2
2
4
2
3
1
1
3
0
07
7
05
0
2
! 1
61
6
42
1
R1
3
2
2
4
2
1
60
R3 6R2 6
! 40
R4 7R2
! 0
2
1
R1 +2R3
!6
R2 R3 6 0
! 40
0
3
0
07
7
05
0
2
3
1
1
3
1
0
0
3
1
0
0
2
1
1
6
0
0
1
0
R3
R4
2
1
! 60
6
2R2 4
! 0
R1
! 0
R2 R1
3
2
0
1
60
07
7
6
3 4
05 R!
0
0
0
3 R1 +3R2 2
0
! 1
60
07
7
6
5
40
0
0
0
3
5
6
7
2
5
5
1
3
1
0
0
2
1
1
6
0
1
0
0
0
0
1
0
3
0
07
7
05
0
3
0
07
7
05
0
3
0
07
7
05
0
2
1
60
!6
40
0
1
5 R2
3
1
6
7
2
1
60
6
40
R4 +6R3
! 0
2
1
5
1
3
1
0
0
3
0
07
7
05
0
2
1
1
0
3
0
07
7
05
0
Since the system has only the trivial solution, the given vectors are linearly independent.
29. There is no obvious dependence relation, so follow Example 2.23. We want scalars c1 , c2 , c3 , c4 so that
2
3
2
1
6 17
6
7
6
c1 6
4 15 + c2 4
0
3
2
1
6
17
7 + c3 6
5
4
0
1
3
2
1
6
07
7 + c4 6
5
4
1
1
3 2 3
0
0
607
17
7=6 7
15 405
1
0
88
CHAPTER 2. SYSTEMS OF LINEAR EQUATIONS
Row-reduce the associated augmented matrix:
2
3
2
1
1
1
0 0
1
1
R
+R
2
1
6 1
1
0
1 07
!6
0
6
7
60
4 1
1 40
0
1
1 05 R3 R!
1
0
1
1
1 0
0
1
2
1
1
60
1
6
40
0
R4 R2
0
! 0
2
1
1 1
60
1 0
6
40
0 1
1
3 R4
0
0 0
!
2
R1 +R2 R3
! 1 0
60 1
6
40 0
0 0
1
1
0
1
1
0
1
1
0
1
1
1
0
0
1
0
0
1
1
1
3
0
07
7
05
0
3
0
07
7
05
0
2
1
1
6
1
R2 $R3 60
! 40
0
0
1
2
0
1
1
60
1
1
6
40
1
0
R4 +R3
2
0
! 0
3
2
0
1
1 1
R2 +R4 6
!
07
0
1 0
7
6
4 40
05 R3 R!
0 1
0
0
0 0
3
0 0
0 07
7
0 05
1 0
1
0
1
1
1
0
1
0
0
0
0
1
0
1
1
1
0
1
1
3
3
0
07
7
05
0
3
0
07
7
05
0
3
0
07
7
05
0
Since the system has only the trivial solution, the given vectors are linearly independent.
30. The vectors
2 3
0
607
7
v1 = 6
405 ,
1
2 3
0
607
7
v2 = 6
425 ,
1
2 3
0
637
7
v3 = 6
425 ,
1
2 3
4
6 37
7
v4 = 6
4 25
1
are linearly dependent. For suppose c1 v1 + c2 v2 + c3 v3 + c4 v4 = 0. Then c4 must be zero in order for
the first component of the sum to be zero. But then the only vector containing a nonzero entry in the
second component is v3 , so that c3 = 0 as well. Now for the same reason c2 = 0, looking at the third
component. Finally, that implies that c1 = 0. So the vectors are linearly independent.
31. By inspection
v1 + v2
2
3 2
3
6 17 6
7 6
v3 + v4 = 6
4 15 + 4
1
so the vectors are linearly dependent.
3
1
37
7
15
1
2 3 2
1
617 6
6 7+6
435 4
1
3
1
17
7 = 0,
15
3
32. Construct a matrix A with these vectors as its rows and try to produce a zero row:



0
2
2
1 3 R1 =R1 +R!
1 1 6
1 1 6
0
0
1
1
2 3
1 2 3 R2 =R2 +2R!
0 3 9
This matrix is in row echelon form and has no zero rows, so the matrix has rank 2 and the vectors are
linearly independent by Theorem 2.7.
33. Construct a matrix A with these vectors as its rows and try to produce a zero row:
2
1
41
1
3
1 1
2 35
1 2
R20 =R2
R30 =R3
2
1
R1
! 40
R1
0
!
3
1 1
1 25
2 1
2
1
40
R300 =R30 +2R20
! 0
1
1
0
3
1
25
5
This matrix is in row echelon form and has no zero rows, so the matrix has rank 3 and the vectors are
linearly independent by Theorem 2.7.
2.3. SPANNING SETS AND LINEAR INDEPENDENCE
89
34. Construct a matrix A with these vectors as its rows and try to produce a zero row:
2
3
2 1
1 25
5 2
2
43
1
2
3
5 2
1 25
2 1
1
R1 $R3 4
! 3
2
2
1
! 40
2R1
0
!
R20 =R2 3R1
R30 =R3
3
2
45
3
5
16
12
R300 =R30
2
1
40
3 0
4 R2
! 0
3
2
35
0
5
12
0
Since there is a zero row, the matrix has rank less than 3, so the vectors are linearly dependent by
Theorem 2.7. Working backwards, we have
[0, 0, 0] = R300 = R30
3 0
R = R3
4 2
3
(R2
4
2R1
1
R1
4
3R1 ) =
3
R 2 + R3
4
2
3
1
1
= 4 55
4
2
2 3 2 3
3
2
34 5 4 5
1 + 2 .
4
2
1
35. Construct a matrix A with these vectors as its rows and try to produce a zero row:
2
0
42
2
1
1
0
3
2
35
1
2
! 2
R20 =R1 40
!
2
R10 =R2
3
3
25
1
1
1
0
R30 =R3
2
2
40
R10
! 0
3
3
25
2
1
1
1
2
2
40
R300 =R30 +R20
! 0
3
3
15
0
1
1
0
This matrix is in row echelon form and has a zero row, so the vectors are linearly dependent by Theorem
2.7. Working backwards, we have
2 3 2 3 2 3
0
2
2
[0, 0, 0] = R300 = R30 + R20 = R3 R10 + R20 = R3 R2 + R1 = R1 R2 + R3 = 415 415 + 405 .
2
3
1
36. Construct a matrix A with these vectors and try to produce a zero row:
2
6
6
6
4
2
4
3
5
3
3 7
1 57
7
7
1 35
0 2
R20 =R2 +2R1
2
!6
6
6
!4
R30 =R3 + 32 R1
R40 =R4 + 52 R1
!
2
0
0
0
···
3
5
11
2
15
2
3
7
197
7
27 7
5
2
39
2
R3000 =R300
2
6
6
6
!4
R300 =2R30
R400 =2R40
2
6
6
6
!4
11 0
5 R2
R4000 =R400 3R20
!
!
2
0
0
0
2
0
0
0
3
3
7
5 197
7
7···
11 275
15 39
3
7
197
7
74 7
5
5
3
5
0
0
18
2
R40000 =R4000
5·18 000
74 R3
!
6
6
6
4
2
0
0
0
3
5
0
0
3
7
197
7
74 7
5
5
0
Since this matrix has a zero row, it has rank less than 2, so the vectors are linearly dependent. Working
backwards, we have
[0, 0, 0] = R40000 = R4000
= R400
90 000
R
74 3
3R20
✓
◆
45
11 0
R300
R2
37
5
90 0
99 0
R + R
37 3 37 2
◆
= 2R40 3R20
✓
5
= 2 R4 + R 1
2
26
12
=
R1
R2
37
37
3(R2 + 2R1 )
90
R3 + 2R4 .
37
90
37
✓
3
R3 + R1
2
◆
+
99
(R2 + 2R1 )
37
90
CHAPTER 2. SYSTEMS OF LINEAR EQUATIONS
Clearing fractions gives
[0, 0, 0] = 26R1
Thus
2
3
2
26 4 35
7
12R2
2
3
4
12 4 15
5
90R3 + 74R4 .
2 3
2 3
3
5
90 415 + 74 405 = 0.
3
2
Note that instead of doing all this work, we could just have applied Theorem 2.8.
37. Construct a matrix A with these vectors and try
2
3
46
0
to produce a zero row:
3
4 5
7 85
0 0
This matrix already has a zero row, so we know by Theorem 2.7 that the vectors are linearly dependent,
and we have an obvious dependence relation: let the coefficients of each of the first two vectors be zero,
and the coefficient of the zero vector be any nonzero number.
38. Construct a matrix A with these vectors and try to produce a zero row:
2
1
4 3
2
1
2
3
2
2
1
3
1
45
1
2
1
!4 0
R30 =R3 +2R1
0
!
R20 =R2 +3R1
1
5
5
2
8
5
3
1
75
1
2
1
4 0
R2
0
!
R300 =R3
1
5
0
2
8
3
3
1
75 .
6
This matrix is in row echelon form, and it has no zero row, so it has rank 4 and thus the vectors are
linearly independent by Theorem 2.7.
39. Construct a matrix A with these vectors and try to produce a zero row:
2
1
6 1
6
4 1
0
1
1
0
1
1
0
1
1
3
0
17
7
15
1
2
1
60
6
40
0
R30 =R3
1
0
1
0
2
1
!6
60
R1 4 0
!
0
R20 =R2 +R1
3
0
17
7
15
2
1
1
0
1
1
0
1
1
1
1
0
1
2
1
!6
60
R300 =R20 40
!
0
R200 =R30
3
0
17
7
15
1
R40 =R4 R30
1
1
0
0
!
3
0
17
7
15
2
1
0
1
1
2
1
60
6
40
00
0
0
R4 =R4 +R3
! 0
3
0
17
7.
15
3
1 1
1 0
0 1
0 0
This matrix is in row echelon form, and it has no zero row, so it has rank 4 and thus the vectors are
linearly independent by Theorem 2.7.
40. Construct a matrix A with these vectors and try to produce a zero row:
2
0
60
6
40
4
0
0
3
3
0
2
2
2
3
1
17
7
15
1
R30 =R3
R40 =R4
2
0
! 60
6
R1 4
! 0
4
R1
!
R20 =R2 R1
0
0
3
3
0
2
2
2
3
1
07
7
05
0
R300 =R30
R400 =R40
2
0
6
0 60
R2
! 40
R20
4
!
0
0
3
3
2
0
60
6
40
4
0
2
0
0
0
0
3
0
3
1
07
7
05
0
0
2
0
0
R4000 =R400 R300
!
3
1
07
7
05
0
R10 =R4000
!2
4
!6
60
R3000 =R20 40
!
0
R40000 =R1
!
R200 =R300
0
3
0
0
0
0
2
0
3
0
07
7.
05
1
2.3. SPANNING SETS AND LINEAR INDEPENDENCE
91
This matrix is in row echelon form, and it has no zero row, so it has rank 4 and thus the vectors are
linearly independent by Theorem 2.7.
41. Construct a matrix A with these vectors and try to produce a zero row:
2
3
6 1
6
4 1
1
1 1
3 1
1 3
1 1
3
1
17
7
15
3
R10 =R3
!
2
1
6 1
6
R30 =R1 4 3
!
1
3
1
17
7
15
3
1 3
3 1
1 1
1 1
R20 =R2 +R10
!
R300 =R30 3R10
!
R40 =R4 +R10
!
2
1
60
6
40
0
1
4
4
0
3
4
8
4
3
1
07
7
45
4
2
1
60
6
R3000 =R300 +R20 +R40 4
! 0
0
1
4
0
0
3
4
0
4
3
1
07
7.
05
4
Since this matrix has a zero row, it has rank less than 4, so the vectors are linearly dependent. Working
backwards, we have
[0, 0, 0, 0] = R3000 = R300 + R20 + R40
Thus
2
= R30
3R10 + R2 + R10 + R4 + R10
= R30
R10 + R2 + R4
= R1
R3 + R 2 + R 4 .
3 2
3
6 17 6
6 7+6
4 15 4
1
3
1
37
7
15
1
2 3 2
1
6 17 6
6 7+6
4 35 4
1
3
1
17
7 = 0.
15
3
42. (a) In this case, rank(A) = n. To see this, recall that v1 , v2 , . . . , vn are linearly independent if and
only if the only solution of [A | 0] = [v1 v2 . . . vn | 0] is the trivial solution. Since the trivial
solution is always a solution, it is the only one if and only if the number of free variables in the
associated system is zero. Then the Rank Theorem (Theorem 2.2 in Section 2.2) says that the
number of free variables is n rank(A), so that n = rank(A).
(b) Let
2
3
v1
6 v2 7
6 7
A = 6 . 7.
4 .. 5
vn
Then by Theorem 2.7, the vectors, which are the rows of A, are linearly independent if and only
if the rank of A is greater than or equal to n. But A has n rows, so its rank is at most n. Thus,
the rows of A are linearly independent if and only if the rank of A is exactly n.
43. (a) Yes, they are. Suppose that c1 (u + v) + c2 (v + w) + c3 (u + w) = 0. Expanding gives
c1 u + c1 v + c2 v + c2 w + c3 u + c3 w = (c1 + c3 )u + (c1 + c2 )v + (c2 + c3 )w = 0.
Since u, v, and w are linearly independent, all three of these coefficients are zero. So to see what
values of c1 , c2 , and c3 satisfy this equation, we must solve the linear system
c1
+ c3 = 0
c1 + c2
=0
c2 + c3 = 0.
92
CHAPTER 2. SYSTEMS OF LINEAR EQUATIONS
Row-reducing the coefficient matrix gives
2
3 2
1 0 1
1
41 1 05 ! 40
0 1 1
0
3
1
15 .
2
0
1
0
Since the rank of the reduced matrix is 3, the only solution is the trivial solution, so c1 = c2 =
c3 = 0. Thus the original vectors u + v, v + w, and u + w are linearly independent.
(b) No, they are not. Suppose that c1 (u
c1 u
c1 v + c2 v
c2 w + c3 u
v) + c2 (v
w) + c3 (u
w) = 0. Expanding gives
c3 w = (c1 + c3 )u + ( c1 + c2 )v + ( c2
c3 )w = 0.
Since u, v, and w are linearly independent, all three of these coefficients are zero. So to see what
values of c1 , c2 , and c3 satisfy this equation, we must solve the linear system
c1
+ c3 = 0
c1 + c2
=0
c2
c3 = 0.
Row-reducing the coefficient matrix gives
2
1
0
4 1
1
0
1
3 2
1
1
05 ! 40
1
0
0
1
0
3
1
15 .
0
So c3 is a free variable, and a solution to the original system is c1 = c2 =
for example, we have
(u v) + (v w) (u w) = 0.
c3 . Taking c3 =
1,
44. Let v1 and v2 be the two vectors. Suppose first that at least one of them, say v1 , is zero. Then they
are linearly dependent since v1 + 0v2 = 0 + 0 = 0; further, v1 = 0v2 , so one is a scalar multiple of the
other.
Now suppose that neither vector is the zero vector. We want to show that one is a multiple of the
other if and only if they are linearly dependent. If one is a multiple of the other, say v1 = kv2 , then
clearly 1v1 kv2 = 0, so that they are linearly dependent. To go the other way, suppose they are
linearly dependent. Then c1 v1 + c2 v2 = 0, where c1 and c2 are not both zero. Suppose that c1 6= 0 (if
instead c2 6= 0, reverse the roles of the first and second subscripts). Then
c1 v1 + c2 v 2 = 0
)
c1 v1 =
c2 v 2
)
v1 =
c2
v2 ,
c1
so that v1 is a multiple of v2 , as was to be shown.
45. Suppose we have m row vectors v1 , v2 , . . . , vm in Rn with m > n. Let
2 3
v1
6 v2 7
6 7
A = 6 . 7.
4 .. 5
vm
By Theorem 2.7, the rows (the vi ) are linearly dependent if and only if rank(A) < m. By definition,
rank(A) is the number of nonzero rows in its row echelon form. But since there can be at most one
leading 1 in each column, the number of nonzero rows is at most n < m. Thus rank(A)  n < m, so
that by Theorem 2.7, the vectors are linearly dependent.
2.4. APPLICATIONS
93
46. Suppose that A = {v1 , v2 , . . . , vn } is a set of linearly independent vectors, and that B is a subset of
A, say B = {vi1 , vi2 , . . . , vik } where 1  i1 < i1 < · · · < ik  n for j = 1, 2, . . . , k. We want to show
that B is a linearly independent set. Suppose that
ci1 vi1 + ci2 vi2 + · · · + cik vik = 0.
Set cr = 0 if r is not in {i1 , i2 , . . . , ik }, and consider
c1 v 1 + c2 v2 + · · · + cn v n .
This sum is zero since the sum of the i1 , i2 , . . . , ik elements is zero, and the other coefficients are zero.
But A forms a linearly independent set, so all the coefficients must in fact be zero. Thus the linear
combination of the vij above must be the trivial linear combination, so B is a linearly independent set.
47. Since S 0 = {v1 , . . . , vk } ⇢ S = {v1 , . . . , vk , v}, Exercise 20(a) proves that span(S 0 ) ✓ span(S). Using
Exercise 21(a), to show that span(S) ✓ span(S 0 ), we must show that each element of S is a linear
combination of elements of S 0 . But clearly each vi is a linear combination of elements of S 0 , since
vi 2 S 0 . Also, we are given that v is linear combination of the vi , so it too is a linear combination
of elements of S 0 . Thus span(S) ✓ span(S 0 ). Since we have proven both inclusions, the two sets are
equal.
48. Suppose bv + b2 v2 + · · · + bk vk = 0. Substitute for v from the given equation, to get
b(c1 v1 + c2 v2 + · · · + ck vk ) + b2 v2 + · · · + bk vk = bc1 v1 + (bc2 + b2 )v2 + · · · + (bck + bk )vk = 0.
Since the vi are a linearly independent set, all of the coefficients must be zero:
bc1 = 0,
bc2 + b2 = 0,
bc3 + b3 = 0,
...
bck + bk = 0.
In particular, bc1 = 0; since we are given that c1 6= 0, it follows that b = 0. Then substituting 0 for b
in the remaining equations above
b2 = 0,
b3 = 0,
...
bk = 0.
Thus b and all of the bk are zero, so the original linear combination was the trivial one. It follows that
{v, v2 , . . . , vk } is a linearly independent set.
2.4
Applications
1. Let x1 , x2 , and x3 be the number of bacteria of strains I, II, and
consumption of A, B, and C, we get the following system:
2
3 2
x1 + 2x2
= 400
1 2 0 400
1
2x1 + x2 + x3 = 600 ) 42 1 1 6005 ! 40
x1 + x2 + 2x3 = 600
1 1 2 600
0
III, respectively. Then from the
2. Let x1 , x2 , and x3 be the number of bacteria of strains I, II, and
consumption of A, B, and C, we get the following system:
2
3 2
x1 + 2x2
= 400
1 2 0 400
1
2x1 + x2 + 3x3 = 500 ) 42 1 3 5005 ! 40
1 1 1 600
0
x1 + x2 + x3 = 600
III, respectively. Then from the
0
1
0
3
160
1205 .
160
0
0
1
So 160 bacteria of strain I, 120 bacteria of strain II, and 160 bacteria of strain III can coexist.
This is an inconsistent system, so the bacteria cannot coexist.
0
1
0
2
1
0
3
0
05 .
1
94
CHAPTER 2. SYSTEMS OF LINEAR EQUATIONS
3. Let x1 , x2 , and x3 be the number of small, medium, and large arrangements. Then from the consumption of flowers in each arrangement we get:
2
3 2
3
x1 + 2x2 + 4x3 = 24
1 2 4 24
1 0 0 2
3x1 + 4x2 + 8x3 = 50 ) 43 4 8 505 ! 40 1 0 35 .
3x1 + 6x2 + 6x3 = 48
3 6 6 48
0 0 1 4
So 2 small, 3 medium, and 4 large arrangements were sold that day.
4. (a) Let x1 , x2 , and x3 be the number of nickels, dimes, and quarters. Then we get
2
3 2
3
x1 + x2 + x3 = 20
1
1 1
20
1 0 0 4
2x1
x2
= 0 ) 42
1
0
05 ! 40 1 0 85 .
5x1 + 10x2 + 25x3 = 300
5 10 25 300
0 0 1 8
So there are 4 nickels, 8 dimes, and 8 quarters.
(b) This is similar to part (a), except that the constraint 2x1 x2 no longer applies, so we get


1
1
1
20
1 0
3
20
!
.
5 10 25 300
0 1
4
40
The general solution is x1 = 20 + 3t, x2 = 40 4t, and x3 = t. From the first equation, we must
have 3t 20 0, so that t 7. From the second, we have t  10. So the possible values for t are
7, 8, 9, and 10, which give the following combinations of nickels, dimes, and quarters:
[x1 , x2 , x3 ] = [1, 12, 7] or [4, 8, 8] or [7, 4, 9] or [10, 0, 10].
5. Let x1 , x2 , and x3 be the number of house, special, and gourmet blends. Then from the consumption
of beans in each blend we get
300x1 + 200x2 + 100x3 = 30,000
200x2 + 200x3 = 15,000
200x1 + 100x2 + 200x3 = 25,000 .
Forming the augmented matrix and reducing
2
300 200 100
4 0
200 200
200 100 200
it gives
3 2
30000
1
150005 ! 40
25000
0
0
1
0
0
0
1
3
65
305 .
45
The merchant should make 65 house blend, 30 special blend, and 45 gourmet blend.
6. Let x1 , x2 , and x3 be the number of house, special, and gourmet blends. Then from the consumption
of beans in each blend we get
300x1 + 200x2 + 100x3 = 30,000
50x1 200x2 + 350x3 = 15,000
150x1 + 100x2 + 50x3 = 15,000 .
Forming the augmented matrix and reducing
2
300 200 100
4 50 200 350
150 100 50
it gives
3 2
30000
1
150005 ! 40
15000
0
0
1
0
1
2
0
3
60
605 .
0
Thus x1 = 60 + t, x2 = 60 2t, and x3 = t. From the first equation, we get t
60; from the second,
we get t  30, and the third gives us t 0. So the range of possible values is 0  t  30. We wish to
maximize the profit
P =
1
3
1
3
x1 + x2 + 2x3 = (60 + t) + (60
2
2
2
2
t) + 2t = 120
1
t.
2
Since 0  t  30, the profit is maximized if t = 0, in which case x1 = x + 2 = 60 and x3 = 0. Therefore
the merchant should make 60 house and special blends, and no gourmet blends.
2.4. APPLICATIONS
95
7. Let x, y, z, and w be the number of FeS2 , O2 , Fe2 O3 , and SO2 molecules respectively. Then compare
the number of iron, sulfur, and oxygen atoms in reactants and products:
so we get the matrix
2
1
6
42
0
0
0
2
Iron:
x =
2z
Sulfur:
2x =
w
Oxygen:
2y = 3z + 2w,
2
0
3
0
1
2
3 2
0
1
7 6
05 ! 40
0
0
0
1
0
1
2
11
8
1
4
0
0
1
3
0
7
05 .
0
1
Thus z = 14 w, y = 11
8 w, and x = 2 w. The smallest positive value of w that will produce integer values
for all four variables is 8, so letting w = 8 gives x = 4, y = 11, z = 2, and w = 8. The balanced
equation is 4FeS2 + 11O2 ! 2Fe2 O3 + 8SO2 .
8. Let x, y, z, and w be the number of CO2 , H2 O, C6 H12 O6 , and O2 molecules respectively. The compare
the number of carbon, oxygen, and hydrogen atoms in reactants and products:
Carbon:
x
=
6z
Oxygen:
2x + y = 6z + 2w
Hydrogen:
2y
=
12z.
This system is easy to solve using back-substitution: the first equation gives x = 6z and the third gives
y = 6z. Substituting 6z for both x and y in the second equation gives 2 · 6z + 6z = 6z + 2w, so that
w = 6z as well. The smallest positive value of z giving integers is z = 1, so that x = y = w = 6 and
z = 1. The balanced equation is 6CO2 + 6H2 O ! C6 H12 O6 + 6O2 .
9. Let x, y, z, and w be the number of C4 H10 , O2 , CO2 , and H2 O molecules respectively. The compare
the number of carbon, oxygen, and hydrogen atoms in reactants and products:
Carbon:
4x =
z
Oxygen:
2y = 2z + w
Hydrogen:
10x =
2w.
This system is easy to solve using back-substitution: the first equation gives z = 4x and the third gives
w = 5x. Substituting these values for z and w in the second equation gives 2y = 2 · 4x + 5x = 13x, so
that y = 13
2 x. The smallest positive value of x giving integers for all four variables is x = 2, resulting
in x = 2, y = 13, z = 8, and w = 10. The balanced equation is 2C4 H10 + 13O2 ! 8CO2 + 10H2 O.
10. Let x, y, z, and w be the number of C7 H6 O2 , O2 , H2 O, and CO2 molecules respectively. The compare
the number of carbon, oxygen, and hydrogen atoms in reactants and products:
Carbon:
7x
=
Oxygen:
2x + 2y = z + 2w
Hydrogen:
6x
=
w
2z.
This system is easy to solve using back-substitution: the first equation gives w = 7x and the third
gives z = 3x. Substituting these values for w and z in the second equation gives 2x + 2y = 3x + 2 · 7x,
so that y = 15
2 x. The smallest positive value for x giving integer values for all four variables is
x = 2, so that the solution is x = 2, y = 15, z = 6, and w = 14. The balanced equation is
2C7 H6 O2 + 15O2 ! 6H2 O+14CO2 .
96
CHAPTER 2. SYSTEMS OF LINEAR EQUATIONS
11. Let x, y, z, and w be the number of C5 H11 OH, O2 , H2 O, and CO2 molecules respectively. The compare
the number of carbon, oxygen, and hydrogen atoms in reactants and products:
Carbon:
5x
Oxygen:
x + 2y = z + 2w
=
w
Hydrogen:
12x
=
2z.
This system is easy to solve using back-substitution: the first equation gives w = 5x and the third
gives z = 6x. Substituting these values for w and z in the second equation gives x + 2y = 6x + 2 · 5x,
so that y = 15
2 x. The smallest positive value for x giving integer values for all four variables is
x = 2, so that the solution is x = 2, y = 15, z = 12, and w = 10. The balanced equation is
2C5 H11 OH+15O2 ! 12H2 O+10CO2 .
12. Let x, y, z, and w be the number of HClO4 , P4 O10 , H3 PO4 , and Cl2 O7 molecules respectively. The
compare the number of hydrogen, chlorine, oxygen, and phosphorus atoms in reactants and products:
This gives the matrix
2
1
61
6
6
44
0
Hydrogen:
x
=
3z
Chlorine:
x
=
2w
Oxygen:
4x + 10y = 4z + 7w
Phosphorus:
4y
0
0
10
4
3
0
4
1
0
2
7
0
=
3 2
0
1
7
6
07 60
7!6
05 40
0
0
1
0
0
0
z.
0
0
1
0
2
1
6
2
3
0
3
0
07
7
7.
05
0
Thus the general solution is x = 2t, y = 16 t, z = 23 t, w = t. The smallest positive value of t giving all
integer answers is t = 6, so the desired solution is x = 12, y = 1, z = 4, and w = 6. The balanced
equation is 12HClO4 +P4 O10 ! 4H3 PO4 + 6Cl2 O7 .
13. Let x1 , x2 , x3 , x4 , and x5 be the number of Na2 CO3 , C, N2 , NaCN, and CO molecules respectively.
The compare the number of sodium, carbon, oxygen, and nitrogen atoms in reactants and products:
Sodium:
2x1
=
x4
Carbon:
x1 + x2 = x4 + x5
Oxygen:
3x1
=
x5
Nitrogen:
2x3
=
x4 .
This gives the matrix
2
2
61
6
6
43
0
0
1
0
0
0
0
0
2
1
1
0
1
0
1
1
0
3 2
0
1
7
6
07 6 0
7!6
05 4 0
0
0
4
x5 ,
3
x3 =
0
1
0
0
0
0
1
0
3
0
07
7
7
05
0
0
0
1
1
3
4
3
1
3
2
3
x4 =
2
x5 .
3
0
Thus x5 is a free variable, and
x1 =
1
x5 ,
3
x2 =
1
x5 ,
3
The smallest value of x5 that produces integer solutions is 3, so the desired solution is x1 = 1, x2 = 4,
x3 = 1, x4 = 2, and x5 = 3. The balanced equation is Na2 CO3 + 4C+N2 ! 2NaCN+3CO.
2.4. APPLICATIONS
97
14. Let x1 , x2 , x3 , x4 , and x5 be the number of C2 H2 Cl4 , Ca(OH)2 , C2 HCl3 , CaCl2 , and H2 O molecules
respectively. The compare the number of carbon, hydrogen, chlorine, calcium, and oxygen atoms in
reactants and products:
Carbon:
2x1
Hydrogen:
2x1 + 2x2 = x3 + 2x5
=
2x3
Chlorine:
4x1
= 3x3 + 2x4
Calcium:
x2
=
x4
Oxygen:
2x2
=
x5 .
This gives the matrix
2
2
6
62
6
64
6
6
40
0
0
2
0
1
2
2
1
3
0
0
0
0
2
1
0
3 2
0
1
7 6
07 60
7 6
6
07
7 ! 60
7 6
05 40
0
0
0
2
0
0
1
Thus x5 is a free variable, and
0
1
0
0
0
0
0
1
0
0
0
0
0
1
0
3
0
7
07
7
07
7
7
05
0
1
1
2
1
1
2
0
1
1
x 5 , x 3 = x 5 , x4 = x 5 .
2
2
The smallest value of x5 that produces integer solutions is 2, so the desired solution is x1 = 2, x2 = 1,
x3 = 2, x4 = 1, and x5 = 2. The balanced equation is 2C2 H2 Cl4 +Ca(OH)2 ! 2C2 HCl3 +CaCl2 +
2H2 O.
x1 = x5 ,
x2 =
15. (a) By applying the conservation of flow rule to each node we obtain the system of equations
f1 + f2 = 20
f2
f3 =
10
Form the augmented matrix and row-reduce it:
2
3 2
1 1
0
20
1
40 1
1
105 ! 40
1 0
1
30
0
f1 + f3 = 30.
0
1
0
3
30
105
0
1
1
0
Letting f3 = t, the possible flows are f1 = 30 t, f2 = t 10, f3 = t.
(b) The flow through AB is f2 , so in this case 5 = t 10 and t = 15. So the other flows are
f1 = f3 = 15.
(c) Since each flow must be nonnegative, the first equation gives t  30, the second gives t 10, and
the third gives t 0. So we have 10  t  30, which means that
0  f1  20
0  f2  20
10  f3  30.
(d) If a negative flow were allowed, it would mean a flow in the opposite direction, so that a negative
flow on f2 would mean a flow from B into A.
16. (a) By applying the conservation of flow rule to each node we obtain the system of equations
f1 + f2 = 20
Form the augmented matrix and
2
1 1
61 0
6
40 1
0 0
f1 + f3 = 25
f2 + f4 = 25
row-reduce it:
3 2
0 0 20
1
60
1 0 257
7!6
0 1 255 40
1 1 30
0
Letting f4 = t, the possible flows are f1 = t
0
1
0
0
5, f2 = 25
0
0
1
0
1
1
1
0
f3 + f4 = 30.
3
5
07
7
05
0
t, f3 = 30
t, f4 = t.
98
CHAPTER 2. SYSTEMS OF LINEAR EQUATIONS
(b) If f4 = t = 10, then the average flows on the other streets will be f1 = 10 5 = 5, f2 = 25 10 = 15,
and f3 = 30 10 = 20.
5, t  25, t  30, and t
(c) Each flow must be nonnegative, so we get the constraints t
5  t  25. This gives the constraints
0  f1  20
0  f2  20
5  f3  25
0. Thus
5  f4  25.
(d) Reversing all of the directions would have no e↵ect on the solution. This would mean multiplying
all rows of the augmented matrix by 1. However, multiplying a row by 1 is an elementary row
operation, so the new matrix would have the same reduced row echelon form and thus the same
solution.
17. (a) By applying the conservation of flow rule to each node we obtain the system of equations
f1 + f2 = 100
Rearrange the equations,
2
1 1
60 1
6
40 0
1 0
f2 + f3 = f4 + 150
f4 + f5 = 150
set up the augmented matrix, and
3 2
0
0
0
100
1 0
60 1
1
1
0
1507
7!6
0
1
1
1505 40 0
1
0
1
200
0 0
f1 + 200 = f3 + f5 .
row reduce it:
1 0
1 0
0 1
0 0
3
200
3007
7
1505
1
1
1
0
Then both f3 = s and f5 = t are free variables, so the flows are
f1 =
200 + s + t,
f2 = 300
s
t,
f3 = s,
f4 = 150
t,
(b) If DC is closed, then f5 = t = 0, so that f1 = 200 + s and f2 = 300
flow through DB, must satisfy 200  s  300 liters per day.
(c) If DB = f3 = S were closed, then f1 = 200 + t, so that t = f5
that f4 = 150 t  50 < 0, which is impossible.
f5 = t.
s. Thus s, which is the
200. But that would mean
(d) Since f4 = 150 t, we must have 0  t  150, so that 0  f4  150. Note that all of these values
for t are consistent with the first three equations as well: if t = 0, then s = 250 produces positive
values for all flows, while if t = 150, then s = 75 produces all positive values. So the flow through
DB is between 0 and 150 liters per day.
18. (a) By applying the conservation of flow rule to each node we obtain:
f3 + 200 = f1 + 100
f1 + 150 = f2 + f4
f2 + f5 = 300
f6 + 100 = f3 + 200
f4 + f7 = f6 + 100
f5 + f7 = 250.
Create and row-reduce the augmented matrix for the system:
2
3 2
1
0
1
0 0
0 0
100
1 0 0
61
7 60 1 0
1
0
1
0
0
0
150
6
7 6
60
6
1
0
0 1
0 0
3007
6
7 ! 60 0 1
60
7
0
1
0 0
1 0
1007 6
6
60 0 0
40
0
0
1 0
1 1
1005 40 0 0
0
0
0
0 1
0 1
250
0 0 0
0
0
0
1
0
0
0
0
0
0
1
0
1
0
1
1
0
0
0
1
0
1
1
0
Let f7 = t and f6 = s, so the flows are
f1 = s
f2 = 50 + t
f3 =
100 + s
f4 = 100 + s
t
f5 = 250
t
3
0
507
7
1007
7
1007
7
2505
0
f6 = s
f7 = t.
(b) From the solution in part (a), since f1 = s = f6 , it is not possible for f1 and f6 to have di↵erent
values.
2.4. APPLICATIONS
99
(c) If f4 = 0, then s
f1 = s
t = 100, so that t = 100 + s. Substituting this for t everywhere gives
f2 = 150 + s
f3 =
100 + s
f4 = 0 f5 = 150
s
f6 = s
f7 = 100 + s.
The constraints imposed by the requirement that all the fi be nonnegative means that 100  s 
150, so that
100  f1  150
50  f5  100
250  f2  300
100  f6  150
0  f3  50 f4 = 0
200  f7  250.
19. Applying the current law to node A gives I1 + I3 = I2 or I1 I2 + I3 = 0. Applying the voltage
law to the top circuit, ABCA, gives I2 I1 + 8 = 0. Similarly, for the circuit ABDA we obtain
I2 + 13 4I3 = 0. This gives the system
2
I1 I2 + I3 = 0
1
I1 + I2
= 8 ) 41
I2 + 4I3 = 13
0
3 2
1
0
85 ! 4 0
13
0
1 1
1 0
1 4
0
1
0
3
3
55
2
0
0
1
Thus I1 = 3 amps, I2 = 5 amps, and I3 = 2 amps.
20. Applying the current law to node B and the voltage law to the upper and lower circuits gives
2
I1
I2 + I3 = 0
1
I1 + 2I2
= 5 ) 41
2I2 + 4I3 = 8
0
3 2
0
1
55 ! 40
8
0
1 1
2 0
2 4
0
1
0
3
1
25
1
0
0
1
Thus I1 = 1 amps, I2 = 2 amps, and I3 = 1 amp.
21. (a) Applying the current and voltage laws to the circuit gives the system:
Node B:
I = I1 + I4
Node C:
I1 = I2 + I3
Node D:
I2 + I5 = I
Node E:
I3 + I4 = I5
Circuit ABEDA:
2I4
I5 + 14 = 0
Circuit BCEB:
I1
I3 + 2I4 = 0
Circuit CDEC:
2I2 + I3 + I5 = 0
Construct the augmented matrix and row-reduce it:
2
1
60
6
61
6
60
6
60
6
40
0
1
1
0
0
0
1
0
0
1
1
0
0
0
2
0
1
0
1
0
1
1
1
0
0
1
2
2
0
0
0
1
1
1
0
1
3 2
1
0
6
07
7 60
6
07
7 60
7
07 ! 6
60
6
147
7 60
05 4 0
0
0
0
1
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
1
0
3
10
67
7
47
7
27
7
47
7
65
0
Thus the currents are I = 10, I1 = 6, I2 = 4, I3 = 2, I4 = 4, and I5 = 6 amps.
(b) From Ohm’s Law, the e↵ective resistance is Re↵ =
V
I
=
14
10
=
7
5
ohms.
(c) To force the current in CE to be zero, we force I3 = 0 and let r be the resistance in BC. This
100
CHAPTER 2. SYSTEMS OF LINEAR EQUATIONS
gives the following system of equations:
Node B:
I = I1 + I4
Node C:
I1 = I2
Node D:
I2 + I5 = I
Node E:
I4 = I5
Circuit ABEDA:
2I4
I5 + 14 = 0
Circuit BCEB:
rI1
+ 2I4 = 0
Circuit CDEC:
2I2
+ I5 = 0
It is easier to solve this system by substitution. Substitute I4 = I5 in the circuit equation for
7
ABEDA to get 3I5 + 14 = 0, so that I4 = I5 = 14
2I2 + 14
3 . Thus
3 = 0 so that I1 = I2 = 3 .
Then the circuit equation for BCED gives us
r·
7
14
+2·
=0
3
3
) r = 4.
So if we use a 4 ohm resistor in BC, the current in CE will be zero.
22. (a) Applying the voltage law to Figure 2.23(a) gives IR1 +IR2 = E. But from Ohm’s Law, E = IRe↵ ,
so that I(R1 + R2 ) = IRe↵ and thus Re↵ = R1 + R2 .
(b) Applying the current and voltage laws to Figure 2.23(b) gives the system of equations I = I1 + I2 ,
I1 R1 + I2 R2 = 0, I1 R1 + E = 0, and I2 R2 + E = 0. Gauss-Jordan elimination gives
2
3 2
3
2 )E
1 0 0 (R1R+R
1
1
1
0
1 R2
60
6
7
E
R1
R2
07
6
7 60 1 0
7
R1
6
7!6
7.
E
40
5
R1
0
E 5 40 0 1
R
2
0
2 )E
Thus I = (R1R+R
, I1 =
1 R2
equation for I gives
I=
0
E
R1 ,
R2
I2 =
E
R2 .
(R1 + R2 )IRe↵
R1 R2
E
0
0
0
0
But E = IRe↵ ; substituting this value of E into the
)
Re↵ =
1
R1 +R2
R1 R2
=
1
R1
1
+
1
R2
.
23. The matrix for this system, with inputs on the left, outputs along the top, is
Farming
Manufacturing
Farming
Manufacturing
1
2
1
2
1
3
2
3
Let the output of the farm sector be f and the output of the manufacturing sector be m. Then
1
1
f+ m= f
2
3
1
2
f + m = m.
2
3
Simplifying and row-reducing the corresponding augmented matrix gives
"
# "
1
1
1
1
2
f
+
m
=
0
0
1
2
3
2
3
3
)
!
1
1
1
1
0
0
0
2f
3m = 0
2
3
Thus f = 23 m, so that farming and manufacturing must be in a 2 : 3 ratio.
#
0
0
2.4. APPLICATIONS
101
24. The matrix for this system, with needed resources on the left, outputs along the top, is
Coal
Steel
Coal Steel
0.3
0.8
0.7
0.2
Let the coal output be c and the steel output be s. Then
0.3c + 0.8s = c
0.7c + 0.2s = s.
Simplifying and row-reducing the corresponding augmented matrix gives


0.7c + 0.8s = 0
0.7
0.8 0
1
!
)
0.7c
0.8s = 0
0.7
0.8 0
0
8
7
0
0
0
Thus c = 87 s, so that coal and steel must be in an 8 : 7 ratio. For a total output of $20 million, coal
8
7
28
output must be 8+7
· 20 = 32
3 ⇡ $10.667 million, while steel output is 8+7 · 20 = 3 ⇡ $9.333 million.
25. Let x be the painter’s rate, y the plumber’s rate, and z the electrician’s rate. From the given schedule,
we get the linear system
2x + y + 5z = 10x
4x + 5y + z = 10y
4x + 4y + 4z = 10z.
Simplifying and row-reducing the corresponding augmented matrix gives
2
3 2
8x + y + 5z = 0
8
1
5 0
1 0
6
7 6
4x 5y + z = 0 ) 4 4
5
1 05 ! 4 0 1
4x + 4y 6z = 0
4
4
6 0
0 0
3
0
7
05 .
0
13
18
7
9
0
7
Thus x = 13
18 z and y = 9 z. Let z = 54; this is a whole number in the required range such that the
others will be whole numbers as well. Then x = 39 and y = 42. The painter should charge $39 per
hour, the plumber $42 per hour, and the electrician $54 per hour.
26. From the given matrix, we get the linear system
1
L+
4
1
1
B+ L+
2
4
1
1
B+ L+
4
4
1
1
B+ L+
4
4
1
T
8
1
T
4
1
T
2
1
T
8
1
+ Z
6
1
+ Z
6
1
+ Z
3
1
+ Z
3
Simplify and row-reduce the associated augmented matrix:
2
1
1
B + 14 L + 18 T + 16 Z = 0
1
4
8
6
1
3
1
1
1
3
1
6 2
2B
4L + 4T + 6Z = 0
4
4
) 6 1
1
1
1
1
1
1
4
B
+
L
T
+
Z
=
0
4
2
2
3
4
2
2
1
4B
+ 18 L + 18 T
2
3Z
=0
1
4
1
8
1
8
= B
= L
= T
= Z.
1
6
1
6
1
3
2
3
3 2
0
1
7
6
07 6 0
7!6
05 4 0
0
0
0
1
0
0
0
0
1
0
2
3
6
5
8
5
0
3
0
07
7
7.
05
0
Thus B = 23 Z, L = 65 Z, and T = 85 Z. Furthermore, B is the smallest price of the four, so B = 23 Z
must be at least 50, in which case Z = 75. With that value of Z, we get B = 50, L = 90, and T = 120.
102
CHAPTER 2. SYSTEMS OF LINEAR EQUATIONS
27. The matrix for this system, with needed resources on the left, outputs along the top, is
Coal
Steel
Coal Steel
0.15
0.2
0.1
0.1
(a) Let the coal output be c and the steel output be s. Then since 45 million of external demand is
required for coal and 124 million for steel, we have
c = 0.15c + 0.25s + 45
s = 0.2c + 0.1s + 124.
Simplify and row-reduce the associated matrix:

0.85c
0.25s = 45
0.85
)
0.2c + 0.9s + 124
0.2
0.25
0.9

45
1
!
124
0
0
1
100
160
So the coal industry should produce $100 million and the steel industry should produce $160
million.
(b) The new external demands for coal and steel are 40 and 130 million respectively. The only thing
that changes in the equations is the constant terms, so we have


0.85c
0.25s = 40
0.85
0.25
40
1 0
95.8
!
)
0.2c + 0.9s + 130
0.2
0.9 130
0 1 165.73
So the coal industry should reduce production by 100 95.8 = $4.2 million, while the steel industry
should increase production by 165.73 160 = $5.73 million.
28. This is an open system with external demand given by the other city departments. From the given
matrix, we get the system
1 + 0.2A + 0.1H + 0.2T = A
1.2 + 0.1A + 0.1H + 0.2T = H
0.8 + 0.2A + 0.4H + 0.3T = T .
Simplifying and row-reducing the associated augmented matrix gives
2
3 2
0.8A
0.1H
0.2T = 1
0.8
0.1
0.2
1
1
0.1A + 0.9H
0.2T = 1.2 ) 4 0.1
0.9
0.2 1.25 ! 40
0.2A
0.4H + 0.7T = 0.8
0.2
0.4
0.7 0.8
0
0
1
0
0
0
1
3
2.31
2.285 .
3.11
So the Administrative department should produce $2.31 million of services, the Health department
$2.28 million, and the Transportation department $3.11 million.
29. (a) Over Z2 , we need to solve x1 a + x2 b + · · · + x5 e = t + s:
2 3
2 3
2
0
0
1 1 0 0 0
607
6 17
61 1 1 0 0
6 7
6 7
6
7
6 7
6
s=6
607 , t = 607 , ) 60 1 1 1 0
405
4 15
40 0 1 1 1
0
0
0 0 0 1 1
3 2
0
1
60
17
7 6
6
07
7 ! 60
5
40
1
0
0
0
1
0
0
0
0
0
1
0
0
0
0
0
1
0
1
1
0
1
0
3
1
17
7
17
7
05
0
Hence x5 is a free variable, so there are exactly two solutions (with x5 = 0 and with x5 = 1).
Solving for the other variables, we get x1 = 1 + x5 , x2 = 1 + x5 , x3 = 1, and x4 = x5 . So the two
solutions are
2 3 2 3
2 3 2 3
x1
1
x1
0
6 x 2 7 6 17
6x2 7 607
6 7 6 7
6 7 6 7
6 x 3 7 = 6 17 ,
6x3 7 = 617 .
6 7 6 7
6 7 6 7
4 x 4 5 4 05
4x4 5 415
x5
0
x5
1
So we can push switches 1, 2, and 3 or switches 3, 4, and 5.
2.4. APPLICATIONS
(b) In this case, t = e2 over Z2 ,
2
1
61
6
60
6
40
0
103
so the matrix becomes
3 2
1 0 0 0 0
1
60
1 1 0 0 17
7 6
6
1 1 1 0 07
7 ! 60
0 1 1 1 05 4 0
0 0 1 1 0
0
0
1
0
0
0
0
0
1
0
0
0
0
0
1
0
1
1
0
1
0
3
0
07
7
07
7.
05
1
This is an inconsistent system, so there is no solution. Thus it is impossible to start with all the
lights o↵ and turn on only the second light.
30. (a) With s = e4 and t = e2 + e4
2
1
61
6
60
6
40
0
over Z2 , so that s + t = e2
3 2
1 0 0 0 0
1 0
60 1
1 1 0 0 17
7 6
6
1 1 1 0 07
7 ! 60 0
5
40 0
0 1 1 1 0
0 0 1 1 0
0 0
and we get the matrix
3
0 0 1 0
0 0 1 07
7
1 0 0 07
7.
0 1 1 05
0 0 0 1
Since the system is inconsistent, there is no solution in this case: it is impossible to start with
only the fourth light on and end up with only the second and fourth lights on.
(b) In this case t = e2 over Z2 , so that s + t = e2 + e4 , and we get
2
3 2
3
1 1 0 0 0 0
1 0 0 0 1 1
6 1 1 1 0 0 17 6 0 1 0 0 1 17
6
7 6
7
6 0 1 1 1 0 07 ! 6 0 0 1 0 0 17 .
6
7 6
7
4 0 0 1 1 1 15 4 0 0 0 1 1 05
0 0 0 1 1 0
0 0 0 0 0 0
So there are two solutions:
2
3 2 3
2 3 2 3
x1
1
x1
0
6 x 2 7 6 17
6x2 7 607
6 7 6 7
6 7 6 7
6 x 3 7 = 6 17 ,
6x3 7 = 617 .
6 7 6 7
6 7 6 7
4 x 4 5 4 05
4x4 5 415
x5
0
x5
1
So we can push switches 1, 2, and 3 or switches 3, 4, and 5.
31. Recall from Example 2.35 that it does not matter what order we push the switches in, and that pushing
a switch twice is the same as not pushing it at all. So with
2 3
2 3
2 3
2 3
2 3
1
1
0
0
0
6 17
6 17
617
607
607
6 7
6 7
6 7
6 7
6 7
7
6 7
6 7
6 7
6 7
a=6
607 , b = 617 , c = 617 , d = 617 , e = 607 ,
4 05
4 05
415
415
415
0
0
0
1
1
there are 25 = 32 possible choices, since there are 5 buttons and we have the choice of pushing or not
pushing each one. If we make a list of the resulting 32 resulting light states and eliminate duplicates,
we see that the list of possible states is
2 3 2 3 2 3 2 3 2 3 2 3 2 3 2 3
0
0
0
0
0
0
0
0
607 607 607 607 617 617 617 617
6 7 6 7 6 7 6 7 6 7 6 7 6 7 6 7
607 , 607 , 617 , 617 , 617 , 617 , 607 , 607 ,
6 7 6 7 6 7 6 7 6 7 6 7 6 7 6 7
405 415 415 405 415 405 405 415
0
1
1
0
0
1
1
0
2 3 2 3 2 3 2 3 2 3 2 3 2 3 2 3
1
1
1
1
1
1
1
1
617 617 617 617 607 607 607 607
6 7 6 7 6 7 6 7 6 7 6 7 6 7 6 7
617 , 617 , 607 , 607 , 607 , 607 , 617 , 617 .
6 7 6 7 6 7 6 7 6 7 6 7 6 7 6 7
405 415 415 405 415 405 405 415
0
1
1
0
0
1
1
0
104
CHAPTER 2. SYSTEMS OF LINEAR EQUATIONS
32. (a) Over Z3 , we need to solve x1 a + x2 b + x3 c = t, where
2 3
2
3 2
0
1 1 0 0
1
41 1 1 25 ! 40
t = 425 , so that
1
0 1 1 1
0
0
1
0
0
0
1
3
1
25 .
2
Thus x1 = 1, x2 = 2, and x3 = 2, so we must push switch A once and switches B and C twice
each.
(b) Over Z3 , we need to solve x1 a + x2 b + x3 c = t, where
2 3
2
3 2
1
1 1 0 1
1
41 1 1 05 ! 40
t = 405 , so that
1
0 1 1 1
0
0
1
0
0
0
1
In other words, we must push each switch twice.
(c) Let x, y, and z be the final states
2
1 1
41 1
0 1
3
2
25 .
2
of the three lights. Then we want to solve
3 2
3
0 x
1 0 0
y + 2z
1 y 5 ! 40 1 0 x + 2y + z 5 .
1 z
0 0 1
2x + y
So we can achieve this configuration if we push switch A y + 2z times, switch B x + 2y + z times,
and switch C 2x + y times (adding in Z3 ).
33. The switches here correspond to the same vectors as in Example 2.35, but now the numbers are
interpreted in Z3 rather than in Z2 . So in Z53 , with s = 0 (all lights initially o↵), we need to solve
x1 a + x2 b + · · · + x5 e = t, where
2 3
2
3 2
3
2
1 1 0 0 0 2
1 0 0 0 1 1
617
61 1 1 0 0 17 60 1 0 0 2 17
6 7
6
7 6
7
7 , so that 60 1 1 1 0 27 ! 60 0 1 0 0 27
2
t=6
6 7
6
7 6
7
415
40 0 1 1 1 15 40 0 0 1 1 25
2
0 0 0 1 1 2
0 0 0 0 0 0
So x5 is a free variable, so there are three solutions, corresponding to x5 = 0, 1, or 2. Solving for the
other variables gives
x1 = 1
x5 ,
x2 = 1
2x5 ,
x3 = 2,
so that the solutions are (remembering that 1 2 = 2 and
2 3 2 3
2 3
x1
1
0
6 x 2 7 6 17
627
6 7 6 7
6 7
6x3 7 = 627 , 627 ,
6 7 6 7
6 7
4 x 4 5 4 25
415
x5
0
1
34. The possible configurations are
2
1
61
6
60
6
40
0
1
1
1
0
0
0
1
1
1
0
0
0
1
1
1
x4 = 2
2 · 2 = 1 in Z3 )
2 3
2
607
6 7
627 .
6 7
405
2
32 3 2
3
0
s1
s1 + s2
6 7 6
7
07
7 6s2 7 6s1 + s2 + s3 7
7
6
7
6
07 6 s 3 7 = 6 s 2 + s 3 + s 4 7
7
15 4 s 4 5 4 s 3 + s 4 + s 5 5
1
s5
s4 + s5
where si 2 {0, 1, 2} is the number of times that switch i is thrown.
x5 ,
2.4. APPLICATIONS
105
35. (a) Let a value of 1 represent white, and 0 represent black. Let ai correspond to touching the number
i; it has a 1 in a given entry if the value of that square changes when i is pressed, and a 0 if it does
not (thus the 1’s correspond to the squares with the blue dots in them in the diagram). Since
each square has two states, white and black, we need to solve, over Z2 , the equation
s + x1 a1 + x2 a2 + · · · + x9 a9 = t.
Here
2 3
2 3
0
0
6 17
607
6 7
6 7
6 17
607
6 7
6 7
6 17
607
6 7
6 7
7 , t = 607 ,
0
s=6
6 7
6 7
6 17
607
6 7
6 7
6 17
607
6 7
6 7
4 15
405
0
0
Row-reducing the
2
1 1
61 1
6
60 1
6
61 0
6
61 0
6
60 0
6
60 0
6
40 0
0 0
so that
x1 a1 + x2 a2 + · · · + x9 a9 = 0
corresponding augmented matrix gives
3 2
0 1 0 0 0 0 0 0
1 0
60 1
1 0 1 0 0 0 0 17
7 6
6
1 0 0 1 0 0 0 17
7 60 0
6
0 1 1 0 1 0 0 17
7 60 0
6
1 0 1 0 1 0 1 07
!
7 60 0
6
1 0 1 1 0 0 1 17
7 60 0
6
0 1 0 0 1 1 0 17
7 60 0
0 0 1 0 1 1 1 15 4 0 0
0 0 0 1 0 1 1 0
0 0
0
0
1
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
1
0
0
s = s.
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
1
3
0
07
7
17
7
07
7
07
7.
07
7
17
7
05
0
So touching the third and seventh squares will turn all the squares black.
(b) In part (a), note that the row-reduced matrix consists of the identity matrix and a column vector
corresponding to the constants. Changing the constants cannot produce an inconsistent system,
since the matrix will still reduce to the identity matrix. So any configuration is reachable, since
we can solve x1 a1 + x2 a2 + · · · + x9 a9 = s for any s.
36. In this version of the puzzle, the squares have three states, so we work over Z3 , with 2 corresponding
to black, 1 to gray, and 0 to white. Then we need to go from
2 3
2 3
2
2
6 27
627
6 7
6 7
6 07
627
6 7
6 7
6 17
627
6 7
6 7
7 to t = 627 .
1
s=6
6 7
6 7
6 07
627
6 7
6 7
6 27
627
6 7
6 7
4 05
425
1
2
Thus we want to solve
x1 a1 + x2 a2 + · · · + x9 a9 = t
2 3
0
607
6 7
627
6 7
617
6 7
7
s=6
617 .
627
6 7
607
6 7
425
1
106
CHAPTER 2. SYSTEMS OF LINEAR EQUATIONS
Row-reducing the
2
1
61
6
60
6
61
6
61
6
60
6
60
6
40
0
corresponding augmented matrix gives
3 2
1 0 1 0 0 0 0 0 0
1 0
60 1
1 1 0 1 0 0 0 0 07
7 6
6
1 1 0 0 1 0 0 0 27
7 60 0
7
0 0 1 1 0 1 0 0 17 6
60 0
6
0 1 0 1 0 1 0 1 17
!
7 60 0
7
0 1 0 1 1 0 0 1 27 6
60 0
6
0 0 1 0 0 1 1 0 07
7 60 0
5
40 0
0 0 0 1 0 1 1 1 2
0 0 0 0 1 0 1 1 1
0 0
0
0
1
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
1
0
0
0
0
showing that a solution exists.
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
1
3
0
17
7
07
7
27
7
27
7,
17
7
07
7
15
2
37. Let Grace and Hans’ ages be g and h. Then g = 3h and g + 5 = 2(h + 5), or g = 2(h + 5) 5 = 2h + 5.
It follows that g = 3h = 2h + 5, so that h = 5 and g = 15. So Hans is 5 and Grace is 15.
38. Let the ages be a, b, and c. Then the first two conditions give us a + b + c = 60 and a b = b c.
For the third condition, Bert will be as old as Annie is now in a b years; at that time, Annie will be
a + (a b) = 2a b years, and we are given that 2a b = 3c. So we get the system
2
3 2
3
a + b + c = 60
31
1
1 60
31 0 0 28
a
2b + c = 0 ) 4 1
2
1 05 ! 4 0 1 0 205 .
2a
b
3c = 0
2
1
3
0
0 0 1 12
Annie is 28, Bert is 20, and Chris is 12.
39. Let the areas of the fields be a and b. We have the two equations
a + b = 1800
2
1
a + b = 1100.
3
2
Multiply the first equation by two and subtract the first equation from the result to get 13 a = 400, so
that a = 1200 square yards and thus b = 1800 a = 600 square yards.
40. Let x1 , x2 , and x3 be the number of bundles of the first, second, and third types of corn. Then from
the given conditions we get
3x1 + 2x2 + x3 = 39
2x1 + 3x2 + x3 = 34
x1 + 2x2 + 3x3 = 26.
Row-reduce the augmented matrix to get
2
3 2 1
6
42 3 1
1 2 3
3 2
39
1
7 6
345 ! 40
26
0
0
1
0
0
0
1
3
37
4
17 7
.
4 5
11
4
Therefore there are 9.25 measures of corn in a bundle of the first type, 4.25 measures of corn in a
bundle of the second type, and 2.75 measures of corn in a bundle of the third type.
41. (a) The addition table gives a + c = 2, a + d = 4, b + c = 3, b + d = 5. Construct and row-reduce the
corresponding augmented matrix:
2
3 2
3
1 0 1 0 2
1 0 0
1
4
61 0 0 1 47 60 1 0
1
57
6
7 6
7.
40 1 1 0 35 ! 40 0 1
1
25
0 1 0 1 5
0 0 0
0
0
2.4. APPLICATIONS
107
So d is a free variable. Solving for the other variables in terms of d gives a = 4 d, b = 5 d, and
c = d 2. So there are an infinite number of solutions; for example, with d = 1, we get a = 3,
b = 4, and c = 1.
(b) We have a + c = 3, a + d = 4, b + c = 6, b + d = 5. Construct and row-reduce the corresponding
augmented matrix:
2
3 2
3
1 0 1 0 3
1 0 0
1 0
61 0 0 1 47 60 1 0
1 07
6
7 6
7.
40 1 1 0 65 ! 40 0 1
1 05
0 1 0 1 5
0 0 0
0 1
So this is an inconsistent system, and there are no such values of a, b, c, and d.
42. The addition table gives a + c = w, a + d = y,
corresponding augmented matrix:
2
3 2
1 0 1 0 w
1
61 0 0 1 y 7 60
6
7 6
40 1 1 0 x 5 ! 40
0 1 0 1 z
0
b + c = x, b + d = z. Construct and row-reduce the
0
1
0
0
0
0
1
0
1
1
1
0
w
x
z
y
x
w
y+w
3
7
7.
5
In order for this to be a consistent system, we must have z x y + w = 0, or z y = x w. Another
way of phrasing this is x + y = z + w; in other words, the sum of the diagonal entries must equal the
sum of the o↵-diagonal entries.
43. (a) The addition table gives the following system of equations:
a+d=3
b+d=2
c+d=1
a+e=5
b+e=4
c+e=3
a+f =4
b+f =3
c + f = 1.
Construct and row-reduce the corresponding augmented matrix:
2
3 2
1 0 0 1 0 0 3
1 0 0 0 0
61 0 0 0 1 0 57 60 1 0 0 0
6
7 6
61 0 0 0 0 1 47 60 0 1 0 0
6
7 6
60 1 0 1 0 0 27 60 0 0 1 0
6
7 6
60 1 0 0 1 0 47 ! 60 0 0 0 1
6
7 6
60 1 0 0 0 1 37 60 0 0 0 0
6
7 6
60 0 1 1 0 0 17 60 0 0 0 0
7 6
6
40 0 1 0 1 0 35 40 0 0 0 0
0 0 1 0 0 1 1
0 0 0 0 0
1
1
1
1
1
0
0
0
0
The sixth row shows that this is an inconsistent system.
(b) The addition table gives the following system of equations:
a+d=1
Construct and row-reduce
2
1 0
61 0
6
61 0
6
60 1
6
60 1
6
60 1
6
60 0
6
40 0
0 0
b+d=2
3
0
07
7
07
7
07
7
07
7.
17
7
07
7
05
0
c+d=3
a+e=3
b+e=4
c+e=5
a+f =4
b+f =5
c + f = 6.
the corresponding augmented matrix:
3 2
0 1 0 0 1
1 0 0 0 0
60 1 0 0 0
0 0 1 0 37
7 6
6
0 0 0 1 47
7 60 0 1 0 0
7
0 1 0 0 27 6
60 0 0 1 0
6
0 0 1 0 47
!
7 60 0 0 0 1
7
0 0 0 1 57 6
60 0 0 0 0
6
1 1 0 0 37
7 60 0 0 0 0
5
40 0 0 0 0
1 0 1 0 5
1 0 0 1 6
0 0 0 0 0
1
1
1
1
1
0
0
0
0
3
4
57
7
67
7
37
7
17
7.
07
7
07
7
05
0
108
CHAPTER 2. SYSTEMS OF LINEAR EQUATIONS
So f is a free variable, and there are an infinite number of solutions of the form a = 4
b = 5 f , c = 6, d = f 3, e = f 1.
44. Generalize the addition table to
+
d
e
f
a
m
p
s
b
n
q
t
f,
c
o
r
u
The addition table gives the following system of equations:
Construct and row-reduce
2
1 0 0
61 0 0
6
61 0 0
6
60 1 0
6
60 1 0
6
60 1 0
6
60 0 1
6
40 0 1
0 0 1
a+d=m
b+d=n
c+d=o
a+e=p
b+e=q
c+e=r
a+f =s
b+f =t
c + f = u.
the corresponding augmented matrix:
3 2
1 0 0 m
1 0 0 0 0 1
60 1 0 0 0 1
0 1 0 p7
7 6
6
0 0 1 s7
7 60 0 1 0 0 1
6
1 0 0 n7
1
7 60 0 0 1 0
60 0 0 0 1
0 1 0 q7
!
1
7 6
60 0 0 0 0 0
0 0 1 t7
7 6
6
1 0 0 o7
7 60 0 0 0 0 0
0 1 0 r 5 40 0 0 0 0 0
0 0 1 u
0 0 0 0 0 0
3
m
7
n
7
7
o
7
7
m p
7
n + p m t7
7.
m n p + q7
7
n o q+r7
7
p q s+t 5
q r t+u
For the table to be valid, the last entry in the zero rows must also be zero, so we need the conditions
m + q = n + p, n + r = o + q, p + t = q + s, and q + u = r + t. In terms of the addition table, this means
that for each 2 ⇥ 2 submatrix, the sum of the diagonal entries must equal the sum of the o↵-diagonal
entries (compare with Exercise 42).
45. (a) Since the three points (0, 1), ( 1, 4), and (2, 1) must satisfy the equation y = ax2 + bx + c,
substitute those values to get the three equations c = 1, a b + c = 4, and 4a + 2b + c = 1.
Construct and row-reduce the corresponding augmented matrix:
2
3 2
3
0
0 1 1
1 0 0
1
41
1 1 45 ! 4 0 1 0
25
4
2 1 1
0 0 1
1
Thus a = 1, b =
2, and c = 1, so that the equation of the parabola is y = x2
2x + 1.
(b) Since the three points ( 3, 1), ( 2, 2), and ( 1, 5) must satisfy the equation y = ax2 + bx + c,
substitute those values to get the three equations 9a 3b + c = 1, 4a 2b + c = 2, and a b + c = 5.
Construct and row-reduce the corresponding augmented matrix:
2
3 2
3
9
3 1 1
1 0 0 1
44
2 1 25 ! 40 1 0 6 5
1
1 1 5
0 0 1 10
Thus a = 1, b = 6, and c = 10, so that the equation of the parabola is y = x2 + 6x + 10.
46. (a) Since the three points (0, 1), ( 1, 4), and (2, 1) must satisfy the equation x2 + y 2 + ax + by + c = 0,
substitute those values to get the three equations
1 + b + c = 0
1 + 16
a + 4b + c = 0
4 + 1 + 2a + b + c = 0
)
b + c = 1
a
4b
c = 17
2a + b + c = 5 .
2.4. APPLICATIONS
109
Construct and row-reduce the corresponding augmented matrix:
2
0
41
2
1
4
1
3 2
1
1
175 ! 40
5
0
1
1
1
0
1
0
3
2
65
5
0
0
1
Thus a = 2, b = 6, and c = 5, so that the equation of the circle is x2 + y 2 2x 6y + 5 = 0.
Completing the square gives (x 1)2 + (y 3)2 +p5 1 9 = 0, so that (x 1)2 + (y 3)2 = 5.
This is a circle whose center is (1, 3), with radius 5. A sketch of the circle follows part (b).
(b) Since the three points ( 3, 1), ( 2, 2), and ( 1, 5) must satisfy the equation x2 +y 2 +ax+by+c = 0,
substitute those values to get the three equations
9 + 1
4 + 4
1 + 25
3a + b + c = 0
2a + 2b + c = 0
a + 5b + c = 0
3a
2a
a
)
b
2b
5b
c = 10
c = 8
c = 26 .
Construct and row-reduce the corresponding augmented matrix:
2
3
42
1
1
2
5
1
1
1
3 2
10
1
85 ! 40
26
0
0
1
0
3
12
105
36
0
0
1
Thus a = 12, b = 10, and c = 36, so that the equation of the circle is x2 +y 2 +12x 10y +36 = 0.
Completing the square gives (x + 6)2 + (y 5)2 + 36 36 25 = 0, so that (x + 6)2 + (y 5)2 = 25.
This is a circle whose center is ( 6, 5), with radius 5. Sketches of both circles are:
6
10
5
9
8
!!1, 4"
4
7
6
3
!1, 3"
!!6, 5"
!!1, 5"
5
4
2
3
1
!2, 1"
!0, 1"
!!2, 2"
2
!!3, 1"
!2
!1
1
2
3
4
!11 !10 !9
!8
!7
!6
!5
!4
1
!3
!2
!1
3x+1
A
B
47. We have x23x+1
1)B =
+2x 3 = (x 1)(x+3) = x 1 + x+3 , so that clearing fractions, we get (x + 3)A + (x
3x + 1. Collecting terms gives (A + B)x + (3A B) = 3x + 1. This gives the linear system A + B = 3,
3A B = 1. Construct and row-reduce the corresponding augmented matrix:

Thus A = 1 and B = 2, so that
1
3
3x+1
x2 +2x 3
=
1
1

3
1
!
1
0
1
x 1
+
2
x+3 .
0
1
1
.
2
110
CHAPTER 2. SYSTEMS OF LINEAR EQUATIONS
48. We have
x2 3x + 3
x3 3x + 3
A
B
C
=
= +
+
x3 + 2x2 + x
x(x + 1)2
x
x + 1 (x + 1)2
x2
x
2
3x + 3 = A(x + 1)2 + Bx(x + 1) + Cx
2
3x + 3 = (A + B)x + (2A + B + C)x + A
A + B = 1,
2A + B + C =
3,
A = 3.
)
)
)
This is most easily solved by substitution: A = 3, so that A + B = 1 gives B =
the remaining equation gives 2 · 3 2 + C = 3, so that C = 7. Thus
x2 3x + 3
3
=
x3 + 2x2 + x
x
2
x+1
2. Substituting into
7
(x + 1)2
49. We have
x 1
A
Bx + C
Dx + E
=
+ 2
+ 2
(x + 1)(x2 + 1)(x2 + 4)
x+1
x +1
x +4
x
)
1 = A(x2 + 1)(x2 + 4) + (Bx + C)(x + 1)(x2 + 4) + (Dx + E)(x + 1)(x2 + 1).
Expand the right-hand side and collect terms to get
x
1 = (A + B + D)x4 + (B + C + D + E)x3 + (5A + 4B + C + D + E)x2
+ (4B + 4C + D + E)x + (4A + 4C + E).
This gives the linear system
A+ B
=
0
B+ C +D+E =
0
5A + 4B + C + D + E =
0
4B + 4C + D + E =
1
4A
+D
+ 4C
Construct and row-reduce the corresponding
2
1 1 0 1 0
6
60 1 1 1 1
6
65 4 1 1 1
6
6
40 4 4 1 1
4 0 4 0 1
Thus A =
1
5,
B = 13 , C = 0, D =
2
15 ,
=
1.
augmented matrix:
3 2
0
1 0 0 0
7 6
07 6 0 1 0 0
7 6
6
07
7 ! 60 0 1 0
7 6
15 4 0 0 0 1
1
0 0 0 0
and E =
x 1
=
(x + 1)(x2 + 1)(x2 + 4)
+E =
1
5,
0
0
0
0
1
3
1
5
17
7
37
07
7.
2 7
15 5
1
5
giving
1
x
+
2
5(x + 1) 3(x + 1)
1
x
+
5(x + 1) 3(x2 + 1)
2
1
15 x + 5
x2 + 4
2x + 3
.
15(x2 + 4)
50. We have
x(x
x3 + x + 1
A
B
Cx + D
Ex + F
Gx + H
Ix + J
= +
+ 2
+ 2
+ 2
+ 2
.
2
2
3
2
1)(x + x + 1)(x + 1)
x
x 1 x +x+1
x +1
(x + 1)
(x + 1)3
2.4. APPLICATIONS
111
1)(x2 + x + 1)(x2 + 1)3 , simplify, and equate coefficients
Multiply both sides of the equation by x(x
of xn to get the system
A+B+C +E =0
B
3A + 4B + 3C
A + 3B
3C + 3D
3A + 6B + 3C
3D + E
3A + 3B
3C + 3D
C +D+F =0
2E + F
A + 4B + C
D + 2E + G = 0
3A + B
E + 2F + H = 0
G+H +J =0
3D
2F
C +D
E
G
I=0
B
F
H
J =1
D
F +G+I =0
A = 1.
Construct and row-reduce the corresponding augmented matrix:
2
3 2
1 1
1
0
1
0
0
0
0
0 0
1 0
6
7 6
6 0 1
1
1
0
1
0
0
0
0 07 60 1
6
7 6
6 3 4
6
3
1
2
0
1
0
0
0 07
6
7 60 0
6
7 6
3
3
1
2
0
1
0
0 07 60 0
6 1 3
6
7 6
6 3 6
6
3
3
1
1
1
0
1
0 07
6
7 ! 60 0
6 3 3
7
3
3
2
1
1
1
0
1 07 6
6
60 0
6
7 6
6 1 4
1
3
0
2
0
1
0
0 17 60 0
6
7 6
6 3 1
6
1
1
1
0
1
0
1
0 07
6
7 60 0
6
7 6
0
1
0
1
0
1
0
1 15 40 0
4 0 1
1 0
0
0
0
0
0
0
0
0 1
0 0
0
0
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
1
0
0
This gives the partial fraction decomposition
x(x
x3 + x + 1
=
1)(x2 + x + 1)(x2 + 1)3
=
H=1
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
1
3
1
17
7
87
17
7
7
07
7
15 7
8 7.
97
87
77
7
47
17
47
17
25
1
2
1
15
9
7
1
1
1
1
x
8 x
8
4x
4
2x + 2
+ 8
+
+
+
x x 1 x2 + x + 1
x2 + 1
(x2 + 1)2
(x2 + 1)3
1
1
x
15x 9
7x 1
x+1
+
+
+
+
.
x 8(x 1) x2 + 1 8(x2 + 1) 4(x2 + 1)2
2(x2 + 1)3
51. Suppose that 1 + 2 + · · · + n = an2 + bn + c, and let n = 0, 1, and 2 to get the system
c=0
a+ b+c=1
4a + 2b + c = 3.
Construct and row-reduce the associated
2
0 0
6
41 1
4 2
Thus a = b =
1
2
and c = 0, so that
augmented matrix:
3 2
1 0
1 0 0
7 6
1 15 ! 40 1 0
1 3
0 0 1
1 + 2 + ··· + n =
3
1
2
17
25
0
1 2 1
1
n + n = n(n + 1).
2
2
2
52. Suppose that 12 + 22 + · · · + n2 = an3 + bn2 + cn + d, and let n = 0, 1, 2, and 3 to get the system
d= 0
a+ b+ c+d= 1
8a + 4b + 2c + d = 5
27a + 9b + 3c + d = 14.
112
CHAPTER 2. SYSTEMS OF LINEAR EQUATIONS
Construct and row-reduce the associated
2
0 0 0
61 1 1
6
6
48 4 2
27 9 3
augmented matrix:
3 2
1 0
1 0
60 1
1 17
7 6
7!6
1 5 5 40 0
1 14
0 0
0
0
1
0
3
1
3
17
27 .
17
5
6
0
0
0
1
0
Thus a = 13 , b = 12 , c = 16 , and d = 0, so that
12 + 22 + · · · + n2 =
1 3 1 2 1
1
1
n + n + n = (2n3 + 3n2 + n) = n(n + 1)(2n + 1).
3
2
6
6
6
53. Suppose that 13 + 23 + · · · + n3 = an4 + bn3 + cn2 + dn + e, and let n = 0, 1, 2, 3, and 4 to get the
system
e=
0
c+ d+e=
1
16a + 8b + 4c + 2d + e =
9
a+
b+
81a + 27b + 9c + 3d + e = 36
256a + 64b + 16c + 4d + e = 100.
Construct and row-reduce the
2
0
6
6 1
6
6 16
6
6
4 81
256
associated augmented matrix:
3 2
0
0 0 1
0
1 0
7 6
7
6
1
1 1 1
1
0 1
7 6
60 0
8
4 2 1
9 7
!
7 6
7 6
27 9 3 1 36 5 40 0
64 16 4 1 100
0 0
0
0
1
0
0
0
0
0
1
0
0
0
0
0
1
Thus a = 14 , b = 12 , c = 14 , and d = e = 0, so that
3
1
4
17
7
27
17
.
47
7
05
0
1
1
1
1
1
1 + 2 + · · · + n = n4 + n3 + n2 = n2 (n2 + 2n + 1) = n2 (n + 1)2 =
4
2
4
4
4
3
2.5
3
3
✓
n(n + 1)
2
◆2
.
Iterative Methods for Solving Linear Systems
1. Start by solving the first equation for x1 and the second for x2 :
6
x1 = +
7
4
x2 = +
5
1
x2
7
1
x1 .
5
Using the initial vector [x1 , x2 ] = [0, 0], we get a sequence of approximations:
n
x1
x2
0
0
0
1
0.857
0.800
2
0.971
0.971
Solving the first equation for x2 gives x2 = 7x1
x1
5(7x1
6) =
4
3
0.996
0.994
4
0.999
0.999
5
1.000
1.000
6
1.000
1.000
6; substituting into the second equation gives
so that
34x1 =
34
)
x1 = 1.
Substituting this value back into either equation gives x2 = 1, so the exact solution is [x1 , x2 ] = [1, 1].
2.5. ITERATIVE METHODS FOR SOLVING LINEAR SYSTEMS
113
2. Solving for x1 and x2 gives
5
2
x2 = 1 +
x1 =
1
x2
2
x1 .
Using the initial vector [x1 , x2 ] = [0, 0], we get a sequence of approximations:
n
x1
x2
n
x1
x2
0
0
0
13
2.008
0.969
1
2.5
1.0
14
2.016
1.008
2
3
1.5
3
1.75
2.0
15
1.996
1.016
4
1.4
0.75
16
1.992
0.996
5
2.125
0.5
17
2.002
0.992
6
2.25
1.125
18
2.004
1.002
7
1.938
1.25
19
1.999
1.004
Using x2 = x1 1 in the first equation gives x1 =
exact solution is [x1 , x2 ] = [2, 1].
5
2
8
1.875
0.938
20
1.998
0.999
1
2 (x1
9
2.031
0.875
21
2.000
0.998
22
2.001
1.000
10
2.062
1.031
23
2.000
1.001
11
1.984
1.062
24
2.000
1.000
12
1.969
0.984
25
2.000
1.000
1), so that x1 = 2 and x2 = 1. So the
3. Solving for x1 and x2 gives
1
x2 +
9
2
x2 = x2 +
7
x1 =
2
9
2
.
7
Using the initial vector [x1 , x2 ] = [0, 0], we get a sequence of approximations:
n
x1
x2
0
0
0
1
0.222
0.286
2
0.254
0.349
3
0.261
0.358
4
0.262
0.360
5
0.262
0.361
6
0.262
0.361
Solving the first equation for x2 gives x2 = 9x1 2. Substitute this into the second equation to get
8
8
x1 3.5(9x2 2) = 1, so that 30.5x1 = 8 and thus x1 = 30.5
⇡ 0.262295. This gives x2 = 9 · 30.5
2⇡
0.360656.
4. Solving for x1 , x2 , and x3 gives
x1 =
x2 =
x3 =
1
x2 +
20
1
x1 +
10
1
x1
10
1
x3 +
20
1
x3
10
1
x2 +
10
17
20
13
10
9
.
5
Using the initial vector [x1 , x2 , x3 ] = [0, 0, 0], we get a sequence of approximations:
n
x1
x2
x3
0
0
0
0
1
0.85
1.3
1.8
2
1.005
1.035
2.015
3
1.002
0.998
2.004
4
1.000
0.999
2.000
To solve the system directly, row-reduce the augmented matrix:
2
3 2
20
1
1 17
1 0 0
4 1
10
1 135 ! 40 1 0
1
1 10 18
0 0 1
giving the exact solution [x1 , x2 , x3 ] = [1, 1, 2].
5
1.000
1.000
2.000
3
1
15
2
6
1.000
1.000
2.000
114
CHAPTER 2. SYSTEMS OF LINEAR EQUATIONS
5. Solve to get
1
1
x2 +
3
3
1
1
1
x1
x3 +
4
4
4
1
1
x2 +
.
3
3
x1 =
x2 =
x3 =
Using the initial vector [x1 , x2 , x3 ] = [0, 0, 0], we get a sequence of approximations
n
x1
x2
x3
0
0
0
0
1
0.333
0.250
0.333
2
0.250
0.083
0.250
3
0.306
0.125
0.306
4
0.292
0.097
0.292
5
0.301
0.104
0.301
To solve the system directly, row-reduce the augmented
2
3 2
3 1 0 1
1
6
7 6
4 1 4 1 15 ! 4 0
0 1 3 1
0
⇥3 1 3⇤
giving the exact solution [x1 , x2 , x3 ] = 10
, 10 , 10 .
6
0.299
0.100
0.299
7
0.300
0.101
0.300
8
0.300
0.100
0.300
9
0.300
0.100
0.300
matrix:
0
1
0
0
0
1
3
3
10
1 7
10 5
3
10
6. Solve to get
1
x2 +
3
1
x2 = x1 +
3
1
x3 = x2 +
3
1
x4 = x3 +
3
1
3
x1 =
1
x3
3
1
1
x4 +
3
3
1
.
3
Using the initial vector [x1 , x2 , x3 , x4 ] = [0, 0, 0, 0], we get a sequence of approximations
n
x1
x2
x3
x4
0
0
0
0
0
1
0.333
0
0.333
0.333
2
0.333
0.222
0.444
0.444
3
0.407
0.259
0.556
0.481
4
0.420
0.321
0.580
0.519
To solve the system directly, row-reduce
2
3
1
6 1
3
6
6
4 0
1
0
0
5
0.440
0.333
0.613
0.527
6
0.444
0.351
0.620
0.538
7
0.450
0.355
0.630
0.540
8
0.452
0.360
0.632
0.543
the augmented matrix:
3 2
0
0 1
1 0 0
7
6
1
0 07 60 1 0
7!6
3
1 15 40 0 1
1
3 1
0 0 0
⇥5 4 7 6⇤
giving the exact solution [x1 , x2 , x3 , x4 ] = 11
, 11 , 11 , 11 .
0
0
0
1
9
0.453
0.361
0.634
0.544
10
0.454
0.363
0.635
0.545
11
0.454
0.363
0.636
0.545
12
0.454
0.363
0.636
0.545
3
5
11
4 7
11 7
7 7
5
11
6
11
7. Using the equations from Exercise 1, we get (to three decimal places, but keeping four during the
calculations)
n 0
1
2
3
4
x1 0 0.857 0.996 1.000 1.000
x2 0 0.971 0.999 1.000 1.000
The Gauss-Seidel method takes 3 iterations, while the Jacobi method takes 5.
2.5. ITERATIVE METHODS FOR SOLVING LINEAR SYSTEMS
115
8. Using the equations from Exercise 2, we get (to three decimal places, but keeping four during the
calculations)
n
x1
x2
0
0
0
1
2.5
1.5
2
1.75
0.75
3
2.125
1.125
4
1.938
0.938
5
2.031
1.031
6
1.984
0.984
7
2.008
1.008
8
1.996
0.996
9
2.002
1.002
10
1.999
0.999
11
2.000
1.000
12
2.000
1.000
The Gauss-Seidel method takes 11 iterations, while the Jacobi method takes 24.
9. Using the equations from Exercise 3, we get (to three decimal places, but keeping four during the
calculations)
n 0
1
2
3
4
x1 0 0.222 0.261 0.262 0.262
x2 0 0.349 0.360 0.361 0.361
The Gauss-Seidel method takes 3 iterations, while the Jacobi method takes 5.
10. Using the equations from Exercise 4, we get (to three
calculations)
n 0
1
2
x1 0
0.850
1.011
x2 0
1.215
0.998
x3 0
2.006
2.001
decimal places, but keeping four during the
3
1.000
1.000
2.000
4
1.000
1.000
2.000
The Gauss-Seidel method takes 3 iterations, while the Jacobi method takes 5.
11. Using the equations from Exercise 5, we get (to three decimal places, but keeping four during the
calculations)
n 0
1
2
3
4
5
6
x1 0 0.333 0.278 0.296 0.299 0.300 0.300
x2 0 0.167 0.111 0.102 0.100 0.100 0.100
x3 0 0.278 0.296 0.299 0.300 0.300 0.300
The Gauss-Seidel method takes 5 iterations, while the Jacobi method takes 8.
12. Using the equations from Exercise
calculations)
n 0
1
x1 0 0.333
x2 0 0.111
x3 0 0.370
x4 0 0.457
6, we get (to three decimal places, but keeping four during the
2
0.370
0.247
0.568
0.523
3
0.416
0.328
0.617
0.539
4
0.433
0.353
0.631
0.544
5
0.451
0.361
0.635
0.545
6
0.454
0.363
0.636
0.545
7
0.454
0.363
0.636
0.545
The Gauss-Seidel method takes 6 iterations, while the Jacobi method takes 11.
1.
13.
0.8
0.6
0.4
0.2
0.2
0.4
0.6
0.8
1.
116
14.
CHAPTER 2. SYSTEMS OF LINEAR EQUATIONS
1.5
1.
0.5
1.
0.5
2.
1.5
2.5
15. Applying the Gauss-Seidel method to x1 2x2
gives
n 0
1
x1 0
3
x2 0
4
= 3, 3x1 + 2x2 = 1 with an initial estimate of [0, 0]
2
5
8
3
19
28
4
53
80
which evidently diverges. If, however, we reverse the two equations to get 3x1 + 2x2 = 1, x1 2x2 = 3,
this is strictly diagonally dominant since |3| > |2| and | 2| > |1|. Applying the Gauss-Seidel method
gives
n
x1
x2
0
0
0
1
0.333
1.333
2
1.222
0.889
3
0.926
1.037
4
1.025
0.988
5
0.992
1.004
6
1.003
0.999
7
0.999
1.000
8
1.000
1.000
9
1.000
1.000
The exact solution is [x1 , x2 ] = [1, 1].
16. Applying the Gauss-Seidel method to x1
initial estimate of [0, 0, 0] gives
n
x1
x2
x3
0
0
0
0
4x2 + 2x3 = 2, 2x2 + 4x3 = 1, 6x1
1
2
0.5
5.25
2
6.5
10
15
3
8
30.5
39.75
x2
2x3 = 1 with an
4
203.5
80
570
which evidently diverges. If, however, we arrange the equations as
6x1
x1
x2
2x3 = 1
4x2 + 2x3 = 2
2x2 + 4x3 = 1,
we get a strictly diagonally dominant system. Applying the Gauss-Seidel method gives
n
x1
x2
x3
17.
0
0
0
0
1
0.167
0.458
0.479
2
0.250
0.198
0.349
⇥
The exact solution is [x1 , x2 , x3 ] = 14 , 14 ,
75
50
25
!50
!40
!30
!20
10
!10
!25
20
3
0.250
0.263
0.382
⇤
3
8 .
4
0.250
0.247
0.373
5
0.250
0.251
0.375
6
0.250
0.250
0.375
7
0.250
0.250
0.375
2.5. ITERATIVE METHODS FOR SOLVING LINEAR SYSTEMS
117
18. Applying the Gauss-Seidel method to the equations
4x1 + 5x2 = 14
x1
3x2 =
7
4
1.18
1.94
5
1.08
1.97
gives
n
x1
x2
0
0
0
1
3.5
1.17
2
2.04
1.65
3
1.43
1.86
This gives an approximate solution (to 0.01) of x1 =
[ 1, 2].
6
1.03
1.99
7
1.01
2.00
8
1.01
2.00
1.01, x2 = 2.00. The exact solution is [x1 , x2 ] =
19. Applying the Gauss-Seidel method to the equations
5x1
2x2 + 3x3 =
x1 + 4x2
2x1
8
4x3 = 102
2x2 + 4x3 = 90
gives
n
x1
x2
x3
0
0
0
0
1
1.6
25.9
10.35
n
x1
x2
x3
2
14.97
11.41
9.31
9
10.00
11.05
11.97
3
8.55
14.05
11.2
10
10.00
11.03
11.98
4
10.74
11.62
11.32
11
10.00
11.01
11.99
5
9.84
11.72
11.72
12
10.00
11.01
12.00
6
10.12
11.25
11.82
13
10.00
11.00
12.00
7
9.99
11.19
11.90
8
10.02
11.08
11.95
14
10.00
11.00
12.00
This gives an approximate solution (to 0.01) of x1 = 10.00, x2 = 11.00, x3 =
solution is [x1 , x2 , x3 ] = [10, 11, 12].
12.00. The exact
20. Redoing the computation from Exercise 18, showing three decimal places, and computing to an accuracy
of 0.001, gives
n
x1
x2
0
0
0
1
3.5
1.167
2
2.042
1.653
n
x1
x2
3
1.434
1.855
9
1.002
1.999
4
1.181
1.940
10
1.001
2.000
5
1.075
1.975
11
1.000
2.000
This gives an approximate solution (to 0.001) of x1 =
[x1 , x2 ] = [ 1, 2].
6
1.031
1.990
7
1.013
1.996
8
1.005
1.998
12
1.000
2.000
1.000, x2 = 2.000. The exact solution is
118
CHAPTER 2. SYSTEMS OF LINEAR EQUATIONS
21. Redoing the computation from Exercise 19, showing three decimal places, and computing to an accuracy
of 0.001, gives
n
x1
x2
x3
0
0
0
0
1
1.6
25.9
10.35
2
14.97
11.408
9.311
3
8.55
14.051
11.199
4
10.74
11.615
11.322
5
9.839
11.718
11.721
6
10.120
11.249
11.816
n
x1
x2
x3
9
10.002
11.052
11.973
10
10.005
11.026
11.985
11
10.001
11.015
11.992
12
10.001
11.008
11.996
13
10.000
11.005
11.998
n
x1
x2
x3
14
10.000
11.002
11.999
15
10.000
11.001
11.999
16
10.000
11.001
12.000
17
10.000
11.000
12.000
18
10.000
11.000
12.000
7
9.989
11.187
11.912
This gives an approximate solution (to 0.001) of x1 = 10.000, x2 = 11.000, x3 =
solution is [x1 , x2 , x3 ] = [10, 11, 12].
8
10.022
11.082
11.948
12.000. The exact
22. With the interior points labeled as shown, we use the temperature-averaging property to get the
following system:
1
1
(0 + 40 + 40 + t2 ) = 20 + t2
4
4
1
5 1
1
t2 = (0 + t1 + t3 + 5) = + t1 + t3
4
4 4
4
1
85 1
t3 = (t2 + 40 + 40 + 5) =
+ t2 .
4
4
4
t1 =
The Gauss-Seidel method gives the following with initial approximation [0, 0, 0]:
n
t1
t2
t3
0
0
0
0
1
20.000
6.250
22.812
2
21.562
12.344
24.336
3
23.086
13.105
24.526
4
23.276
13.201
24.550
5
23.300
13.213
24.553
6
23.303
13.214
24.554
7
23.304
13.214
24.554
8
23.304
13.214
24.554
This gives equilibrium temperatures (to 0.001) of t1 = 23.304, t2 = 13.214, and t3 = 24.554.
23. With the interior points labeled as shown, we use the temperature-averaging property to get the
following system:
1
(t2 + t3 )
4
1
t2 = (t1 + t4 )
4
1
t3 = (t1 + t4 + 200)
4
1
t4 = (t2 + t3 + 200)
4
t1 =
The Gauss-Seidel method gives the following with initial approximation [0, 0, 0]:
n
t1
t2
t3
t4
0
0
0
0
0
1
0
0
50
62.5
2
12.5
18.75
68.75
71.875
3
21.875
23.438
73.438
74.219
4
24.219
24.609
74.609
74.805
5
24.805
24.902
74.902
74.951
6
24.951
24.976
74.976
74.988
7
24.988
24.994
74.994
74.997
8
24.997
24.998
74.998
74.998
This gives equilibrium temperatures (to 0.001) of t1 = t2 = 25 and t3 = t4 = 75.
9
24.999
25.000
75.000
75.000
10
25.000
25.000
75.000
75.000
2.5. ITERATIVE METHODS FOR SOLVING LINEAR SYSTEMS
119
24. With the interior points labeled as shown, we use the temperature-averaging property to get the
following system:
1
(t2 + t3 )
4
1
t2 = (t1 + t4 + 40)
4
1
t3 = (t1 + t4 + 80)
4
1
t4 = (t2 + t3 + 200)
4
t1 =
The Gauss-Seidel method gives the following with initial approximation [0, 0, 0]:
n
t1
t2
t3
t4
0
0
0
0
0
1
0
10
20
57.5
2
7.5
26.25
36.25
65.625
3
15.625
30.312
40.312
67.656
4
17.656
31.328
41.328
68.164
5
18.164
31.582
41.582
68.291
6
18.291
31.646
41.646
68.323
7
18.323
31.661
41.661
68.331
8
18.331
31.665
41.665
68.333
9
18.333
31.666
41.666
68.333
10
18.333
31.667
41.667
68.333
This gives equilibrium temperatures (to 0.001) of t1 = 18.333, t2 = 31.667, t3 = 41.667, and t4 = 68.333.
25. With the interior points labeled as shown, we use the temperature-averaging property to get the
following system:
t1 =
t2 =
t3 =
t4 =
t5 =
t6 =
1
(t2 + 80)
4
1
(t1 + t3 + t4 )
4
1
(t2 + t5 + 80)
4
1
(t2 + t5 + 5)
4
1
(t3 + t4 + t6 + 5)
4
1
(t5 + 85).
4
The Gauss-Seidel method gives the following with initial approximation [0, 0, 0]:
n
t1
t2
t3
t4
t5
t6
n
t1
t2
t3
t4
t5
t6
0
0
0
0
0
0
0
1
2
20
21.25
5
11.25
21.25 24.609
2.5
5.859
7.188 14.629
23.047 24.097
7
23.809
15.282
28.058
9.308
16.963
25.491
8
23.821
15.297
28.065
9.315
16.968
25.492
3
22.812
13.32
26.987
8.237
16.283
25.321
9
23.824
15.301
28.067
9.317
16.969
25.492
4
23.33
14.639
27.730
8.980
16.758
25.439
10
23.825
15.302
28.068
9.318
16.970
25.492
5
23.66
15.093
27.963
9.213
16.904
25.476
11
23.826
15.303
28.068
9.318
16.970
25.492
6
23.773
15.237
28.035
9.285
16.949
25.487
12
23.826
15.303
28.068
9.318
16.970
25.492
This gives equilibrium temperatures (to 0.001) of t1 = 23.826, t2 = 15.303, t3 = 28.068, t4 = 9.318,
t5 = 16.970, and t6 = 25.492.
120
CHAPTER 2. SYSTEMS OF LINEAR EQUATIONS
26. With the interior points labeled as shown, we use the temperature-averaging property to get the
following system:
t1 =
t2 =
t3 =
t4 =
t5 =
t6 =
t7 =
t8 =
n
t1
t2
t3
t4
t5
t6
t7
t8
t9
t10
t11
t12
t13
t14
t15
t16
n
t1
t2
t3
t4
t5
t6
t7
t8
t9
t10
t11
t12
t13
t14
t15
t16
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
1
(t2 + t5 )
4
1
(t1 + t3 + t6 )
4
1
(t2 + t4 + t7 + 20)
4
1
(t3 + t8 + 40)
4
1
(t1 + t6 + t9 )
4
1
(t2 + t5 + t7 + t10 )
4
1
(t3 + t6 + t8 + t11 )
4
1
(t4 + t7 + t12 + 20)
4
1
2
0
0
0
1.25
5.
8.438
11.25 14.141
0
2.5
0
1.875
1.25
4.844
8.125 16.562
10.
16.875
2.5
8.984
0.938 17.598
27.266 49.575
22.5
28.281
16.25 26.641
29.297 52.095
64.141 75.417
10
7.274
13.794
23.179
24.998
16.522
26.029
34.83
37.142
34.088
40.284
53.83
69.006
40.452
48.051
71.733
85.185
11
7.579
14.197
23.506
25.162
16.924
26.559
35.259
37.357
34.415
40.714
54.178
69.18
40.617
48.266
71.907
85.272
3
0.938
2.812
10.449
16.753
4.922
5.391
12.5
24.707
20.547
17.544
32.928
58.263
31.797
35.359
60.926
79.797
12
7.78
14.461
23.721
25.269
17.189
26.906
35.54
37.497
34.63
40.995
54.406
69.294
40.724
48.406
72.021
85.329
t9 =
t10 =
t11 =
t12 =
t13 =
t14 =
t15 =
t16 =
4
1.934
4.443
13.424
19.533
6.968
10.364
20.356
29.538
24.077
25.682
41.307
62.661
34.859
40.367
65.368
82.007
13
7.912
14.635
23.861
25.34
17.362
27.133
35.724
37.589
34.77
41.179
54.554
69.368
40.794
48.498
72.095
85.366
5
2.853
6.66
16.637
21.544
9.323
15.505
25.747
32.488
27.466
31.161
46.234
65.182
36.958
43.372
67.903
83.271
14
7.999
14.748
23.953
25.386
17.476
27.282
35.845
37.65
34.862
41.299
54.652
69.417
40.84
48.559
72.144
85.39
1
(t5 + t10 + t13 + 40)
4
1
(t6 + t9 + t11 + t14 )
4
1
(t7 + t10 + t12 + t15 )
4
1
(t8 + t11 + t16 + 100)
4
1
(t9 + t14 + 80)
4
1
(t10 + t13 + t15 + 40)
4
1
(t11 + t14 + t16 + 100)
4
1
(t12 + t15 + 200).
4
6
3.996
9.035
19.081
22.892
11.742
19.421
29.306
34.345
29.965
34.748
49.285
66.725
38.334
45.246
69.451
84.044
15
8.056
14.823
24.013
25.416
17.55
27.379
35.923
37.689
34.922
41.378
54.716
69.449
40.87
48.598
72.176
85.406
7
5.194
10.924
20.781
23.781
13.645
22.156
31.642
35.537
31.682
37.092
51.227
67.702
39.232
46.444
70.429
84.533
16
8.093
14.871
24.053
25.435
17.599
27.443
35.975
37.715
34.962
41.43
54.757
69.47
40.89
48.624
72.197
85.417
8
6.142
12.27
21.923
24.365
14.995
24.
33.172
36.31
32.83
38.625
52.482
68.331
39.818
47.218
71.058
84.847
17
8.118
14.903
24.078
25.448
17.631
27.485
36.009
37.732
34.988
41.463
54.785
69.483
40.903
48.641
72.21
85.423
9
6.816
13.185
22.68
24.748
15.911
25.223
34.174
36.813
33.589
39.628
53.298
68.74
40.202
47.722
71.467
85.052
18
8.133
14.924
24.095
25.457
17.651
27.512
36.031
37.743
35.004
41.485
54.803
69.492
40.911
48.652
72.219
85.428
19
8.144
14.938
24.106
25.462
17.665
27.53
36.045
37.75
35.015
41.5
54.814
69.498
40.917
48.659
72.225
85.431
2.5. ITERATIVE METHODS FOR SOLVING LINEAR SYSTEMS
n
t1
t2
t3
t4
t5
t6
t7
t8
t9
t10
t11
t12
t13
t14
t15
t16
20
8.151
14.947
24.114
25.466
17.674
27.541
36.055
37.755
35.023
41.509
54.822
69.502
40.92
48.664
72.229
85.433
21
8.155
14.953
24.118
25.468
17.68
27.549
36.061
37.758
35.027
41.516
54.827
69.504
40.923
48.667
72.232
85.434
22
8.158
14.956
24.121
25.47
17.684
27.554
36.065
37.76
35.03
41.52
54.83
69.506
40.924
48.669
72.233
85.435
n
t1
t2
t3
t4
t5
t6
t7
t8
t9
t10
t11
t12
t13
t14
t15
t16
23
8.16
14.959
24.123
25.471
17.686
27.557
36.068
37.761
35.033
41.522
54.832
69.507
40.925
48.67
72.234
85.435
30
8.163
14.963
24.127
25.473
17.691
27.563
36.072
37.764
35.036
41.527
54.836
69.509
40.927
48.673
72.236
85.436
24
8.161
14.961
24.125
25.471
17.688
27.56
36.069
37.762
35.034
41.524
54.834
69.508
40.926
48.671
72.235
85.436
31
8.164
14.963
24.127
25.473
17.691
27.563
36.073
37.764
35.036
41.527
54.836
69.509
40.927
48.673
72.236
85.436
25
8.162
14.962
24.126
25.472
17.689
27.561
36.071
37.763
35.035
41.525
54.835
69.508
40.926
48.672
72.235
85.436
32
8.164
14.964
24.127
25.473
17.691
27.563
36.073
37.764
35.036
41.527
54.836
69.509
40.927
48.673
72.236
85.436
121
26
8.163
14.962
24.126
25.472
17.69
27.562
36.071
37.763
35.035
41.526
54.835
69.509
40.927
48.672
72.236
85.436
33
8.164
14.964
24.127
25.473
17.691
27.564
36.073
37.764
35.036
41.527
54.836
69.509
40.927
48.673
72.236
85.436
27
8.163
14.963
24.127
25.472
17.69
27.562
36.072
37.763
35.036
41.526
54.836
69.509
40.927
48.672
72.236
85.436
28
8.163
14.963
24.127
25.472
17.69
27.563
36.072
37.763
35.036
41.527
54.836
69.509
40.927
48.672
72.236
85.436
29
8.163
14.963
24.127
25.473
17.691
27.563
36.072
37.763
35.036
41.527
54.836
69.509
40.927
48.673
72.236
85.436
34
8.164
14.964
24.127
25.473
17.691
27.564
36.073
37.764
35.036
41.527
54.836
69.509
40.927
48.673
72.236
85.436
Column 34 gives the equilibrium temperature to an accuracy of 0.001.
27. (a) Let x1 correspond to the left end of the paper and x2 to the right end, and let n be the number
of folds. Then the first six values of [x1 , x2 ] are
n
x1
x2
0
0
1
1
0
1
2
2
3
4
5
6
1
4
1
2
1
4
3
8
5
16
3
8
5
16
11
32
21
64
11
32
The graph below shows these points, together with the two lines determined in part (b):
1.
0.9
0.8
0.7
0.6
0.5
0.4
0.3
0.2
0.1
0.
0.
0.1
0.2
0.3
0.4
0.5
0.6
0.7
0.8
0.9
1.
122
CHAPTER 2. SYSTEMS OF LINEAR EQUATIONS
(b) From the table, we see that x1 = 0 gives x2 = 12 in the next iteration, and x1 = 14 gives x2 = 38 in
the next iteration. Similarly, x2 = 1 gives x1 = 0 in the next iteration, and x2 = 12 gives x1 = 14
in the next iteration. Thus
3
8
1
4
1
4
1
2
1
=
2
x2
x1
0=
1
2
(x1
0), or x2 =
1
1
x1 +
2
2
0
(x2
1
1), or x1 =
1
1
x2 +
2
2
0
(c) Switching to decimal representations, we apply Gauss-Seidel to these equations after n = 6,
11
starting with the estimates x1 = 21
64 and x2 = 32 , to obtain an approximate point of convergence:
n
x1
x2
0
0
1
1
0
1
2
2
3
4
5
6
1
4
1
2
1
4
3
8
5
16
3
8
5
16
11
32
21
64
11
32
7
0.328
0.336
8
0.332
0.334
9
0.333
0.333
10
0.333
0.333
(d) Row-reduce the augmented matrix for this pair of equations, giving
"
# "
#
1
1 12
1 0 13
2
.
!
1 12 12
0 1 13
⇥
⇤
The exact solution to the system of equations is [x1 , x2 ] = 13 , 13 . The ends of the piece of paper
converge to 13 .
28. (a) Let x1 record the positions of the left endpoints and x2 the positions of the right endpoints at the
end of each walks. Then the first six values of [x1 , x2 ] are
n
x1
x2
0
0
1
2
1
2
3
4
5
6
1
4
1
2
1
4
5
8
5
16
5
8
5
16
21
32
21
64
21
32
21
64
85
128
The graph below shows these points, together with the two lines determined in part (b):
1.
0.9
0.8
0.7
0.6
0.5
0.4
0.3
0.2
0.1
0.
0.
0.1
0.2
0.3
0.4
0.5
0.6
0.7
0.8
0.9
1.
1
2
(b) From the table, we see that x1 = 0 gives x2 = in the next iteration, and x1 = 14 gives x2 = 58 in
5
the next iteration. Similarly, x2 = 12 gives x1 = 14 in the next iteration, and x2 = 58 gives x1 = 16
in the next iteration. Thus
x2
1
=
2
x1
1
=
4
5
8
1
4
5
8
5
16
1
2
0
(x1
1
2
1
4
✓
0), or x2 =
x2
1
2
◆
1
1
x1 +
2
2
, or x1 =
1
x2 .
2
(c) Switching to decimal representations, we apply Gauss-Seidel to these equations after n = 6,
85
starting with the estimates x1 = 21
64 and x2 = 128 , to obtain an approximate point of convergence:
n
x1
x2
0
0
1
2
1
2
3
4
5
6
1
4
1
2
1
4
5
8
5
16
5
8
5
16
21
32
21
64
21
32
21
64
85
128
7
0.332
0.666
8
0.333
0.667
9
0.333
0.667
Chapter Review
123
(d) Row-reduce the augmented matrix for this pair of equations, giving
"
# "
#
1
1
1
1 0 13
2
2
.
!
1
1
0
0 1 23
2
⇥
⇤
The exact solution to the system of equations is [x1 , x2 ] = 13 , 23 . So the ant eventually oscillates
between 13 and 23 .
Chapter Review
1. (a) False. Some linear systems are inconsistent, and have no solutions. See Section 2.1. One example
is x + y = 0 and x + y = 1.
(b) True. See Section 2.2. Since in a homogeneous system, the lines defined by the equations all
pass through the origin, the system has at least one solution; namely, the solution in which all
variables are zero.
(c) False. See Theorem 2.6. This is guaranteed only if the system is homogeneous, but not if it is not
(for example, x + y = 0, x + y = 1 is inconsistent in R3 ).
(d) False. See the discussion of homogeneous systems in Section 2.2. It can have either a unique
solution, infinitely many solutions, or no solutions.
(e) True. See Theorem 2.4 in Section 2.3.
(f ) False. If u and v are linearly dependent (i.e., one is a multiple of the other), we get a line, not
a plane. And if u and v are both zero, we get only the origin. If the two vectors are linearly
independent, then their span is indeed a plane through the origin.
(g) True. If u and v are parallel, then v = cu, so that cu + v = 0, so that u and v are linearly
dependent.
(h) True. If they can be drawn that way, then the sum of the vectors is 0, since tracing the vectors
in turn returns you to the starting point.
(i) False. For example, u = [1, 0, 0], v = [0, 1, 0] and w = [1, 1, 0] are linearly dependent since
u + v w = 0, yet none of these is a multiple of another.
(j) True. See Theorem 2.8 in Section 2.3. m vectors in Rn are linearly dependent if m > n.
2. The rank is the
2
1
2
63
1
6
43
4
0
5
number of nonzero rows in row
3
2
0
3 2
1
2
60
1
3 47
5
R
$R
4 6
7 2 !
43
2
3 25
4
1
6 2
3
1
echelon form:
3
0
3 2
1
6 27
7 R
2
3 25 3
1
3 4 R4
2
1
60
3R1 +2R2 6
! 40
3R1 +R2
! 0
2
5
0
0
This row echelon form of A has two nonzero rows, so that rank(A) = 2.
3. Construct the augmented matrix of the
2
3
2
1 1
2 4
1
1
R R1
41 3
40
1 75 2 !
2
1
2 1
5 7 R3 2R!
0
1
R1 +2R3 2
! 1 1 0
R2 +R3 40 2 0
!
0 0 1
Thus the solution is
system and row-reduce
3
2
2
4
1 1
40 2
1
35
2R3
1
1
! 0 2
3
2
2 1
1 1 0
2 4
45 2 R!
0 1 0
1
0 0 1
2
3 2 3
x1
0
4 x 2 5 = 4 25 .
x3
1
3
0 3 2
1 6 27
7
0 0 05
0 0 0
it:
3
2
2 4
1 1
40 2
1 35
2
2 2 R3 R!
0 0
3
2
2
2 R1 R!
1 0 0
40 1 0
25
1
0 0 1
2
1
1
3
0
25 .
1
3
4
35
1
124
CHAPTER 2. SYSTEMS OF LINEAR EQUATIONS
4. Construct the augmented matrix and row-reduce it:
2
3
41
1
8
2
3
18 1
4 0
7 1
3 R2 2
35 ! 1
3 4
115 R!
1
1
10 R!
3
R3
R2
2
3
8
2
3
11
105
35
4 0
7 1
18 1
1 2
40 1
2R2
! 0 0
2
1 2
R3 4
! 0 1
0 0
2
1 2
R2 R1
! 40 1
R3 3R1
! 0 2
3
2
4
0 11
1
40
3
1
15
R3
0
1
4
! 0
3 R1 2R2 2
4 0 11
! 1
40
3 0
35
0 1
4
0
Then y = t is a free variable, and the general solution
2 3 2
3 2
w
5 2t
6 x 7 63 + 3t7 6
6 7=6
7 6
4y 5 4 t 5 = 4
z
4
is w = 5
3
2
5
6
37
7 + t6
4
05
4
5. Construct the augmented matrix and row-reduce it over Z7 :



1
2 3 4 4R!
1 5 2
1 5 2
1
1 2 3
1 2 3 R2 +6R!
0 4 1


x
6
Thus the solution is
=
.
y
2
6. Construct the augmented matrix and row-reduce over Z5 :


3 2 1
3 2
1
1 4 2 R2 +3R!
0 0
4
3
6
0
1
1
3
11
15
2
2
1
0
4 0
3 1
0 1
0
1
0
2 0
3 0
0 1
3
11
15
4
3
5
35
4
2t, x = 3 + 3t, y = t, z =
3
2
37
7.
15
0

! 1
0
R1 +4R2
0
4
4:
6
.
2
1
0
So the solutions are all of the form 3x + 2y = 1. Setting y = 0, 1, 2, 3, 4 in turn, we get


x
2
y = 0 ) 3x = 1 ) x = 2 )
=
y
0


x
3
y = 1 ) 3x + 2 = 1 ) 3x = 4 ) x = 3 )
=
y
1


x
4
y = 2 ) 3x + 4 = 1 ) 3x = 2 ) x = 4 )
=
y
2


x
0
y = 3 ) 3x + 1 = 1 ) x = 0 )
=
y
3


x
1
y = 4 ) 3x + 3 = 1 ) 3x = 3 ) x = 1 )
=
.
y
4
7. Construct the augmented matrix and row-reduce it:


k 2 1 R2 $R1 1 2k 1
! k 2 1
1 2k 1

1
! 0
R2 kR1
2
2k
2k 2
1
1
k
The system is inconsistent when 2 2k 2 = 2(k + 1)(k 1) = 0 but 1 k 6= 0. The first of these is
zero for k = 1 and for k = 1; when k = 1 we have 1 k = 2 6= 0. So the system is inconsistent
when k = 1.
Chapter Review
125
8. Form the augmented matrix of the system and row reduce it (see Example 2.14 in Section 2.2):

1
5
2
6
3
7
4
8
R2

1
5R1
! 0
2
4
3
8
4
12

1
! 0
1
4 R2
2
1
3
2

! 1
0
R1 2R2
4
3
0
1
1
2
2
3
Thus z = t is a free variable, and x = 2 + t, y = 3 2t, so the parametric equation of the line is
2 3 2
3 2 3
2 3
x
2+t
2
1
4y 5 = 4 3 2t 5 = 4 35 + t 4 25 .
z
t
0
1
9. As in Example 2.15 in Section 2.2, we wish to find x = [x, y, z] that satisfies both equations simultaneously; that is, x = p + su = q + tv. This means su tv = q p. Substituting the given vectors
into this equation gives
2 3
2 3 2 3 2 3 2 3
1
1
5
1
4
s + t = 4
s
t = 4
s 4 15 t 4 15 = 4 25 425 = 4 45 )
2
1
4
3
7
2s
t = 7
Form the augmented matrix of this system and row-reduce it to determine s and t:
2
1
4 1
2
Thus s =
1
1
1
3
4
45
7
2
1
R2 $R3 4
2
!
1
1
1
1
3
4
75
4
2
1
R2 2R1
! 40
R3 +R1
! 0
1
3
0
3
4
155
0
2
1
1
3 R2 4 0
!
0
3
4
55
0
1
1
0
2
! 1
40
0
R1 R2
0
1
0
1 and t = 5, so the point of intersection is
2 3 2 3 2 3 2 3
x
1
1
0
4y 5 = 425 4 15 = 435 .
z
3
2
1
3
1
55
0
(We could equally well have substituted t = 5 into the equation for the other line; check that this gives
the same answer.)
10. As in Example 2.18 in Section 2.3, we want to find scalars x and y such
2 3
2 3 2 3
1
1
3
x + y =
x + 2y =
x 415 + y 4 25 = 4 55 )
3
2
1
3x
2y =
that
3
5
1
Form the augmented matrix of this system and row-reduce it to determine x and y:
2
1
41
3
1
2
2
3
3
55
1
R2
R3
2
1
R1
! 40
3R1
! 0
1
1
5
3
3
25
10
So this system has a solution, and thus so the vector is in
2 3 2 3
2
3
1
4 55 = 415 + 2 4
1
3
2
1
R1 R2
! 40
R3 +5R3
! 0
0
1
0
3
1
25
0
the span of the other two vectors:
3
1
2 5.
2
126
CHAPTER 2. SYSTEMS OF LINEAR EQUATIONS
11. As in Example 2.21 in Section 2.3, the equation of the plane we are looking for is
2 3
2 3
2 3
x
1
3
s + 3t = x
4y 5 = s 415 + t 425 ) s + 2t = y
z
1
1
s + t = z
Form the augmented matrix and row-reduce it to find conditions for x, y, and z:
2
1
41
1
3
2
1
3
x
y5
z
R2
R3
2
1
R1
! 40
R2
! 0
3
1
2
x
y
z
3
2
1
R2 4
! 0
0
x5
x
3
1
2
x
x
z
3
2
1
40
y5
2
x R3 +2R!
0
3
1
0
x
x
x
3
y 5
2y + z
We know the system is consistent, since the two given vectors span some plane. So we must have
x 2y + z = 0, and this is the general equation of the plane we were seeking. (An alternative method
would be to take the cross product of the two given vectors, giving [ 1, 2, 1], so that the plane is of
the form x + 2y z = d; plugging in x = y = z = 0 gives x + 2y z = 0, or x 2y + z = 0.)
12. As in Example 2.23 in Section 2.3, we want to find scalars
2 3
2 3
2
2
1
c1 4 15 + c2 4 15 + c3 4
3
2
Form the augmented matrix of the resulting
2
2
1
4 1
1
3
2
c1 , c2 , and c3 so that
3 2 3
3
0
95 = 405 .
2
0
system and row-reduce it:
3 2
3
3 0
1 0
4 0
9 05 ! 40 1
5 05
2 0
0 0
0 0
Since c1 = 4c3 , c2 = 5c3 is a solution, the vectors are linearly dependent. For example, setting
c3 = 1 gives
2 3
2 3 2 3 2 3
2
1
3
0
4 4 15 5 4 15 4 95 = 405 .
3
2
2
0
13. By part (a) of Exercise 21 in Section 2.3, it is sufficient to prove that ei is in span(u, v, w) ⇢ R3 , since
then we have R3 = span(e1 , e2 , e3 ) ⇢ span(u, v, w) ⇢ R3 .
(a) Begin with e1 . We want to find scalars x, y, and z such that
2 3 2 3 2 3
2 3
1
1
0
1
x 415 + 405 + 415 = e1 = 405 .
0
1
1
0
Form the augmented matrix and row
2
3 2
1 1 0 1
1
6
7 6
4 1 0 1 05 ! 4 0
0 1 1 0
0
Similarly, for e2 ,
2
1
6
41
0
1
0
1
0
1
1
3 2
0
1
7 6
15 ! 4 0
0
0
reduce:
3
)
e1 =
1
1
u+ v
2
2
3
)
e1 =
1
u
2
0
1
0
0
0
1
1
2
1 7
2 5
1
2
0
1
0
0
0
1
1
2
17
25
1
2
1
w.
2
1
1
v + w.
2
2
Chapter Review
127
Similarly, for e3 ,
2
1
6
41
0
1
0
1
0
1
1
Thus R3 = span(u, v, w).
3 2
0
1
7 6
05 ! 40
1
0
0
1
0
0
0
1
3
1
2
1 7
2 5
1
2
)
e1 =
1
1
1
u + v + w.
2
2
2
(b) Since w = u + v, these vectors are not linearly independent. But any set of vectors that spans
R3 must contain 3 linearly independent vectors, so span(u, v, w) 6= R3 .
14. All three statements are true, so (d) is the correct answer. If the vectors are linearly independent, and
b is any vector in R3 , then the augmented matrix [A | b], when reduced to row echelon form, cannot
have a zero row (since otherwise, the portion of the matrix corresponding to A would have a zero row,
so the three vectors would be linearly dependent). Thus [A | b] has a unique solution, which can be
read o↵ from the row-reduced matrix. This proves statement (c). Next, since there is not a zero row,
continuing and putting the system in reduced row-echelon form results in a matrix with 3 leading ones,
so it is the identity matrix. Hence the reduced row echelon form of A is I3 , and thus A has rank 3.
15. Since A is a 3 ⇥ 3 matrix, rank(A)  3. Since the vectors are linearly dependent, rank(A) cannot be
3; since they are not all zero, rank(A) cannot be zero. Thus rank(A) = 1 or rank(A) = 2.
16. By Theorem 2.8 in Section 5.3, the maximum rank of a 5 ⇥ 3 matrix is 3: since the rows are vectors in
R3 , any set of four or more of them are linearly dependent. Thus when we row reduce the matrix, we
must introduce at least two zero rows. The minimum rank is zero, achieved when all entries are zero.
17. Suppose that c1 (u + v) + c2 (u v) = 0. Rearranging gives (c1 + c2 )u + (c1 c2 )v = 0. Since u and v
are linearly independent, it follows that c1 + c2 = 0 and c1 c2 = 0. Adding the two equations gives
2c1 = 0, so that c1 = 0; then c2 = 0 as well. It follows that u + v and u v are linearly independent.
18. Using Exercise 21 in Section 2.3, we show first that u and v are linear combinations of u and u + v.
But this is clear: u = u, and v = (u + v) u. This shows that span(u, v) ✓ span(u, u + v).
Next, we show that u and u + v are linear combinations of u and v. But this is also clear. Thus
span(u, u + v) ✓ span(u, v). Hence the two spans are equal.
19. Note that rank(A)  rank([A | b]), since any zero rows in an echelon form of [A | b] are also zero rows
in a row echelon form of A, so that a row echelon form of A has at least as many zero rows as a row
echelon form of [A | b]. If rank(A) < rank([A | b]), then a row echelon form of [A | b] will contain at least
one row that is zero in the columns corresponding to A but nonzero in the last column on that row,
so that the system will be inconsistent. Hence the system is consistent only if rank(A) = rank([A | b]).
20. Row-reduce both A and B:
2
1
4 2
1
2
1
41
0
1
3
4
0
1
1
3
2
1
1 1
R 2R1
40 1
15 2 !
1
1 R3 +R!
0 5
3
2
1
1 0
1 4
15 R2 R!
0 1
3
0 1
3
2
3
1
1 1
1
40 1
35
35
R3 5R2
2
! 0 0 17
3
2
3
1
1 0
1
40 1
25
25 .
2
3 R3 R!
0 0
1
Since in row echelon form each of these matrices has no zero rows, it follows that both matrices have
rank 3, so that their reduced row echelon forms are both I3 . Thus the matrices are row-equivalent.
The sequence of row operations required to derive B from A is the sequence of row operations required
to reduce A to I3 , followed by the sequence of row operations, in reverse, required to reduce B to I3 .
Chapter 3
Matrices
3.1
Matrix Operations
1. Since A and D have the same shape, this operation makes sense.






3 0
0
3
3 0
0
6
3+0 0 6
3
A + 2D =
+2
=
+
=
=
1 5
2
1
1 5
4
2
1 4 5+2
5
2. Since D and A have the same shape, this operation makes sense.





0
3
3 0
0
9
6
0
0 6
3D 2A = 3
2
=
=
2
1
1 5
6
3
2 10
6 ( 2)
9
3
0
=
10
6
7

6
4
9
7
3. Since B is a 2 ⇥ 3 matrix and C is a 3 ⇥ 2 matrix, they do not have the same shape, so they cannot
be subtracted.
4. Since both C and B T are 3 ⇥ 2 matrices, this
2
3
2

1 2
1
T
4
2
1
C B T = 43 45
= 43
0
2 3
5 6
5
operation makes sense.
3 2
3 2
2
4 0
1 4
45 4 2 25 = 43 ( 2)
6
1 3
5 1
2
4
6
3 2
0
3
25 = 4 5
3
4
5. Since A has two columns and B has two rows, the product makes sense, and




3 0 4
2 1
3·4+0·0
3 · ( 2) + 0 · 2
3·1+0·3
12
AB =
=
=
1 5 0
2 3
1·4+5·0
1 · ( 2) + 5 · 2
1·1+5·3
4
6
12
3
2
25
3
3
.
14
6. Since the number of columns of B (3) does not equal the number of rows of D (2), this matrix product
does not make sense.
7. Since B has three columns and C has three rows, the multiplication BC makes sense, and results in a
2 ⇥ 2 matrix (since B has 2 rows and C has 2 columns). Since D is also a 2 ⇥ 2 matrix, they can be
subtracted, so this formula makes sense. First,
2
3



1 2
4
2 1 4
4·1 2·3+1·5 4·2 2·4+1·6
3
6
5
3 4 =
BC =
=
.
0
2 3
0·1+2·3+3·5 0·2+2·4+3·6
21 26
5 6
Then
D + BC =

0
2

3
3
+
1
21
6
=
26
129

0+3
2 + 21

3+6
3
=
1 + 26
19
3
.
27
130
CHAPTER 3. MATRICES
8. Since B has three columns and B T has three rows, the product makes sense, and

4
2 1
BB =
0
2 3

4
2 1
=
0
2 3

4·4 2·(
=
0·4+2·(
T

4
0
2
2 1
2 3
3
4 0
4 2 25
1 3
2) + 1 · 1
2) + 3 · 1
T
4·0 2·2+1·3
=
0·0+2·2+3·3

21
1
1
13
9. Since A has two columns and F has two rows, AF makes sense; since A has two rows and F has one
column, the result is a 2 ⇥ 1 matrix. So this matrix has two rows, and E has two columns, so we can
multiply by E on the left. Thus this product makes sense. Then
AF =

3
1
⇥
E(AF ) = 4



0
1
3 · ( 1) + 0 · 2
3
=
=
5
2
1 · ( 1) + 5 · 2
11

⇤ 3
⇥
⇤ ⇥ ⇤
2
= 4 · ( 3) + 2 · 11 = 10 .
11
and then
10. F has shape 2 ⇥ 1. D has shape 2 ⇥ 2, so DF has shape 2 ⇥ 1. Since the number of columns (1) of F
does not equal the number of rows (2) of DF , the product does not make sense.
11. Since F has one column and E has one row, the product makes sense, and
FE =

1 ⇥
4
2
⇤
2 =

1·4
2·4
1·2
=
2·2

4
8
2
.
4
12. Since E has two columns and F has one row, the product makes sense, and
⇥
EF = 4
⇤
2

⇥
⇤ ⇥ ⇤
1
= 4 · ( 1) + 2 · 2 = 0 .
2
(Note that F E from Exercise 11 and EF are not the same matrix. In fact, they don’t even have the
same shape! In general, matrix multiplication is not commutative.)
13. Since B has two rows, B T has two columns; since C has two columns, C T has two rows. Thus B T C T
makes sense, and the result is a 3 ⇥ 3 matrix (since B T has three rows and C T has three columns).
Next, CB is defined (3 ⇥ 2 multiplied by 2 ⇥ 3) and also yields a 3 ⇥ 3 matrix, so its transpose is also
a 3 ⇥ 3 matrix. Thus B T C T (BC)T is defined.
2
3
2
4 0 
4·1+0·2
4·3+0·4
1
3
5
2·3+2·4
B T C T = 4 2 25
=4 2·1+2·2
2 4 6
1 3
1·1+3·2
1·3+3·4
2
3
2
1 2 
1 · 4 + 2 · 0 1 · ( 2) + 2 · 2
4
2
1
CB = 43 45
= 43 · 4 + 4 · 0 3 · ( 2) + 4 · 2
0
2 3
5 6
5 · 4 + 6 · 0 5 · ( 2) + 6 · 2
2
3T 2
3
4 2 7
4 12 20
2 5.
(CB)T = 412 2 155 = 42 2
20 2 23
7 15 23
Thus B T C T = (CB)T , so their di↵erence is the zero matrix.
3 2
3
4·5+0·6
4 12 20
2 · 5 + 2 · 65 = 42 2
25
1·5+3·6
7 15 23
3 2
3
1·1+2·3
4 2 7
3 · 1 + 4 · 35 = 412 2 155
5·1+6·3
20 2 23
3.1. MATRIX OPERATIONS
131
14. Since A and D are both 2 ⇥ 2 matrices, the products in both orders are defined, and
2 ⇥ 2 matrix, so their di↵erence is defined as well.




0
3
3 0
0 · 3 3 · ( 1)
0·0 3·5
3
DA =
=
=
2
1
1 5
2 · 3 + 1 · ( 1)
2·0+1·5
7




3 0
0
3
3 · 0 + 0 · ( 2)
3 · ( 3) + 0 · 1
AD =
=
=
1 5
2
1
1 · 0 + 5 · ( 2)
1 · ( 3) + 5 · 1




3
15
0
9
3 0
15 ( 9)
3
DA AD =
=
=
7
5
10
8
7 ( 10)
5 8
3
15. Since A is 2 ⇥ 2, A2 is defined and


3 0
3
2
A =
1 5
1


3 0
9
A3 = AA2 =
1 5
8
is also a 2 ⇥ 2 matrix, so we can multiply by A


0
3 · 3 + 0 · ( 1)
3·0+0·5
9
=
=
5
1 · 3 + 5 · ( 1)
1·0+5·5
8


0
3 · 9 + 0 · ( 8)
3 · 0 + 0 · 25
=
=
25
1 · 9 + 5 · ( 8)
1 · 0 + 5 · 25
each results in a
15
5
0
10
9
8
6
.
3
again to get A3 .
0
25
27
49
0
.
125
16. Since D is a 2 ⇥ 2 matrix, I2 A makes sense and is also a 2 ⇥ 2 matrix. Since it has the same number
of rows as columns, squaring it (which is multiplying it by itself) also makes sense.



1 0
0
3
1 3
.
I2 D =
=
0 1
2
1
2 0
Thus
(I2
17. Suppose that A =

a
c

1
D) =
2
2
3
0
2

1·1+3·2
=
2·1+0·2

1·3+3·0
7
=
2·3+0·0
2
3
6
b
. Then if A2 = 0,
d
A2 =

a
c
b
d

a
c
 2
b
a + bc
=
d
ac + cd

ab + bd
0
=
bc + d2
0
0
0
If c = 0, so that the upper right-hand
entry is zero, then we must also have a = 0 and d = 0 so that

0 0
the diagonal entries are zero. So
is one solution. Alternatively, if c 6= 0, then ac = cd and
1 0
thus a = d. This will make the lower left entry zero as well, and the two diagonal entries will both
be equal to a2 + bc. So any values of a, b, and c 6= 0 with a2 = bc will work. For example, a = b = 1,
c = 1, d = 1:




1
1
1
1
1 1
1 1
0 0
=
=
.
1
1
1
1
1+1
1+1
0 0
18. Note that the first column of A is twice its second column. Thus



1 0
0 0
0 0
A
=
=A
.
2 0
0 0
0 0

1.50 1.00 2.00
, where the first row
1.75 1.50 1.00
gives the costs of shipping each product by truck, and the second row the costs when shipping by train.
Then BA gives the costs of shipping all the products to each of the warehouses by truck or train:
2
3


200 75
1.50 1.00 2.00 4
650.00 462.50
5
150 100 =
BA =
.
1.75 1.50 1.00
675.00 406.25
100 125
19. The cost of shipping one unit of each product is given by B =
132
CHAPTER 3. MATRICES
For example, the upper left entry in the product is the sum of 1.50 · 200, the cost of shipping doohickies
to warehouse 1 by truck, 1.00 · 150, the cost of shipping gizmos to warehouse 1 by truck, and 2.00 · 100,
the cost of shipping widgets to warehouse 1 by truck. So 650.00 is the cost of shipping all of the items
to warehouse 1 by truck. Comparing, we see that it is cheaper to ship items to warehouse 1 by truck
($650.00 vs $675.00), but to warehouse 2 by train ($406.25 vs $462.50).

0.75 0.75 0.75
20. The cost of distributing one unit of the product is given by C =
, where the entry
1.00 1.00 1.00
in row i, column j is the cost of shipping one unit of product j from warehouse i. Then the cost of
distribution is
2
3


200 75
0.75 0.75 0.75 4
337.50 225.00
150 1005 =
CA =
1.00 1.00 1.00
450.00 300.00
100 125
In this product, the entry in row i, column i represents the total cost of shipping all the products from
warehouse i. The o↵-diagonal entries do not represent anything in particular, since they are a result
of multiplying shipping costs from one warehouse with products from the other warehouse.
2 3


x
1
2
3 4 15
0
x2 =
21.
, so that Ax = b, where
2
1
5
4
x3
2 3


x1
1
2
3
0
4
5
.
A=
, x = x2 , b =
2
1
5
4
x3
2
1
22. 4 1
0
32 3 2 3
0 2 x1
1
1 05 4x2 5 = 4 25, so that Ax = b, where
1 1 x3
1
2
3
2 3
1
0 2
x1
1 05 , x = 4x2 5 ,
A=4 1
0
1 1
x3
2
3
1
b = 4 25 .
1
23. The column vectors of B are
2
3
2
b1 = 4 15 ,
1
so the matrix-column representation is
2
Ab1 = 2 4
2
Ab2 = 3 4
2
Thus
Ab3 = 0 4
2
3
3
b 2 = 4 15 ,
6
3
2 3
2
1
0
35 + 1 415 1 4
2
0
3
2 3
2
1
0
35 1 415 + 6 4
2
0
3
2 3
2
1
0
35 + 1 415 + 4 4
2
0
2
4
AB = 4 6
5
9
4
0
2 3
0
b3 = 415 ,
4
3 2
2
15 = 4
1
3 2
2
15 = 4
1
3 2
2
15 = 4
1
3
8
55 .
4
3
4
65
5
3
9
45
0
3
8
55 .
4
3.1. MATRIX OPERATIONS
133
24. The row vectors of A are
⇥
A1 = 1
0
⇤
2 ,
A2 =
so the row-matrix representation is
⇥
⇤
⇥
A1 B = 1 2 3 0 + 0 1
⇥
⇤
⇥
A2 B = 3 2 3 0 + 1 1
⇥
⇤
⇥
A3 B = 2 2 3 0 + 0 1
Thus
⇥
3
4
AB = 4 6
5
25. The outer product expansion is
2
3
1 ⇥
a1 B1 + a2 B2 + a3 B3 = 4 35 2
2
2
2
3
9
=4 6
4
6
2
4
9
4
=4 6
5
0
1
⇤
2
9
4
0
⇥
⇥
A3 = 2
0
⇤
1 ,
⇤ ⇥
⇤
9
8
4 = 4
⇤
⇥
⇤ ⇥
⇤
1 1 +1 1 6 4 = 6
4 5
⇤
⇥
⇤ ⇥
⇤
1 1
4 .
1 1 6 4 = 5 0
1
2
⇤
1 ,
1
1
6
3
8
55 .
4
2 3
2
0 ⇥
⇤
⇤
1 1 +4
3 0 + 415 1
0
3 2
3 2
0
0
0 0
2
05 + 41
1 15 + 4 1
0
0
0 0
1
3
8
55 .
4
3
2 ⇥
15 1
1
12
6
6
6
3
8
45
4
26. We have
2
Thus
3
2 3
2 3 2
3
2
3
0
7
Ba1 = 1 4 15 3 4 15 + 2 415 = 4 65
1
6
4
11
2 3
2 3
2 3 2 3
2
3
0
3
Ba2 = 0 4 15 + 1 4 15 + 0 415 = 4 15
1
6
4
6
2 3
2 3
2 3 2 3
2
3
0
1
Ba3 = 2 4 15 + 1 4 15 1 415 = 4 45 .
1
6
4
4
2
7
BA = 4 6
11
27. We have
Thus
⇥
B1 A = 2 1 0
⇥
B2 A = 1 1 0
⇥
B3 A = 1 1 0
3
1
6
3
1
45 .
4
⇤
⇥
⇤
⇥
2 +3 3 1 1 +0 2 0
⇤
⇥
⇤
⇥
2
1 3 1 1 +1 2 0
⇤
⇥
⇤
⇥
2 +6 3 1 1 +4 2 0
2
7
BA = 4 6
11
3
1
6
3
1
45 .
4
⇤ ⇥
1 = 7 3
⇤ ⇥
1 = 6
1
⇤ ⇥
1 = 11 6
⇤
1
⇤
4
⇤
4 .
⇤
4
134
CHAPTER 3. MATRICES
28. The outer product expansion is
2
3
2
2 ⇥
⇤
2 +4
b1 A1 + b2 A2 + b3 A3 = 4 15 1 0
1
2
3 2
2 0
4
9
25 + 4 3
=4 1 0
1 0
2
18
2
3
7
3
1
1
45 .
=4 6
11
6
4
3
3 ⇥
15 3
6
3
1
6
2 3
0 ⇥
⇤
1 1 + 415 2 0
4
3 2
3
3
0 0
0
15 + 42 0
15
6
8 0
4
⇤
1
⇥
⇤
29. Assume that the columns of B = b1 b2 . . . bn are linearly dependent. Then there exists a
solution to x1 b1 + x2 b2 + · · · + xn bn = 0 with at least one xi 6= 0. Now, we have
⇥
⇤ ⇥
⇤
AB = A b1 b2 . . . bn = Ab1 Ab2 . . . Abn ,
so that A(x1 b1 + x2 b2 + · · · + xn bn ) = x1 (Ab1 ) + x2 (Ab2 ) + · · · + xn (Abn ) = 0, so that the columns
of AB are linearly dependent.
30. Let
2
3
a1
6 a2 7
6 7
A = 6 . 7,
4 .. 5
so that
an
2
3
a1 B
6 a2 B 7
6
7
AB = 6 . 7 .
4 .. 5
an B
Since the rows of A are linearly dependent, there are xi , not all zero, such that x1 a1 +x2 a2 +· · ·+xn an =
0. But then (x1 a1 + x2 a2 + · · · + xn an )B = x1 (a1 B) + x2 (a2 B) + · · · + xn (an B) = 0, so that the rows
of AB are linearly dependent.
31. We have the block structure


A11 A12 B11
AB =
A21 A22 B21

B12
A11 B11 + A12 B21
=
B22
A21 B11 + A22 B21
A11 B12 + A12 B22
.
A21 B12 + A22 B22
Now, the blocks A12 , A21 , B12 , and B21 are all zero blocks, so any products involving those block are
zero. Thus AB reduces to

A11 B11
0
AB =
.
0
A22 B22
But
A11 B11 =
So finally,

1
0
⇥
A22 B22 = 2


1
2 3
1 · 2 1 · ( 1)
=
1
1 1
0 · 2 + 1 · ( 1)

⇤ 1
⇥
⇤ ⇥ ⇤
3
= 2·1+3·1 = 5 .
1
AB =
32. We have
⇥
AB = A1

A11 B11
0

⇤ B11
A2
B21
0
A22 B22
1·3 1·1
=
0·3+1·1
2
3
=4 1
0
⇥
B12
= A1 B11 + A2 B21
B22
2
1
0

3
1
2
1
3
0
05 .
5
⇤
A1 B12 + A2 B22 .
But B11 is the zero matrix, and both A2 and B12 are the identity matrix I2 , so this expression simplifies
to

⇥
⇤ ⇥
⇤
1 7 7
AB = I2 B21 A1 I2 + I2 B22 = B21 A1 + B22 =
.
2 7 7
3.1. MATRIX OPERATIONS
135
33. We have the block structure


A11 A12 B11
AB =
A21 A22 B21

B12
A11 B11 + A12 B21
=
B22
A21 B11 + A22 B21
A11 B12 + A12 B22
.
A21 B12 + A22 B22
Now, B21 is the zero matrix, so products involving that block are zero. Further, A12 , A21 , and B11 are
all the identity matrix I2 . So this product simplifies to


A11 I2 A11 B12 + I2 B22
A11 A11 B12 + B22
AB =
=
.
I2 I2 I2 B12 + A22 B22
I2 B12 + A22 B22
The remaining products are



1
2 1
0
1
2 0
+ B22 =
+
=
0
4 3
1
0
5 3





1
1 0
1
0 1
1
1
1
= B12 +
=
+
=
1
1 1
0
1 0
1
1
0
A11 B12 + B22 =
B12 + A22 B22

1
3
2
4

0
1
Thus
2
34. We have the block structure


A11 A12 B11
AB =
A21 A22 B21
1
63
AB = 6
41
0
2
4
0
1
2
.
1
3
0
37
7.
25
1
2
5
1
0

B12
A11 B11 + A12 B21
=
B22
A21 B11 + A22 B21
A11 B12 + A12 B22
.
A21 B12 + A22 B22
Now, A11 is the identity matrix I3 , and A21 is the zero matrix. Further, B22 is a 1 ⇥ 1 matrix, as is
A22 , so they both behave just like scalar multiplication. Thus this product reduces to


I3 B11 + A12 B21 I3 B12 A12
B11 + A12 B21 B12 A12
AB =
=
.
4B21
4 · ( 1)
4B21
4
Compute the matrices in the first row:
2 3
1 ⇥
B11 + A12 B21 = B11 + 425 1 1
3
2 3 2 3 2
1
1
B12 A12 = 415 425 = 4
1
3
So
2
1
1 = 40
0
3
⇤
0
15 .
2
2
2
62
AB = 6
43
4
3
3
3
4
4
6
4
4
2
1
0
3 2
3
1
45 + 42
1
3
1
2
3
3
0
17
7.
25
4
35. (a) Computing the powers of A as required, we get



1 1
1
0
0
2
3
4
A =
, A =
, A =
1 0
0
1
1



1
1
1 0
0
A5 =
, A6 =
= I2 , A7 =
1
0
0 1
1
Note that A6 = I2 and A7 = A.
3 2
1
2
25 = 42
3
3
1
,
1
1
= A.
1
3
3
3
3
4
65
4
136
CHAPTER 3. MATRICES
k
(b) Since A6 = I2 , we know that A6k = A6
2015 = 2010 + 5 = 6 · 370 + 5, we have
A
36. We have
"
B2 =
#
0
1
1
0
2015
=A
B3 = B2B =
,
6·370+5
"
=A
p1
2
p1
2
p1
2
p1
2
= I2k = I2 for any positive integer value of k. Since
6·370
#
,
5
5
5
A = I2 A = A =
B 4 = (B 2 )2 =
"

1
0
1
1
1
0
#
0
1
,
B8 =
"
1
0
#
0
1
= I2
Thus any power of B whose exponent is divisible by 8 is equal to I2 . Now, 2015 = 8 · 251 + 7, so that
"
# "
# "
#
p1
p1
p1
p1
0
1
2015
8·251
7
7
4
3
2
2
2
2
B
=B
·B =B =B ·B =
·
=
p1
p1
p1
p1
1
0
2
2
2
2

1 n
for n 1. The base case, n = 1, is just the definition of
0 1
1
A = A. Now assume that this equation holds for k. Then
n
37. Using induction, we show that A =
=

So by induction, the formula holds for all n
1.
A
38. (a)

k+1
cos ✓
A =
sin ✓
2
But cos2 ✓
1
=A A
sin ✓
cos ✓
inductive
k hypothesis

cos ✓
sin ✓
1
0
1
1

1
0

k
1
=
1
0
k+1
.
1
 2
sin ✓
cos ✓ sin2 ✓
=
cos ✓
2 cos ✓ sin ✓
2 cos ✓ sin ✓
.
cos2 ✓ sin2 ✓
sin2 ✓ = cos 2✓ and 2 cos ✓ sin ✓ = sin 2✓, so that we get

cos 2✓
sin 2✓
2
A =
.
sin 2✓
cos 2✓
(b) We have already proven the cases n = 1 and n = 2, which serve as the basis for the induction.
Now assume that the given formula holds for k. Then

inductive 
hypothesis cos ✓
sin ✓ cos k✓
sin k✓
Ak+1 = A1 Ak
=
sin ✓
cos ✓ sin k✓
cos k✓

cos ✓ cos k✓ sin ✓ sin k✓
(cos ✓ sin k✓ + sin ✓ cos k✓)
=
.
sin ✓ cos k✓ + cos ✓ sin k✓
cos ✓ cos k✓ sin ✓ sin k✓
But cos ✓ cos k✓ sin ✓ sin k✓ = cos(✓ + k✓) = cos(k + 1)✓, and sin ✓ cos k✓ + cos ✓ sin k✓ = sin(✓ +
k✓) = sin(k + 1)✓, so this matrix becomes

cos(k + 1)✓
sin(k + 1)✓
k+1
A
=
.
sin(k + 1)✓
cos(k + 1)✓
So by induction, the formula holds for all k
39. (a) aij = ( 1)i+j means that if i + j is even, then
2
1
6 1
A=6
4 1
1
1.
the corresponding entry is 1; otherwise, it is
3
1
1
1
1
1
17
7.
1
1
15
1
1
1
1.
3.1. MATRIX OPERATIONS
137
(b) Each entry is its row number minus its column number:
2
0
1
2
61
0
1
A=6
42
1
0
3
2
1
(c)
(d)
2
(1
6(2
A=6
4(3
(4
2
1)1
1)1
1)1
1)1
(1
(2
(3
(4
sin (1+14
1)⇡
1)2
1)2
1)2
1)2
(1
(2
(3
(4
sin (1+24
3
3
27
7.
15
0
3 2
1)4
0
61
1)4 7
7=6
1)4 5 42
1)4
3
1)3
1)3
1)3
1)3
(1
(2
(3
(4
1)⇡
sin (1+34
6
6sin (2+14 1)⇡ sin (2+24 1)⇡ sin (2+34
A=6
6sin (3+1 1)⇡ sin (3+2 1)⇡ sin (3+3
4
4
4
4
sin (4+14 1)⇡ sin (4+24 1)⇡ sin (4+34
2
3
sin ⇡4
sin ⇡2 sin 3⇡
sin ⇡
4
6
7
6 sin ⇡2 sin 3⇡
7
sin ⇡ sin 5⇡
4
4 7
6
=6
3⇡
5⇡
3⇡ 7
sin ⇡ sin 4 sin 2 5
4sin 4
sin ⇡ sin 5⇡
sin 3⇡
sin 7⇡
4
2
4
p
p
2
3
2
2
1
0
2
p
p 7
6 2
2
27
61
0
2
2 7
p
=6
.
6 p2
2
0
17
4 2
5
2
p
p
2
2
0
1
2
2
0
1
4
9
0
1
8
27
1)⇡
sin (1+44
1)⇡
sin (2+44
1)⇡
sin
1)⇡
sin
3
0
17
7.
165
81
1)⇡
3
1)⇡ 7
7
7
(3+4 1)⇡ 7
5
4
(4+4 1)⇡
4
40. (a) This matrix has zeros whenever the row number exceeds the column number, and the nonzero
entries are the sum of the row and column numbers:
2
3
2 3 4 5 6 7
60 4 5 6 7 8 7
6
7
60 0 6 7 8 9 7
6
7.
A=6
7
60 0 0 8 9 107
40 0 0 0 10 115
0 0 0 0 0 12
(b) This matrix has 1’s on the main diagonal
zeros elsewhere:
2
1
61
6
60
A=6
60
6
40
0
and one diagonal away from the main diagonal, and
3
1 0 0 0 0
1 1 0 0 07
7
1 1 1 0 07
7.
0 1 1 1 07
7
0 0 1 1 15
0 0 0 1 1
(c) This matrix is zero except where the row number
2
0 0 0
60 0 0
6
60 0 1
A=6
60 1 1
6
41 1 1
1 1 0
plus the column number is 6, 7, or 8:
3
0 1 1
1 1 17
7
1 1 07
7.
1 0 07
7
0 0 05
0 0 0
138
CHAPTER 3. MATRICES
41. Let A be an m ⇥ n matrix with rows a1 , a2 , . . . , am . Since ei has a 1 in the ith position and zeros
elsewhere, we get
ei A = 0 · a1 + 0 · a2 + · · · + 1 · ai + · · · + 0 · am = ai ,
which is the ith row of A.
3.2
Matrix Algebra
1. Since X
2A + 3B = O, we have X = 2A 3B, so



1 2
1 0
2 · 1 3 · ( 1)
X=2
3
=
3 4
1 1
2·3 3·1
2. Since 2X = A
B, we have X = 12 (A
B), so
✓
1
1 2
B) =
3 4
2
1
X = (A
2

1
1
0
1
◆
2·2
2·4

1 2
=
2 2

3·0
5
=
3·1
3

1
2
=
3
1
4
.
5
1
3
2
3. Solving for X gives
"
2
2
X = (A + 2B) =
3
3
4. The given equation is 2A
1
3
#
"
2
1
+2
4
1
#!
"
0
1
2
=
3
1
5
2B + 2X = 3X
X = 5A
3A; solving for X gives X = 5A


1 2
1 0
7 10
2
=
2B = 5
3 4
1 1
13 18
Thus

2
0
5
=2
3

1
1

2
0
+
1
2
6. As in Example 3.16, we form a matrix whose columns consist
and row-reduce it:
2
3
2
1
0 1
2
1 0
60
7
60 1
1
1
3
6
7!6
40
40 0
1 0
45
1
0 1
2
0 0

2
4
2
3
10
3
4
3
#
4
.
2B.

5. As in Example 3.16, we form a matrix whose columns consist
and row-reduce it:
2
3 2
1 0 2
1 0
6 2 1 57 60 1
6
7 6
4 1 2 05 ! 40 0
1 1 3
0 0
so that
# "
2
=
6

3
1
=3
2
0
0
1

0
4
1
1
0
7. As in Example 3.16, we form a matrix whose columns consist
and row-reduce it:
2
3
2
1
1 1 3
1 0
6 0
60 1
2 1 17
6
7
6
6
6 1
0 1 17
6
7 ! 60 0
6 0
7
60 0
0 0 07
6
6
4 1
5
40 0
1 0 1
0
0 0 0
0 0
of the entries of the matrices in question
3
2
17
7
05
0
1
.
1
of the entries of the matrices in question
3
0
3
0
47
7
1
15
0
0

1
0
1
1
of the entries of the matrices in question
3
0 0
0 07
7
1 07
7.
0 17
7
0 05
0 0
So the system is inconsistent, and B is not a linear combination of A1 , A2 , and A3 .
3.2. MATRIX ALGEBRA
139
8. As in Example 3.16, we form a matrix whose columns consist of the entries of the matrices in question
and row-reduce it:
2
3
2
3
1 0
1
1
2
1 0 0 0 1
60 1
6 0 1 0 0 27
0
1
27
6
7
6
7
60 1
7
6 0 0 1 0 37
1
1
3
6
7
6
7
60 0
6 0 0 0 1 47
0
0
07
6
7
6
7
61 0
6
7
1
1
07
6
7 ! 6 0 0 0 0 07
60 1
7
6 0 0 0 0 07
0
1
2
6
7
6
7
60 0
6 0 0 0 0 07
0
0
07
6
7
6
7
40 0
4 0 0 0 0 05
0
0
05
1 0
1
1
2
0 0 0 0 0
Since the linear system is consistent, we conclude that B = A1 + 2A2 + 3A3 + 4A4 .

w x
9. As in Example 3.17, let
be an element in the span of the matrices. Then form a matrix whose
y z
columns consist of the entries of the matrices in question and row-reduce it:
2
3 2
3
1 0 w
1 0
w
6 2 1 x7 6 0 1
x 2w 7
6
7 6
7
4 1 2 y 5 ! 40 0 y + 5w 2x5 .
1 1 z
0 0
z+w x
This gives the restrictions y + 5w 2x = 0, so that y = 2x 5w, and z + w
So the span consists of matrices of the form

w
x
.
2x 5w x w
x = 0, so that z = x
w.

w x
be an element in the span of the matrices. Then form a matrix whose
y z
columns consist of the entries of the matrices in question and row-reduce it:
2
3
2
3
1
0 1 w
1 0 0 w + x2y
60
60 1 0
7
x+y
1 1 x7
6
7
6
7
2
6
7!6
7
y x
40
40 0 1
5
1 0 y5
2
1
0 1 z
0 0 0
z w
10. As in Example 3.17, let
This gives the single restriction z

w = 0, so that z = w. So the span consists of matrices of the form

w x
.
y w
w x
be an element in the span of the matrices. Then form a matrix whose
y z
columns consist of the entries of the matrices in question and row-reduce it:
2
3
2
3
1
1 1 a
1 0 0
e
6 0
60 1 0
7
2 1 b7
c+e
6
7
6
7
6 1
7
60 0 1
7
0
1
c
b
2c
2e
6
7!6
7
6 0
7
60 0 0 a + 3b 4c 5e7 .
0
0
d
6
7
6
7
4 1
40 0 0
5
1 0 e5
d
0
0 0 f
0 0 0
f
11. As in Example 3.17, let
We get the restrictions a + 3b 4c 5e = 0, so that a = 3b + 4c + 5e, as well as d = f = 0. Thus the
span consists of matrices of the form

3b + 4c + 5e b c
.
0
e 0
140
CHAPTER 3. MATRICES
2
x1
12. As in Example 3.17, let 4x4
x7
whose columns consist of the
2
1 0
60 1
6
6
60 1
6
60 0
6
6
61 0
6
60 1
6
60 0
6
6
40 0
1 0
We get the restrictions
3
x2 x3
x5 x6 5 be an element in the span of the matrices. Then form a matrix
x8 x9
entries of the matrices in question and row-reduce it:
3
2
3
x1 +x5
1
1 x1
1 0 0 0
2
60 1 0 0
7
0
1 x2 7
x3 + x5 2 x1
7
6
7
7
6
7
60 0 1 0 x3 + x5 x1 x2 7
1
1 x3 7
7
6
7
60 0 0 1 x3 x2 + x5 x1 7
0
0 x4 7
7
6
7
2
7
6
7
1
1 x5 7 ! 60 0 0 0
x6 x2
7
7
6
7
60 0 0 0
7
0
1 x6 7
x
4
7
6
7
7
6
7
0
0 x7 7
x7
60 0 0 0
7
7
6
7
40 0 0 0
5
0
0 x8 5
x8
1
1
x6 = x2 ,
x9
0
x9 = x 1 ,
Thus the span consists of matrices of the form
2
x1
40
0
x2
x5
0
0
0
0
x9
x1
x4 = x7 = x8 = 0.
3
x3
x6 5 .
x1
13. Following Example 3.18, we construct an augmented matrix whose columns are the entries of the given
matrices and with zeros in the constant column, and row-reduce it:
2
3
2
3
1 4 0
1 0 0
62 3 07
6 0 1 07
6
7
6
7
43 2 05 ! 40 0 05
4 1 0
0 0 0
Since the only solution is the trivial solution, we conclude that these matrices are linearly independent.
14. Following Example 3.18, we construct an augmented matrix whose columns are the entries of the given
matrices and with zeros in the constant column, and row-reduce it:
2
3
2
3
1
2 1 0
1 0 13 0
62
60 1 1 07
1 1 07
6
7
6
7
3
6
7!6
7
44
40 0 0 05
1 1 05
3
0 1
0
0
0
0
0
Since the system has a nontrivial solution, the original matrices are not linearly independent; in fact,
from the solutions we get from the row-reduced matrix,



1 1 2
1
2 1
1 1
+
=
1 0
1 1
3 4 3
3
15. Following Example 3.18, we construct an augmented matrix whose columns are the entries of the given
matrices and with zeros in the constant column, and row-reduce it:
2
3
2
3
0 1
2
1 0
1 0 0 0 0
6 1 0
6 0 1 0 0 07
1
3 07
6
7
6
7
6 5 2
7
6
7
0
1
0
7 ! 6 0 0 1 0 07
6
6 2 3
7
6 0 0 0 1 07
1
9
0
6
7
6
7
4 1 1
4 0 0 0 0 05
0
4 05
0 1
2
5 0
0 0 0 0 0
Since the only solution is the trivial solution, we conclude that these matrices are linearly independent.
3.2. MATRIX ALGEBRA
141
16. Following Example 3.18, we construct an augmented matrix whose columns are the entries of the given
matrices and with zeros in the constant column, and row-reduce it:
2
6
6
6
6
6
6
6
6
6
6
6
6
4
1
1
0
0
2
0
0
2
6
2
1
0
0
3
0
0
4
9
1
2
0
0
1
0
0
3
5
1
1
0
0
1
0
0
0
4
3
2
0
1
60
07
7
6
60
07
7
6
7
60
07
6
7
07 ! 6
60
60
07
7
6
7
60
07
6
5
40
0
0
0
0
1
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
1
0
0
0
0
0
3
0
07
7
07
7
07
7
07
7
07
7
07
7
05
0
The only solution is the trivial solution, so the matrices are linearly independent.
17. Let A, B, and C be m ⇥ n matrices.
(a)
2
a11
6 ..
A+B =4 .
2
am1
...
..
.
...
a11 + b11
..
.
6
=4
am1 + bm1
6
=4
bm1 + am1
2
2
b11 + a11
..
.
b11
6 ..
=4 .
bm1
= B + A.
...
..
.
...
3 2
a1n
b11
.. 7 + 6 ..
. 5 4 .
amn
...
..
.
...
bm1
3
. . . b1n
.. 7
..
.
. 5
. . . bmn
3
a1n + b1n
7
..
5
.
amn + bmn
3
b1n + a1n
7
..
5
.
...
..
.
. . . bmn + amn
3 2
b1n
a11 . . .
.. 7 + 6 ..
..
.
. 5 4 .
bmn
am1
...
3
a1n
.. 7
. 5
amn
142
CHAPTER 3. MATRICES
(b)
02
a11
B6 ..
(A + B) + C = @4 .
2
6
=4
2
6
=4
3 2
a1n
b11
.. 7 + 6 ..
. 5 4 .
...
..
.
...
am1
amn
a11 + b11
..
.
...
..
.
...
am1 + bm1
bm1
a1n + b1n
..
.
amn + bmn
a11 + b11 + c11
..
.
...
..
.
...
2
31 2
. . . b1n
c11
.. 7C + 6 ..
..
.
. 5A 4 .
. . . bmn
cm1
3 2
3
c11 . . . c1n
7 6 ..
.. 7
..
5+4 .
.
. 5
am1 . . . amn
02
b11 . . .
B6 ..
..
= A + @4 .
.
bm1 . . .
...
..
.
...
bm1 + cm1
3 2
b1n
c11 . . .
.. 7 + 6 ..
..
.
. 5 4 .
bmn
cm1 . . .
= A + (B + C).
3
c1n
.. 7
. 5
cmn
. . . cmn
3
a1n + b1n + c1n
7
..
5
.
amn + bmn + cmn
am1 + bm1 + cm1
3
a11 . . . a1n
b11 + c11
6 ..
7
6
.
..
.
..
.. 5 + 4
=4 .
.
2
cm1
...
..
.
...
3
b1n + c1n
7
..
5
.
bmn + cmn
31
c1n
.. 7C
. 5A
cmn
(c)
2
a11
6 ..
A+O =4 .
am1
...
..
.
...
3 2
a1n
0
.. 7 + 6 ..
. 5 4.
amn
3 2
0
a11 + 0
.. 7 = 6
..
.5 4
.
...
..
.
...
0
0
am1 + 0
...
..
.
...
3
a1n + 0
7
..
5
.
amn + 0
2
a11
6 ..
=4 .
am1
(d)
2
a11
6 ..
A + ( A) = 4 .
2
am1
...
..
.
...
3 0 2
31
a1n
a11 . . . a1n
.. 7 + B 6 ..
.. 7C
..
.
. 5 @ 4 .
. 5A
amn
am1 . . . amn
3 2
3
a1n
a11 . . .
a1n
.. 7 + 6 ..
.. 7
..
.
. 5 4 .
. 5
amn
am1 . . .
amn
3
. . . a1n a1n
7
..
..
5
.
.
a11 . . .
6 ..
..
=4 .
.
am1 . . .
2
a11 a11
6
..
=4
.
am1 am1 . . .
2
3
0 ... 0
6
7
= 4 ... . . . ... 5
0
= O.
...
0
amn
amn
...
..
.
...
3
a1n
.. 7 = A.
. 5
amn
3.2. MATRIX ALGEBRA
143
18. Let A and B be m ⇥ n matrices, and c and d be scalars.
(a)
02
a11
B6 ..
c(A + B) = c @4 .
02
B6
= c @4
2
6
=4
2
3 2
a1n
b11
.. 7 + 6 ..
. 5 4 .
...
..
.
...
am1
amn
a11 + b11
..
.
c(a11 + b11 )
..
.
c(am1 + bm1 )
bm1
...
..
.
...
am1 + bm1
...
..
.
...
a1n + b1n
7C
..
5A
.
amn + bmn
3
c(a1n + b1n )
7
..
5
.
c(amn + bmn )
3
ca1n + cb1n
7
..
5
.
...
..
.
cam1 + cbm1 . . . camn + cbmn
2
3 2
ca11 . . . ca1n
cb11 . . .
6 ..
7 6 ..
.
.
..
.
.
=4 .
.
.
. 5+4 .
6
=4
ca11 + cb11
..
.
31
. . . b1n
.. 7C
..
.
. 5A
. . . bmn
31
cam1
2
a11
6 ..
= c4 .
...
am1
= cA + cB.
3
cb1n
.. 7
. 5
camn
cbm1 . . . cbmn
3
2
3
. . . a1n
b11 . . . b1n
.. 7 + c 6 ..
.. 7
..
..
4 .
.
.
. 5
. 5
. . . amn
bm1 . . . bmn
(b)
2
a11
6
(c + d)A = (c + d) 4 ...
am1
2
...
..
.
...
(c + d)a11
6
..
=4
.
...
..
.
...
am1
amn
2
(c + d)am1
3
a1n
.. 7
. 5
amn
3
(c + d)a1n
7
..
5
.
(c + d)amn
3
ca11 + da11 . . .
ca1n + d1n
6
7
..
..
..
=4
5
.
.
.
cam1 + dam1 . . . camn + damn
2
3 2
3
ca11 . . . ca1n
da11 . . . da1n
6
.. 7 + 6 ..
.. 7
..
..
= 4 ...
.
.
. 5 4 .
. 5
cam1 . . . camn
dam1 . . . damn
2
3
2
3
a11 . . . a1n
a11 . . . a1n
6
.. 7 + d 6 ..
.. 7
..
..
= c 4 ...
4 .
.
.
. 5
. 5
= cA + dA.
...
am1
...
amn
144
CHAPTER 3. MATRICES
(c)
2
da11
6 ..
c(dA) = c 4 .
...
..
.
...
dam1
(d)
2
a11
6 ..
1A = 1 4 .
damn
amn
cdamn
1amn
...
..
.
...
am1
3 2
1a1n
a11
.. 7 = 6 ..
. 5 4 .
...
..
.
...
1am1
3
2
cda1n
a11
.. 7 = cd 6 ..
4 .
. 5
...
..
.
...
cdam1
3 2
a1n
1a11
.. 7 = 6 ..
. 5 4 .
...
..
.
...
am1
3 2
da1n
cda11
.. 7 = 6 ..
. 5 4 .
am1
...
..
.
...
3
a1n
.. 7 .
. 5
amn
3
a1n
.. 7 = A.
. 5
amn
19. Let A and B be r ⇥ n matrices, and C be an n ⇥ s matrix, so that the product makes sense. Then
31 2
3
02
3 2
a11 . . . a1n
b11 . . . b1n
c11 . . . c1s
B6
.. 7 + 6 ..
.. 7C 6 ..
.. 7
..
..
..
(A + B)C = @4 ...
.
.
.
. 5 4 .
. 5A 4 .
. 5
ar1 . . . arn
br1 . . . brn
cn1 . . . cns
2
32
3
a11 + b11 . . . a1n + b1n
c11 . . . c1s
6
7 6 ..
..
..
.. 7
..
..
=4
54 .
.
.
.
.
. 5
2
ar1 + br1
...
arn + brn
cn1
...
(a11 + b11 )c11 + · · · + (a1n + b1n )cn1
6
..
=4
.
2
6
=4
...
..
.
...
(ar1 + br1 )c11 + · · · + (arn + brn )cn1
a11 c11 + · · · + a1n cn1
..
.
...
..
.
...
cns
3
(a11 + b11 )c1s + · · · + (a1n + b1n )cns
7
..
5
.
(ar1 + br1 )c1s + · · · + (ar1 + br1 )c1s
3
a11 c1s + · · · + a1n cns
7
..
5+
.
ar1 c11 + · · · + arn cn1
ar1 c1s + · · · + ar1 cns
2
3
b11 c11 + · · · + b1n cn1 . . . b11 c1s + · · · + b1n cns
6
7
..
..
..
4
5
.
.
.
br1 c11 + · · · + brn cn1 . . . br1 c1s + · · · + br1 cns
= AC + BC.
20. Let A be an r ⇥ n matrix and C be an n ⇥ s matrix, so that the product makes sense. Let k be a
scalar. Then
02
31
a11 c11 + · · · + a1n cn1 · · · a11 c1s + · · · + a1n cns
B6
7C
..
..
..
k(AB) = k @4
5A
.
.
.
2
ar1 c11 + · · · + arn cn1
k(a11 c11 + · · · + a1n cn1 )
6
..
=4
.
2
k(ar1 c11 + · · · + arn cn1 )
···
···
..
.
···
(ka11 )c11 + · · · + (ka1n )cn1
6
..
=4
.
(kar1 )c11 + · · · + (karn )cn1
32
ka11 . . . ka1n
c11 . . .
6 ..
7
6
.
.
..
..
.. 5 4 ...
=4 .
.
2
kar1
= (kA)B.
...
karn
cn1
...
ar1 c1s + · · · + arn cns
3
k(a11 c1s + · · · + a1n cns )
7
..
5
.
k(ar1 c1s + · · · + arn cns )
···
..
.
···
3
(ka11 )c1s + · · · + (ka1n )cns
7
..
5
.
(kar1 )c1s + · · · + (karn )cns
3
c1s
.. 7
. 5
cns
3.2. MATRIX ALGEBRA
145
That proves the first equality. Regrouping the expression on the second line above di↵erently, we get
2
3
a11 (kc11 ) + · · · + a1n (kcn1 ) · · · a11 (kc1s ) + · · · + a1n (kcns )
6
7
..
..
..
k(AB) = 4
5
.
.
.
2
ar1 (kc11 ) + · · · + arn (kcn1 ) · · · ar1 (kc1s ) + · · · + arn (kcns )
32
3
. . . a1n
kc11 . . . kc1s
.. 7 6 ..
.. 7
..
..
.
.
. 54 .
. 5
ar1 . . . arn
kcn1 . . . kcns
a11
6 ..
=4 .
= A(kB).
21. If A is m ⇥ n, then to prove Im A = A, note that
Im
2
3
e1
6 e2 7
6 7
= 6 . 7,
4 .. 5
em
where ei is a standard (row) unit vector; then
2 3
2
3 2
3
e1
e1 A
A1
6 e2 7
6 e2 A 7 6 A 2 7
6 7
6
7 6
7
Im A = 6 . 7 A = 6 . 7 = 6 . 7 = A.
4 .. 5
4 .. 5 4 .. 5
em
em A
Am
22. Note that
(A
B)(A + B) = A(A + B)
so that (A
B(A + B) = A2 + AB
B)(A + B) = A2 + B 2 if and only if AB
BA + B 2 = (A2 + B 2 ) + (AB
BA),
BA = O; that is, if and only if AB = BA.
23. Compute both AB and BA:
AB =
BA =


1
0
a
c

a b
a+c b+d
=
c d
c
d


b 1 1
a a+b
=
.
d 0 1
c c+d
1
1

So if AB = BA, then a + c = a, b + d = a + b, c = c, and d = c + d. The first equation gives c = 0 and
the second gives d = a. So the matrices that commute with A are matrices of the form

a b
.
0 a
24. Compute both AB and BA:
AB =


a
BA =
c
1
1
b
d
1
1


1
1
a
c
b
=
d

a c
a+c
b d
b+d
b
d
a+b
.
c+d

1
a
=
1
c
So if AB = BA, then a c = a b, b d = a + b, a + c = c d, and b + d = c + d. The first
(or fourth) equation gives b = c and the second (or third) gives d = a. So the matrices that commute
with A are matrices of the form

a b
.
b a
146
CHAPTER 3. MATRICES
25. Compute both AB and BA:
AB =
BA =



a b
a + 2c
b + 2d
=
c d
3a + 4c 3b + 4d


b 1 2
a + 3b 2a + 4b
=
.
d 3 4
c + 3d 2c + 4d
1
3

2
4
a
c
So if AB = BA, then a + 2c = a + 3b, b + 2d = 2a + 4b, 3a + 4c = c + 3d, and 3b + 4d = 2c + 4d. The
first (or fourth) equation gives 3b = 2c and the third gives a = d c. Substituting a = d c into the
second equation gives 3b = 2c again. Thus those are the only two conditions, and the matrices that
commute with A are those of the form
"
#
d c 23 c
.
c
d
26. Let A1 be the first matrix and A4 be the second. Then


1 0 a b
A1 B =
0 0 c d


a b 1 0
BA1 =
c d 0 0


0 0 a b
A4 B =
0 1 c d


a b 0 0
BA4 =
c d 0 1
=
=
=
=

a
0
b
0

a
c
0
0
0
c
0
d
0
0
b
.
d


So the conditions imposed by requiring that A1 B = BA1 and A4 B = BA4 are b = c = 0. Thus the
matrices that commute with both A1 and A4 are those of the form

a 0
0 d
27. Since we want B to commute with every2 ⇥ 2 matrix, in particular it must commute with A1 and A4
a 0
from the previous exercise, so that B =
. In addition, it must commute with
0 d
A2 =

0
0
1
0

0
1
and
A3 =


0
0
=
d
0

1
0
=
0
0

0
0
=
d
a

0
0
=
0
d
0
.
0
Computing the products, we get
A2 B =
BA2 =
A3 B =
BA3 =

0
0
1
0

a
0
0
d
0
1
0
0
a
0
0
d


a
0

0
0

a
0

0
1
The condition that arises from these equalities is a = d. Thus

a 0
B=
.
0 a
d
0
a
0
0
0
0
.
0
3.2. MATRIX ALGEBRA
147
Finally, note that for any 2 ⇥ 2 matrix M ,

x y
M=
= xA1 + yA2 + zA3 + wA4 ,
z w
so that if B commutes with A1 , A2 , A3 , and A4 , then it will commute with M . So the matrices B
that satisfy this condition are the matrices above: those where the o↵-diagonal entries are zero and
the diagonal entries are the same.
28. Suppose A is an m ⇥ n matrix and B is an a ⇥ b matrix. Since AB is defined, the number of columns
(n) of A must equal the number of rows (a) of B; since BA is defined, the number of columns (b) of
B must equal the number of rows (m) of A. Since m = b and n = a, we see that A is m ⇥ n and B is
n ⇥ m. But then AB is m ⇥ m and BA is n ⇥ n, so both are square.
29. Suppose A = (aij ) and B = (bij ) are both upper triangular n ⇥ n matrices. This means that aij =
bij = 0 if i > j. But if C = (cij ) = AB, then
ckl = ak1 b1l + ak2 b2l + · · · + akl bll + ak(l+1) b(l+1)l + · · · + akn bnl .
If k > l, then ak1 b1l + ak2 b2l + · · · + akl bll = 0 since each of the akj is zero. But also ak(l+1) b(l+1)l +
· · · + akn bnl = 0 since each of the bil is zero. Thus ckl = 0 for k > l, so that C is upper triangular.
30. Let A and B whose sizes permit the indicated operations, and let k be a scalar. Denote the ith row of
a matrix X by rowi (X) and its j th column by colj (X). Then
Theorem 3.4(a): For any i and j,
h
i
⇥ ⇤
⇥ ⇤
T
= AT ji = A ij , so that (AT )T = A.
AT
ij
Theorem 3.4(b): For any i and j,
⇥
⇤
⇥
⇤
⇥ ⇤
⇥ ⇤
⇥ ⇤
⇥ ⇤
(A + B)T ij = A + B ji = A ji + B ji = AT ij + B ij , so that (A + B)T = AT + B T .
Theorem 3.4(c): For any i and j,
⇥
⇤
⇥ ⇤
⇥ ⇤
⇥ ⇤
(kA)T ij = kA ji = k A ji = k A ij , so that (kA)T = k(AT ).
31. We prove that (Ar )T = (AT )r by induction on r. For r = 1, this says that (A1 )T = (AT )1 , which is
clear. Now assume that (Ak )T = (AT )k . Then
(Ak+1 )T = (A · Ak )T = (Ak )T AT = (AT )k AT = (AT )k+1 .
(The second equality is from Theorem 3.4(d), and the third is the inductive assumption).
32. We prove that (A1 + A2 + · · · + An )T = AT1 + AT2 + · · · + ATn by induction on n. For n = 1, this says
that AT1 = AT1 , which is clear. Now assume that (A1 + A2 + · · · + Ak )T = AT1 + AT2 + · · · + ATk . Then
(A1 + A2 + · · · + Ak + Ak+1 )T = ((A1 + A2 + · · · + Ak ) + Ak+1 )T
= (A1 + A2 + · · · + Ak )T + ATk+1
=
AT1
+
AT2
+ ··· +
ATk
+
ATk+1
(Thm. 3.4(b))
(Inductive assumption).
33. We prove that (A1 A2 · · · An )T = ATn ATn 1 · · · AT2 AT1 by induction on n. For n = 1, this says that
AT1 = AT1 , which is clear. Now assume that (A1 A2 · · · Ak )T = ATk ATk 1 · · · AT2 AT1 . Then
(A1 A2 · · · Ak Ak+1 )T = ((A1 A2 · · · Ak )Ak+1 )T
= ATk+1 (A1 A2 · · · Ak )T
= ATk+1 ATk · · · AT2 AT1
(Thm. 3.4(d))
(Inductive assumption).
148
CHAPTER 3. MATRICES
34. Using Theorem 3.4, we have (AAT )T = (AT )T AT = AAT and (AT A)T = AT (AT )T = AT A. So each
of AAT and AT A is equal to its own transpose, so both matrices are symmetric.
35. Let A and B be symmetric n ⇥ n matrices, and let k be a scalar.
(a) By Theorem 3.4(b), (A + B)T = AT + B T ; since both A and B are symmetric, this is equal to
A + B. Thus A + B is equal to its transpose, so is symmetric.
(b) By Theorem 3.4(c), (kA)T = kAT ; since A is symmetric, this is equal to kA. Thus kA is equal
to its own transpose, so is symmetric


1 1
0 1
36. (a) For example, let n = 2, with A =
and B =
. Then both A and B are symmetric,
1 0
1 0
but



1 1 0 1
1 1
AB =
=
,
1 0 1 0
0 1
which is not symmetric.
(b) Suppose A and B are symmetric. Then if AB is symmetric, we have
AB = (AB)T = B T AT = BA
using Theorem 3.4(d) and the fact that A and B are symmetric. In the other direction, if A and
B are symmetric and AB = BA, then
(AB)T = B T AT = BA = AB,
37. (a)
(b)
(c)
(d)
so that AB

1
T
A =
2

0
AT =
1
2
0
AT = 4 3
1
2
0
AT = 4 1
2
is symmetric.
2
while
3
1
=
0
3
0
2
A=

1
2
2
, so AT 6=
3
A and A is not skew-symmetric.
A, so A is skew-symmetric.
3
1
25 = A, so A is skew-symmetric.
0
3
1 2
0 55 =
6
5 0
A, so A is not skew-symmetric.
38. A matrix A is skew-symmetric if AT = A. But this means that for every i and j, (AT )ij = Aji =
Thus the components must satisfy aij = aji for every i and j.
39. By the previous exercise, we must have aij =
which implies that aii = 0 for all i.
40. If A and B are skew-symmetric, so that AT =
aji for every i and j. So when i = j, we get aii =
A and B T =
(A + B)T = AT + B T =
A
B=
Aij .
aii ,
B, then
(A + B),
so that A + B is also skew-symmetric.


0 a
0 b
41. Suppose that A =
and B =
are skew-symmetric. Then
a 0
b 0


ab
0
ab
0
T
AB =
, and thus (AB) =
.
0
ab
0
ab
So AB = (AB)T only if ab =
be the zero matrix.
ab, which happens if either a or b is zero. Thus either A or B must
3.2. MATRIX ALGEBRA
149
42. We must show that A AT is skew-symmetric, so we must show that (A AT )T =
But
(A AT )T = AT (AT )T = AT A
(A AT ) = AT
A.
as desired, proving the result.
43. (a) If A is a square matrix, let S = A + AT and S 0 = A AT . Then clearly A = 12 S + 12 S 0 . S is
symmetric by Theorem 3.5, so 12 S is as well. S 0 is symmetric by Exercise 42, so 12 S 0 is as well.
Thus A is a sum of a symmetric and a skew-symmetric matrix.
(b) We have
2
Thus
1
S = A + AT = 4 4
7
2
1
S = A AT = 4 4
7
3 2
3
1
65 + 4 2
9
3
3 2
3
1
65 4 2
9
3
2
5
8
2
5
8
2
1
1
1 0 4
A= S+ S = 3
2
2
5
3
5
7
4
5
6
4
5
6
3 2
5
0
75 + 41
9
2
3 2
3
7
2 6 10
85 = 4 6 10 145
9
10 14 18
3 2
3
7
0
2
4
85 = 42 0
25 .
9
4 2
0
1
0
1
44. Let A and B be n ⇥ n matrices, and k be a scalar. Then
3 2
2
1
15 = 44
0
7
2
5
8
3
3
65 .
9
(a)
tr(A + B) = (a11 + b11 ) + (a22 + b22 ) + · · · + (ann + bnn )
= (a11 + a22 + · · · + ann ) + (b11 + b22 + · · · + bnn ) = tr(A) + tr(B).
(b) tr(kA) = ka11 + ka22 + · · · + kann = k(a11 + a22 + · · · + ann ) = k tr(A).
45. Let A and B be n ⇥ n matrices. Then
tr(AB) = (AB)11 + (AB)22 + · · · + (AB)nn
= (a11 b11 + a12 b21 + · · · + a1n bn1 ) + (a21 b12 + a22 b22 + · · · + a2n bn2 ) + · · ·
(an1 b1n + an2 b2n + · · · + ann bnn )
= (b11 a11 + b12 a21 + · · · + b1n an1 ) + (b21 a12 + b22 a22 + · · · + b2n an2 ) + · · ·
(bn1 a1n + bn2 a2n + · · · + bnn ann )
= tr(BA).
46. Note that the ii entry of AAT is equal to
(AAT )ii = (A)i1 (AT )1i + (A)i2 (AT )2i + · · · (A)in (AT )ni = a2i1 + a2i2 + · · · + a2in ,
so that
tr(AAT ) = (a211 + a212 + · · · + a21n ) + (a221 + a222 + · · · + a22n ) + (a2n1 + a2n2 + · · · + a2nn ).
That is, it is the sum of the squares of the entries of A.
47. Note that by Exercise 45, tr(AB) = tr(BA). By Exercise 44, then,
tr(AB
BA) = tr(AB)
Since tr(I2 ) = 2, it cannot be the case that AB
tr(BA) = tr(AB)
BA = I2 .
tr(AB) = 0.
150
3.3
CHAPTER 3. MATRICES
The Inverse of a Matrix
1. Using Theorem 3.8, note that
det
so that
As a check,


4
1
7
2


4
1
4
1
7
= ad
2
1
7
2
1
=
1

bc = 4 · 2
2
1
7
=
4

7
4 · 2 + 7 · ( 1)
=
4
1 · 2 + 2 · ( 1)
2
1
1 · 7 = 1,

2
1
7
.
4

4 · ( 7) + 7 · 4
1
=
1 · ( 7) + 2 · 4
0
0
.
1
2. Using Theorem 3.8, note that
det
so that
As a check,

"
"
4
2
2
= ad
0
4
2
2
0
# "
2
0
·
1
0
2
4
2
1
2
#
1
#
=
1
"
1
=
4
bc = 4 · 0
"
4·0
0
2
2·
2·0+0·
#
2
4
( 2) · 2 = 4,
=
1
2
1
2
"
4·
#
0
1
2
1
2
1
1
2
1
2
2·1
#
"
#
1
0
0
1

0 · ( 1) + 1 · 0
1
=
1 · ( 1) + 0 · 0
0
0
1
2·
+0·1
=
3. Using Theorem 3.8, note that
det

3
6
4
= ad
8
bc = 3 · 8
4 · 6 = 0.
Since the determinant is zero, the matrix is not invertible.
4. Using Theorem 3.8, note that
det
so that
As a check,

0
1

0
1
1
= ad
0

0
1
1
0

1
0
·
0
1
5. Using Theorem 3.8, note that
"
det
3
4
5
6
1
=
0
3
5
2
3
#
1

= ad
bc = 0 · 0

0
=1·
1
0·0+1·1
1·0+0·1
bc =
3 2
·
4 3
1 · ( 1) = 1,

1
0
=
0
1
3 5
1
· =
5 6
2
Since the determinant is zero, the matrix is not invertible.
1
0
1
= 0.
2
3.3. THE INVERSE OF A MATRIX
6. Using Theorem 3.8, note that
"
p1
2
p1
2
det
so that
"
As a check,
"
p1
2
p1
2
p1
2
p1
2
# "
p1
2
p1
2
#
p1
2
p1
2
p1
2
p1
2
·
151
p1
2
p1
2
p1
2
p1
2
#
=
#

1.5
0.5
As a check,


1.5
4.2
1.6
0.5
2.4 0.333

=

As a check,


3.55 0.25
1.408
8.52 0.50
24
p1
2
p1
2
1.5 · 2.4
1
·
1.5

=
1 1
1 1
p ·p = + =1
2 2
2
2
#
p1
2
p1
2
"
=
p1
2
p1
2
p1
2
p1
2

#
p1 · p1 +
p1 · p1
2
2
2
2
p1 · p1 + p1 · p1
2
2
2
2
1
0
#
0
1
1.5.

1.5 · ( 2.8) 4.2 · 1
1
⇡
0.5 · ( 2.8) + 2.4 · 1
0
0.25
⇡
3.55

0.704
3.55 · ( 1.408) + 0.25 · 24
=
10
8.52 · ( 1.408) + 0.50 · 24
=
"
2.8
1
0.25 · 8.52 = 1.775
0.5
8.52
#
3.6 + 2.1 =

4.2
1.6
=
1.5
0.333
1.5 · ( 1.6) 4.2 · 0.333
0.5 · ( 1.6) + 2.4 · 0.333
1
·
0.335
p1
2
p1
2
( 4.2) · 0.5 =
2.4
0.5
bc = 3.55 · 0.5
1
0.25
0.50
"
p1 · p1 +
p1 ·
2
2
2
p1 · p1 + p1 ·
2
2
2
1
4.2
2.4
2.8
=
1
3.55
8.52
=1·
bc =
8. Using Theorem 3.8, note that

3.55 0.25
det
= ad
8.52 0.50
Then
1
"
7. Using Theorem 3.8, note that

1.5
4.2
= ad
det
0.5
2.4
Then
1
1
bc = p · p
2
2
= ad

2.13 =
1.408
24
0
.
1
0.335.
0.704
10

3.55 · 0.704 + 0.25 · ( 10)
1
⇡
8.52 · 0.704 + 0.50 · ( 10)
0
0
.
1
9. Using Theorem 3.8, note that
det
Then
"
a
b
b
a
#

1
a
b
b
=a·a
a
1
·
= 2
a + b2
"
( b) · b = a2 + b2 .
a
b
# "
b
=
a
a
a2 +b2
b
a2 +b2
b
a2 +b2
a
a2 +b2
#
As a check,
"
a
b
b
a
#"
a
a2 +b2
b
a2 +b2
b
a2 +b2
a
a2 +b2
#
2
=4
a·
b·
a
a2 +b2
a
a2 +b2
b·
+a·
⇣
⇣
b
a2 +b2
b
a2 +b2
⌘
⌘
a·
b·
b
b·
a2 +b2
b
a2 +b2
+a·
"
=
a
a2 +b2
a
a2 +b2
a2 +b2
a2 +b2
ab ab
a2 +b2
3
5
ab ab
a2 +b2
a2 +b2
a2 +b2
#
"
1
=
0
#
0
.
1
152
CHAPTER 3. MATRICES
10. Using Theorem 3.8, note that
"
det
Then
"
1
a
1
c
1
b
1
d
#
1
a
1
c
1
1
b
1
d
#
=
1 1
·
a d
abcd
=
·
bc ad
"
1 1
1
· =
b c
ad
1
d
1
c
1
b
1
a
#
=
1
bc ad
=
bc
abcd
"
abc
bc ad
abd
bc ad
acd
bc ad
bcd
bc ad
#
.
As a check,
"
1
a
1
c
1
b
1
d
#"
abc
bc ad
abd
bc ad
acd
bc ad
bcd
bc ad
#
=
"
bc
bc ad
ab
bc ad
ad
bc ad
ab
bc ad
cd
bc ad
ad
bc ad
+
+
#
cd
bc ad
bc
bc ad
"
1
=
0
#
0
.
1
11. Following Example 3.25, we first invert the coefficient matrix

2 1
.
5 3
The determinant of this matrix is 2 · 3
1 · 5 = 1, so its inverse is


1
3
1
3
1
=
.
5
2
5
2
1
Thus the solution to the system is given by



1 
x
2 1
1
3
=
·
=
y
5 3
2
5
so the solution is x =
1
·
2

1
=
2

3 · ( 1) 1 · 2
=
5 · ( 1) + 2 · 2

5, y = 9.
12. Following Example 3.25, we first invert the coefficient matrix

1
1
.
2
1
The determinant of this matrix is 1 · 1
( 1) · 2 = 3, so its inverse is
"
# "
#
1
1
1 1
1
3
3
·
=
2
1
3
2 1
3
3
and thus the solution to the system is given by
" #
x
y
=
"
1
2
#
1
1
1
·
" #
1
2
=
"
1
3
2
3
1
3
1
3
# " # "
1
·
=
2
#
1
1
3 ·1+ 3 ·2
2
1
3 ·1+ 3 ·2
and the solution is x = 1, y = 0.
13. (a) Following Example 3.25, we first invert the coefficient matrix

1 2
.
2 6
The determinant of this matrix is 1 · 6 2 · 2 = 2, so its inverse is
"
# "
#
6
2
3
1
1
·
=
.
1
2
2
1
1
2
=
" #
1
0
5
,
9
3.3. THE INVERSE OF A MATRIX
153
Then the solutions for the three vectors given

x1 = A 1 b1 =

x2 = A 1 b2 =

x3 = A 1 b3 =
are
3
1

1
1
2

3
=
5

4
1
2

1
5
=
1
2
2
2


1 2
6
=
.
1
0
2
2
3
1
1
3
1
⇥
(b) To solve all three systems at once, form the augmented matrix A |
reduce to solve.


1 0
4
5
6
1 2 3
1 2
!
1
2 6 5
2 0
0 1
2
2
2
(c) The method of constructing A
1
b1
b2
⇤
b3 and row
requires 7 multiplications, while row reductions requires only 6.
14. To show the result, we must show that (cA) · 1c A
1
(cA) · A
c
1
=
✓
15. To show the result, we must show that AT (A
AT (A
1
= I. But
◆
1
· c AA
c
1
= AA
1
= I.
1 T
) = I. But
1 T
) = (A
1
A)T = I T = I,
where the first equality follows from Theorem 3.4(d).
16. First we show that In 1 = In . To see that, we need to prove that In · In = In , but that is true. Thus
In 1 = In , and it also follows that In is invertible, since we found its inverse.
17. (a) There are many possible counterexamples. For instance, let

1
2
0
1
1
AB =
2
0
1
A=
Then

Then det(AB) = 1 · ( 1)
so that
0·3 =
(AB)
A
1
B
1
Finally,
A
and (AB)
1
6= A
1
B
1
.
B
1

=

1
1
0
.
1

0
1
=
1
3
1
1
1, det A = 1 · 1
1
1
B=
0
.
1
0 · 2 = 1, and det B = 1 · ( 1)


1
1 0
1
=
=
3
1
3
1


1
1 0
1
=
=
2
1
2
1


1
1 0
1
=
=
1
1
1
1

1
2
0
1

1
1
0
=
1

0
1
0
1
0
.
1
1
1
0
,
1
0·1 =
1,
154
CHAPTER 3. MATRICES
(b) Claim that (AB)
1
1
=A
1
B
AB = ((AB)
if and only if AB = BA. First, if (AB)
1
1
= BA.
)
1
= (A
1
1
B
1
)
1
= (B
1
)
1
(A
1
)
1
=A
1
B
, then
To prove the reverse, suppose that AB = BA. Then
1
(AB)
1
= (BA)
=A
1
B
1
.
This proves the claim.
1
18. Use induction on n. For n = 1, the statement says that (A1 )
that the statement holds for n = k. Then
1
(A1 A2 · · · Ak Ak+1 )
= ((A1 A2 · · · Ak )Ak+1 )
1
= A1 1 , which is clear. Now assume
1
= Ak+1
(A1 A2 · · · Ak )
1
by Theorem 3.9(c). But then by the inductive hypothesis, this becomes
1
Ak+1
(A1 A2 · · · Ak )
1
1
1
= Ak+1
(Ak 1 · · · A2 1 A1 1 ) = Ak+1
Ak 1 · · · A2 1 A1 1 .
So by induction, the statement holds for all k
1.
19. There are many examples. For instance, let

1 0
A=
,
0 1
Then A
exist.
1
20. XA2 = A
1
= A and B
1
) (XA2 )A
21. AXB = (BA)2 ) A
X = A 1 (BA)2 B 1 .
2
A, so that A
1
A
(AXB)B
1
=A
2
A=
1
1
+B
) X(A2 A
1+( 2)
) XI = A
A)X(BB
22. By Theorem 3.9, (A 1 X) 1 = X 1 (A 1 ) 1 = X 1 A and (B
original equation becomes X 1 A = AA 1 B 2 = B 2 . Then
2
A)
1
A = B 2 ) XX
23. ABXA
1
B
1
B
1
1
1
0
.
1
A = XB 2 ) A = XB 2 ) AB
2
1
1
=A
= XB 2 B
2
1
A
1
AB)X(A
1
B
3
)X=A
1
) = A
1
(B
= X,
2
)
3
1
=A
so that
1
1
(ABXA
1
B
1
)BA = B
1
A
1
(I + A)BA = B
BA + B
1
A
B
1
A
1
(ABXA
1
B
1
)BA = (B
1
A
1
1
BA) = X,
and
B
1
A
1
1
A
1
BA + B
BA + A = (AB)
1
1
A
1
ABA = B
1
A
1
BA + A.
BA + A.
2
0
24. To get from A to B, reverse the first and third rows. Thus E = 40
1
2
0
25. To get from B to A, reverse the first and third rows. Thus E = 40
1
2
0
1
0
0
1
0
1
26. To get from A to C, add the first row to the third row. Thus E = 40
1
3
1
05.
0
3
1
05.
0
0
1
0
3
0
05 .
1
1
. Thus
B 2 , so the
X = AB
= I + A, so that
A
does not
.
(BA)2 B
But
Thus X = B
1
= O. But A + B = O, so (A + B)
)=A
(BA)2 B
1
0
1
= A
1
2

) (A
X
1
=B=
B=
1
ABA
2
.
3.3. THE INVERSE OF A MATRIX
155
2
1
27. To get from C to A, subtract the first row from the third row. Thus E = 4 0
1
28. To get from C to D, subtract twice the third row from the second row. Thus E
2
1
29. To get from D to C, add twice the third row to the second row. Thus E = 40
0
30. The row operations R2 2R1 2R3 ! R2 ,
2
3
2
1 0
0
25 satisfies EA = 4
E=4 2 1
1 0
1
R3 + R1 ! R3 will turn matrix
32
3 2
1 0
0 1
2
1
2 1
25 41
1
15 = 4
1 0
1 1
1
0
3
0
05.
1
2
1 0
= 40 1
0 0
3
0 0
1 25.
0 1
0
1
0
3
0
25.
1
A into matrix D. Thus
3
1
2
1
3
1
35 = D.
2
1
1
However, E is not an elementary matrix since it includes several row operations, not just one.
31. This matrix multiplies the first row by 3, so its inverse multiplies the first row by 13 :

3
0
0
1
1
=
1
0
.
1
3
0
32. This matrix adds twice the second row to the first row, so its inverse subtracts twice the second row
from the first row:


1
1 2
1
2
=
.
0 1
0
1
33. This matrix reverses rows 1 and 2, so its inverse does the same thing. Hence this matrix is its own
inverse.
34. This matrix adds
the second row:
1
2
times the first row to the second row, so its inverse adds

1
1
2
1
0
1
=

1
1
2
times the first row to
0
.
1
1
2
35. This matrix subtracts twice the third row from the second row, so its inverse adds twice the third row
to the second row:
2
3 1 2
3
1 0
0
1 0 0
40 1
25 = 40 1 25 .
0 0
1
0 0 1
36. This matrix reverses rows 1 and 3, so its inverse does the same thing. Hence this matrix is its own
inverse.
37. This matrix multiplies row 2 by c. Since c 6= 0, its inverse multiplies the second row by 1c :
2
1
40
0
0
c
0
38. This matrix adds c times the third row to
row from the second row:
2
1 0
40 1
0 0
3
0
05
1
1
2
1
= 40
0
0
1
c
0
3
0
05 .
1
the second row, so its inverse subtracts c times the third
3
0
c5
1
1
2
1
= 40
0
0
1
0
3
0
c5 .
1
156
CHAPTER 3. MATRICES
39. As in Example 3.29, we first row-reduce A:


1
0
1
A=
1
1
2 R2 +R!
0
0
2

1
! 0
1
2 R2
0
.
1
The first row operation adds the first row to the second row, and the second row operation multiplies
the second row by 12 . These correspond to the elementary matrices


1
0
1 0
E1 =
and E2 =
1 .
1 1
0
2
So E2 E1 A = I and thus A = E1 1 E2 1 . But E1 1 is R2 R1 ! R2 , and E2 1 is


1 0
1
0
1
1
E1 =
and E2 =
,
1 1
0
2
so that
1
A = E 1 E2
1
=

1
1
0
1

1
0
2R2 ! R2 , so
0
.
2
Taking the inverse of both sides of the equation A = E1 1 E2 1 gives
A
1
= (E1 1 E2 1 )
so that
A
1
1
= (E2 1 )
"
1
= E2 E 1 =
0
40. As in Example 3.29, we first row-reduce A:


2 4 R1 $R2 1 1
A=
! 2 4 R2
1 1
0
1
2
#"
1
1
2R1
! 0
1
= E2 E1 ,
# "
0
1
=
1
1
2
1
1

(E1 1 )
1
2

1
! 0
1
2 R2
Converting each of these steps to an elementary matrix gives



1 0
0 1
1 0
E1 =
, E2 =
, E3 =
,
1 0
2 1
0 12
0
1
2
1
1
#
.

! 1
0
R1 R2
E4 =

1
0
0
.
1
1
.
1
So E4 E3 E2 E1 A = I and thus A = E1 1 E2 1 E3 1 E4 1 . We can invert each of the elementary matrices
above by considering the row operation which reverses its action, giving




0 1
1 0
1 0
1 1
1
1
1
1
E1 =
, E2 =
, E3 =
, E4 =
,
1 0
2 1
0 2
0 1
so that
1
1
A = E1 E 2 E4
1

0
=
1
1
0

1
2
0
1

1
0
0
2

1
0
1
.
1
Taking the inverse of both sides of the equation A = E1 1 E2 1 E3 1 E4 1 gives A 1 = E4 E3 E2 E1 , so that
"
#"
#"
#"
# "
#
1
1
1 1 0
1 0 0 1
2
1
2
A = E4 E3 E 2 E1 =
=
.
1
0
1 0 12
2 1 1 0
1
2
41. Suppose AB = I and consider the equation Bx = 0. Left-multiplying by A gives ABx = A0 = 0.
But AB = I, so ABx = Ix = x. Thus x = 0. So the system represented by Bx = 0 has the unique
solution x = 0. From the equivalence of (c) and (a) in the Fundamental Theorem, we know that B is
invertible (i.e., B 1 exists and satisfies BB 1 = I = B 1 B.) So multiply both sides of the equation
AB = I on the right by B 1 , giving
ABB
This proves the theorem.
1
= IB
1
) AI = IB
1
)A=B
1
.
3.3. THE INVERSE OF A MATRIX
157
42. (a) Let A be invertible and suppose that AB = O. Left-multiply both sides of this equation by A 1 ,
giving A 1 AB = A 1 O. But A 1 A = I and A 1 O = O, so we get IB = B = O. Thus B is the
zero matrix.
(b) For example, let
A=
Then

1
AB =
2

2
4
1
2
2
,
4

2
1
B=

2
1

2
0
=
1
0
2
.
1
0
, but B 6= O.
0
1
43. (a) Since A is invertible, we can multiply both sides of BA = CA on the right by A
BAA 1 = CAA 1 , so that BI = CI and thus B = C.
(b) For example, let

1
A=
2
Then

1
BA =
1
1
1

1
2
2
,
4

1
,
1

6
3
=
6
1
0
1
1
B=
1

2
3
=
4
3
44. (a) There are many possibilities. For example:


1 0
1
A=
= I, B =
0 1
1

C=

0
,
0
1
2
3
1
, giving
0
.
1
2
= CA, but B 6= C.
4
C=

1
0
1
.
0
Then
A2 = I 2 = I = A


1 0 1
2
B =
1 0 1


1 1 1
2
C =
0 0 0

0
1
=
0
1

1
1
=
0
0
0
=B
0
1
= C,
0
so that A, B, and C are idempotent.
(b) Suppose that A and n ⇥ n matrix that is both invertible and idempotent. Then A2 = A since A is
idempotent. Multiply both sides of this equation on the right by A 1 , giving AAA 1 = AA 1 = I.
Then AAA 1 = A, so we get A = I. Thus A is the identity matrix.
45. To show that A 1 = 2I A, it suffices to show that A(2I A) = I. But A2 2A + I = O means that
I = 2A A2 = A(2I A). This equation says that A multiplied by 2I A is the identity, which is
what we wanted to prove.
46. Let A be an invertible symmetric matrix. Then AA
equation:
(AA
1 T
) = I T ) (A
1 T
) AT = I ) (A
1
= I. Take the transpose of both sides of this
1 T
) A = I ) (A
Thus (A 1 )T AA 1 = A 1 . But also clearly (A 1 )T AA
A 1 . This says that A 1 is its own transpose, so that A
1
1
1 T
) AA
1
so that A is invertible with A
1
= B(AB)
B (AB)
1
1
= A(B(AB)
1
) = I,
. But this means that
A = B(AB)
so that also B is invertible, with inverse (AB)
1
A.
= IA
= (A 1 )T I = (A
is symmetric.
47. Let A and B be square matrices and assume that AB is invertible. Then
AB(AB)
1
1
A = I,
1 T
1
=A
1
.
) , so that (A
1 T
)
=
158
CHAPTER 3. MATRICES
⇥
48. As in Example 3.30, we adjoin the identity matrix to A then row-reduce A

1
1
5
4
1
0
0
1

1
R1
! 0
R2
5
1
1
1

0
1
1
R2
! 0
5
1
1
1
⇤
⇥
I to I

! 1
0
R1 5R2
0
1
0
1
A
1
4
1
⇤
.
5
1
Thus the inverse of the matrix is

4
1
5
.
1
⇥
49. As in Example 3.30, we adjoin the identity matrix to A then row-reduce A
"
2
3
4
1
1
0
#
0
1
1
2 R1
"
! 1
3
#
0
1
1
2
2
1
0
"
1
3R1
! 0
"
R2
1
2
3
2
2
5
1
0
0
1
1
2
3
10
2
1
#
0
1
5
⇤
⇥
I to I
1
5 R2
#
!
A
1
⇤
.
···
"
! 1
0
R1 +2R2
1
10
3
10
0
1
2
5
1
5
#
Thus the inverse of the matrix is
"
1
10
3
10
2
5
1
5
#
.
⇥
50. As in Example 3.30, we adjoin the identity matrix to A then row-reduce A
"
4
2
2
0
1
0
#
0
1
"
! 1
2
1
4 R1
1
2
1
4
0
0
#
0
1
R2
"
1
2
1
2R1
! 0
1
4
1
2
1
#
0
1
⇤
⇥
I to I
"
! 1
0
R1 + 12 R2
A
0
1
1
⇤
.
#
0
1
2
1
2
1
Thus the inverse of the matrix is
"
#
0
1
2
1
2
1
.
⇥
51. As in Example 3.30, we adjoin the identity matrix to A then row-reduce A
"
1
a
a
1
1
0
#
0
1
"
1
aR1 +R2
! 0
a
a2 + 1
1
a
#
0
1
R2 · · ·
1 a2 +1 !
"
1 a
1
0 1 a2a+1
0
1
a2 +1
Thus the inverse of the matrix is
"
1
a2 +1
a
a2 +1
a
a2 +1
1
a2 +1
#
.
#
"
! 1
0
R1 aR2
⇤
⇥
I to I
0
1
1
a2 +1
a
a2 +1
A
1
⇤
.
a
a2 +1
1
a2 +1
#
.
3.3. THE INVERSE OF A MATRIX
159
52. As in Example 3.30, we adjoin the identity matrix
2
3
2
2
3
0 1 0 0
1
2
1
6
7 R1 $R2 6
61
!6
2
1 0 1 07
3
0
5
42
4
2
0
1 0 0 1
2
0
1
2
1
2
1 0
6
1
2
1
7 R2 6
1
! 40
7
7
0
4
1 0
2
1
2
1 0
6
2
1
60
1
4
7
7
7R3
0
1 4
! 0
2
R1 +2R2
2
! 1 0 0
6
60 1 0
1
4
0 0 1
4
The inverse of the matrix is
2
2
4 1
4
⇥
⇤
to A then row-reduce A I to
3
2
0 1 0
1
2
1
7 R2 2R1 6
1 0 07
!6
7
2
5
40
R3 2R1
0 0 1
4
1
! 0
3
2
1
2
1
1 0
7
6
2
2
7
6
05
1
40
7
7
R3 4R2
1
2 1
0
0
!
7
3 R +R 2
1
3
!
1
0
1
2 0
7 R2 2 R3 6
2
7
6
7
05
1 0
! 40
7
6
7
0
0 1
3
3
3
7
2
27
5
6
7
3
3
25 .
7
3
2
6
⇥
53. As in Example 3.30, we adjoin the identity matrix to A then row-reduce A
2
1
43
2
1
1
3
2
2
1
1
0
0
0
1
0
3
0
05
1
R3
2
1
! 40
2R1
! 0
R2 3R1
1
4
5
2
4
5
1
3
2
0
1
0
3
0
05
1
4R3
2
1
40
5R2
! 0
⇥
I
1
1
0
0
1
7
4
7
4
2
4
0
1
⇤
.
3
0
7
2 07
5
2 1
3
1 0
7
2
07
5
7
6
1
7
0
⇤
⇥
I to I
1
4
0
A
3
7
7
27
5
7
7
1
2
4
6
A
1
3
7
1
⇤
.
3
0 0
1 05 .
5 4
Since the matrix on the left has a zero row, it cannot be reduced to the identity matrix, so that the
given matrix is not invertible.
⇥
⇤
⇥
⇤
54. As in Example 3.30, we adjoin the identity matrix to A then row-reduce A I to I A 1 .
2
3
2
3
2
3
1 1 0 1 0 0
1
1 0
1 0 0
1 1
0 1
0 0
6
7
6
7 R2 6
7
1 6
61 0 1 0 1 07 R2 R!
6
1 1
1 1 07
1 1
1 07
4
5
40
5 ! 40 1
5
0 1 1 0 0 1
0
1 1
0 0 1
0 1
1 0
0 1
2
3
2
3
1 1
0
1
0 0
1 1
0
1
0 0
6
7
6
7
60 1
60 1
1
1
1 07
1
1
1 07
4
5 1
4
5
R3
R3 R2
1
1
1
2
1
1 1 2 !
0 0
1
! 0 0
2
2
2
2
3
2
3
1
1
1
1 1 0
1
0 0 R1 R2 1 0 0
2
2
27
6
7
!6
R2 +R3 6
1
1
17
1
1
17
60 1 0
! 40 1 0
4
2
2
25
2
2
25
1
1
1
1
1
1
0 0 1
0
0
1
2
2
2
2
2
2
The inverse of the matrix is
2
6
6
4
1
2
1
2
1
2
1
2
1
2
1
2
3
1
27
17
25
1
2
160
CHAPTER 3. MATRICES
⇥
55. As in Example 3.30, we adjoin the identity matrix to A then row-reduce A
2
a
6
41
0
0
a
1
0
0
a
1
0
0
0
1
0
3
0
7
05
1
2
! 1
1
61
a R2
! 4a
1
0
a R3
!
1
a R1
0
1
0
0
1
1
a
1
a
0
0
0
1
a
3
0
7
05
R2
1
a
0
2
1
1
6
a R1
! 40
0
R3
Thus the inverse of the matrix is
2
6
4
1
a
1
a2
1
a3
0
1
⇤
⇥
I to I
0
0
1
1
a
2
1
6
40
1
a R2
! 0
1
a
1
a2
0
0
0
0
1
0
0
0
1
1
a
A
1
3
0
7
05
⇤
.
1
a
1
a
1
a2
1
a3
3
0
7
05 .
0
1
a
1
a2
1
a
3
0
7
05 .
0
1
a
1
a2
1
a
Note that these calculations assume a 6= 0. But if a = 0, then the original matrix is not invertible since
the first row is zero.
⇥
⇤
⇥
⇤
56. As in Example 3.30, we adjoin the identity matrix to A then row-reduce A I to I A 1 . Note
that if a = 0, then the first row is zero, so the matrix is clearly not invertible. So assume a 6= 0. Then
2
3
2
3
0 a 0 1 0 0
0 a 0
1 0 0
6
7
6
7
0 1 05 .
4 b 0 c 0 1 05 .
4b 0 c
d
R3 a R1
d
0 d 0 0 0 1
0 1
! 0 0 0
a
Since the matrix on the left has a zero row, it cannot be reduced to the identity matrix, so the original
matrix is not invertible.
⇥
⇤
⇥
⇤
57. As in Example 3.30, we adjoin the identity matrix to A then row-reduce A I to I A 1 .
2
3
2
3
0
1 1
0 1 0 0 0
1 0 0 0
11
2
5
4
6
62
1 0
2 0 1 0 07
4
1
2
27
6
7 ! 60 1 0 0
7.
41
5
4
1 3
0 0 0 1 0
0 0 1 0
5
1
2
25
0
1 1
1 0 0 0 1
0 0 0 1
9
2
4
3
Thus the inverse of the matrix is
2
11
6 4
6
4 5
9
2
1
1
2
3
4
27
7.
25
3
5
2
2
4
⇥
58. As in Example 3.30, we adjoin the identity matrix to A then row-reduce A
2 p
2
6 4p2
6
6
4 0
0
p
0 2 2 0
p
2
0
0
0
1
0
0
3
1
Thus the inverse of the matrix is
1
0
0
0
0
1
0
0
2
0
0
1
0
p
2
2
6 p
62 2
6
4 0
0
3
2
0
1
6
07
7
60
7!6
40
05
1
0
0
p
2
2
0
0
0
1
0
0
3
2 0
7
8 07
7.
1 05
3 1
0
0
1
0
0
0
0
1
p
2
2
p
2 2
0
0
0
p
2
2
0
0
⇤
⇥
I to I
A
3
2 0
7
8 07
7.
1 05
3 1
1
⇤
.
3.3. THE INVERSE OF A MATRIX
59. As in Example 3.30,
2
1
60
6
6
40
a
161
we adjoin the identity matrix to A
3
2
0 0 0 1 0 0 0
1 0
7
6
1 0 0 0 1 0 07
60 1
7!6
5
40 0
0 1 0 0 0 1 0
b c d 0 0 0 1
0 0
Thus the inverse of the matrix is
2
1
6 0
6
4 0
0
1
0
a
d
⇥
then row-reduce A
0
0
1
0
1
0
0
a
d
0
1
0
b
d
c
d
3
0
07
7.
05
0
0
1
b
d
0
0
0
1
c
d
⇤
⇥
I to I
3
0
0
0
07
7
7.
1
05
A
1
⇤
.
1
d
1
d
⇥
⇤
⇥
60. As in Example 3.30, we adjoin the identity matrix to A then row-reduce A I over Z2 to I



2
0 1 1 0 R1 +R!
1 0 1 1
1 0 1 1
.
1
1 1 0 1
1 1 0 1 R2 +R!
0 1 1 0
Thus the inverse of the matrix is

1
1
A
1
A
1
⇤
.
1
.
0
⇥
⇤
⇥
61. As in Example 3.30, we adjoin the identity matrix to A then row-reduce A I over Z5 to I



1
4 2 1 0 4R!
1 3 4 0
1 3 4 0
.
1
3 4 0 1
3 4 0 1 R2 +2R!
0 0 3 1
⇤
.
Since the left-hand matrix has a zero row, it cannot be reduced to the identity matrix, so the original
matrix is not invertible.
⇥
⇤
⇥
⇤
62. As in Example 3.30, we adjoin the identity matrix to A then row-reduce A I over Z3 to I A 1 .
2
2
41
0
1
1
2
0
2
1
1
0
0
0
1
0
3 2R1 2
0
! 1
1 40
05 R2 +R!
1
0
2
1
40
R3 +R2
! 0
2
1
R2 +2R3 4
! 0
0
Thus the inverse of the matrix is
63. As in Example 3.30, we adjoin
2
1
41
3
3
2
0
1 2
2 4
05 2R!
0 1
1
0 2
3
2
2 0 2 0 0
1 2
40 1
1 1 2 2 05
3
0 2 2 2 1 2R!
0 0
3 R1 +R2 2
2 0 2 0 0
! 1
40
1 0 1 1 15
0 1 1 1 2
0
2
2
2
0
2
1
2
0
41
1
2
1
0
1
1
1
0
1
0
3
1
15 .
2
the identity matrix to A
3
2
5 0 1 0 0
1
2 4 0 1 05 ! 40
6 1 0 0 1
0
Thus the inverse of the matrix is
2
4
45
0
6
3
6
0
1
1
2
2
0
0
2
0
0
1
1
2
2
1
0
2
1
0
1
0
0
0
1
0
1
1
⇥
then row-reduce A
3
0 0 4 6 4
1 0 5 3 25 .
0 1 0 6 5
3
4
25 .
5
3
0
05
1
3
0
05
2
1
1
1
3
1
15 .
2
⇤
⇥
I over Z7 to I
A
1
⇤
.
162
CHAPTER 3. MATRICES
64. Using block multiplication gives

A
O
B
D

A 1
O
A
1
D
BD
1
1
=
=
=

AA
OA

I
O

AA
O
1
1
+ BO
+ DO
1
AA
OA
1
1
BD
BD
BD 1 + BD
DD 1
1
1
+ BD
+ DD
1
1
1
O
.
I
65. Using block multiplication gives

O
C

(BC) 1
(BC) 1 B
1
C(BC)
I C(BC) 1 B

O(BC) 1 + BC(BC) 1 O(BC) 1 B + B(I C(BC) 1 B)
=
C(BC) 1 + IC(BC) 1 C(BC) 1 B + I(I C(BC) 1 B)

BC(BC) 1
BI BC(BC) 1 B
=
1
1
C(BC) + C(BC)
C(BC) 1 B + I C(BC) 1 B

I O
=
.
O I
B
I
66. Using block multiplication gives

I
C
B
I

(I BC) 1
(I BC) 1 B
1
C(I BC)
I + C(I BC) 1 B

I(I BC) 1 BC(I BC) 1
I(I BC) 1 B + B(I + C(I BC)
=
1
1
C(I BC)
I(C(I BC) )
C(I BC) 1 B + I(I + C(I BC)

(I BC)(I BC) 1
(BC I)(I BC) 1 B + B
=
1
1
C(I BC)
C(I BC)
C(I BC) 1 B + I + C(I BC) 1 B

I
IB + B
=
O
I

I O
=
.
O I
1
1
B)
B)
67. Using block multiplication gives

O
C
B
D

(BD 1 C) 1
(BD 1 C) 1 BD 1
1
1
1
D C(BD C)
D
D 1 C(BD 1 C) 1 BD 1
2
BD 1 C(BD 1 C) 1
D 1 D 1 CBD 1 C) 1 BD 1
4
C(BD 1 C) 1 BD 1
=
C(BD 1 C) 1 + C(BD 1 C) 1
1
+D(D
D 1 C(BD 1 C) 1 BD

I
D 1 D 1 CC 1 DB 1 BD 1
=
1
O C(BD C) 1 BD 1 + I DD 1 C(BD 1 C) 1 BD 1

I
D 1 D 1
=
1
1
O C(BD C) BD 1 + I C(BD 1 C) 1 BD 1

I O
=
.
O I
1
1
)
3
5
3.3. THE INVERSE OF A MATRIX
163
68. Using block multiplication gives


⇥
⇤
A B P Q
= AP + BR AQ + BS CP + DR DQ + DS
C D R S

AP + B( D 1 CP ) A( P BD 1 + B(D 1 + D 1 CP BD 1 )
=
CP + D( D 1 CP ) C( P BD 1 ) + D(D 1 + D 1 CP BD 1 )

(A BD 1 C)P
(A BD 1 C)P BD 1 + BD 1
=
CP CP
CP BD 1 + I + CP BD 1
 1
P P
P 1 P BD 1 + P BD 1
=
O
I


I
BD 1 + BD 1
I O
=
=
.
O
I
O I
69. Partition the matrix
2
1
60
A=6
42
1
0
1
3
2
3
0

07
7= I
05
C
1
0
0
1
0
B
, where B = O.
I
Then using Exercise 66, (I BC) 1 = (I OC) 1 = I 1 = I, so that



(I BC) 1
(I BC) 1 B
I
IB
I
1
A =
=
=
C(I BC) 1 I + C(I BC) 1 B
CI I + CIB
C
Substituting back in for C gives
A
70. Partition the matrix
1
6 0
6
=4
2
1
2 p
2
p
6 4 2
M =6
4 0
0
and use Exercise 64. Then
A
1
2
A
1
D
1
1
 p
2
p
=
4 2

1 0
=
3 1
"p
2
BD =
2
p
2 2
0
1
3
2
p
3

p0 2 2 0
2
0
07
7= A
O
0
1
05
0
3
1
M
71. Partition the matrix
1
=
A 1
O
A
2
1
D
2
0
60
A=6
40
1
B
D
"p
#
1
2
0
0
2
p
p
= p
2
2 2 22

1
1 0
=
3 1
# p


0
1 0
2
2 2 0
p
=
2
3 1
8
0
0
Then
"
3
0
07
7.
05
1
0
0
1
0
BD
1
1
#
2
p
2
2
6 p
62 2
=6
4 0
0
3
0 1 1

0 1 07
7= O
1 1 05
C
1 0 1
0
p
2
2
0
0
B
I
0
.
0
3
2 0
7
8 07
7.
1 05
3 1
O
.
I
164
CHAPTER 3. MATRICES
and use Exercise 65. Then
1
(BC)
1
(BC
)B =
1
C(BC)
I
1
C(BC)
✓
=
=


1
0
0
1
0
1
1
1

B=I
Thus
1
A
=

(BC)
C(BC)
72. Partition the matrix
1
⇥
1
1
(BD
D
1
C)
1
1
BD
1
1
1
C(BD
1
C(BD
C)
1
⇤
1 ,
C)
(BD
1
1
6 0
(BC) 1 B
=6
4 0
I C(BC) 1 B
1
0
A=4 1
1
and use Exercise 67. Then
D
2
1
B= 1
1
0
1

◆ 1

1 0
1
1
=
0 1
1
0


1 1
1 1
=
1 0
1 0


1
0
0
1
=
0
1
1
1


1 1 1
0 0
=
.
1 1 0
0 0
2
where
D
1
1
C)
BD
1
3

1
O
5
1 =
C
2
1
3
5
C=

1
,
1
0
=
1
0
1
1
1
A
3.4
1
=
1
D=

3
5
1
,
2
"
# " #1 1
i 3 1 1 1
A =
=
1
5 2
1
"
#
i 3 1 1 h
i
1h
2
=
= 35
1 1
5
5
5 2
"
# 1" #✓
"
#
◆
3
3 1
1
1
5
=
=
8
5
5 2
1
5
"
# 1 " #
"
#
i 3 1
3 h
3 1
5
=
1 1
8
5 2
5 2
5
(BD C)
D C(BD 1 C)
1
1
D
1
1
(BD C) BD
D 1 C(BD 1 C)
0
1
3
1
07
7.
05
0
0
h
@ 1
1
1
1
1
0
0
1
0
B
,
D
Thus
"

1
1
BD
1
#
2
6
=4
h
5
=
"
1
1
5
3
5
8
5
i
1
=
1
5
1
5
3
5
1
5
1
5
h i
1
5
1
5
1
5
#
.
3
2
5
17
.
55
1
5
The LU Factorization
1. As in the method outlined after Theorem 3.15, we want to solve Ax = (LU )x = b. Let y = U x and
first solve Ly = b for y by forward substitution. Then solve U x = y by backward substitution. We
have




2 1
1 0
2 1
5
A=
=
= LU, and b =
.
2 5
1 1
0 6
1
Then
Ly = b )

1
1
0
1


y1
5
=
)
y2
1

y1 = 5
y1
= 5
5
)
)y=
.
y1 + y2 = 1
6
y 2 = y1 + 1
3.4. THE LU FACTORIZATION
165
Likewise,
Ux = y )

2
0
1
6


x1
5
=
)
x2
6
2x1 + x2 = 5
)
6x2 = 6
x2 = 1
2x1 = 5
x2
)

x1
=
x2

2
.
1
2. As in the method outlined after Theorem 3.15, we want to solve Ax = (LU )x = b. Let y = U x and
first solve Ly = b for y by forward substitution. Then solve U x = y by backward substitution. We
have
"
# "
#"
#
" #
4
2
1 0 4
2
0
A=
= 1
= LU, and b =
.
2
3
1 0
4
8
2
Then
Ly = b )

1
1
2
0
1



y
= 0
y1
0
0
=
)1 1
)y=
.
y2
8
8
y
+
y
=
8
1
2
2
Likewise,

4
Ux = y )
0
2
4


x1
0
4x
=
) 1
x2
8


x2 = 2
2x2 = 0
x1
1
)
)
=
.
4x2 = 8
x
2
4x1 = 2x2
2
3. As in the method outlined after Theorem 3.15, we want to solve Ax = (LU )x = b. Let y = U x and
first solve Ly = b for y by forward substitution. Then solve U x = y by backward substitution. We
have
2
3 2
32
3
2 3
2
1
2
1
0 0
2 1
2
3
6
7 6
76
7
6 7
A=4 2
3
45 = 4 1
1 05 4 0 4
65 = LU, and b = 4 15 .
5
7
4
3
0
2
1 0 0
0
4
2
Then
2
1
Ly = b ) 4 1
2
0
1
5
4
32 3 2 3
0
y1
y1
3
05 4y2 5 = 4 15 ) y1 +
y3
0
1
2y1
=
y2
=
5
y
+
y
=
2
3
4
3
1
0
2 3 2 3
3
y1
) y2 = y 1 + 1
) 4 y 2 5 = 4 25 .
7
5
y3
2
y3 = 2y1 + y2
4
y1 =
3
Likewise,
2
2
6
U x = y ) 40
0
1
4
0
32 3 2 3
2 x1
3
2x1 + x2
76 7 6 7
65 4x2 5 = 4 25 )
4x2
7
7
x3
2
2
2x3 =
6x3 =
7
2 x3 =
x3 =
3
2
7
2
1
) 4x2 = 6x3 2
2x1 = x2 + 2x3
2
3 2 3
3
x1
2
6 7 6 7
) 4 x 2 5 = 4 25 .
x3
1
3
4. As in the method outlined after Theorem 3.15, we want to solve Ax = (LU )x = b. Let y = U x and
first solve Ly = b for y by forward substitution. Then solve U x = y by backward substitution. We
have
2
3 2
32
3
2 3
2
4 0
1 0 0
2
4 0
2
6
7 6 3
76
7
6 7
A=4 3
1 45 = 4 2 1 05 4 0
5 45 = LU, and b = 4 05 .
1
1
2 2
0 1
0
0 2
5
2
166
CHAPTER 3. MATRICES
Then
2
6
Ly = b ) 4
1
3
2
1
2
32 3 2 3
0 0
y1
2
76 7 6 7
1 05 4 y 2 5 = 4 05 )
0 1
y3
5
y1
3
2 y1
1
2 y1
+ y2
=
=
+ y3 =
2
0
5
y1 = 2
) y2 =
y3 =
3
y1
2
1
y1
2
5
2 3 2 3
y1
2
6 7 6 7
) 4 y 2 5 = 4 35 .
y3
4
Likewise,
2
2
U x = y ) 40
0
32 3 2 3
4 0 x1
2
2x1
5 45 4x2 5 = 4 35 )
0 2 x3
4
4x2
=
5x2 + 4x3 =
2x3 =
2
3
4
3 2 3
x1
3
6 7 6 7
) 5x2 = 4x3 3 ) 4x2 5 = 4 15 .
x3
2
2x1 = 4x2 + 2
x3 =
2
2
5. As in the method outlined after Theorem 3.15, we want to solve Ax = (LU )x = b. Let y = U x and
first solve Ly = b for y by forward substitution. Then solve U x = y by backward substitution. We
have
2
3 2
32
3
2 3
2
1 0
0
1
0 0 0 2
1 0
0
1
66
7 63
7 60
7
627
4
5
3
1
0
0
1
5
3
7=6
76
7 = LU, and b = 6 7 .
A=6
48
425
4 1
05 44
0 1 05 40
0 1
05
4
1 0
7
2
1 5 1
0
0 0
4
1
Then
2
1
63
Ly = b ) 6
44
2
32 3 2 3
0 0 0 y1
1
y1
6y2 7 627
1 0 07
3y
7 6 7 = 6 7 ) 1 + y2
0 1 05 4y3 5 425
4y1
+ y3
1 5 1 y4
1
2y1
y2 + 5y3 + y4
=
=
=
=
y1 = 1
)
y2 =
3y1 + 2
y3 =
4y1 + 2
y4 =
2y1 + y2
1
2
2
1
2 3 2
y1
6 y2 7 6
7 6
)6
4 y3 5 = 4
y4
5y3 + 1
3
1
17
7.
25
8
Likewise,
2
2
60
Ux = y ) 6
40
0
1 0
1 5
0 1
0 0
32 3 2
0 x1
6 7 6
37
7 6x2 7 = 6
05 4x3 5 4
4 x4
3
1
2x1
17
7)
25
8
x2
x2 + 5x3
x3
=
3x4 =
=
4x4 =
1
1
2
8
3 2
3
x1
7
6x2 7 6 157
x3 = 2
7 6
7
)
)6
4x3 5 = 4 25 .
x2 = 5x3 3x4 + 1
x4
2
2x1 = x2 + 1
x4 = 2
2
6. As in the method outlined after Theorem 3.15, we want to solve Ax = (LU )x = b. Let y = U x and
first solve Ly = b for y by forward substitution. Then solve U x = y by backward substitution. We
3.4. THE LU FACTORIZATION
have
2
1
A=4 2
3
4
5
6
3
1
3
Then
2
1
6 2
Ly = b ) 6
4 3
5
167
2
3
1
0
6 2
25 = 6
4 3
4
5
0
1
2
4
32 3 2
0 0 y1
6 7 6
0 07
7 6 y2 7 = 6
5
1 0 4 y3 5 4
2 1 y4
0
1
2
4
32
0 1
6
07
7 60
5
0 40
1
0
0
0
1
2
3
1
37
7)
15
0
y1
2y1 + y2
3y1
2y2 + y3
5y1 + 4y2
2y3 + y4
1
60
Ux = y ) 6
40
0
4
3
0
0
32 3 2
3 0 x1
6 7 6
5 27
7 6x2 7 = 6
2 05 4x3 5 4
0 1 x4
x4 =
)
The multiplier was
3, so
L=
L=

9. We use the multiplier method from Example 3.35:
2
3
2
1 2 3
1
2
R 4R1
40
3
A = 44 5 65 2 !
1
8 7 9 R3 8R!
0
9
1
3
2
The multipliers were `21 = 4, `31 = 8, and `32 = 3,
2
1
L = 44
8
1
3
1
0
3 2
x1
6x 7 6
6 27 6
)6 7=6
4x3 5 4
1
5x3
2x4
x1 =
4x2
3x3 + 1

2
x4
2
= U.
5
0
.
1
8. We use the multiplier method from Example 3.35:


2
4
2
3
A=
1
3
1 R3 2 R!
0
The multiplier was 32 , so
3
1
1
6
3
3x2 =
1
! 0
1
3
=
=
=
=
x3 = 3
( 3R1 )

3
1
37
7.
15
0
2 3 2
y1
6 y2 7 6
y2 = 2y1 3
7 6
)
)6
4 y3 5 = 4
y3 = 3y1 + 2y2 1
y4
y4 = 5y1 4y2 + 2y3
3
1
x1 + 4x2 + 3x3
17
3x2 + 5x3 + 2x4
7)
65
2x3
3
x4
7. We use the multiplier method from Example 3.35:

1
2
A=
3
1 R3
=
=
=
=
y1 = 1
Likewise,
2
3
2
3 0
6
5 27
7 = LU, and b = 6
5
4
2 0
0 1
4
3
0
0
4
= U.
7
0
.
1
3
3
65
15
R3
so we get
3
0 0
1 05
3 1
2
1
40
3R2
! 0
2
3
0
3
3
65 = U.
3
3
1
17
7.
65
3
3
16
3
10 7
3 7.
7
35
3
168
CHAPTER 3. MATRICES
10. We use the multiplier
2
2
A = 44
3
method from Example 3.35:
3
2
2
2
2
1 R 2R
2
1
! 40
4
0
45
3
1
4
4 R3 2 R!
0
1
The multipliers were `21 = 2, `31 = 32 , and `32 =
2
1
L = 42
3
1
65
11
2
1
4,
3
2
11. We use the multiplier method from Example 3.35:
2
1
6 2
A=6
4 0
1
2
6
6
2
3
3
6
9
3
1
07
7
75
0
R3
R4 (
2
1
! 60
6
0R1
! 40
1R1 )
0
!
R2 2R1
2
2
6
0
R3
2
2
40
( 14 R2 ) 0
!
3
1
65 = U.
7
2
4
0
so we get
3
0 0
1 05
1
1
4
3
1
27
7
75
1
3
3
6
6
R3
R4
2
1
60
3R2 6
! 40
0R2
! 0
R4 (
2
2
0
0
3
1
27
7
15
1
3
3
3
6
2
1
60
6
40
2R3 )
! 0
The multipliers were `21 = 2, `31 = 0, `41 = 1, `32 = 3, `42 = 0, and `43 =
2
3
1 0
0 0
6 2 1
0 07
7.
L=6
4 0 3
1 05
1 0
2 1
2
2
0
0
3
1
27
7 = U.
15
1
3
3
3
0
2, so
12. We use the multiplier method from Example 3.35:
2
2
6 2
6
A=4
4
6
2
4
4
9
3
2 1
1 27
7
7 35
5 8
R3
R4
2
2
!6
0
6
2R1 4
! 0
3R1
0
!
R2 ( R1 )
2
6
0
3
3
2 1
1 37
7
3 15
1 5
R3
R4
2
2
60
6
0R2
! 40
1
2 R2
! 0
2
6
0
0
R4
The multipliers were `21 =
2
1
3
3
2
3
1
37
7
15
7
2
2
2
60
6
40
1
( 2 R3 ) 0
!
1, `31 = 2, `41 = 3, `32 = 0, `42 = 12 , and `43 =
2
3
1 0
0 0
6 1 1
0 07
7.
L=6
4 2 0
1 05
1
3 12
1
2
1
2,
2
6
0
0
2
1
3
0
3
1
37
7 = U.
15
4
so
13. We can use the multiplier method as before to make the matrix “upper triangular”, i.e., in row echelon
form. In this case, however, the matrix is already in row echelon form, so it is equal to U , and L = I3 :
2
3 2
32
3
1 0 1
2
1 0 0
1 0 1
2
40 3 3
15 = 40 1 05 40 3 3
15 .
0 0 0
5
0 0 1
0 0 0
5
3.4. THE LU FACTORIZATION
169
14. We can use the multiplier method as before to make the matrix “upper triangular”, i.e., in row echelon
form.
2
1
6 2
6
4 1
0
2
7
1
3
0
3
3
3
3
1
27
7
25
0
1
8
5
6
R3
R4
2
1
!6
60
1R1
! 40
0R1
0
!
R2 ( 2R1 )
2
3
1
3
0
3
3
3
2
15. In Exercise 1, we have
A=
L
"
1 1
=
1 1

2
2
1
= LU =
5
# "
0
1
=
1
1
#
0
,
1
U
so that
A
1
2
R3
R4 (
1
60
1
6
3 R2
! 40
1R2 )
0
!
= (LU )
1
=U
1
L
1
2 0
3 3
0 2
0 0
1
6
4
0
3
1
07
7 = U.
15
0
1, so
3
0 0 0
1 0 07
7.
1
1 05
3
1 0 1
1
6 2
L=6
4 1
0
1
3
1
07
7
15
0
2, `31 = 1, `41 = 0, `32 = 13 , and `42 =
The multipliers were `21 =
Then
1
6
6
6
"
=
1

1
1
"
1 6
12 0
=
1
2
0

0
1
1
12
1
6
#"
1
1
2
0
1
.
6
# "
1
1
2
=
2
0
# "
0
=
1
5
12
1
6
1
12
1
6
1
12
1
6
#
#
,
.
16. In Exercise 4, we have
2
3
2
4 0
1
7
6
1 45 = LU = 4 32
1
2 2
2
2
6
A=4 3
1
To compute L 1 , we adjoin the identity matrix
2
3
2
1 0 0 1 0 0
1 0
6 3
7
6
!
1
0
0
1
0
4 2
5
40 1
1
0 1 0 0 1
0 0
2
To compute U 1 ,
2
2
6
40
0
Then
A
1
32
0 0
2
76
1 05 40
0 1 0
to L and row-reduce:
3
0
1 0 0
7
3
0
1 05 ) L
2
1
1
0 1
2
3
4 0
7
5 45 .
0 2
1
2
6
=4
3
2
1
2
we adjoin the identity matrix to U and row-reduce:
3
2
3
2
4
1
4 0 1 0 0
1 0 0 12 25
5
2
7
6
7
6
1
2
) U = 40
5 4 0 1 05 ! 40 1 0 0 15
55
1
0 2 0 0 1
0 0 1 0 0
0
2
= (LU )
1
=U
1
L
1
2
1
2
6
= 40
0
2
5
1
5
0
32
4
5
27 6
55 4
1
2
1
3
2
1
2
3 2
0 0
7 6
1 05 = 4
0 1
3
0 0
7
1 05 .
0 1
1
1
2
1
2
1
4
2
5
1
5
0
2
5
1
5
0
3
4
5
27
.
55
1
2
3
4
5
27
.
55
1
2
17. Using the LU decomposition from Exercise 1, we compute A 1 one column at a time using the method
outlined in the text. That is, we first solve Lyi = ei , and then solve U xi = yi . Column 1 of A 1 is
170
CHAPTER 3. MATRICES
found as follows:
"
#" # " #
0 y11
1
y11
=
Ly1 = e1 )
=
)
1 y21
0
y11 + y21 =
"
#" # " #
2 1 x11
1
2x11 + x21
U x1 = y1 )
=
)
0 6 x21
1
6x21
1
1
Use the same procedure to find
"
1
Ly2 = e2 )
1
"
2
U x2 = y2 )
0
" #
1
1
) y1 =
0
1
h i
= 1
) x =
1
= 1
the second column:
#" # " #
0 y12
0
y12
=
=
)
1 y22
1
y12 + y22 =
#" # " #
1 x12
0
2x12 + x22
=
)
6 x22
1
6x21
So
A
1
h
= x1
This matches what we found in Exercise 15.
i
x2 =
"
5
12
1
6
1
12
1
6
#
"
" #
0
0
) y2 =
1
1
h i
= 0
) x =
1
= 1
"
5
12
1
6
1
12
1
6
#
#
.
.
.
18. Using the LU decomposition from Exercise 4, we compute A 1 one column at a time using the method
outlined in the text. That is, we first solve Lyi = ei , and then solve U xi = yi . Column 1 of A 1 is
found as follows:
2 3 2 3
2
32 3 2 3
1 0 0
y11
1
y11
= 1
y11
1
6 3
76 7 6 7
6 7 6 37
3
Ly1 = e1 = 4 2
1 05 4y21 5 = 405 ) 2 y11 + y21
= 0 ) 4y21 5 = 4 2 5
1
1
1
0
1
y
0
y
+
y
y31
31 = 0
31
2
2 11
2
2
32 3 2 3
2 3 2 3
1
2
4 0 x11
1
2x11
4x21
= 1
x11
2
6
7 6 7 6 37
6
7
6
7
U x 1 = y1 = 40
5 45 4x21 5 = 4 2 5 )
5x21 + 4x31 = 32 ) 4x21 5 = 4 12 5
1
1
0
0 2 x31
2x31 = 12
x31
2
4
Repeat this procedure to find the second column:
2
32 3 2 3
2 3 2 3
1 0 0
y12
0
y12
= 0
y12
0
6 3
76 7 6 7
6 7 6 7
3
Ly2 = e2 = 4 2
1 05 4y22 5 = 415 ) 2 y12 + y22
= 1 ) 4y22 5 = 415
1
1
0
1
y
0
y
+
y
y32
0
32
32 = 0
2
2 12
2
32 3 2 3
2 3 2 3
2
2
4 0 x12
0
2x12
4x22
= 0
x12
5
6
76 7 6 7
6 7 617
U x 2 = y2 = 40
5 45 4x22 5 = 415 )
5x22 + 4x32 = 1 ) 4x22 5 = 4 5 5
0
0 2 x32
0
2x32 = 0
x32
0
Do this again to find the third column:
2
32 3 2 3
2 3 2 3
1 0 0
y13
0
y13
= 0
y13
0
76 7 6 7
6
6 7 6 7
Ly3 = e3 = 4 32
1 05 4y23 5 = 405 ) 32 y13 + y23
= 0 ) 4y23 5 = 405
1
1
0 1 y33
1
+ y33 = 1
y33
1
2
2 y13
2
32 3 2 3
2 3 2 3
4
2
4 0
x13
0
2x13
4x23
= 0
x13
5
6
76 7 6 7
6 7 6 27
U x 3 = y 3 = 40
5 45 4x23 5 = 405 )
5x23 + 4x33 = 0 ) 4x23 5 = 4 5 5
1
0
0 2 x33
1
2x33 = 1
x33
2
2
3
1
2
4
h
i
2
5
5
6 1 1
1
27
Thus A = x1 x2 x3 = 4 2 5
; this matches what we found in Exercise 16.
55
1
1
0
4
2
3.4. THE LU FACTORIZATION
171
19. This matrix results from the identity matrix by first interchanging rows 1 and 2, and then interchanging
the (new) row 1 with row 3, so it is
2
32
3 2
3
0 0 1 0 1 0
0 0 1
E13 E12 = 40 1 05 41 0 05 = 41 0 05 .
1 0 0 0 0 1
0 1 0
Other products of elementary matrices are possible.
20. This matrix results from the identity matrix by first interchanging rows 1 and 4, and then interchanging
rows 2 and 3, so it is
2
32
3 2
3
0 0 0 1
1 0 0 0
0 0 0 1
60 1 0 07 60 0 1 07 60 0 1 07
76
7 6
7
E14 E23 = 6
40 0 1 05 40 1 0 05 = 40 1 0 05
1 0 0 0
0 0 0 1
1 0 0 0
Other products of elementary matrices are possible; for example, also P = E13 E23 E34 .
21. To get from
2
0
60
6
41
0
the given matrix to the
3
2
1 0 0
1
60
0 0 17
R
$R
3 6
7 1 !
40
0 0 05
0 1 0
0
identity matrix, perform the
3
2
0 0 0
1 0 0
60 1 0
0 0 17
R
$R
3 6
7 2 !
40 0 0
1 0 05
0 1 0
0 0 1
following exchanges:
3
2
0
1 0 0
60 1 0
07
R
$R
4 6
7 3 !
40 0 1
15
0
0 0 0
3
0
07
7
05
1
Thus we can construct the given matrix P by multiplying these elementary matrices together in the
reverse order:
P = E13 E23 E34 .
Other combinations are possible.
22. To get from the given matrix to the identity matrix, perform the following exchanges:
2
0
61
6
60
6
40
0
0
0
0
0
1
1
0
0
0
0
0
0
1
0
0
3
0
07
7
07
7
15
0
2
1
60
6
R1 $R2 6
! 60
40
0
0
0
0
0
1
0
1
0
0
0
0
0
1
0
0
3
0
07
7
07
7
15
0
2
1
60
6
R2 $R5 6
! 60
40
0
2
1
60
6
60
6
40
0
0
1
0
0
0
0
0
0
0
1
0
0
1
0
0
0
1
0
0
0
0
0
1
0
0
0
0
0
0
1
3
0
07
7 R $R
3
5
07
!
7
5
1
0
3
2
0
1
60
07
7 R $R 6
4
5 6
07
!
7
60
5
40
1
0
0
0
1
0
0
0
0
0
1
0
0
0
0
0
1
0
3
0
07
7
07
7
05
1
Thus we can construct the given matrix P by multiplying these elementary matrices together in the
reverse order:
P = E12 E25 E35 E45 .
Other combinations are possible.
23. To find a P T LU factorization of A, we begin by permuting the rows of A to get it
will result in a small number of operations to reduce to row echelon form. Since
2
3
2
32
3 2
0 1 4
0 1 0
0 1 4
1 2
A = 4 1 2 15 , we have P A = 41 0 05 4 1 2 15 = 4 0 1
1 3 3
0 0 1
1 3 3
1 3
into a form that
3
1
45 .
3
172
CHAPTER 3. MATRICES
Now follow the multiplier method from Example 3.35 to find the LU factorization of this matrix:
2
1
PA = 4 0
1
Then `21 = 0, `31 =
2
1
3
3
1
45
3
R3 (
2
1
4 0
R1 )
0
!
R3
2
1
4 0
5R2
0
!
3
1
45 = U.
16
2
1
0
1, and `32 = 5, so that
2
1
L=4 0
1
Thus
3
1
45
4
2
1
5
2
0
A = P T LU = 41
0
1
0
0
3
0
05
1
0
1
5
32
0
1
05 4 0
1
1
0
1
5
32
0
1
05 4 0
1
0
2
1
0
3
1
45 .
16
24. To find a P T LU factorization of A, we begin by permuting the rows of A to get
will result in a small number of operations to reduce to row echelon form. Since
2
3
2
32
3 2
0 0
1 2
0 0 0 1
0 0
1 2
6 1 1
6
76
6
3 27
3 27
7 , we have P A = 60 0 1 07 6 1 1
7=6
A=6
4 0 2
4 1 0 0 05 4 0 2
1 15
1 15 4
1 1
1 0
0 1 0 0
1 1
1 0
it into a form that
1
0
0
1
3
1 0
1 17
7.
1 25
3 2
1
2
0
1
Now follow the multiplier method from Example 3.35 to find the LU factorization of this matrix:
2
1
6 0
PA = 6
4 0
1
1
2
0
1
3
1 0
1 17
7
1 25
3 2
R4 (
2
1
60
6
40
R1 )
! 0
1
2
0
2
3
1 0
1 17
7
1 25
2 2
R4
2
1
60
6
40
R2
! 0
1
2
0
0
R4
Then `41 =
Thus
1, `42 = 1, `42 = 1, and the other `ij
2
1
6 0
L=6
4 0
1
2
0
60
T
A = P LU = 6
40
1
0
0
1
0
1
0
0
0
32
0
6
17
76
5
0 4
0
are zero, so that
3
0 0 0
1 0 07
7.
0 1 05
1 1 1
1
0
0
1
0
1
0
1
0
0
1
1
32
0
1
60
07
76
05 4 0
1
0
1
2
0
0
3
1 0
1 17
7
1 25
1 1
2
1
60
6
40
R3
! 0
1
1
1
0
1
2
0
0
3
0
17
7 = U.
25
1
1
1
1
0
3
0
17
7.
25
1
25. To find a P T LU factorization of A, we begin by permuting the rows of A to get
will result in a small number of operations to reduce to row echelon form. Since
2
3
2
32
3 2
0
1
1 3
0 1 0 0
0
1
1 3
6 1
6
76
6
1
1 27
1
1 27
7 , we have P A = 61 0 0 07 6 1
7=6
A=6
4 0
5
4
5
4
5
4
1
1 1
0 0 0 1
0
1
1 1
0
0
1 1
0 0 1 0
0
0
1 1
it into a form that
1
0
0
0
1
1
0
1
3
1 2
1 37
7.
1 15
1 1
3.4. THE LU FACTORIZATION
173
Now follow the multiplier method
2
1
6 0
6
PA = 4
0
0
Then `42 =
from Example 3.35 to find the
3
2
1
1 2
1
6 0
1
1 37
7
6
0
1 15 R4 ( R2 ) 4 0
!
1
1 1
0
1 and the other `ij are zero, so that
2
1
60
L=6
40
0
Thus
2
0
6
1
A = P T LU = 6
40
0
1
0
0
0
0
0
0
1
32
0
1
60
07
76
15 4 0
0
0
0
1
0
1
0
0
1
0
3
0
07
7.
05
1
0
1
0
1
0
0
1
0
LU factorization of this matrix:
3
1 1 2
1 1 37
7 = U.
0 1 15
0 0 4
32
0
6
07
76
05 4
1
1
0
0
0
3
1 1 2
1 1 37
7.
0 1 15
0 0 4
26. From the text, a permutation matrix consists of the rows of the identity matrix written in some order.
So each row and each column of a permutation matrix has exactly one 1 in it, and the remaining entries
in that row or column are zero. Thus there are n choices for the column containing the 1 in the first
row; once that is chosen, there are n 1 choices for the column to contain the 1 in the second row.
Next, there are n 2 choices in the third row. Continuing, we decrease the number of choices by one
each time, until there is just one choice in the last row. Multiplying all these together gives the total
number of choices, i.e., the number of n⇥n permutation matrices, which is n·(n 1)·(n 2) · · · 2·1 = n!.
27. To solve Ax = P T LU x = b, multiply both sides by P to get LU x = P b. Now compute P b, which is a
permutation of the rows of b, and solve the resulting equation by the method outlined after Theorem
3.15:
2
3
2
3
2
32 3 2 3
0 1 0
0 1 0
0 1 0
1
1
P T = 41 0 05 ) P = 41 0 05 ) P b = 41 0 05 415 = 415 .
0 0 1
0 0 1
0 0 1
5
5
Then
2
1
6
Ly = P b ) 4 0
0
1
1
2
1
2
32 3 2 3
0
y1
1
y1
76 7 6 7
0 5 4 y 2 5 = 4 15 )
1
1
y3
5
2 y1
= 1
y2
= 1
1
2 y 2 + y3 = 5
y1 = 1
) y2 = 1
y3 = 5
2 3 2 3
y1
1
6 7 6 7
) 4 y 2 5 = 4 15 .
1
1
y3
5
y1 + y2
2
2
Likewise,
2
2
U x = y ) 40
0
3
1
0
32 3 2 3
2
2x1 + 3x2 + 2x3 = 1
x1
1
15 4x2 5 = 415 )
x2
x3 = 1
5
5
x
5
3
2
2 x3 = 5
x3 =
2
) x2 = x3 + 1
2x1 = 1 3x2
2
3 2 3
x1
4
6 7 6 7
) x = 4x2 5 = 4 15 .
x3
2
2x3
174
CHAPTER 3. MATRICES
28. To solve Ax = P T LU x = b, multiply both sides by P to get LU x = P b. Now compute P b, which is a
permutation of the rows of b, and solve the resulting equation by the method outlined after Theorem
3.15:
2
3
2
3
2
32 3 2 3
0 0 1
0 1 0
0 1 0
16
4
P T = 41 0 05 ) P = 40 0 15 ) P b = 40 0 15 4 45 = 4 45 .
0 1 0
1 0 0
1 0 0
4
16
Then
2
1
Ly = P b ) 41
2
32 3 2 3
0 0 y1
4
y1
= 4
1 05 4y2 5 = 4 45 ) y1 + y2
= 4
1 1
y3
16
2y1
y2 + y3 = 16
2 3 2 3
y1
4
6 7 6 7
) y2 = 4 y1
) 4y2 5 = 4 85 .
y3
32
y3 = 16 2y1 + y2
y1 =
Likewise,
2
4
U x = y ) 40
0
4
32 3 2 3
1 2 x1
4
4x1 + x2 + 2x3 = 4
1 15 4x2 5 = 4 85 )
x2 + x3 = 8
0 2 x3
32
2x3 = 32
x3 = 16
) x2 = x3 8
4x1 = x2 2x3
2
3 2
3
x1
11
6 7 6
7
) x = 4x2 5 = 4 85 .
x3
16
4
29. Let A = (aij ) and B = (bij ) be unit lower triangular n ⇥ n matrices, so that
8
>
<0 i < j
aij = bij = 1 i = j
>
:
⇤ i>j
(where of course the ⇤’s maybe di↵erent for A and B). Then if C = (cij ) = AB, we have
cij = ai1 b1j + ai2 b2j + · · · + ai(j
If i < j, then all the terms up through ai(j 1) b(j
term to the end are zero since the a’s are zero.
1) b(j 1)j
1)j
+ aij bjj + · · · + ain bnj .
since the b’s are zero, and all terms from the next
If i = j, then all the terms up through ai(j 1) b(j 1)j since the b’s are zero, the aij bjj = ajj bjj term is
1, and all terms from the next term to the end are zero since the a’s are zero.
Finally, if i > j, multiple terms in the sum may be nonzero.
Thus we get
8
>
<0
cij = 1
>
:
⇤
so that C is lower unit triangular as well.
⇥
⇤
30. To invert L, we row-reduce L I :
2
1 0 ···
6⇤ 1 · · ·
6
6 .. .. . .
4. .
.
⇤ ⇤ ···
i<j
i=j
i>j
0
0
..
.
1
0
..
.
0
1
..
.
···
···
..
.
1
0
0
···
3
0
07
7
.. 7 .
.5
1
3.4. THE LU FACTORIZATION
175
To do this, we first add multiples of the first row to each of the lower rows to make the first column
zero except for the 1 in row 1. Note that the result of this is a unit lower triangular matrix on the right
(since we added multiples of the leading 1 in the first row to the lower rows). Now do the same with
row 2: add multiples of that row to each of the lower rows, clearing column 2 on the left except for the
1 in row 2. Again, for the same reason, the result on the right remains unit lower triangular. Continue
this process for each row. The result is the identity matrix on the left, and a unit lower triangular
matrix on the right. Thus L is invertible and its inverse is unit lower triangular.
31. We start by finding D. First, suppose D is a diagonal 2 ⇥ 2 matrix and A is any 2 ⇥ 2 matrix with 1’s
on the diagonal. Then

a 0
DA =
A
0 b
has the e↵ect of multiplying the first row of A by a and the second row of A by b, so results in a matrix
with a and b in the diagonal entries. In our example, we want to write

2 1
U=
0 6
as unit upper triangular, say U = DU1 . So we must turn the 2 and 6 into 1’s, and thus


1
2 0 1
2 = DU .
U=
1
0 6 0
1
Thus
A = LU =

1
1
0
1

2
0
0
=
6

1
1
0
1

2
0
0
6

1
0
1
2
1
= LDU1 .
32. We start by finding D. First, suppose D is a diagonal 3 ⇥ 3 matrix and A is any 3 ⇥ 3 matrix with 1’s
on the diagonal. Then
2
3
a 0 0
DA = 40 b 05 A
0 0 c
has the e↵ect of multiplying the first row of A by a, the second row of A by b, and the third row of A
by c, so the result is a matrix with a, b, and c on the diagonal. In our example, we want to write
2
3
2
4 0
5 45
U = 40
0
0 2
as unit upper triangular, say U = DU1 . So we must turn the 2, 5, and 2 into 1’s, and thus
2
32
3
1
2 0
2 0 0
1 45 5 = DU1 .
U = 4 0 5 05 4 0
0 0 2
0
0 1
Thus
2
6
A = LU = 4
1
3
2
1
2
32
0 0
2
76
1 05 4 0
0 1
0
3 2
4 0
1
7 6 3
5 45 = 4 2
1
0 2
2
32
0 0
2
76
1 05 4 0
0 1
0
0
5
0
32
0 1
76
05 40
2 0
2
1
0
3
0
47
= LDU1 .
55
1
33. We must show that if A = LDU is both symmetric and invertible, then U = LT . Since A is symmetric,
L(DU ) = LDU = A = AT = (LDU )T = U T DT LT = U T (DT LT ) = U T (DLT )
since the diagonal matrix D is symmetric. Since L is lower triangular, it follows that LT is upper
triangular. Since any diagonal matrix is also upper triangular, Exercise 29 in Section 3.2 tells us that
176
CHAPTER 3. MATRICES
DLT and DU are both upper triangular. Finally, since U is upper triangular, it follows that U T is
lower triangular. Thus we have A = U T (DLT ) = L(DU ), giving two LU -decompositions of A. But
since A is invertible, Theorem 3.16 tells us that the LU -decomposition is unique, so that U T = L.
Taking transposes gives U = LT .
34. Suppose A = L1 D1 LT1 = L2 D2 LT2 where L1 and L2 are unit lower triangular, D1 and D2 are diagonal,
and A is symmetric and invertible. We want to show that L1 = L2 and D1 = D2 . First, since
L2 (D2 LT2 ) = A = L1 (D1 LT1 )
gives two LU -factorizations of an invertible matrix A, it follows from Theorem 3.16 that these two
factorizations are the same, so that L1 = L2 and D2 LT2 = D1 LT1 . Since L1 = L2 , this equation
becomes D2 LT2 = D1 LT2 . From Exercise 47 in Section 3.3, since A = L2 (D2 LT2 ) and A is invertible,
we see that both L2 and D2 LT2 are invertible; thus LT2 is invertible as well ((LT2 ) 1 = (L2 1 )T ). So
multiply D2 LT2 = D1 LT2 on the right by (LT2 ) 1 , giving D2 = D1 , and we are done.
3.5
Subspaces, Basis, Dimension, and Rank
1. Geometrically, x = 0 is a line through the origin in R2 , so you might conclude from Example 3.37 that
x
this is a subspace. And in fact, the set of vectors
with x = 0 is
y


0
0
=y
.
y
1
✓ ◆
0
Since y is arbitrary, S = span
, and this is a subspace of R2 by Theorem 3.19.
1



1
1
1
2. This is not a subspace. For example,
is in S, but
=
is not in S. This violates property
0
0
0
3 of the definition of a subspace.
2
3. Geometrically, y = 2x is a line through the origin in R
 , so you might conclude from Example 3.37
x
that this is a subspace. And in fact, the set of vectors
with y = 2x is
y


x
1
=x
.
2x
2
✓ ◆
1
Since y is arbitrary, S = span
, and this is a subspace of R2 by Theorem 3.19.
2


3
2
4. This is not a subspace. For example, both
and
are in S, since 3 · ( 1) and 2 · 2 are both
1
2
nonnegative. But their sum,



3
2
1
+
=
,
1
2
1
is not in S since
1 · 1 < 0. So S is not a subspace (violates property 2 of the definition).
5. Geometrically, x = y = z is a line through the origin in R3 , so you2could
3 conclude from Exercise 9 in
x
this section that this is a subspace. And in fact, the set of vectors 4y 5 with x = y = z is
z
2 3
2 3
x
1
4x5 = x 415 .
x
1
3.5. SUBSPACES, BASIS, DIMENSION, AND RANK
177
02 3 1
1
Since x is arbitrary, S = span @415A, and this is a subspace of R3 by Theorem 3.19.
1
6. Geometrically, this is a line in R3 lying in the xz-plane and passing
2 3 through the origin, so Exercise 9
x
in this section implies that it is a subspace. The set of vectors 4y 5 satisfying these conditions is
z
2
3
2 3
x
1
4 0 5 = x 405 .
2x
2
02 3 1
1
Since x is arbitrary, S = span @405A, and this is a subspace of R3 by Theorem 3.19.
2
7. This is a plane in R3 , but it does not pass through the origin since (0, 0, 0) does not satisfy the equation.
So condition 1 in the definition is violated, and this is not a subspace.
8. This is not a subspace. For example
since for the first vector |1
since |1
1| =
6 |1
0| = |0
3|.
2 3
2 3
1
0
405 2 S and 415 2 S
1
2
1| and for the second |0
2 3 2 3 2 3
1
0
1
405 + 415 = 415 2
/S
1
2
3
1| = |1
2|. But their sum,
9. If a line ` passes through the origin, then it has the equation x = 0 + td = td, so that the set of points
on ` is just span(d). Then Theorem 3.19 implies that ` is a subspace of R3 . (Note that we did not use
the assumption that the vectors lay in R3 ; in fact this conclusion holds in any Rn for n 1.)
10. This is not a subspace of R2 . For example, (1, 0) 2 S and (0, 1) 2 S, but their sum, (1, 1), is not in S
since it does not lie on either axis.
11. As in Example 3.41, b is in col(A) if and only if the augmented matrix

⇥
⇤
1 0
1 3
A b =
1 1
1 2
is consistent as a linear system. So we row-reduce that augmented matrix:

1
1
0
1
1
1
3
2

1
! 0
R2 R1
0
1
1
2
3
.
1
This matrix is in row echelon form and has no zero rows, so the system is consistent and thus b 2 col(A).
Using Example 3.41, w is in row(A) if and only if the matrix
2
3

1 0
1
A
1 5
=4 1 1
w
1 1
1
178
CHAPTER 3. MATRICES
can be row-reduced to a matrix whose last row is zero by operations excluding row interchanges
involving the last row. Thus
2
4
1
1
1
3
1
1 5
1
0
1
1
2
1
R2 R1
!4 0
R3 +R1
! 0
3
1
2 5
0
0
1
1
R3
2
1
4 0
R2
! 0
3
1
2 5
2
0
1
0
This matrix is in row echelon form and has no zero rows, so we cannot make the last row zero. Thus
w2
/ row(A).
12. Using Example 3.41, b is in col(A) if and only if the
2
1
⇥
⇤
A b = 40
1
augmented matrix
3
1
3 1
2
1 15
1
4 0
is consistent as a linear system. So we row-reduce that augmented matrix:
2
1
6
40
1
1
2
1
3
1
7
15
0
3
1
4
R3
2
1
6
40
R1
! 0
1
2
2
3
1
7
15
1
3
1
1
2
1
6
40
R3 +R2
! 0
1
2
0
3
1
0
3
1
7
15
0
2
1
1
6
2 R2
! 40
0
1
1
0
3
1
2
0
3
1
17
25
0
The linear system has a solution (in fact, an infinite number of them), so is consistent. Thus b is in
col(A).
Using Example 3.41, w is in row(A) if and only if the
2
1

6 0
A
=6
4 1
w
2
matrix
3
3
1 7
7
4 5
5
1
2
1
4
can be row-reduced to a matrix whose last row is zero by operations excluding row interchanges
involving the last row. Thus
2
1
6 0
6
4 1
2
1
2
1
4
3
3
1 7
7
4 5
5
R3
R4
2
1
6 0
R1 6
!4 0
2R1
! 0
1
2
2
2
3
3
1 7
7
1 5
1
2
1
6 0
6
4 0
R4 +R3
! 0
1
2
2
0
3
3
1 7
7
1 5
0
Thus w is in row(A).
13. As in the remarks following Example 3.41, we determine whether w is in row(A) using col(AT ). To
say w is in row(A) means that wT is in col(A), so we form the augmented matrix and row-reduce to
see if the system has a solution:
⇥
AT
w
⇤
T
2
1
=4 0
1
1
1
1
3
1
15
1
2
1
40
R3 +R2
! 0
1
1
2
3
1
15
0
R2
2
1
4
1
0
2 R3
! 0
1
0
2
3
1
15
0
2
1
R2 $R3 4
! 0
0
1
2
0
3
1
05
1
This matrix has a row that is all zeros except for the final column, so the system is inconsistent. Thus
wT 2
/ col(AT ) and thus w 2
/ row(A).
14. As in the remarks following Example 3.41, we determine whether w is in row(A) using col(AT ). To
say w is in row(A) means that wT is in col(A), so we form the augmented matrix and row-reduce to
3.5. SUBSPACES, BASIS, DIMENSION, AND RANK
179
see if the system has a solution:
⇥
AT
w
⇤
T
2
1
=4 1
3
0
2
1
1
1
4
3
2
45
5
2
1
R2 R1
! 40
R3 +3R1
! 0
0
2
1
1
2
1
2
3
2
25
1
1
1
2 R2 4 0
!
0
0
1
1
1
1
1
3
2
15
1
R3
2
1
40
R2
! 0
0
1
0
1
1
0
3
2
15 .
0
This is in row reduced form, and the final row consists of all zeros. This is a consistent system, so that
wT 2 col(AT ) and thus w 2 row(A).
15. To check if v is in null(A), we need only multiply A by v:
2 3


1
1 0
1 4 5
0
3 =
Av =
6= 0,
1 1
1
1
1
so that v 2
/ null(A).
16. To check if v is in null(A), we need only multiply A by v:
2
32 3 2 3
1
1
3
7
0
2
15 4 15 = 405 = 0,
Av = 40
1
1
4
2
0
so that v 2 null(A).
17. We follow the comment after Example 3.48. Start by finding the reduced row echelon form of A:

1
1
0
1
1
1

1
! 0
R2 R1
0
1
1
2
Thus a basis for row(A) is given by
⇥
{ 1
0
⇤ ⇥
1 , 0
1
⇤
2 }.
A basis for col(A) is given by the column vectors of the original matrix A corresponding to the leading
1s in the row-reduced matrix, so a basis for col(A) is
⇢

1
0
,
.
1
1
Finally, from the row-reduced form, the solutions to Ax = 0 where
2 3
x1
x = 4x2 5
x3
can be found by noting that x3 is a free variable; setting it to t and then solving for the other two
variables gives x1 = t and x2 = 2t. Thus a basis for the nullspace of A is given by
2 3
1
x = 4 25 .
1
180
CHAPTER 3. MATRICES
18. We follow the comment after
2
3
2
1
1
3
1
6
7
6
60
7
6
2
15
4
40
R3 R1
1
1
4
! 0
Example 3.48. Start
3
2
1
3
1
7
6
7
6
2
15
40
R3 +R2
2
1
! 0
Thus a basis for row(A) is given by
⇥
⇤ ⇥
⇤
7
1
1 0
2 , 0 1
2
by finding the reduced row echelon form
3
2
3
2
R1 R2
1
3
1 1
3
! 1
7 1 R2 6
7
6
17
2
6
60
2
17
5 ! 40 1
4
25
0
0
0 0
0
0
⇥
or, clearing fractions,
2
⇤ ⇥
7 , 0
0
2
of A:
3
7
27
17
25
0
1
0
0
⇤
1 .
Since the row-reduction did not involve any row exchanges, a basis for row(A) is also given by the rows
of A corresponding to nonzero rows in the reduced matrix:
⇥
⇤ ⇥
⇤
1 1
3 , 0 2 1 .
A basis for col(A) is then given by the original column vectors of A corresponding to the columns
containing the leading 1s in the row-reduced matrix; thus a basis for col(A) is given by
8 2 3 2 39
1 =
< 1
405 , 4 25
:
;
1
1
Finally, from the row-reduced form, the solutions of Ax = 0 where
2 3
x1
x = 4x2 5
x3
can be found by noting that x3 is a free variable; we set it to t, and then solving for the other two
variables in terms of t to get x1 = 72 t and x2 = 12 t. Thus a basis for the nullspace of A is given by
2 3
2 3
7
7
6 27
6 7
1
6
7
6
x = 4 2 5 or 4 17
5
1
2
19. We follow the comment after Example 3.48. Start by finding the reduced row echelon form of A:
2
1
40
0
1
1
1
0
1
1
3
1
15
1
R3
2
1
40
R2
! 0
1
1
0
Thus a basis for row(A) is given by
⇥
1 0 1
0
1
0
3
1
15
2
⇤ ⇥
0 , 0
2
1
40
1
2 R3
! 0
1
1
1
1
0
2
1
40
0
⇤ ⇥
0 , 0
3
0 1
1 15
0 1
!
R2 R3
!
3
0 0
1 05
0 1
1
1
0
0
R1 R3
0
1
⇤
2
! 1
40
0
R1 R2
0
1
0
3
1 0
1 05
0 1
.
Since the row-reduction did not involve any row exchanges, a basis for row(A) is also given by the rows
of A corresponding to nonzero rows in the reduced matrix; since all the rows are nonzero, the rows of
A form a basis for row(A).
A basis for col(A) is then given by the original column vectors of A corresponding to the columns
containing the leading 1s in the row-reduced matrix; thus a basis for col(A) is given by
8 2 3 2 3 2 39
1
1 =
< 1
4 05 , 4 15 , 4 15
:
;
0
1
1
3.5. SUBSPACES, BASIS, DIMENSION, AND RANK
181
Finally, from the row-reduced form, the solutions of Ax = 0 where
2 3
x1
6x2 7
7
x=6
4x3 5
x4
can be found by noting that x3 is a free variable; we set it to t. Also, x4 = 0; solving for the other two
variables in terms of t gives x1 = t and x2 = t. Thus a basis for the nullspace of A is given by
2 3
1
6 17
7
x=6
4 15 .
0
20. We follow
2
2
6
4 1
1
the comment after Example
3
2
4 0 2 1
1
7 R $R3 6
2 1 2 35 1 !
4 1
2 1 4 4
2
2
1
2 1
6
0 2
40
0
0 0
3.48. Start by finding the reduced row echelon form of A:
3
2
3
2 1 4 4
1
2
1
4
4
7 R +R1 6
7
2 1 2 35 2 !
0
2
6
75
40
R3 +R2
1
4 0 2 1 R3 2R!
0
0
2
6
7
!
3
2
3 R R 2
3
1
2
4 4
1
2 1 4 4
1
2 0 1 12
!
7 1 R2 6
7
6
7
6 75 2 !
0 1 3 72 5
0 1 3 72 5
40
40
0 0
0
0 0 0 0
0
0 0 0 0
Thus a basis for row(A) is given by the nonzero rows of the reduced matrix, which after clearing
fractions are
⇥
⇤ ⇥
⇤
2
4 0 2 1 , 0 0 2 6 7 .
A basis for col(A) is then given by the original column vectors of A corresponding to the columns
containing the leading 1s in the row-reduced matrix; thus a basis for col(A) is given by
8 2 3 2 39
0 =
< 2
4 15 , 415 .
:
;
1
1
Finally, from the row-reduced form, the solutions of Ax = 0 where
2 3
x1
6x2 7
6 7
7
x=6
6x3 7
4x4 5
x5
can be found by noting that x2 = s, x4 = t, and x5 = u are free variables. Solving for the other two
variables gives x1 = 2s t 12 u and x3 = 3t 72 u. Thus the nullspace of A is given by
82
39 8 2 3
2 3
2 1 39
02 3 2 3 2 1 31
2s t + 12 u >
2
1
2
1
>
>
>
2
2
>
>
>
>
>6
>
>
7>
>
>
617
6 07
6 07>
>
B617 6 07 6 07C
s
<6
=
<
=
7
6
7
6
7
6
7
B
6
7
6
7
6
7C
6 3t 7 u 7 = s 607 + t 6 37 + u 6 7 7 = span B607 , 6 37 , 6 7 7C .
2
2
2
6
7
6
7
6
7
6
7
B
6
7
6
7
6
7C
>4
> >
> 4 5
>
>
5>
4 15
4 05>
@405 4 15 4 05A
0
t
>
>
>
>
>
>
>
>
:
; :
;
0
0
0
0
u
1
1
Clearing fractions, we get for a basis for the nullspace of A
8 2 3 2 3 2 39
1
1 >
>
> 2
>
>
>
7 6 07 6 07>
>
1
<6
6 7 6 7 6 7=
607 , 6 37 , 6 77 .
6 7 6 7 6 7>
>
>
4 5 4 15 4 05>
>
>
> 0
>
;
:
0
0
2
182
CHAPTER 3. MATRICES
21. Since row(A) = col(AT ) and col(A) = row(AT ),
Then
2
1
AT = 4 0
1
it suffices to find the row and column spaces of AT .
3
2
3
1
1 0
15 ! 40 15 ,
1
0 0
so that the column space of AT is given by the columns of AT corresponding to columns of this matrix
containing a leading 1, so
8 2 3 2 39
1 =
< 1
⇥
⇤ ⇥
⇤
1 , 1 1 1 .
col(AT ) = 4 05 , 415 ) row(A) = 1 0
:
;
1
1
The row space of AT is given by the nonzero rows of the reduced matrix, so
⇢

⇥
⇤ ⇥
⇤
1
0
T
row(A ) = 1 0 , 0 1 ) col(A) =
,
.
0
1
22. Since row(A) = col(AT ) and col(A) = row(AT ),
Then
2
1 0
AT = 4 1 2
3 1
it suffices to find the row and column spaces of AT .
3
2
3
1
1 0
1
15 ! 40 1
15 ,
4
0 0
0
so that the column space of AT is given by the columns of AT corresponding to columns of the reduced
matrix containing a leading 1, so
82 3 2 39
0 =
< 1
⇥
⇤ ⇥
⇤
T
4
5
4
1
3 , 0 2 1 .
col(A ) =
, 25 ) row(A) = 1 1
:
;
3
1
The row space of AT is given by the nonzero rows of the reduced matrix, so
8 2 3 2 39
0 =
< 1
⇥
⇤ ⇥
⇤
1 ) col(A) = 405 , 4 15 .
row(AT ) = 1 0 1 , 0 1
:
;
1
1
23. Since row(A) = col(AT ) and col(A) = row(AT ), it suffices to find the row and column spaces of AT .
Then
2
3
2
3
1
0
1
1 0 0
61
6
7
1
17
7 ! 60 1 07 ,
AT = 6
40
5
4
1
1
0 0 15
1
1
1
0 0 0
so that the column space of AT is given by the columns of AT corresponding to columns of the reduced
matrix containing a leading 1, so
8 2 3 2 3 2 39
1
0
0 >
>
>
< 6 7 6 7 6 7>
=
⇥
⇤ ⇥
⇤ ⇥
⇤
17 6 17 6 17
T
6
1 1 , 0 1
1
1 .
col(A ) = 4 5 , 4 5 , 4 5 ) row(A) = 1 1 0 1 , 0 1
0
1
1 >
>
>
>
:
;
1
1
1
The row space of AT is given by the nonzero rows of the reduced matrix, so
82 3 2 3 2 39
0
0 =
< 1
⇥
⇤
⇥
⇤
⇥
⇤
row(AT ) = 1 0 0 , 0 1 0 , 0 0 1 ) col(A) = 405 , 415 , 405 .
:
;
0
0
1
3.5. SUBSPACES, BASIS, DIMENSION, AND RANK
183
24. Since row(A) = col(AT ) and col(A) = row(AT ), it suffices to
Then
2
3
2
2
1
1
1
6 4
7
60
2
2
6
7
6
6
1
17
AT = 6
6 0
7 ! 60
4 2
40
2
45
1
3
4
0
find the row and column spaces of AT .
3
0 1
1 17
7
0 07
7,
0 05
0 0
so that the column space of AT is given by the columns of AT corresponding to columns of the reduced
matrix containing a leading 1, so
82 3 2 39
2
1 >
>
>
>
>
>
>
6
7
6
>
4
<6 7 6 27
=
7
⇥
⇤ ⇥
⇤
T
7 , 6 17 ) row(A) = 2
0
4 0 2 1 , 1 2 1 2 3 .
col(A ) = 6
6
7
6
7
>
>
>
4 25 4 25>
>
>
>
>
:
;
1
3
The row space of AT is given by the nonzero rows of the reduced matrix, so
8 2 3 2 39
0 =
< 1
⇥
⇤
⇥
⇤
row(AT ) = 1 0 1 , 0 1 1 ) col(A) = 405 , 415 .
:
;
1
1
25. There is no reason to expect the bases for the row and column spaces found in Exercises 17 and 21 to
be the same, since we used di↵erent methods to find them. What is important⇥is that they
span
⇤ should
⇥
⇤
1 0
1 , 0 1 2 },
the same subspaces. The
basis
for
the
row
space
found
in
Exercise
17
was
{
⇥
⇤ ⇥
⇤
1 , 1 1 1 }. Since the
and Exercise 21 gave { 1 0
vector is the same, in order to
⇥ first basis
⇤
0 1 2 is in the span of the second basis,
show the ⇥subspaces
are
the
same
it
suffices
to
see
that
⇤
and that 1 1 1 is in the span of the first basis. But
⇥
⇤
⇥
⇤ ⇥
⇤
⇥
⇤ ⇥
⇤ ⇥
⇤
0 1 2 =
1 0
1 + 1 1 1 ,
1 1 1 = 1 0
1 + 0 1 2 .
So the two bases span the same row space. Similarly, the basis for the two column spaces were
⇢

⇢

1
0
1
0
Exercise 17:
,
,
Exercise 21:
,
.
1
1
0
1
Here the second basis vectors are the same, and since





1
1
0
1
1
=
+
,
=
1
0
1
0
1

0
,
1
each basis vector is a linear combination of the vectors in the other basis, so they span the same column
space.
26. There is no reason to expect the bases for the row and column spaces found in Exercises 18 and 22 to
be the same, since we used di↵erent methods to find them. What is important⇥is that they
span
⇤ should
⇥
⇤
2
0
7
0
2
1
the same subspaces. The
basis
for
the
row
space
found
in
Exercise
18
was
{
,
},
⇥
⇤ ⇥
⇤
3 , 0 2 1 }. Since the second basis vector is the same, in order to
and Exercise 22 gave { 1 1
⇥
⇤
7 is in the span of the second basis,
show the ⇥subspaces ⇤are the same it suffices to see that 2 0
3 is in the span of the first basis. But
and that 1 1
⇥
2
0
⇤
⇥
7 =2 1
1
⇤
3
⇥
0
2
⇤
1 ,
⇥
1
1
⇤ 1⇥
3 =
2
2
0
⇤ 1⇥
7 +
0
2
2
⇤
1 .
So the two bases span the same row space. Similarly, the basis for the two column spaces were
82 3 2 39
8 2 3 2 39
1 =
0 =
< 1
< 1
405 , 4 25 ,
405 , 4 15 .
Exercise 18:
Exercise 22:
:
;
:
;
1
1
1
1
184
CHAPTER 3. MATRICES
Here the first basis vectors are the same, and since
2
3 2 3
2 3
1
0
1
4 25 = 405 + 2 4 15 ,
1
1
1
2
3
2 3
0
1
4 15 = 1 4 25
2
1
1
2 3
1
14 5
0 ,
2
1
each basis vector is a linear combination of the vectors in the other basis, so they span the same column
space.
27. As in Example 3.46, a basis for the span of the given set of vectors is the same as a basis for the
row
space of
⇥
⇤ ⇥the matrix⇤that has
⇥ rows equal
⇤ to these vectors. Thus we form a matrix whose rows are
1
1 0 ,
1 0 1 , and 0 1
1 and row-reduce it:
2
1
4 1
0
1
0
1
3
0
15
1
2
1
R2 +R1 4
! 0
0
3
0
15
1
1
1
1
2
1
R2 4
! 0
0
1
1
1
3
0
15
1
2
! 1
40
R2
! 0
R1 +R2
R3
0
1
0
3
1
15
0
A basis for the row space of the matrix, and thus a basis for the span of the given vectors, is therefore
given by the nonzero rows of the reduced matrix:
8 2 3 2 39
0 =
< 1
4 05 , 4 15 .
:
;
1
1
Note that since no row interchanges were involved, the rows of the original matrix corresponding to
nonzero rows in the reduced matrix also form a basis for the span of the given vectors.
28. As in Example 3.46, a basis for the span of the given set of vectors is the same as a basis for the
row
space of
to these
vectors. Thus we form a matrix whose rows are
⇥
⇤ ⇥the matrix
⇤ ⇥that has ⇤rows equal
⇥
⇤
1
1 1 , 1 2 0 , 0 1 1 , and 2 1 2 and row-reduce it:
2
1
61
6
40
2
3
1 1
R2
2 07
7
1 15
1 2 R4
R3
R4
2
1
!6
60
40
2R1
! 0
2
1
60
R2 6
! 40
2R2
! 0
R1
1
3
1
3
1
1
0
0
A basis for the row space of the matrix, and
given by
82
<
4
:
3
1
17
7
15
0
3
1
07
7
15
1
2
1
0
R2 $R4 6
!6
40
0
2
1
60
6
40
R4 +R3
! 0
1
3
1
3
3
1 1
1 07
7
0 15
0 0
3
1
07
7
15
1
2
1
60
!6
40
0
1
3 R2
1
1
1
3
3
1
07
7
15
1
thus a basis for the span of the given vectors, is therefore
3 2 3 2 39
1
0
0 =
15 , 415 , 405 .
;
1
0
1
(Note that this matrix could be further reduced, and we would get a di↵erent basis).
29. As in Example 3.46, a basis for the span of the given set of vectors is the same as a basis for the
row space of the matrix that has rows equal to these vectors. Thus we form a matrix whose rows are
3.5. SUBSPACES, BASIS, DIMENSION, AND RANK
⇥
2
3
⇤ ⇥
1 , 1
2
2
41
4
1
3
1
4
185
⇤ ⇥
⇤
0 , 4
4 1 and row-reduce it:
3
2
3
2
1
1
1 0
1
R 2R1
2 4
40
05 R1 $R!
2
3 15 2 !
1
1
4
4 1 R3 4R!
0
2
3
2
1
1
0
1
40
40
1
15
R3 +2R2
R3
0
1
! 0
! 0
2
3
R1 +R2
! 1 0 0
4 0 1 05
0 0 1
1
1
2
1
1
0
3
0
15
1
3
0
15
1
2
1
R2 4
! 0
0
1
1
2
2
1
! 40
0
R2 +R3
3
0
15
1
3
1 0
1 05
0 1
A basis for the row space of the matrix, and thus a basis for the span of the given vectors, is therefore
given by
⇥
⇤ ⇥
⇤ ⇥
⇤
{ 1 0 0 , 0 1 0 , 0 0 1 }
(Note that as soon as we got to the first matrix on the second line above, we could have concluded that
the reduced matrix would have no zero rows; since it was 3 ⇥ 3, it would thus be the identity matrix,
so that we could immediately write down the basis above, or simply use the basis given by the rows of
that matrix).
30. As in Example 3.46, a basis for the span of the given set of vectors is the same as a basis for the
row
space of the
⇥
⇤ ⇥matrix that has
⇤ ⇥rows equal to
⇤ these vectors. Thus we form a matrix whose rows are
1 0 , 2 1 5 1 and row-reduce it:
0 1
2 1 , 3 1
2
3
2
3 1
2
3
R1
1
0 1
2 1
3 1
1 0 3 !
1 13
0
3
6
7 R1 $R2 6
7
6
7
63 1
7
60 1
7
60 1
!
1
0
2
1
2 17
4
5
4
5
4
5
2 1
5 1
2 1
5 1
2 1
5 1
2
3
2
3
1
1
1 13
0
1 13
0
3
3
6
7
6
7
60 1
60 1
2 17
2 17
4
5
4
5
R3 13 R2
R3 2R1
17
19
2
1
0
0
! 0 13
!
3
3
3
2
3
1
1 13
0
3
6
7
60 1
2
17
4
5
3
2
19 R3
1 19
! 0 0
Although this matrix could be reduced further, we already have a basis for the row space:
⇥
⇤
⇥
⇤
⇥
⇤
1
2
0 1
2 1 ,
1 13
0 ,
0 0 1 19
,
3
or
⇥
3
1
1
⇤
0 ,
⇥
0
1
2
⇤
1 ,
⇥
0
0
19
⇤
2 .
31. If, as in Exercise 29, we form the matrix whose rows are those vectors and row-reduce it, we get
the identity matrix (see Exercise 29). Because there were no row exchanges, the rows of the original
matrix corresponding to nonzero rows in the row-reduced matrix form a basis for the span of the
vectors (because there were no row exchanges). Thus a basis consists of all three given vectors:
⇥
⇤ ⇥
⇤ ⇥
⇤
2
3 1 , 1
1 0 , 4
4 1 .
32. If, as in Exercise 30, we form the matrix whose rows are those vectors and row-reduce it, we get a
matrix with no zero rows. Thus the rank of the matrix is 3, so the rows are linearly independent, so a
basis consists of all three row vectors:
⇥
⇤ ⇥
⇤ ⇥
⇤
0 1
2 1 , 3 1
1 0 , 2 1 5 1 .
186
CHAPTER 3. MATRICES
33. If R is a matrix in echelon form, then row(R) is the span of the rows of R, by definition. But nonzero
rows of R are linearly independent, since their first entries are in di↵erent columns, so those nonzero
rows form a basis for row(R) (they span, and are linearly independent).
34. col(A) is the span of the columns of A by definition. If the columns of A are also linearly independent,
then they form a basis for col(A) by definition of “basis”.
35. The rank of A is the number of vectors in a basis for either row(A) or col(A), by definition. From
Exercise 17, row(A) has two vectors in its basis, so rank(A) = 2. Then nullity(A) is the number of
columns minus the rank, or 3 2 = 1.
36. The rank of A is the number of vectors in a basis for either row(A) or col(A), by definition. From
Exercise 18, row(A) has two vectors in its basis, so rank(A) = 2. Then nullity(A) is the number of
columns minus the rank, or 3 2 = 1.
37. The rank of A is the number of vectors in a basis for either row(A) or col(A), by definition. From
Exercise 19, row(A) has three vectors in its basis, so rank(A) = 3. Then nullity(A) is the number of
columns minus the rank, or 4 3 = 1.
38. The rank of A is the number of vectors in a basis for either row(A) or col(A), by definition. From
Exercise 20, row(A) has two vectors in its basis, so rank(A) = 2. Then nullity(A) is the number of
columns minus the rank, or 5 2 = 3.
⇥
⇤
T
39. Since nullity(A) > 0, the equation
P Ax = 0 has a nontrivial solution x = c1 c2 · · · cn , where at
least one ci 6= 0. Then Ax =
ci ai = 0. Since at least one ci 6= 0, this shows that the columns of A
are linearly dependent.
40. Since rank(A) is at most the number of columns of A and also at most the number of rows of A, in
this case rank(A)  2. Thus any row-reduced form of A will have at least 4 2 = 2 zero rows, which
means that some row of A is a linear combination of the other rows, so that the rows of A are linearly
dependent.
41. Since rank(A) is at most the number of columns of A and also at most the number of rows of A, in
this case rank(A)  5 and rank(A)  3, so that rank(A)  3 and its possible values are 0, 1, 2, and 3.
Using the Rank Theorem, we know that
nullity(A) = n
rank(A) = 5
rank(A), so that nullity(A) = 5, 4, 3, or 2.
42. Since rank(A) is at most the number of columns of A and also at most the number of rows of A, in
this case rank(A)  2 and rank(A)  4, so that rank(A)  2 and its possible values are 0, 1, and 2.
Using the Rank Theorem, we know that
nullity(A) = 2
43. First row-reduce A:
2
3
1
2 a
4 2 4a 25
a
2 1
If a =
1, then we have
If a = 2, then we have
rank(A) = 1
R3
2
1
! 40
aR1
! 0
R2 +2R1
rank(A), so that nullity(A) = 2, 1, or 0.
3
2
a
4a + 4 2a + 25
2a 2 1 a2
2
1
40
0
2
0
0
2
1
40
0
2
12
0
2
1
40
R2 + 21 R2
! 0
2
4a + 4
0
3
1
05 so that rank(A) = 1.
0
3
2
65 so that rank(A) = 2.
0
Otherwise, the row-reduced matrix has all rows nonzero, so that rank(A) = 3.
3
a
2a + 2 5
a2 + a + 2
3.5. SUBSPACES, BASIS, DIMENSION, AND RANK
44. First row-reduce A:
2
3
2
a
2
1
2
3 4
4 3
3
25 R1 $R!
3
2
1
a
a
2
1
R2 3R1
! 40
R3 aR2
! 0 2
2
R3 (2
3
a
25
1
187
2
! 1 2
43 3
a 2
3
2
a 2
3
4 3a 5
2
2a
a + 2a 1
3
1 2
a 2
40 1
5
a 43
2a)R2
5
0
0
(a
1)
a
!
3
1
3
2
R1 +R2
3
2
25
1
2
1
2
1
3 R2 4 0
1
!
0 2 2a
a
a 2
a 43
a2 + 2a
3
5
1
The first two rows are always nonzero. If a = 1 or a = 53 , then the bottom row is zero, so that
rank(A) = 2. Otherwise, all three rows are nonzero and rank(A) = 3.
45. As in example 3.52,
form a basis for R3 if and only
that matrix:
2
1 1
41 0
0 1
2 3
1
415 ,
0
2 3
1
405 ,
1
2 3
0
4 15
1
if the matrix with those vectors as its rows has rank 3. Row-reduce
3
0
15
1
R2
2
3
1 0
1 15
1 1
1
R1 4
! 0
0
2
3
1 0
1 15
0 2
1
40
R3 +R2
! 0
Since the matrix has rank 3 (it has no nonzero rows and is in echelon form), the vectors are linearly
independent; since there are three of them, they form a basis for R3 .
46. As in example 3.52,
form a basis for R3 if and
that matrix:
2
1
4 1
1
2
3
1
4 15 ,
3
2
3
1
4 55 ,
1
2
3
1
4 35
1
only if the matrix with those vectors as its rows has rank 3. Row-reduce
3
1 3
5 15
3 1
2
1
R2 +2R1
! 40
R3 R1
! 0
1
4
2
3
3
45
2
2
1
R2 +2R3 4
! 0
0
1
0
2
3
3
05
2
Since the matrix does not have rank 3, it follows that these vectors do not form a basis for R3 . (In
fact,
2 3 2 3
2 3 2 3
1
1
1
0
4 15 + 4 55 + 2 4 35 = 405 ,
3
1
1
0
so the vectors are linearly dependent.)
47. As in example 3.52,
2 3
1
6 17
6 7,
4 15
0
2 3
1
617
6 7,
405
1
2 3
1
607
6 7,
415
1
2 3
0
6 17
6 7
4 15
1
188
CHAPTER 3. MATRICES
form a basis for R4 if and only if the matrix with those vectors as its rows has rank 4. Row-reduce
that matrix:
2
1
61
6
41
0
1
1
0
1
1
0
1
1
3
0
17
7
15
1
R3
2
1
!6
60
R1 4 0
!
0
R2 R1
1
0
1
1
1
1
0
1
3
0
17
7
15
1
2
1
6
R4 $R2 60
! 40
0
1
1
1
0
2
1
60
6
R3 +R2 4
! 0
0
3
1 0
1 17
7
0 15
1 1
1
1
0
0
3
1 0
1 17
7
1 25
1 1
2
1
60
6
40
R4 +R3
! 0
1
1
0
0
1
1
1
0
3
0
17
7
25
3
This matrix is in echelon form and has no zero rows, so it has rank 4 and thus these vectors form a
basis for R4 .
48. As in example 3.52,
form a basis for R4
that matrix:
2
1
6 0
6
4 0
1
2
3
1
6 17
6 7,
4 05
0
2
3
0
6 17
6 7,
4 05
1
2
3
0
6 07
6 7,
4 15
1
2
3
1
6 07
6 7
4 15
0
if and only if the matrix with those vectors as its rows has rank 4. Row-reduce
1
1
0
0
0
0
1
1
3
0
17
7
15
0
2
1
60
6
40
R4 +R1
! 0
1
1
0
1
0
0
1
1
3
2
0
1
60
17
7
6
40
15
R4 +R2
0
! 0
2
1
60
6
40
R4 +R3
! 0
1
1
0
0
0
0
1
1
1
1
0
0
0
0
1
0
3
0
17
7
15
1
3
0
17
7
15
0
Since the matrix does not have rank 4, these vectors do not form a basis for R4 . (In fact, the sum of
the four vectors is zero, so they are linearly dependent.)
49. As in example 3.52,
form a basis for Z32 if and only if
that matrix over Z2 :
2
1 1
40 1
1 0
2 3
1
415 ,
0
2 3
0
415 ,
1
2 3
1
4 05
1
the matrix with those vectors as its rows has rank 3. Row-reduce
3
0
15
1
2
3
0
15
1
1
40
R3 +R1
! 0
1
1
1
2 3
1
415 ,
0
2 3
0
415 ,
1
2
1
40
R3 +R2
! 0
1
1
0
3
0
15
0
Since the matrix does not have rank 3, these vectors do not form a basis for Z32 .
50. As in example 3.52,
2 3
1
4 05
1
3.5. SUBSPACES, BASIS, DIMENSION, AND RANK
189
form a basis for Z33 if and only if the matrix with those vectors as its rows has rank 3. Row-reduce
that matrix over Z3 :
2
3
2
3
2
3
1 1 0
1 1 0
1 1 0
40 1 15
4 0 1 15
40 1 15
1
2
1 0 1 R3 +2R!
0 2 1 R3 +R!
0 0 2
Since the matrix has rank 3, these vectors form a basis for Z33 .
51. We want to solve
2 3
2 3 2 3
1
1
1
c1 425 + c2 4 05 = 465 ,
0
1
2
so set up the augmented matrix of the
2
3
2
1
1
1 1
1 4
42
0 65 R2 2R!
0
0
1 2
0
Thus c1 = 3 and c2 =
2, so that
Then the coordinate vector [w]B =

52. We want to solve
linear system and row-reduce to solve it:
3
2
3 R1 R2 2
1 1
1
1
1
! 1 0
1
2 4
40 1
2 45 2 R!
0
1
25
R
+R
3
2
1 2
0
1
2
! 0 0
2 3
1
3 425
0
3
.
2
2
3 2 3
1
1
2 4 05 = 465 .
1
2
2 3
2 3 2 3
3
5
1
c1 415 + c2 415 = 435 ,
4
6
4
so set up the augmented matrix of the linear system
3
2
3
2
2
3 5 1
1 1 3
1 1
R
3R
2
1
R
$R
2 4
4 1 1 35 1 !
3 5 15
! 40 2
1
4 6 4
4 6 4 R3 4R!
0 2
Thus c1 = 7 and c2 =
4, so that
Then the coordinate vector [w]B =

2 3
3
7 415
4
7
.
4
3
3
25
0
and row-reduce to
3
2
3 1
1 1
R
2 4
85 2 !
0 1
8
0 2
solve it:
3 R1 R2 2
3
! 1
40
45
2
8 R3 2R!
0
0
1
0
3
7
45
0
2 3 2 3
5
1
4 415 = 435 ,
6
4
53. Row-reduce A over Z2 :
2
1
A = 40
1
1
1
0
3
0
15
1
2
1
40
R3 +R1
! 0
1
1
1
3
0
15
1
2
1
40
R3 +R2
! 0
1
1
0
3
0
15 .
0
This matrix is in echelon form with one zero row, so it has rank 2. Thus nullity(A) = 3
3 2 = 1.
54. Row-reduce A over Z3 :
2
1
A = 42
2
1
1
0
3
2
25
0
2
1
R2 +R1
! 40
R3 +R1
! 0
1
2
1
3
2
15
2
2
1
40
R3 +R2
! 0
1
2
0
3
2
15 .
0
rank(A) =
190
CHAPTER 3. MATRICES
This matrix is in echelon form with one zero row, so it has rank 2. Thus nullity(A) = 3
3 2 = 1.
55. Row-reduce A over Z5 :
2
1
A = 42
1
3
3
0
3
4
15
0
1
0
4
2
1
R2 +3R1
! 40
R3 +4R1
! 0
3
2
2
1
3
3
3
4
35
1
2
1
40
R3 +4R2
! 0
3
2
0
rank(A) =
3
4
35
3
1
3
0
This matrix is in echelon form with no zero rows, so it has rank 3. Thus nullity(A) = 4
4 3 = 1.
rank(A) =
56. Row-reduce A over Z7 :
2
2
66
A=6
41
1
4
3
0
1
0
5
2
1
0
1
2
1
3
1
07
7
55
1
2
2
! 60
6
R3 +3R1 4
! 0
R4 +3R1
! 0
R2 +4R1
4
5
5
6
0
5
2
1
3
1
47
7
15
4
0
1
2
1
R3 +6R2
!
R4 +3R2
!
2
2
60
6
40
0
4
5
0
0
0
5
4
2
0
1
1
4
3
1
47
7
45
2
2
2
60
6
40
R4 +3R3
! 0
4
5
0
0
This matrix is in echelon form with one zero row, so it has rank 3. Thus nullity(A) = 5
5 3 = 2.
0
5
4
0
0
1
1
0
3
1
47
7.
45
0
rank(A) =
57. Since x is in null(A), we know that Ax = 0. But if Ai is the ith row of A, then
2
3 2
3 2 3
a11 x1 + a12 x2 + · · · + a1n xn
A1 · x
0
6 a21 x1 + a22 x2 + · · · + a2n xn 7 6 A2 · x 7 607
6
7 6
7 6 7
Ax = 6
7 = 6 .. 7 = 6 .. 7 .
..
4
5
4
.
. 5 4.5
am1 x1 + am2 x2 + · · · + amn xn
Am · x
0
Thus Ai · x = 0 for all i, so that every row of A is orthogonal to x.
58. By the Fundamental Theorem, if A and B have rank n, then they are both invertible. Therefore, by
Theorem 3.9(c), AB is invertible as well. Applying the Fundamental Theorem again shows that AB
also has rank n.
59. (a) Let bk be the k th column of B. Then Abk is the k th column of AB. Now suppose that some set
of columns of AB, say Abk1 , Abk2 , . . . , Abkr , are linearly independent. Then
c k 1 b k1 + c k2 b k 2 + · · · + c k r b kr = 0 )
A (ck1 bk1 + ck2 bk2 + · · · + ckr bkr ) = 0 )
ck1 (Abk1 ) + ck2 (Abk2 ) + · · · + ckr (Abkr = 0).
But the Abki are linearly independent, so all the cki are zero and thus the bki are linearly
independent. So the number of linearly independent columns in B is at least as large as the
number of linearly independent columns of AB; it follows that rank(AB)  rank(B).
(b) There are many examples. For instance,


1 1
1 0
A=
, B=
.
0 0
1 1
Then B has rank 2, and
AB =

2
0
1
0
has rank 1.
3.5. SUBSPACES, BASIS, DIMENSION, AND RANK
191
60. (a) Recall that by Theorem 3.25, if C is any matrix, then rank(C T ) = rank(C). Thus
rank(AB) = rank((AB)T ) = rank(B T AT )  rank(AT ) = rank(A),
using Exercise 59(a) with AT in place of B and B T in place of A.
(b) For example, we can use the transposes of the matrices in part (b) of Exercise 59:


1 1
1 0
A=
, B=
.
0 1
1 0
Then A has rank 2 and
AB =
61. (a) Since U is invertible, we have A = U
1

2
1
0
0
has rank 1.
(U A), so that using Exercise 59(a) twice,
rank(A) = rank U
1
(U A)  rank(U A)  rank(A).
Thus rank(A) = rank(U A).
(b) Since V is invertible, we have A = (AV )V
1
, so that using Exercise 60(a) twice,
rank(A) = rank (AV )V
1
 rank(AV )  rank(A).
Thus rank(A) = rank(AV ).
62. Suppose first that A has rank 1. Then col(A) = span(u) for some
⇥ vector u. It ⇤follows that if ai is the
ith column of A, then for each i, we have ai = ci u. Let vT⇥ = c1 c2 · · · ⇤ cn . Then A = uvT . For
the converse, suppose that A = uvT . Suppose that vT = c1 c2 · · · cn . Then
⇥
⇤
A = uvT = c1 u · · · cn u .
Thus every column of A is a scalar multiple of u, so that col(A) = span(u) and thus rank(A) = 1.
63. If rank(A) = r, then a basis for col(A) consists of r vectors, say u1 , u2 , . . . , ur 2 Rm . So if ai is the
ith column of A, then for each i we have
⇥
Now let viT = c1i c2i
⇥
A = a1 a2 · · ·
ai = ci1 u1 + ci2 u2 + · · · + cir ur , i = 1, 2, . . . , n.
⇤
· · · cni , for i = 1, 2, . . . , r. Then so that
⇤ ⇥
⇤
an = c11 u1 + c12 u2 + · · · + c1r ur · · · cn1 u1 + cn2 u2 + · · · + cnr ur
⇥
⇤ ⇥
⇤ ⇥
⇤
= c11 u1 · · · cn1 u1 + c12 u2 · · · cn2 u2 + c1r ur · · · cnr ur
= u1 · v1T + u2 · v2T + · · · + ur · vrT .
By exercise 62, each ui · viT is a matrix of rank 1, so A is a sum of r matrices of rank 1.
64. Let A = {u1 , . . . , ur } be a basis for col(A) and B = {v1 , . . . , vs } be a basis for col(B). Then rank(A) =
r and rank(B) = s. Let ai and bi be the columns of A and B respectively, for i = 1, 2, . . . , n. Then if
x 2 col(A + B),
0
1
!
!
n
n
r
s
r
n
s
n
X
X
X
X
X
X
X
X
@
A=
x=
c1 (ai + bi ) =
↵ij uj +
ci ↵ij uj +
ci ik vk ,
ik vk
i=1
i=1
j=1
k=1
j=1
i=1
k=1
i=1
so that x is a linear combination of A and B. Thus A [ B spans col(A + B). But this implies that
col(A + B) ✓ span(A [ B) = span(A) + span(B) = col(A) + col(B).
Since rank(A + B) is the dimension of col(A + B) and rank(A) + rank(B) is the dimension of col(A)
plus the dimension of col(B), we are done.
192
CHAPTER 3. MATRICES
P
65. We show that col(A) ✓ null(A) and then use the Rank Theorem. Suppose x =
ci ai 2 col(A). Let
ai be the ith column of A; then ai = Aei where the ei are the standard basis vectors for Rn . Then
⇣X
⌘ X
X
X
X
Ax = A
ci ai =
ci (Aai ) =
ci (A (Aei )) =
c i A2 e i =
ci (Oei ) = 0.
Thus any vector in the column space of A is taken to zero by A, so it is in the null space of A. Thus
col(A) ✓ null(A). But then
rank(A) = dim(col(A))  dim(null(A)) = nullity(A),
so by the Rank Theorem,
rank(A) + rank(A)  rank(A) + nullity(A) = n
)
2 rank(A)  n
)
rank(A) 
66. (a) Note first that since x is n ⇥ 1, then xT is 1 ⇥ n and thus xT Ax is 1 ⇥ 1. Thus xT Ax
Now since A is skew-symmetric,
xT Ax = xT Ax
T
= x T AT x T
T
= xT ( A) x =
T
n
.
2
= xT Ax.
xT Ax.
So xT Ax is equal to its own negation, so it must be zero.
(b) By Theorem 3.27, to show I + A is invertible it suffices to show that null(I + A) = 0. Choose x
with (I + A)x = 0. Then xT (I + A)x = xT 0 = x · 0 = 0, so that
0 = xT (I + A)x = xT Ix + xT Ax = xT Ix + xT 0 = xT Ix = xT x = x · x
since Ax = 0. But x · x = 0 implies that x = 0. So any vector in null(I + A) is the zero vector
and thus I + A has a zero null space, so it is invertible.
3.6
Introduction to Linear Transformations
1. We must compute T (u) and T (v), so we take the product of the matrix A of T and each of these two
vectors:



2
1 1
0
T (u) = Au =
=
3
4 2
11



2
1 1
8
T (v) = Av =
=
.
3
4 2
1
2. We must compute T (u) and T (v), so we
vectors:
2
3
T (u) = Au = 41
1
2
3
T (v) = Av = 41
1
take the product of the matrix A of T and each of these two
3
2
3 2 3
1 
3·1 1·2
1
1
25
= 41 · 1 + 2 · 25 = 455
2
4
1·1+4·2
9
3
2
3 2 3

1
3 · 3 1 · ( 2)
11
3
25
= 41 · 3 + 2 · ( 2)5 = 4 15
2
4
1 · 3 + 4 · ( 2)
5
3. Use the remark following Example 3.55. T is a linear transformation if and only if


x
x
T (c1 v1 + c2 v2 ) = c1 T (v1 ) + c2 T (v2 ), where v1 = 1 , v2 = 2 .
y1
y2
3.6. INTRODUCTION TO LINEAR TRANSFORMATIONS
193
Then
T (c1 v1 + c2 v2 ) = T
✓
c1



◆
x1
x
+ c2 2
y1
y2
c x + c 2 x2
=T 1 1
c 1 y1 + c 2 y2

(c1 x1 + c2 x2 ) + (c1 y1 + c2 y2 )
=
(c1 x1 + c2 x2 ) (c1 y1 + c2 y2 )

(c1 x1 + c1 y1 ) + (c2 x2 + c2 y2 )
=
(c1 x1 c1 y1 ) + (c2 x2 c2 y2 )


c 1 x 1 + c 1 y1
c x + c 2 y2
=
+ 2 2
c 1 x 1 c 1 y1
c 2 x 2 c 2 y2


x + y1
x + y2
= c1 1
+ c2 2
x 1 y1
x 2 y2
= c1 T (v1 ) + c2 T (v2 )
and thus T is linear.
4. Use the remark following Example 3.55. T is a linear transformation if and only if
T (c1 v1 + c2 v2 ) = c1 T (v1 ) + c2 T (v2 ), where v1 =


x1
x
, v2 = 2 .
y1
y2
Then
T (c1 v1 + c2 v2 ) = T
✓

c1


◆
x1
x2
+ c2
y1
y2
c x + c 2 x2
=T 1 1
c 1 y1 + c 2 y2
2
3
(c1 y1 + c2 y2 )
= 4 (c1 x1 + c2 x2 ) + 2(c1 y1 + c2 y2 ) 5
3(c1 x1 + c2 x2 ) 4(c1 y1 + c2 y2 )
2
3 2
3
c 1 y1
c 2 y2
= 4 c1 x1 + 2c1 y1 5 + 4 c2 x2 + 2c2 y2 5
3c1 x1 4c1 y1
3c2 x2 4c2 y2
2
3
2
3
y1
y2
= c1 4 x1 + 2y1 5 + c2 4 x2 + 2y2 5
3x1 4y1
3x2 4y2
= c1 T (v1 ) + c2 T (v2 )
and thus T is linear.
5. Use the remark following Example 3.55. T is a linear transformation if and only if
2
3
2 3
x1
x2
T (c1 v1 + c2 v2 ) = c1 T (v1 ) + c2 T (v2 ), where v1 = 4 y1 5 , v2 = 4 y2 5 .
z1
z2
194
CHAPTER 3. MATRICES
Then
0
2
3
2 31
x1
x2
T (c1 v1 + c2 v2 ) = T @c1 4 y1 5 + c2 4 y2 5A
z1
z2
2
3
c 1 x1 + c 2 x2
= T 4 c 1 y1 + c 2 y2 5
c 1 z1 + c 2 z2

(c1 x1 + c2 x2 ) (c1 y1 + c2 y2 ) + (c1 z1 + c2 z2 )
=
2(c1 x1 + c2 x2 ) + (c1 y1 + c2 y2 ) 3(c1 z1 + c2 z2 )

(c1 x1 c1 y1 + c1 z1 ) + (c2 x2 c2 y2 + c2 z2 )
=
(2c1 x1 + c1 y1 3c1 z1 ) + (2c2 x2 + c2 y2 3c2 z2 )


c 1 x 1 c 1 y 1 + c 1 z1
c 2 x 2 c 2 y 2 + c 2 z2
=
+
2c1 x1 + c1 y1 3c1 z1
2c2 x2 + c2 y2 3c2 z2


x 1 y 1 + z1
x 2 y 2 + z2
= c1
+ c2
2x1 + y1 3z1
2x2 + y2 3z2
= c1 T (v1 ) + c2 T (v2 )
and thus T is linear.
6. Use the remark following Example 3.55. T is a linear transformation if and only if
2 3
2 3
x1
x2
T (c1 v1 + c2 v2 ) = c1 T (v1 ) + c2 T (v2 ), where v1 = 4 y1 5 , v2 = 4 y2 5 .
z1
z2
Then
0
2
3
2 31
x1
x2
T (c1 v1 + c2 v2 ) = T @c1 4 y1 5 + c2 4 y2 5A
z1
z2
2
3
c 1 x1 + c 2 x2
= T 4 c 1 y1 + c 2 y2 5
c 1 z1 + c 2 z2
2
3
(c1 x1 + c2 x2 ) + (c1 z1 + c2 z2 )
= 4 (c1 y1 + c2 y2 ) + (c1 z1 + c2 z2 ) 5
(c1 x1 + c2 x2 ) + (c1 y1 + c2 y2 )
2
3 2
3
c 1 x 1 + c 1 z1
c 2 x 2 + c 2 z2
= 4 c 1 y 1 + c 1 z1 5 + 4 c 2 y 2 + c 2 z2 5
c 1 x 1 + c 1 y1
c 2 x 2 + c 2 y2
2
3
2
3
x 1 + z1
x 2 + z2
= c 1 4 y 1 + z1 5 + c 2 4 y 2 + z2 5
x 1 + y1
x 2 + y2
= c1 T (v1 ) + c2 T (v2 )
and thus T is linear.
7. For example,












1
0
0
2
0
0
1+2
3
0
0
0
0
T
= 2 =
, T
= 2 =
, but T
=T
= 2 =
6=
+
.
0
1
1
0
2
4
0
0
3
9
1
4
8. For example,












1
| 1|
1
1
|1|
1
1+1
0
|0|
0
1
1
T
=
=
, T
=
=
, but T
=
=
=
6=
+
.
0
|0|
0
0
|0|
0
0
0
|0|
0
0
0
3.6. INTRODUCTION TO LINEAR TRANSFORMATIONS
195
9. For example
T









2
2·2
4
2+2
4
4·4
16
4
4
, but T
=T
=
=
6=
+
.
=
=
2
2+2
4
2+2
4
4+4
8
4
4
10. For example,









1
1+1
2
1+1
2
2+1
3
2
2
T
=
=
, but T
=T
=
=
6=
+
.
0
0 1
1
0+0
0
0 1
1
1
1
11. By Theorem 3.31, the standard matrix of T is the matrix A whose ith column is T (ei ). Since






1
1+0
1
0
0+1
1
T (e1 ) = T
=
=
, T (e2 ) = T
=
=
,
0
1 0
1
1
0 1
1
the matrix of T is

A=
1
1
1
.
1
12. By Theorem 3.31, the standard matrix of T is the matrix A whose ith column is T (ei ). Since
2
3 2 3
2
3 2 3


0
0
1
1
1
0
T (e1 ) = T
= 4 1 + 2 · 0 5 = 415 , T (e2 ) = T
= 4 0 + 2 · 1 5 = 4 25 ,
0
1
3·1 4·0
3
3·0 4·1
4
the matrix of T is
2
3
1
25 .
4
0
A = 41
3
13. By Theorem 3.31, the standard matrix of T is the matrix A whose ith column is T (ei ). Since
2 3


1
1 0+0
1
T (e1 ) = T 405 =
=
,
2·1+0 3·0
2
0
2 3


0
0 1+0
1
T (e2 ) = T 415 =
=
,
2·0+1 3·0
1
0
2 3


0
0 0+1
1
T (e3 ) = T 405 =
=
,
2·0+0 3·1
3
1
the matrix of T is
A=

1
2
1
1
1
.
3
14. By Theorem 3.31, the standard matrix of T is the matrix A whose ith column is T (ei ). Since
2 3 2
3 2 3
1
1+0
1
T (e1 ) = T 405 = 40 + 05 = 405 ,
0
1+0
1
2 3 2
3 2 3
0
0+0
0
T (e2 ) = T 415 = 41 + 05 = 415 ,
0
0+1
1
2 3 2
3 2 3
0
0+1
1
T (e3 ) = T 405 = 40 + 15 = 415 ,
1
0+0
0
196
CHAPTER 3. MATRICES
the matrix of T is
2
1
A = 40
1
3
1
15 .
0
0
1
1
15. Reflecting a vector in the x-axis means negating the x-coordinate. So using the method of Example
3.56,




x
x
1
0
,
F
=
=x
+y
y
y
0
1
and thus F is a matrix transformation with matrix

1
F =
0
0
.
1
It follows that F is a linear transformation.
16. Example 3.58 shows that a
" #
1
to
0
rotation of 45 about the origin sends
"
# "p #
" #
"
# "
2
cos 45
0
cos
135
= p22 and
to
=
sin 45
1
sin 135
2
so that R is a matrix transformation with matrix
"p
R=
p
2
p2
2
2
2
p2
2
2
#
p
2
p2
2
2
#
,
.
It follows that R is a linear transformation.
17. Since D stretches by a factor of 2 in the x-direction and a factor of 3 in the y-direction, we see that




x
2x
2 0 x
D
=
=
.
y
3y
0 3 y
Thus D is a matrix transformation with matrix

2
0
0
.
3
It follows that D is a linear transformation.
18. Since P = (x, y) projects onto the line whose direction vector is h1, 1i, it follows that the projection of
P onto that line is
✓
◆
⌧
hx, yi · h1, 1i
x+y x+y
projh1,1i hx, yi =
h1, 1i =
,
.
h1, 1i · h1, 1i
2
2
Thus
" # "
#
" #
" # "
x+y
1
1
1
x
2
2
T
= x+y = x 1 + y 21 = 21
y
2
2
2
2
1
2
1
2
#" #
x
.
y
Thus T is a matrix transformation with the matrix above, and it follows that T is a linear transformation.
19. Let
A1 =

k
0
0
,
1
A2 =

1
0
0
,
k
B=

0
1
1
,
0
C1 =

1
0
k
,
1
C2 =

1
k
0
1
3.6. INTRODUCTION TO LINEAR TRANSFORMATIONS
197
Then


x
k
A1
=
y
0


x
1
A2
=
y
0


x
0
B
=
y
1


x
1
C1
=
y
0


x
1
C2
=
y
k
0
1
0
k
1
0
k
1
0
1


x
kx
: A1 stretches vectors horizontally by a factor of k
=
y
y


x
x
=
: A2 stretches vectors vertically by a factor of k
y
ky


x
y
=
: B reflects vectors in the line y = x
y
x


x
x + ky
=
: C1 extends vectors horizontally by ky
y
y


x
x
=
: C1 extends vectors vertically by kx
y
y + kx
(C1 and C2 are called shears.) The e↵ect of each of these transformations, with k = 2, on the unit
square is
Original
A1
3
3
2
2
2
1
1
1
0
0
1
2
0
3
0
1
B
2
0
3
C1
3
3
2
2
2
1
1
1
0
0
1
2
0
3
0
1
A2
3
0
1
0
3
3
2
3
C2
3
2
2
0
1
Note that reflection across the line y = x leaves the square unchanged, since it is symmetric about that
line.
20. Using the formula from Example 3.58, we compute the matrix for a counterclockwise rotation of 120 :
R120 (e1 ) =
"
cos 120
sin 120
#
=
"
1
p2
3
2
#
and R120 (e2 ) =
"
sin 120
cos 120
So by Theorem 3.31, we have
h
A = R120 (e1 )
i
R120 (e2 ) =
"
1
p2
3
2
p
3
2
1
2
#
.
#
=
"
p
3
2
1
2
#
.
198
CHAPTER 3. MATRICES
21. A clockwise rotation of 30 is a rotation of
compute the matrix for that rotation:
R
30
(e1 ) =
"
cos( 30 )
sin( 30 )
#
=
30 . Then using the formula from Example 3.58, we
"p #
3
2
1
2
and R120 (e2 ) =
"
sin( 30 )
cos( 30 )
#
=
"
1
p2
3
2
#
.
So by Theorem 3.31, we have
h
A= R
30
(e1 )
22. A direction vector of the line y = 2x is d =
P` (e1 ) =
✓
d · e1
d·d
◆
R

30
i
(e2 ) =
"p
3
2
1
2
1
p2
3
2
#
.
1
, so
2
" # " #
" # " #
✓
◆
1
2
1
1
1
d · e2
2
5
d=
= 2 and P` (e2 ) =
d=
= 54 .
1+4 2
d·d
1+4 2
5
5
Thus by Theorem 3.31, we have
h
i
A = P` (e1 )
23. A direction vector of the line y =
P` (e1 ) =
✓
d · e1
d·d
◆
1
d=
1+1
"
x is d =
# "
1
=
1

1
2
1
2
P` (e2 ) =
"
1
5
2
5
2
5
4
5
#
.
1
, so
1
#
and P` (e2 ) =
✓
d · e2
d·d
◆
1
d=
1+2
Thus by Theorem 3.31, we have
h
i
A = P` (e1 )
P` (e2 ) =
"
1
2
1
2
24. Reflection in the line y = x is given by





x
y
0
1
0
R
=
=x
+y
=
y
x
1
0
1
so that the matrix of R is

0
1
1
2
1
2
1
0
#
.

x
,
y
1
.
0
25. Reflection in the line y =
x is given by





x
y
0
1
0
R
=
=x
+y
=
y
x
1
0
1
so that the matrix of R is

0
1
1
.
0
1
0

x
,
y
"
# "
1
=
1
1
2
1
2
#
.
3.6. INTRODUCTION TO LINEAR TRANSFORMATIONS
199
26. (a) Consider the following diagrams:
a!b
!
c F! !b"
a
!
F! !b"
F! !b"
cb
b
F! !a"
F! !a ! b"
b
F! !a"
From the left-hand graph, we see that F` (cb) = cF` (b), and the right-hand graph shows that
F` (a + b) = F` (a) + F` (b).
(b) Let P` (x) be the projection of x on the line `; this is the image of x under the linear transformation
P` . Then from the figure in the text, since the diagonals of the parallelogram shown bisect each
other, and the point furthest from the origin on ` is x + F` (x), we have
x + F` (x) = 2P` (x), so that F` (x) = 2P` (x)
x.
Now, Example 3.59 gives us the matrix of P` , where ` has direction vector d =
x
"
#
"
#
d21
d1 d2
1 0
2
= 2
x
x
d1 + d22 d1 d2
d22
0 1
"
#
"
d21
d1 d2
d21 + d22
2
1
=
d21 + d22 d1 d2
d21 + d22
d22
0
"
#
d21 d22
2d1 d2
1
= 2
x,
2
d1 + d2 2d1 d2
d21 + d22

d1
, so we have
d2
F` (x) = 2P` (x)
0
2
d1 + d22
#!
x
which shows that the matrix of F` is as claimed in the exercise statement.
(c) If ✓ is the angle that ` forms with the positive x-axis, we may take


d
cos ✓
d= 1 =
d2
sin ✓
as the direction vector. Substituting those values into the formula for F` from part (b) gives
 2
1
d1 d22
2d1 d2
F` = 2
2
2d
d
d21 + d22
d1 + d2
1 2
 2
1
cos ✓ sin2 ✓
2 cos ✓ sin ✓
=
2
2
2
2
cos
✓
sin
✓
cos
✓ + sin2 ✓
cos ✓ + sin ✓
 2
cos ✓ sin2 ✓
2 cos ✓ sin ✓
=
,
2 cos ✓ sin ✓
cos2 ✓ + sin2 ✓
200
CHAPTER 3. MATRICES
since cos2 ✓ + sin2 ✓ = 1. Now, cos2 ✓
above simplifies to
sin2 ✓ = cos 2✓, and 2 cos ✓ sin ✓ = sin 2✓, so the matrix
F` =

cos 2✓
sin 2✓
sin 2✓
.
cos 2✓
27. By part (b) of the previous exercise, since the line y = 2x has direction vector d =
the standard matrix of this reflection
"
1 4
1
F` (x) =
1+4 2·2
# "
2·2
=
1+4
3
5
4
5
4
5
3
5
#

1
, we have for
2
.

p
p
1
3
sin ✓
28. The direction vector of this line is p , so that cos
=
3, and thus ✓ = ⇡3 = 60 . So by part
✓
1 =
3
(c) of the previous exercise, the standard matrix of this reflection is
"
# "
# " p
#
3
1
cos 2✓
sin 2✓
cos 120
sin 120
2
p2
F` (x) =
=
=
.
3
1
sin 2✓
cos 2✓
sin 120
cos 120
2
2
29. Since
T

2
3
x1
x1
= 4 2x1 x2 5 ,
x2
3x1 + 4x2
2
3
2 3
2y1 + y3
y1
6 3y2 y3 7
7
S 4 y2 5 = 6
4 y1 y2 5 ,
y3
y1 + y 2 + y3
we substitute the result of T into the equation for S:
2
3

x1
x
(S T ) 1 = S 4 2x1 x2 5
x2
3x1 + 4x2
2
3
2x1 + (3x1 + 4x2 )
6 3(2x1 x2 ) (3x1 + 4x2 ) 7
7
=6
4
5
x1 (2x1 x2 )
x1 + (2x1 x2 ) + (3x1 + 4x2 )
2
3
5x1 + 4x2
63x1 7x2 7
7
=6
4 x1 + x2 5
6x1 + 3x2
2
3
5
4 
6 3
77
7 x1 .
=6
4 1
15 x2
6
3
Thus the standard matrix of S
T is
2
3
4
77
7.
15
3
5
6 3
A=6
4 1
6
30. (a) By direct substitution, we have

✓ 
◆



x
x
x
x2
2(x1 x2 )
2x1 2x2
(S T ) 1 = S T 1
=S 1
=
=
.
x2
x2
x1 + x2
(x1 + x2 )
x1 x2
Thus
[S
T] =

2
1
2
.
1
3.6. INTRODUCTION TO LINEAR TRANSFORMATIONS
201
(b) Using matrix multiplication, we first find the standard matrices for S and T , which are


2
0
1
1
[S] =
, [T ] =
.
0
1
1
1
Then

2
[S][T ] =
0
Then [S
T ] = [S][T ].
0
1

1
1
1
=
1

2
1
2
.
1
31. (a) By direct substitution, we have

✓ 
◆



x1
x1
x1 + 2x2
(x1 + 2x2 ) + 3( 3x1 + x2 )
8x1 + 5x2
(S T )
=S T
=S
=
=
.
x2
x2
3x1 + x2
(x1 + 2x2 ) ( 3x1 + x2 )
4x1 + x1
Thus
[S
T] =

8
4
5
.
1
(b) Using matrix multiplication, we first find the standard matrices for S and T , which are


1
3
1 2
[S] =
, [T ] =
.
1
1
3 1
Then
[S][T ] =
Then [S
T ] = [S][T ].

1
1
3
1

1
3
2
=
1

8
4
5
.
1
32. (a) By direct substitution, we have
(S
Thus
2
3 2
3
✓ 
◆

x2 + 3( x1 )
3x1 + x2
x1
x1
x2
T)
=S T
=S
= 4 2x2 x1 5 = 4 x1 + 2x2 5 .
x2
x2
x1
x2 ( x1 )
x1 + x2

[S
2
3
T] = 4 1
1
3
1
25 .
1
(b) Using matrix multiplication, we first find the standard matrices for S and T , which are
2
3

1
3
0 1
4
5
1 , [T ] =
[S] = 2
.
1 0
1
1
Then
2
Then [S
1
[S][T ] = 42
1
T ] = [S][T ].
3
3 
0
15
1
1
2
3
1
=4 1
0
1
3
1
25 .
1
33. (a) By direct substitution, we have
(S
2
3
0
x1
T ) 4x2 5 = S @T
x3
2
31

x1
4 x2 5 A = S x1 + x2 x3
2x1 x2 + x3
x3

4(x1 + x2 x3 ) 2(2x1
=
(x1 + x2 x3 ) + (2x1

x2 + x3 )
6x2 6x3
=
x2 + x3 )
x1 2x2 + 2x3
202
CHAPTER 3. MATRICES
Thus
[S

T] =
0
1
6
2
6
.
2
(b) Using matrix multiplication, we first find the standard matrices for S and T , which are


4
2
1
1
1
[S] =
, [T ] =
.
1
1
2
1
1
Then
[S][T ] =
Then [S

4
1
2
1

1
2

1
0
=
1
1
1
1
6
2
6
.
2
T ] = [S][T ].
34. (a) By direct substitution, we have
2 3
0 2 31
2
3 2
3

x1
x1
(x1 + 2x2 ) (2x2 x3 )
x1 + x3
x + 2x2
(S T ) 4x2 5 = S @T 4x2 5A = S 1
= 4 (x1 + 2x2 ) + (2x2 x3 ) 5 = 4x1 + 4x2 x3 5
2x2 x3
x3
x3
(x1 + 2x2 ) + (2x2 x3 )
x1 x3
Thus
[S
2
1
T] = 4 1
1
(b) Using matrix multiplication, we first find
2
1
[S] = 4 1
1
Then
Then [S
2
T ] = [S][T ].
the standard matrices for S and T , which are
3

1
1 2
0
5
1 , [T ] =
.
0 2
1
1
3
1 
1
15
0
1
1
[S][T ] = 4 1
1
3
1
15 .
1
0
4
0
2
1
0
=4 1
1
1
2
2
3
1
15 .
1
0
4
0
35. (a) By direct substitution, we have
2 3
0 2 31
2
3 2
3 2
3
x1
x1
x1 + x2
(x1 + x2 ) (x2 + x3 )
x1 x3
(S T ) 4x2 5 = S @T 4x2 5A = S 4x2 + x3 5 = 4 (x2 + x3 ) (x1 + x3 ) 5 = 4 x1 + x2 5 .
x3
x3
x1 + x3
(x1 + x2 ) + (x1 + x3 )
x2 + x3
Thus
2
1
T] = 4 1
0
[S
(b) Using matrix multiplication, we first find the
2
1
1
1
[S] = 4 0
1
0
Then
Then [S
2
1
[S][T ] = 4 0
1
T ] = [S][T ].
1
1
0
0
1
1
3
1
05 .
1
standard matrices for S and T , which are
3
2
3
0
1 1 0
15 , [T ] = 40 1 15 .
1
1 0 1
32
0 1
15 40
1 1
1
1
0
3 2
0
1
15 = 4 1
1
0
0
1
1
3
1
05 .
1
3.6. INTRODUCTION TO LINEAR TRANSFORMATIONS
203
36. A counterclockwise rotation T through 60 is given by (see Example 3.58)
"
# "
p #
3
1
cos 60
sin 60
2
2
[T ] =
= p3
,
1
sin 60
cos 60
2
2
and reflection S in y = x is reflection in a line with direction vector d =
(see Exercise 26(b))

1 0
[S] =
2 2

2
0
=
0
1

1
, which has the matrix
1
1
.
0
Thus by Theorem 3.32, the composite transformation is given by
"
#"
"p
p #
3
3
1
0 1
2
2
p2
[S T ] = [S][T ] =
=
3
1
1
1 0
2
2
2
1
p2
3
2
#
.
37. Reflection T in the y-axis is taking the negative of the x-coordinate, so it has matrix

1 0
[T ] =
,
0 1
while clockwise rotation S through 30 has matrix (see Example 3.58)
"
# "p
#
3
1
cos( 30 )
sin( 30 )
2
2
p
[S] =
=
.
3
1
sin( 30 ) cos( 30 )
2
2
Thus by Theorem 3.32, the composite transformation is given by
"p
#"
# "
3
1
1
0
[S T ] = [S][T ] = 21 p23
=
0 1
2
2
38. Clockwise rotation T through 45 has matrix (see Example 3.58)
"
# " p
2
cos 45
sin 45
p2
[T ] =
=
2
sin 45 cos 45
2
p
3
2
1
2
p
2
p2
2
2
#
1
p2
3
2
#
.
,
while projection S onto the y-axis has matrix (see Example 3.59, or note that this transformation
simply sets the x-coordinate to zero)

0 0
[S] =
.
0 1
Thus by Theorem 3.32, the composite transformation is given by
" p
#" p
p #"
2
2
2
0 0
2
2
p
p
p2
[T S T ] = [T ][S][T ] =
2
2
2
0 1
2
2
2
39. Reflection T in the line y = x, with direction vector d =

1 0
[T ] =
2 2


2
0
=
0
1
p
2
p2
2
2
#
=
"
1
2
1
2
1
2
1
2
#
.
1
, is given by (see Exercise 26(b))
1
1
.
0
Counterclockwise rotation S through 30 has matrix (see Example 3.58)
"
# "p
#
3
1
cos 30
sin 30
[S] =
= 21 p23 .
sin 30 cos 30
2
2
204
CHAPTER 3. MATRICES
Finally, reflection R in the line y =
x, with direction vector d =
26(b))

1
, is given by (see Exercise
1


1
0
2
0
1
=
.
2
0
1
0
2
Then by Theorem 3.32, the composite transformation is given by
"
#" p
#"
# "
3
1
0
1
0 1
2
2
p
[R S T ] = [R][S][T ] =
=
3
1
1
0
1 0
2
2
[R] =
p
3
2
1
2
1
p2
3
2
#
.
40. Using the formula for rotation through an angle ✓, from Example 3.58, we have


cos ↵
sin ↵ cos
sin
[R↵ R ] = [R↵ ][R ] =
sin ↵
cos ↵ sin
cos

cos ↵ cos
sin ↵ sin
cos ↵ sin
sin ↵ cos
=
sin ↵ cos + cos ↵ sin
sin ↵ sin + cos ↵ cos

cos(↵ + )
sin(↵ + )
=
sin(↵ + )
cos(↵ + )
= R↵+ .


cos ↵
cos
41. Assume line ` has direction vector l =
and line m has direction vector
. Let ✓ =
sin ↵
sin
be the angle from ` to m (note that if ↵ > , then ✓ is negative). Then by Exercise 26(c),


cos 2
sin 2
cos 2↵
sin 2↵
F m F` =
sin 2
cos 2
sin 2↵
cos 2↵

cos 2 cos 2↵ + sin 2 sin 2↵ cos 2 sin 2↵ sin 2 cos 2↵
=
sin 2 cos 2↵ cos 2 sin 2↵ cos 2 cos 2↵ + sin 2 sin 2↵

cos(2(↵
)) sin(2(↵
))
=
sin(2(↵
)) cos(2(↵
))

cos(2(
↵))
sin(2(
↵))
=
sin(2(
↵)) cos(2(
↵))

cos 2✓
sin 2✓
=
sin 2✓
cos 2✓
↵
= R2✓ .
42. (a) Let P be the projection onto a line with direction vector d =
P is
P =
and thus
P
 2
1
d1
d21 + d22 d1 d2

d1
. Then the standard matrix of
d2
d1 d2
,
d22
 2
 2
1
1
d1
d1 d2
d1
d1 d2
d22 d21 + d22 d1 d2
d22
d21 + d22 d1 d2
 4
1
d1 + d21 d22
d31 d2 + d1 d32
= 2
3
3
2
2
d21 d22 + d42
(d1 + d2 ) d1 d2 + d1 d2
 2 2
1
d1 (d1 + d22 ) d1 d2 (d21 + d22 )
= 2
2
2
(d1 + d2 ) d1 d2 (d21 + d22 ) d22 (d21 + d22 )
 2
1
d1
d1 d2
= 2
d22
d1 + d22 d1 d2
P = P2 =
= P.
3.6. INTRODUCTION TO LINEAR TRANSFORMATIONS
205
Note that we have proven this fact before, in a di↵erent form. Since P (v) = projd v, then
(P
P )(v) = projd (projd v) = projd v = P (v)
by Exercise 70(a) in Section 1.2.
(b) Let P be any projection matrix. Using the form of P from the previous exercise, we first prove
that P 6= I. Suppose it were. Then d1 d2 = 0, so that either d1 = 0 or d2 = 0. But then either d21
or d22 would be zero, so that P could not be the identity matrix. So P is a non-identity matrix
such that P 2 = P . Then P 2 P = P (P I) = O. Since P 6= I, we see that P I 6= O, so
there is some k such that (P I)ek 6= 0 (otherwise, every row of P I would be orthogonal
to every standard basis vector, so would be zero, so that P I would be the zero matrix). Let
x = (P I)ek . But then
P x = P (P I)ek = Oek = 0,
so that x 2 null(P ). Since the null space of P contains a nonzero vector, P is not invertible by
Theorem 3.27.
43. Let `, m, and n be lines through the origin with angles from the x-axis of ✓, ↵, and respectively.
Then



cos 2✓
sin 2✓ cos 2↵
sin 2↵ cos 2
sin 2
Fn F m F` =
sin 2✓
cos 2✓ sin 2↵
cos 2↵ sin 2
cos 2


cos 2✓ cos 2↵ + sin 2✓ sin 2↵ cos 2✓ sin 2↵ sin 2✓ cos 2↵ cos 2
sin 2
=
sin 2✓ cos 2↵ cos 2✓ sin 2↵ sin 2✓ sin 2↵ + cos 2✓ cos 2↵ sin 2
cos 2


cos 2(✓ ↵)
sin 2(✓ ↵) cos 2
sin 2
=
sin 2(✓ ↵) cos 2(✓ ↵)
sin 2
cos 2

cos 2(✓ ↵) cos 2
sin 2(✓ ↵) sin 2
cos 2(✓ ↵) sin 2 + sin 2(✓ ↵) cos 2
=
sin 2(✓ ↵) cos 2 + cos 2(✓ ↵) sin 2
sin 2(✓ ↵) sin 2
cos 2(✓ ↵) cos 2

cos 2(✓ ↵ + )
sin 2(✓ ↵ + )
=
.
sin 2(✓ ↵ + )
cos 2(✓ ↵ + )
This is a reflection through a line making an angle of ✓
↵+
with the x-axis.
44. If ` is a line with direction vector d through the point P , then it has equation x = p + td. Since T is
linear, we have
T (x) = T (p + td) = T (p) + tT (d),
so that the equation of the resulting graph has equation T (x) = T (p) + tT (d). If T (d) = 0, then the
image of ` has equation T (x) = T (p) so that it is the single point T (p). Otherwise, the image of ` is
the line `0 with direction vector T (d) passing through T (p).
45. Suppose that we start with two parallel lines
` : x = p + td,
m : x = q + sd.
Then the images of these lines have equations
`0 : T (p) + tT (d),
m0 : T (q) + sT (d).
Then
• If T (d) 6= 0 and T (p) 6= T (q), then `0 and m0 are distinct parallel lines.
• If T (d) 6= 0 and T (p) = T (q), then `0 and m0 are the same line.
• If T (d) = 0 and T (p) 6= T (q), then `0 and m0 are distinct points T (p) and T (q).
• If T (d) = 0 and T (p) 6= T (q), then `0 and m0 are the single point T (p).
206
CHAPTER 3. MATRICES
46. Using the linear transformation T from Exercise 3, we have
T

1
=
1



1
2
T
=
,
1
0
0
,
2
T


1
0
=
,
1
2
T

1
=
1

2
.
0
The image is the square below:
2
C
1
D
B
!2
1
!1
2
!1
A
!2
47. Using the linear transformation D from Exercise 17, we have
D

1
=
1

2
,
3
D


1
2
=
,
1
3
D

1
=
1

2
,
3
D

1
=
1

2
.
3
The image is the rectangle below:
3
A
B
2
1
!2
1
!1
2
!1
!2
D
C
48. Using the linear transformation P from Exercise 18, we have
P


1
0
=
,
1
0
The image is the line below:


1
1
P
=
,
1
1
P


1
0
=
,
1
0
P

1
=
1

1
.
1
3.6. INTRODUCTION TO LINEAR TRANSFORMATIONS
207
1.
B
0.5
A
!1.
1.
0.5
C
!0.5
!0.5
D
!1.
49. Using the projection P from Exercise 22, we have
" # " #
" # " #
" # "
1
3
1
1
1
5
5
P
= 2 , P
= 6 , P
=
1
1
1
5
5
1
5
2
5
#
,
P
"
# "
1
=
1
3
5
6
5
#
.
The image is the line below:
B
1.
0.5
A
!1.
1.
0.5
!0.5
C
!0.5
!1.
D
2
, we have
1




1
1
1
3
T
=
, T
=
.
1
4
1
2
50. Using the transformation T from Exercise 31, with matrix
T


1
1
=
,
1
4

1
T
=
1

3
,
2

1
3
The image is the parallelogram below:
4
A
3
2
D
1
!4
!3
!2
1
!1
2
3
4
!1
B
!2
!3
C
!4
51. Using the transformation resulting from Exercise 37, which we call Q, we have
" # "p
#
" # " p
#
" # " p
#
" # "
3+1
3+1
3 1
1
1
1
1
2
p2
p2
Q
= p3+1
, Q
=
, Q
=
, Q
=
3 1
3 1
1
1
1
1
2
2
2
The image is the parallelogram below:
p
3 1
p2
3+1
2
#
.
208
CHAPTER 3. MATRICES
A
1
D
1
!1
B
!1
C
52. If ` has direction vector d, then using standard properties of the dot product and of scalar multiplication, we get
✓
◆
✓
◆
✓✓
◆ ◆
d · cv
c(d · v)
d·v
P` (cv) = projd (cv) =
d=
d=c
d = cP` (v).
d·d
d·d
d·d
53. First suppose that T is linear. Then applying property 1 first, followed by property 2, we have
T (c1 v1 + c2 v2 ) = T (c1 v1 ) + T (c2 v2 ) = c1 T (v1 ) + c2 T (v2 ).
For the reverse direction, assume that T (c1 v1 + c2 v2 ) = c1 T (v1 ) + c2 T (v2 ). To prove property 1, set
c1 = c2 = 1; this gives
T (v1 + v2 ) = T (v1 ) + T (v2 ).
For property 2, set c1 = c, v1 = v, and c2 = 0; this gives
T (cv) = T (cv + 0v2 ) = cT (v) + 0T (v2 ) = cT (v).
Thus T is linear.
54. Let range(T ) be the range of T and col([T ]) be the column space of [T ]. We show that range(T ) =
col([T ]) by showing that they are both equal to span({T (ei )}).
To see
P that range(T ) = span({T (ei )}), note that x 2 range(T ) if and only if x = T (v) for some
v=
ci ei , so that
X
X
x = T (v) = T (
c i ei ) =
ci T (ei ),
which is the case if and only if x 2 span({T (ei )}). Thus range(T ) = span({T (ei )}).
⇥
Next, we show
that col([T ]) =⇤ span({T (ei )}). By Theorem 2, [T ] = T (e1 ) T (e2 )
⇥
Thus if vT = v1 v2 · · · vn , then
···
⇤
T (en ) .
T (v) = v1 T (e1 ) + v2 T (e2 ) + · · · + vn T (en ).
Then
x 2 col([T ]) () x = T (v) = T (
so that col([T ]) = span({T (ei )}).
X
vi T (ei )) =
X
vi T (ei ),
55. The Fundamental Theorem of Invertible Matrices implies that any invertible 2 ⇥ 2 matrix is a product
of elementary matrices. The matrices in Exercise 19 represent the three types of elementary 2 ⇥ 2
matrices, so TA must be a composition of the three types of transformations represented in Exercise
19.
3.7. APPLICATIONS
3.7
209
Applications
1. x1 = P x0 =

0.5
0.5
0.3
0.7


0.5
0.4
=
,
0.5
0.6
x2 = P x1 =

0.5
0.5
0.3
0.7


0.4
0.38
=
.
0.6
0.62
2. Since there are two steps involved, the proportion we seek will be [P 2 ]21 — the probability of a state
1 population member being in state 2 after two steps:

0.5
P =
0.5
2
0.3
0.7
2
=

0.4
0.6
0.36
.
0.64
The proportion of the state 1 population in state 2 after two steps is 0.6.
3. Since there are two steps involved, the proportion we seek will be [P 2 ]22 — the probability of a state
2 population member remaining in state 2 after two steps:

0.5
P =
0.5
2
0.3
0.7
2
=

0.4
0.6
0.36
.
0.64
The proportion of the state 2 population in state 2 after two steps is 0.64.

x
4. The steady-state vector x = 1 has the property that P x = x, or (I P )x = 0. So we form the
x2
augmented matrix of that system and row-reduce it:



⇥
⇤
1 0.5
0.3
0
0.5
0.3 0
1
0.6 0
I P 0 :
=
!
0.5 1 0.7 0
0.5
0.3 0
0
0 0
Thus x2 = t is free and x1 = 35 t. Since we also want x to be a probability vector, we also have
x1 + x2 = 35 t + t = 85 t = 1, so that t = 58 . Thus the steady-state vector is
" #
x=
2
1
2
6
5. x1 = P x0 = 4 0
1
2
1
3
1
3
1
3
32
3
1
120
3
7
27 6
1805
5
4
3
0
90
2
3
150
6
7
= 41205 ,
120
3
8
5
8
.
2
1
2
1
3
1
3
1
3
6
x2 = P x1 = 4 0
1
2
32
3
3
155
6
7
= 41205 .
120
115
1
150
3
7
27 6
1205
5
4
3
0
2
6. Since there are two steps involved, the proportion we seek will be [P 2 ]11 — the probability of a state
1 population member being in state 1 after two steps:
2
1
2
6
P 2 = 40
1
2
1
3
1
3
1
3
32
1
3
27
35
=
0
2
5
12
61
43
1
4
7
18
1
3
5
18
3
7
18
2 7
.
9 5
7
18
The proportion of the state 1 population in state 1 after two steps is
5
12 .
7. Since there are two steps involved, the proportion we seek will be [P 2 ]32 — the probability of a state
2 population member being in state 3 after two steps:
2
1
2
6
P 2 = 40
1
2
1
3
1
3
1
3
32
1
3
27
35
0
=
2
5
12
61
43
1
4
7
18
1
3
5
18
3
7
18
2 7
.
9 5
7
18
The proportion of the state 2 population in state 3 after two steps is
5
18 .
210
CHAPTER 3. MATRICES
2
3
x1
8. The steady-state vector x = 4x2 5 has the property that P x = x, or (I
x3
augmented matrix of that system and row-reduce it:
h
I
P
i
2
1
6
0 : 4 0
1
2
1
3
1
3
1
1
2
1
3
2
3
1
3
1
3 2
1
0
2
7 6
05 = 4 0
1
0
2
1
3
2
3
1
3
P )x = 0. So we form the
3
2
0
1
7
6
05 ! 4 0
0
0
1
3
2
3
1
0
1
0
4
3
1
0
3
0
7
05
0
Thus x3 = t is free and x1 = 43 t, x2 = t. Since we want x to be a probability vector, we also have
x1 + x2 + x3 =
4
10
3
t+t+t=
t = 1, so that t =
.
3
3
10
Thus the steady-state vector is
x=
2
3
2
5
637
4 10 5 .
3
10
9. Let the first column and row represent dry days and the second column and row represent wet days.
(a) Given the data in the exercise, the transition matrix is

0.750
P =
0.250
0.338
,
0.662
with x =

x1
,
x2
where x1 and x2 denote the probabilities of dry and wet days respectively.

1
(b) Since Monday is dry, we have x0 =
. Since Wednesday is two days later, the probability vector
0
for Wednesday is given by




0.750 0.338 0.750 0.338 1
0.647
x2 = P (P x0 ) =
=
.
0.250 0.662 0.250 0.662 0
0.353
Thus the probability that Wednesday is wet is 0.353.
(c) We wish to find the steady state probability vector; that is, over
that a given day will be dry or wet. Thus we need to solve P x =
the augmented matrix and row-reduce:


⇥
⇤
1 0.750
0.338
0
0.250
0.338
I P 0 :
=
0.750 1 0.662 0
0.750 0.338
the long run, the probability
x, or (I P )x = 0. So form

0
1
!
0
0
Let x2 = t, so that x1 = 1.352t; then we also want x1 + x2 = 1, so that
x1 + x2 = 2.352t = 1, so that t =
1
⇡ 0.425.
2.352
Thus the steady state probability vector is


1.352 · 0.425
0.575
x=
⇡
.
0.425
0.425
So in the long run, about 57.5% of the days will be dry and 42.5% will be wet.
1.352
0
0
.
0
3.7. APPLICATIONS
211
10. Let the three rows (and columns) correspond to, in order, tall, medium, or short.
(a) From the given data, the transition matrix is
2
0.6
P = 40.2
0.2
0.1
0.7
0.2
3
0.2
0.45 .
0.4
2 3
0
(b) Starting with a short person, we have x0 = 405. That person’s grandchild is two generations
1
later, so that
2
32
32 3 2
3
0.6 0.1 0.2 0.6 0.1 0.2
0
0.24
x2 = P (P x0 ) = 40.2 0.7 0.45 40.2 0.7 0.45 405 = 40.485 .
0.2 0.2 0.4 0.2 0.2 0.4
1
0.28
So the probability that a grandchild will be tall is 0.24.
2 3
0.2
(c) The current distribution of heights is x0 = 40.55, and we want to find x3 :
0.3
2
0.6
x2 = P 3 x0 = 40.2
0.2
0.1
0.7
0.2
33 2 3 2
3
0.2
0.2
0.2457
0.45 40.55 = 40.50395 .
0.4
0.3
0.2504
So roughly 24.6% of the population is tall, 50.4% medium, and 25.0% short.
(d) We want to solve (I P )x = 0, so form the augmented matrix and row-reduce:
2
3 2
3
2
1 0.6
0.1
0.2
0
0.4
0.1
0.2 0
1
⇥
⇤
0.4
05 = 4 0.2
0.3
0.4 05 ! 40
I P 0 : 4 0.2 1 0.7
0.2
0.2 1 0.4 0
0.2
0.2
0.6 0
0
0
1
0
1
2
0
So x3 = t is free, and x1 = t, x2 = 2t. Since
x1 + x2 + x3 = t + 2t + t = 4t = 1, we get t =
so that the steady state vector is
3
0
05 .
0
1
,
4
2 3
1
4
6 7
x = 4 12 5 .
1
4
So in the long run, half the population is medium and one quarter each are tall and short.
11. Let the rows (and columns) represent good, fair, and poor crops, in order.
(a) From the data given, the transition matrix is
2
0.08
P = 40.07
0.85
0.09
0.11
0.80
3
0.11
0.055 .
0.84
(b) We want to find the probability that starting from a good crop in 1940 we end up with a good
crop in each of the five succeeding years, so we want (P i )11 for i = 1, 2, 3, 4, 5. Computing P i and
212
CHAPTER 3. MATRICES
looking at the upper left entries, we get
1941 :
P11
1942 :
2
= 0.0800
(P )11 = 0.1062
1943 :
(P 3 )11 ⇡ 0.1057
(P 4 )11 ⇡ 0.1057
1944 :
(P 5 )11 ⇡ 0.1057.
1945 :
(c) To find the long-run proportion, we want to solve
2
1 0.08
0.09
⇥
⇤
I P 0 = 4 0.07 1 0.11
0.85
0.80 1
⇥
P x = x. So row-reduce I P
3
2
0.11
0
1 0
0.126
0.05
05 ! 40 1
0.066
0.84 0
0 0
0
So x3 = t is free, and x1 = 0.126t, x2 = 0.066t. Since x1 + x2 + x3 = 1, we have
x1 + x2 + x3 = 0.126t + 0.066t + t = 1.192t = 1, so that t =
⇤
0 :
3
0
05 .
0
1
⇡ 0.839.
1.192
So the steady state vector is
2
3 2
3 2
3
x1
0.126 · 0.839
0.106
x = 4x2 5 ⇡ 40.066 · 0.8395 ⇡ 40.0555 .
x3
0.839
0.839
In the long run, about 10.6% of the crops will be good, 5.5% will be fair, and 83.9% will be poor.
12. (a) Looking at the connectivity of the graph, we get for the transition matrix
2
0
61
6
P = 6 21
42
0
1
3
1
4
1
4
0
1
3
1
3
0
3
17
37 .
27
5
3
0
1
2
0
Thus for example from state 2, there is a 13 probability of getting to state 3, since there are three
paths out of state 2 one of which leads to state 3. Thus P32 = 13 .
(b) First find the steady-state probability distribution by solving P x = x:
h
I
P
2
i 6
6
0 =6
4
1
1
2
1
2
0
1
3
1
1
3
1
3
1
4
1
4
0
1
1
3
2
3
1
2
1
3
2
0
1
60
07
7
6
7!6
40
05
0
0
0
1
0
0
0
0
1
0
2
3
1
4
3
0
3
0
07
7
7.
05
0
So x4 is free; setting x4 = 3t to avoid fractions gives x1 = 2t, x2 = 3t, x3 = 4t, and x4 = 3t. Since
this is a probability vector, we also have
x1 + x2 + x3 + x3 = 2t + 3t + 4t + 3t = 12t = 1, so that t =
So the long run probability vector is
213
6
617
6 7
x = 6 41 7 .
435
1
4
1
.
12
3.7. APPLICATIONS
213
Since we initially started with 60 total robots, the steady-state distribution will be
2 3
10
6157
7
60x = 6
4205 .
15
⇥
⇤
13. First suppose that P = p1 · · · pn is a stochastic matrix (so that the elements of each pi sum to
1). Let j be a row vector consisting entirely of 1’s. Then
⇥
⇤ ⇥P
⇤ ⇥
⇤
P
(p1 )i · · ·
(pn )i = 1 · · · 1 = j.
jP = j · p1 · · · j · pn =
P
Conversely, suppose that jP = j. Then j · pi = 1 for each i, so that j (pi )j = 1 for all i, showing that
the column sums of P are 1 so that P is stochastic.


x 1 y1
w 1 z1
14. Let P =
and Q =
be stochastic matrices. Then
x 2 y2
w 2 z2



x y 1 w 1 z1
x w + y 1 w 2 x 1 z 1 + y 1 z2
PQ = 1
= 1 1
.
x 2 y 2 w 2 z2
x 2 w 1 + y 2 w 2 x 2 z 1 + y 2 z2
Now, both P and Q are stochastic, so that x1 + x2 = y1 + y2 = z1 + z2 = w1 + w2 = 1. Then summing
each column of P Q gives
x1 w1 + y1 w2 + x2 w1 + y2 w2 = (x1 + x2 )w1 + (y1 + y2 )w2 = w1 + w2 = 1
x1 z1 + y1 z2 + x2 z1 + y2 z2 = (x1 + x2 )z1 + (y1 + y2 )z2 = z1 + z2 = 1.
Thus each column of P Q sums to 1 so that P Q is also stochastic.
15. The transition matrix from Exercise 9 is

P =
0.750
0.250
0.338
.
0.662
To find the expected number of days until
a ⇤wet day, we delete the second row and column of P
⇥
(corresponding to wet days), giving Q = 0.750 . Then the expected number of days until a wet day is
(I
Q)
1
= (1
0.750)
1
= 4.
We expect four days until the next wet day.
16. The transition matrix from Exercise 10 is
2
0.6
P = 40.2
0.2
0.1
0.7
0.2
3
0.2
0.45 .
0.4
We delete the row and column of this matrix corresponding to “tall”, which is the first row and column,
giving

0.7 0.4
Q=
.
0.2 0.4
Then
(I
Q)
1
=

0.3
0.2
0.4
0.6
1
=

6.0
2.0
4.0
.
3.0
The column of this matrix corresponding to “short” is the second column, since this was the third
column in the original matrix. Summing the values in the second column shows that the expected
number of generations before a short person has a tall descendant is 7.
214
CHAPTER 3. MATRICES
17. The transition matrix from Exercise 11 is
2
0.08
P = 40.07
0.85
0.09
0.11
0.80
3
0.11
0.055 .
0.84
Since we want to know how long we expect it to be before a good crop, we remove the first row and
column, leaving

0.11 0.05
Q=
.
0.80 0.84
Then
(I
Q)
1
=

0.89
0.80
1
0.05
0.16
=

1.562
7.812
0.488
.
8.691
Since we are starting from a fair crop, we sum the entries in the first column (which was the “fair”
column in the original matrix), giving 1.562 + 7.812, or 9 or 10 years expected until the next good crop.
18. The transition matrix from Exercise 12 is
2
0
61
6
P = 6 21
42
0
1
3
1
4
1
4
0
1
3
1
3
0
3
17
37 .
27
5
3
0
1
2
0
The expected number of moves from any junction other than 4 to reach junction 4 is found by removing
the fourth row and column of P , giving
2
3
0 13 14
6
7
Q = 4 12 0 14 5 .
1
1
0
2
3
Then
(I
Q)
1
2
6
=4
1
1
3
1
2
1
2
1
3
1
3
1
4
17
45
1
1
=
2
22
13
6 15
4 13
16
13
10
13
21
13
12
13
3
8
13
9 7
,
13 5
20
13
15 16
53
so that a robot is expected to reach junction 4 starting from junction 1 in 22
13 + 13 + 13 = 13 ⇡ 4.08 moves,
10
21
12
43
8
9
37
from junction 2 in 13 + 13 + 13 = 13 ⇡ 3.31 moves, and from junction 3 in 13 + 13 + 20
13 = 13 ⇡ 2.85
moves.
19. Since the sums of the columns of the matrix E are both 1, and all numbers are nonnegative, this matrix
is an exchange matrix. To find a solution to Ex = x, or equivalently (E I)x = 0, we row-reduce the
augmented matrix of that system:
"
#
"
#
"
#
2R1
h
i
1
1
1
1
1
0
0
1
0
!
2
4
2
4
2
.
E I 0 :
R2 +R1
1
1
0
0 0 0
0
0 0
!
2
4
Thus
the solutions are of the form x1 = 12 t, x2 = t. Set t = 2 to avoid fractions, giving a price vector

1
. Of course, other values of t lead to equally valid price vectors.
2
20. Since neither column sum is 1, this is not an exchange matrix.
21. While the first column sums to 1, the second does not, so this is not an exchange matrix.
3.7. APPLICATIONS
215
22. Since the sums of the columns of the matrix E are both 1, and all numbers are nonnegative, this matrix
is an exchange matrix. To find a solution to Ex = x, or equivalently (E I)x = 0, we row-reduce the
augmented matrix of that system:
⇥
E
⇤
0 :
I

0.9
0.9
0.6
0.6
0
0
R2 +R1
!

0.9 0.6
0
0

! 1
0
10
9 R1
0
0
2
3
0
0
.
0
Thus
the solutions are of the form x1 = 23 t, x2 = t. Set t = 3 to avoid fractions, giving a price vector

2
. Of course, other values of t lead to equally valid price vectors.
3
23. Since the matrix contains a negative entry, it is not an exchange matrix.
24. Since the sums of the columns of the matrix E are both 1, and all numbers are nonnegative, this matrix
is an exchange matrix. To find a solution to Ex = x, or equivalently (E I)x = 0, we row-reduce the
augmented matrix of that system:
h
E
I
2
1
i
2
6
0 : 4 0
1
2
1
1
0
0
1
3
1
3
3
0
7
05
0
2
1
2
6
! 4 0
R3 +R1
0
!
R2
1
1
1
0
2
1
3
1
3
1
2
6
4 0
0
3
0
7
05
0
R1 R2
!
R3 R2
!
1
3
1
3
0
1
0
0
3
0
7
05
0
2
! 1
6
40
0
2R1
0
1
0
2
3
1
3
0
3
0
7
05 .
0
Thus the
solutions are of the form x1 = 23 t, x2 = 13 t, x3 = t. Set t = 3 to avoid fractions, giving a price
2 3
2
vector 415. Of course, other values of t lead to equally valid price vectors.
3
25. Since the sums of the columns of the matrix E are both 1, and all numbers are nonnegative, this matrix
is an exchange matrix. To find a solution to Ex = x, or equivalently (E I)x = 0, we row-reduce the
augmented matrix of that system:
h
E
I
2
0.7
i
6
0 : 4 0.3
0.4
R2
R3
0
0.5
0.5
2
1
0.3R1 6
! 40
0.4R1
! 0
0.2
0.3
0.5
0
1
2
1
2
3
0
7
05
0
2
7
27
70
27
70
10
7 R1
!
3
0
7
05
0
2
1
6
40.3
0.4
0
0.5
0.5
2
1
6
40
R3 +R2
! 0
3
0
7
0.3 05
0.5 0
3
2
0
0
7
7
1
27
05
2
70
0
0 0
2
7
2
1
2R2 6
! 40
0
0
1
0
2
7
27
35
0
3
0
7
05 .
0
2
27
Thus the solutions
2 3 are of the form x1 = 7 t, x2 = 35 t, x3 = t. Set t = 35 to avoid fractions, giving a
10
price vector 4275. Of course, other values of t lead to equally valid price vectors.
35
26. Since the sums of the columns of the matrix E are both 1, and all numbers are nonnegative, this matrix
is an exchange matrix. To find a solution to Ex = x, or equivalently (E I)x = 0, we row-reduce the
216
CHAPTER 3. MATRICES
augmented matrix of that system:
h
E
I
2
0.50
i
6
0 : 4 0.25
0.25
2
1
1.40
40
0.35
0
0.35
0.70
0.70
0
0.70
0.425
0.425
2
3 2R 2
1
0.35 0
1
!
7
6
0.25 05
40.25
0.6 0
0.25
3 R1 4R2 2
0
! 1
0
40
05
0.35
2
0 R3 +R!
0
0
1.40
0.70
0
2.40
0.425
0
0.70
0.25
0.6
3
0
05
0
3
0
7
05
0
R2 0.25R1
!
R3 0.25R1
!
2
1
4
! 0
0
1
3.5 R2
0
1
0
2.40
1.214
0
3
0
05
0
3
2.4
So one possible solution is 41.2145. Multiplying through by 70 will roughly clear fractions, giving
1
2
3
168
⇡ 4 85 5 if an integral solution is desired.
70
27. This matrix is a consumption matrix since all of its entries are nonnegative. By Corollary 3.36, it is
productive since the sum of the entries in each column is strictly less than 1. (Note that Corollary 3.35
does not apply since the sum of the entries in the second row exceeds 1).
28. This matrix is a consumption matrix since all of its entries are nonnegative. By Corollary 3.35, it is
productive since the sum of the entries in each row is strictly less than 1. (Note that Corollary 3.36
does not apply since the sum of the entries in the third column exceeds 1).
29. This matrix is a consumption matrix since all of its entries are nonnegative. Since the entries in the
second row as well as those in the second column sum to a number greater than 1, neither Corollary
3.35 nor 3.36 applies. So to determine if C is productive, we must use the definition, which involves
seeing if I C is invertible and if its inverse has all positive entries.
2
3 2
3 2
3
1 0 0
0.35 0.25
0
0.65
0.25
0
0.45
0.355 .
I C = 40 1 05 40.15 0.55 0.355 = 4 0.15
0 0 1
0.45 0.30 0.60
0.45
0.30
0.40
Using technology,
(I
C)
1
2
13.33
⇡ 4 38.67
44.00
17.78
46.22
54.67
3
15.56
40.445 .
45.33
Since this is not a positive matrix, it follows that C is not productive.
30. This matrix is a consumption matrix since all of its entries are nonnegative. Since the entries in the
first row as well as those in the each column sum to a number at least equal to 1, neither Corollary
3.35 nor 3.36 applies. So to determine if C is productive, we must use the definition, which involves
seeing if I C is invertible and if its inverse has all positive entries.
2
3 2
3 2
3
1 0 0 0
0.2 0.4 0.1 0.4
0.8
0.4
0.1
0.4
60 1 0 07 60.3 0.2 0.2 0.17 6 0.3
0.8
0.2
0.17
7 6
7 6
7.
I C=6
40 0 1 05 4 0 0.4 0.5 0.35 = 4
0
0.4
0.5
0.35
0 0 0 1
0.5 0 0.2 0.2
0.5
0
0.2
0.8
Using technology reveals that I
31. To find a solution x to (I
C is not invertible, so that C is not productive.
C)x = d, we invert I C and compute x = (I
"
# "
# "
#
1
1
1
1
1 0
2
4 =
2
4 .
I C=
1
1
1
1
0 1
2
2
2
2
C)
1
d. Here
3.7. APPLICATIONS
Then det(I
217
C) =
1
2
·
1
2
1
4
1
2
= 18 , so that
"
1 12
1
C) =
1/8 12
(I
Thus a feasible production vector is
x = (I
32. To find a solution x to (I
Then det(I
C)
1
1
4
1
2

4
d=
4
#
"
4
=
4

2
4
#
2
.
4

1
10
=
.
3
16
C)x = d, we invert I C and compute x = (I



1 0
0.1 0.4
0.9
0.4
I C=
=
.
0 1
0.3 0.2
0.3
0.8
C) = 0.9 · 0.8
( 0.4)( 0.3) = 0.60, so that
"
# "
4
0.8 0.4
1
1
(I C) =
= 31
0.60 0.3 0.9
2
Thus a feasible production vector is
x = (I
1
C)
d=
"
4
3
1
2
2
3
3
2
2
3
3
2
C)x = d, we invert I C and compute
2
3 2
3 2
1 0 0
0.5 0.2 0.1
0.5
I C = 40 1 05 4 0 0.4 0.25 = 4 0
0 0 1
0
0 0.5
0
⇥
⇤
Then row-reduce I C I to compute (I C) 1 :
2
3
2
0.5
0.2
0.1 1 0 0
1
h
i
6
7
6
I C I =4 0
0.6
0.8 0 1 05 ! 40
0
0
0.5 0 0 1
0
x = (I
34. To find a solution x to (I
I
Compute (I
C)
1
2
2
6
1
C) d = 40
0
2
3
5
3
0
(I
Thus a feasible production vector is
x = (I
C)
2
1
2
1.238
⇡ 40.143
0.571
1.238
6
1
C) d = 40.143
0.571
0.714
1.429
0.714
d. Here
.
32 3
2
0.2
0.6
0
0
1
0
0
0
1
3
10
6 7
= 4 6 5.
4
8
2
3
3
27 6 7
3 5 425
0.714
1.429
0.714
C) 1 d. Here
3
0.1
0.85 .
0.5
x = (I
2
0
0
3
2
3
5
3
2
3
27
.
35
0
2
2
C)x = d, we invert I C and compute x = (I
2
3 2
3 2
1 0 0
0.1 0.4 0.1
0.9
0.4
0
0.8
C = 40 1 05 4 0 0.2 0.25 = 4
0 0 1
0.3 0.2 0.3
0.3
0.2
using technology to get
1
#" # " #
10
2
= 35 .
1
2
33. To find a solution x to (I
Thus a feasible production vector is
#
C)
C)
1
d. Here
3
0.1
0.25 .
0.7
3
0.381
0.4295 .
1.714
32 3 2
3
0.381
1.1
4.624
76 7 6
7
0.4295 43.55 = 46.0145 .
1.714
2.0
6.557
218
CHAPTER 3. MATRICES
35. Since x
0 and A
O, then Ax has all nonnegative entries as well, so that Ax
x > Ax
Ox = 0. But then
Ox = 0,
so that x > 0.
36. (a) Since all entries of all matrices are positive, it is clear that AC
O and BD O. To see that
AC
BD, let A = (aij ), B = (bij ), C = (cij ), and D = (dij ). Then each aij
bij and each
cij dij . So
n
n
n
X
X
X
(AC)ij =
aik ckj
bik ckj
bik dkj = (BD)ij .
k=1
k=1
k=1
So every entry of AC is greater than or equal to the corresponding entry of BD, and thus
AC BD.
⇥
⇤
(b) Let xT = x1 x2 · · · xn . Then for any i and j, we know that aij xj
bij xj since aij > bij
and xj 0. Since x 6= 0, we know that some component of x, say xk , is strictly positive. Then
aik xk > bik xk since xk > 0. But then
(Ax)i = ai1 x1 + · · · + aik xk + · · · + ain xn > bi1 x1 + · · · + bik xk + · · · + bin xn = (Bx)i .
37. Multiplying, we get

2
x1 = Lx0 =
0.6

2
x2 = Lx1 =
0.6

2
x3 = Lx2 =
0.6
5
0
5
0
5
0


10
45
=
5
6


45
120
=
6
22.5


120
375
=
27
72
38. Multiplying, we get
2
0
x1 = Lx0 = 40.2
0
2
0
x2 = Lx1 = 40.2
0
2
0
x3 = Lx2 = 40.2
0
1
0
0.5
1
0
0.5
1
0
0.5
32 3 2 3
2 10
10
05 4 4 5 = 4 2 5
0
3
2
32 3 2 3
2 10
6
05 4 2 5 = 425
0
2
1
32 3 2 3
2 6
4
05 425 = 41.25
0 1
1
39. Multiplying, we get
2
1
x1 = Lx0 = 40.7
0
2
1
x2 = Lx1 = 40.7
0
2
1
x3 = Lx2 = 40.7
0
1
0
0.5
1
0
0.5
1
0
0.5
32
3 2
3
3 100
500
05 41005 = 4 70 5
0 100
50
32
3 2
3
3 500
720
05 4 70 5 = 43505
0
50
35
32
3 2
3
3 720
1175
05 43505 = 4 504 5 .
0
35
175
3.7. APPLICATIONS
219
40. Multiplying, we get
2
0
60.5
x1 = Lx0 = 6
40
0
2
0
60.5
x2 = Lx1 = 6
40
0
2
0
60.5
x3 = Lx2 = 6
40
0
1
0
0.7
0
1
0
0.7
0
1
0
0.7
0
32 3 2 3
5 10
80
6107 6 5 7
07
76 7 = 6 7
05 4105 4 7 5
0
10
3
32 3 2 3
2 5 80
34
6 5 7 6 40 7
0 07
76 7 = 6 7
0 05 4 7 5 43.55
0.3 0
3
2.1
32 3 2
3
2 5
34
57.5
6 7 6
7
0 07
7 6 40 7 = 6 17 7 .
0 05 43.55 4 28 5
0.3 0
2.1
1.05
2
0
0
0.3
41. (a) The first ten generations of the species using each of the two models are as follows:
L1
L2
x0
10
10

x1
50
8
10
10
50
8

64184.3
12252.2
L1
L2
 x6
640
640

x2
40
40


 x3
200
32
208
40
 x7
3200
512
268989
51347.5

872
166.4


 x4
160
160

3654.4
697.6
 x8
2560
2560

1.127 ⇥ 106
215192
 x5
800
128
15315.2
2923.52
 x9
12800
2048

4.724 ⇥ 106
901844
 x10
10240
10240
1.980 ⇥ 107
.
3.780 ⇥ 106
(b) Graphs of the first ten generations using each model are:
Model L1
Percentage
1.0
Class 1
0.8
0.8
0.6
0.6
0.4
0.4
0.2
0.2
Class 2
0
2
4
6
Model L2
Percentage
1.0
8
10
Time!yrs"
Class 2
Class 1
0
2
4
6
8
10
Time!yrs"
The graphs suggest that the first model does not have a steady state, while the second Leslie
matrix quickly approaches a steady state ratio of class 1 to class 2 population.
220
CHAPTER 3. MATRICES
2
3
20
42. Start with x0 = 4205. Then
20
2
0
x1 = 40.1
0
2
0
x2 = 40.1
0
2
0
x3 = 40.1
0
0
0
0.5
0
0
0.5
0
0
0.5
32 3 2
3
20 20
400
0 5 4205 = 4 2 5
0
20
10
32
3 2
3
20 400
200
0 5 4 2 5 = 4 40 5
0
10
1
32
3 2 3
20 200
20
0 5 4 40 5 = 4205 .
0
1
20
Thus after three generations the population is back to where it started, so that the population is cyclic.
2 3
20
43. Start with x0 = 4205. Then
20
2
32 3 2
3
0 0 20 20
400
0 5 4205 = 420s5
x1 = 4 s 0
0 0.5 0
20
10
2
32
3 2
3
0 0 20 400
200
0 5 420s5 = 4400s5
x2 = 4 s 0
0 0.5 0
10
10s
2
32
3 2
3
0 0 20
200
200s
0 5 4400s5 = 4200s5 .
x3 = 4 s 0
0 0.5 0
10s
200s
Thus after three generations the population is still evenly distributed among the three classes, but the
total population has been multiplied by 10s. So
• If s < 0.1, then the overall population declines after three years;
• If s = 0.1, then the overall population is cyclic every three years;
• If s > 0.1, then the overall population increases after three years.
44. The table gives us seven separate age classes, each of duration two years; from the table, the Leslie
matrix of the system is
2
3
0 0.4 1.8 1.8 1.8 1.6 0.6
60.3 0
0
0
0
0
07
6
7
6 0 0.7 0
0
0
0
07
6
7
0 0.9 0
0
0
07
L=6
60
7.
60
7
0
0
0.9
0
0
0
6
7
40
0
0
0 0.9 0
05
0
0
0
0
0 0.6 0
Since in 1990,
2
3
2
3
2
3
10
46.4
42.06
627
6 3 7
613.927
6 7
6
7
6
7
687
6 1.4 7
6 2.1 7
6 7
6
7
6
7
7
6
7
6
7
x0 = 6
6 5 7 , so that x1 = Lx0 = 6 7.2 7 , and x2 = Lx1 = 6 1.26 7 .
6127
6 4.5 7
6 6.48 7
6 7
6
7
6
7
405
410.85
4 4.05 5
1
0
6.48
3.7. APPLICATIONS
221
Then x1 and x2 represent the population distribution in 1991 and 1992. To project the population for
the years 2000 and 2010, note that xn = Ln x0 . Thus
2
3
2
3
69.12
167.80
618.617
6 46.08 7
6
7
6
7
612.247
6 29.54 7
6
7
6
7
20
7
6
7
x10 = L10 x0 ⇡ 6
610.597 , x20 = L x0 ⇡ 6 24.35 7 .
6 8.29 7
6 20.06 7
6
7
6
7
4 6.51 5
4 16.48 5
3.57
9.05
45. The adjacency matrix contains a 1 in row i, column j
2
0 1 0
61 0 1
6
40 1 0
1 0 1
if there is an edge between vi and vj :
3
1
07
7.
15
0
46. The adjacency matrix contains a 1 in row i, column j
2
1 0 0
60 1 1
6
40 1 0
1 0 0
if there is an edge between vi and vj :
3
1
07
7.
05
0
47. The adjacency matrix contains a 1 in row i, column
2
0 1 1
61 0 1
6
61 1 0
6
41 0 1
1 0 0
j if there is an edge between vi and vj :
3
1 1
0 07
7
1 07
7.
0 15
1 0
48. The adjacency matrix contains a 1 in row i, column
2
0 0 0
60 0 0
6
60 0 0
6
41 1 1
1 0 0
j if there is an edge between vi and vj :
3
1 1
1 07
7
1 07
7.
0 05
0 0
49.
v2
v4
v1
v3
222
CHAPTER 3. MATRICES
50.
v4
v3
v1
v2
51.
v4
v1
v2
v3
v5
52.
v2
v4
v1
v5
v3
53. The adjacency matrix contains a 1 in row i, column j if there is an edge from vi to vj :
2
0
60
6
40
1
1
0
1
0
1
0
0
0
3
0
07
7.
15
0
54. The adjacency matrix contains a 1 in row i, column j if there is an edge from vi to vj :
2
0
60
6
41
0
0
0
1
1
1
0
0
1
3
1
17
7.
05
0
3.7. APPLICATIONS
223
55. The adjacency matrix contains a 1 in row i, column
2
0 1 0
61 0 0
6
61 1 0
6
41 0 0
1 0 0
j if there is an edge from vi to vj :
3
1 0
1 07
7
0 07
7.
0 15
0 0
56. The adjacency matrix contains a 1 in row i, column
2
0 1 0
60 0 0
6
60 1 0
6
40 0 0
0 0 1
j if there is an edge from vi to vj :
3
1 1
0 17
7
1 07
7.
0 15
0 0
57.
v4
v3
v1
v2
58.
v3
v4
v1
v2
59.
v1
v5
v3
v4
v2
224
CHAPTER 3. MATRICES
60.
v5
v2
v1
v3
v4
61. In Exercise 50,
2
0
61
A=6
40
1
1
1
1
1
0
1
0
1
3
2
1
2
62
17
2
7 , so A = 6
42
15
0
1
2
4
2
3
3
1
37
7.
15
3
2
2
2
1
The number of paths of length 2 between v1 and v2 is given by A2
length 2 between v1 and v2 .
62. In Exercise 52,
2
0
60
6
A=6
60
41
1
0
0
0
1
1
0
0
0
1
1
1
1
1
0
0
3
2
1
2
62
17
7
6
2
6
17
7 , so A = 62
5
40
0
0
0
2
2
2
0
0
12
2
2
2
0
0
2
0
61
A=6
40
1
1
1
1
1
0
1
0
1
3
2
1
3
6
17
7 , so A3 = 67
43
15
0
6
12
2
0
60
6
A=6
60
41
1
0
0
0
1
1
0
0
0
1
1
1
1
1
0
0
3
2
1
12
612
17
7
6
4
6
17
7 , so A = 612
40
05
0
0
12
12
12
0
0
= 2. So there are two paths of
3
3 6
7 87
7.
3 65
6 5
7
11
7
8
The number of paths of length 3 between v1 and v3 is given by A3
of length 3 between v1 and v3 .
64. In Exercise 52,
3
0
07
7
07
7.
35
3
0
0
0
3
3
The number of paths of length 2 between v1 and v2 is given by A2
length 2 between v1 and v2 .
63. In Exercise 50,
= 2. So there are two paths of
13
= 3. So there are three paths
12 0
12 0
12 0
0 18
0 18
The number of paths of length 4 between v2 and v2 is given by A4
of length 4 between v2 and itself.
22
3
0
07
7
07
7.
185
18
= 12. So there are twelve paths
3.7. APPLICATIONS
225
65. In Exercise 57,
2
0
61
A=6
40
1
1
0
1
0
0
0
0
1
3
2
0
1
6
17
7 , so A2 = 61
41
05
1
1
The number of paths of length 2 from v1 to v3 is A2
to v3 .
66. In Exercise 57,
2
0
61
A=6
40
1
1
0
1
0
0
0
0
1
2
0
60
6
A=6
61
41
1
1
0
0
0
1
0
0
0
1
0
0
1
1
0
0
2
0
60
6
A=6
61
41
1
1
0
0
0
1
0
0
0
1
0
0
1
1
0
0
41
14
1
2
1
2
1
1
1
1
3
1
27
7.
15
3
1
1
3
3
1
1
0
0
1
1
1
1
3
1
1
3
1
27
7
37
7.
15
0
= 3. There are three paths of length 3 from
3
2
1
3
63
07
7
6
4
6
17
7 , so A = 66
43
05
0
2
The number of paths of length 4 from v1 to v4 is A3
to v4 .
3
1
17
7.
15
1
= 3. There are three paths of length 3 from
3
2
1
1
61
07
7
6
3
6
17
7 , so A = 62
5
43
0
0
2
The number of paths of length 3 from v4 to v1 is A3
v4 to v1 .
68. In Exercise 60,
41
0
1
0
1
= 0. There are no paths of length 2 from v1
3
2
0
1
6
17
7 , so A3 = 62
41
05
1
3
The number of paths of length 3 from v4 to v1 is A3
v4 to v1 .
67. In Exercise 60,
13
0
1
0
2
2
3
5
4
2
1
1
3
1
1
2
1
3
4
2
3
2
17
7
27
7.
45
3
= 2. There are two paths of length 4 from v1
69. (a) If there are no ones in row i, then no edges join vertex i to any of the other vertices. Thus G is a
disconnected graph, and vertex i is an isolated vertex.
(b) If there are no ones in column j, then no edges join vertex j to any of the other vertices. Thus G
is a disconnected graph, and vertex j is an isolated vertex.
70. (a) If row i of A2 is all zeros, then there are no paths of length 2 from vertex i to any other vertex.
(b) If column j of A2 is all zeros, then there are no paths of length 2 from any vertex to vertex j.
71. The adjacency matrix for the digraph is
2
0
61
6
61
A=6
61
6
40
1
0
0
0
1
0
0
0
1
0
0
0
1
0
0
1
0
0
1
1
1
1
1
0
0
3
0
17
7
07
7.
07
7
15
0
226
CHAPTER 3. MATRICES
The number of wins that each player had is given by
2 3 2 3
1
1
617 647
6 7 6 7
617 637
7 6 7
A6
617 = 637 ,
6 7 6 7
415 415
1
3
so the ranking is: first, P2 ; second, P3 , P4 , P6 (tie);
wins and indirect wins, we compute
2 3 2
1
0 0 0
617 63 0 2
6 7 6
617 62 1 0
7 6
(A + A2 ) 6
617 = 62 1 1
6 7 6
415 41 0 1
1
3 1 1
third, P1 , P5 (tie). Ranking instead by combined
0
2
1
0
1
2
1
3
3
3
0
3
32 3 2 3
1
1
2
617 6127
27
76 7 6 7
6 7 6 7
17
7 617 = 6 8 7 ,
7
7 6 7
27 6
617 6 9 7
5
4
1
15 4 4 5
0
1
10
so the ranking is: first, P2 ; second, P6 ; third, P4 ; fourth, P3 ; fifth, P5 ; sixth, P1 .
72. Let vertices 1 through 7 correspond to, in order, rodent, plant, insect, bird, fish, fox, and bear. Then
the adjacency matrix is
2
3
0 1 0 0 0 0 0
6 0 0 0 0 0 0 07
6
7
6 0 1 0 0 0 0 07
6
7
7
A=6
6 0 1 1 0 1 0 07 .
6 0 1 1 0 0 0 07
6
7
4 1 0 0 1 0 0 05
1 0 0 0 1 1 0
(a) The number of direct sources of food is the number of ones in each row, so it is
2 3 2 3
1
1
617 607
6 7 6 7
617 617
6 7 6 7
7 6 7
A6
617 = 637 .
617 627
6 7 6 7
415 425
1
3
Thus birds and bears have the most direct sources of food, 3.
(b) The number of species for which a given species is a direct source of food is the number of ones in
its column. Clearly plants, with a value of 4, are a direct source of food for the greatest number
of species. Note the analogy of parts (a) and (b) with the previous exercise, where eating a given
species is equivalent to winning that head-to-head match, and being eaten is equivalent to losing
it.
(c) Since A2 ij gives the number of paths of length 2 from vertex i to vertex j, it follows that A2 ij
is the number of ways in which species j is an indirect food source for species i. So summing the
entries of the ith row of A2 gives the total number of indirect food sources:
2 3 2 3
1
0
617 607
6 7 6 7
617 607
6 7 6 7
26 7
7
A 6 17 = 6
637 .
617 617
6 7 6 7
415 445
1
5
3.7. APPLICATIONS
227
Thus bears have the most indirect food sources. Combined direct and indirect food sources are
given by
2 3 2 3
1
1
617 607
6 7 6 7
617 617
6 7 6 7
2 6 7
7
(A + A ) 617 = 6
6 67 .
617 637
6 7 6 7
415 465
1
8
(d) The matrix A⇤ is the matrix A after removing
2
0 0
60 0
6
60 1
A⇤ = 6
60 1
6
41 0
1 0
Then we have
row 2 and column 2:
3
0 0 0 0
0 0 0 07
7
0 1 0 07
7.
0 0 0 07
7
1 0 0 05
0 1 1 0
2 3 2 3
1
0
617 607
6 7 6 7
6 7 627
⇤ 617
7
A 6 7=6
6 7,
617 617
415 425
1
3
so the species with the most direct food sources is the bear, with three. The species that is a
direct food source for the most species is the species for which the column sum is the largest;
three species, rodents, insects, and birds, each are prey for two species. Note that the column
sums can be computed as
2 3
1
617
6 7
617
⇥
⇤
7
1 1 1 1 1 1 A⇤ or (A⇤ )T 6
617 .
6 7
415
1
The number of indirect food sources each species has is
2 3 2 3
1
0
617 607
6 7 6 7
6 7
2 617
7 617
(A⇤ ) 6
617 = 607 ,
6 7 6 7
415 425
1
3
and the number of combined direct and indirect food sources is
2 3 2 3
1
0
617 607
7
⇣
⌘6 7 6
6 7
2 617
7 = 637 .
A⇤ + (A⇤ ) 6
617 617
6 7 6 7
415 445
1
6
Thus birds have lost the most combined direct and indirect sources of food (5). The bear, fox,
and fish species have each lost two combined sources of food, and rodents and insects have each
lost one combined food source.
228
CHAPTER 3. MATRICES
5
(e) Since (A⇤ ) is the zero matrix, after a while, no species will have a direct food source, so that the
ecosystem will be wiped out.
73. (a) Using the relationships given, we get the digraph
Bert
Ehaz
Carla
Dana
Ann
Letting vertices 1 through 5 correspond to the people in alphabetical order, the adjacency matrix
is
2
3
0 0 1 0 1
6 0 0 1 1 07
6
7
7
A=6
6 0 0 0 0 17 .
4 1 0 1 0 05
0 1 0 0 0
⇥ k⇤
(b) The element A ij is 1 if there is a path of length k from i to j, so there is a path of length at
⇥
⇤
most k from i to j if A + A2 + · · · + Ak ij 1. If Bert hears a rumor and we want to know how
many steps it takes before everyone else has heard the rumor, then we must calculate A, A + A2 ,
A + A2 + A3 , and so forth, until every entry in row 2 (except for the one in column 2, which
corresponds to Bert himself) is nonzero. This is not true for A, since for example A21 = 0. Next,
2
3
0 1 1 0 2
6 1 0 2 1 17
6
7
7
A + A2 = 6
6 0 1 0 0 17 ,
4 1 0 2 0 25
0 1 1 1 0
so everyone else will have heard the rumor, either directly or indirectly, from Bert after two steps
(Carla will have heard it twice).
(c) Using the same logic as in part (b), we must compute successive sums of powers of A until every
entry in the first row, except for the entry in the first column corresponding
to ⇤Ann herself, is
⇥
nonzero. From part (b), we see that this is not the case for A + A2 since A + A2 14 = 0. Next,
2
3
0 2 2 1 2
6 1 1 3 1 37
6
7
2
3
7
A+A +A =6
6 0 1 1 1 17 ,
4 1 2 2 0 35
1 1 2 1 1
so everyone else will have heard the rumor from Ann after three steps. In fact, everyone but Dana
will have heard it twice.
⇥ ⇤
(d) j is reachable from i by a path of length k if Ak ij is nonzero. Now, if j is reachable from i at
all, it is reachable by a path of length at most n, where n is the number of vertices in the digraph,
so we must compute
⇥
⇤
A + A2 + A3 + · · · + An ij
and see if it is nonzero. If it is, the two vertices are connected.
3.7. APPLICATIONS
229
74. Let G be a graph with m vertices, and A = (aij ) its adjacency matrix.
(a) The basis for the induction is n = 1, where the statement says that aij is the number of direct
paths between vertices i and j. But this is just the definition of the adjacency matrix, since
aij = 1 if i and j have an edge between them and is zero otherwise.
Next, assume [Ak ]ij is the
Pm
k+1
k
number of paths of length k between i and j. Then [A
]ij = l=1 [A ]il [A]lj . For a given vertex
l, [Ak ]il gives the number of paths of length k from i to l, and [A]lj gives the number of paths
of length 1 from l to j (this is either zero or one). So the product gives the number of paths of
length k + 1 from i to j passing through l. Summing over all vertices l in the graph gives all paths
of length k + 1 from i to j, proving the inductive step.
(b) If G is a digraph, the statement of the result must be modified to say that the (i, j) entry of An
is equal to the number of n-paths from i to j; the proof is valid since the adjacency matrix of a
digraph takes arrow directions into account.
Pm
75. [AAT ]ij = k=1 aik ajk . So this matrix entry is the number of vertices k that have edges from both
vertex i and vertex j.
76. From the diagram in Exercise 49, this graph is bipartite, with U = {v1 } and V = {v2 , v3 , v4 }.
77. From the diagram in Exercise 52, this graph is bipartite, with U = {v1 , v2 , v3 } and V = {v4 , v5 }.
78. The graph in Exercise 51 is not bipartite. For suppose v1 2 U . Then since there are edges from v1 to
both v3 and v4 , we must have v3 , v4 2 V . Next, there are edges from v3 to v5 , and from v4 to v2 , so
that v2 , v5 2 U . But there is an edge from v2 to v5 . So this graph cannot be bipartite.
79. Examining the matrix, we see that vertices 1, 2, and 4 are connected to vertices 3, 5, and 6 only, so
that if we define U = {v1 , v2 , v4 } and V = {v3 , v5 , v6 }, we have shown that this graph is bipartite.
80. Suppose that G has n vertices.
(a) First suppose that the adjacency matrix of G is of the form shown. Suppose that the zero blocks
have p and q = n p rows respectively. Then any edges at vertices v1 , v2 , . . . , vp have their other
vertex in vertices vp+1 , vp+2 , . . . , vn , since the upper left p ⇥ p matrix is O. Similarly, any edges
at vertices vp+1 , vp+2 , . . . , vn have their other vertex in vertices v1 , v2 , . . . , vp , since the lower right
q ⇥ q matrix is O. So G is bipartite, with U = {v1 , v2 , . . . , vp } and V = {vp+1 , vp+2 , . . . , vn }.
Next suppose that G is bipartite. Label the vertices so that U = {v1 , v2 , . . . , vp } and V =
{vp+1 , vp+2 , . . . , vn }. Now consider the adjacency matrix A of G. aij = 0 if i, j  p, since then
both vi and vj are in U ; similarly, aij = 0 if i, j > p, since then both vi and vj are in V . So the
upper left p ⇥ p submatrix as well as the bottom right q ⇥ q submatrix are the zero matrix. The
remaining two submatrices are transposes of each other since this is an undirected graph, so that
aij = 1 if and only if aji = 1. Thus A is of the form shown.
(b) Suppose G is bipartite and A its adjacency matrix, of the form given in part (a). It suffices to
show that odd powers of A have zero blocks at upper left of size p ⇥ p and at lower right of size
q ⇥ q, since then there are no paths of odd length between any vertex and itself. We do this by
induction. Clearly A1 = A is of that form. Now,



OO + BB T
OB + BO
BB T
O
D1 O
A2 =
=
=
O D2
B T O + OB T B T B + OO
O
BT B
where D1 = BB T and D2 = B T B. Suppose that A2k 1 has the required zero blocks. Then


O C1 D1 O
A2k+1 = A2k 1 A2 =
C2 O
O D2

OD1 + C1 O OO + C1 D2
=
C2 D1 + OO D1 O + OD2

O
C1 D2
=
.
C2 D1
O
230
CHAPTER 3. MATRICES
Thus A2k+1 is of this form as well. So no odd power of A has any entries in the upper left or
lower right block, so there are no odd paths beginning and ending at the same vertex. Thus all
circuits must be of even length. (Note that in the matrix A2k+1 above, we must in fact have
(C1 D2 )T = C2 D1 since powers of a symmetric matrix are symmetric, but we do not need this
fact.)
Chapter Review
1. (a) True. See Exercise 34 in Section 3.2. If A is an m ⇥ n matrix, then AT is an n ⇥ m matrix. Then
AAT is defined and is an m ⇥ m matrix, and AT A is also defined and is an n ⇥ n matrix.
(b) False. This is true only when A is invertible. For example, if



1 0
0 0
0
A=
, B=
, then AB =
0 1
0 1
0
but A 6= O.
(c) False. What is true is that XAA 1 = BA 1 so that X = BA
commutative, so that BA 1 6= A 1 B in general.
1
0
,
0
. Matrix multiplication is not
(d) True. This is Theorem 3.11 in Section 3.3.
(e) True. If E performs kRi , then E is a matrix with all zero entries o↵ the diagonal, so it is its
own transpose and thus E T is also the elementary matrix that performs kRi . If E performs
Ri + kRj , then the only o↵-diagonal entry in E is [E]ij = k, so the only o↵-diagonal entry of E T
is [E T ]ji = k and thus E T is the elementary matrix performing Rj + kRi . Finally, if E performs
Ri $ Rj , then [E]ii = [E]jj = 0 and [E]ij = [E]ji = 1, so that E T = E and thus E T is also an
elementary matrix exchanging Ri and Rj .
(f ) False. Elementary matrices perform a single row operation, while their products can perform multiple row operations. There are many examples. For instance, consider the elementary matrices
corresponding to R1 $ R2 and 2R1 :



0 1 2 0
0 1
=
,
1 0 0 1
2 0
which is not an elementary matrix.
(g) True. See Theorem 3.21 in Section 3.5.
(h) False. For this to be true, the plane must pass through the origin.
(i) True, since
T (c1 u + c2 v) =
(c1 u + c2 v) =
c1 u
c2 v = c1 ( u) + c2 ( v) = c1 T (u) + c2 T (v).
(j) False; the matrix is 5 ⇥ 4, not 4 ⇥ 5. See Theorems 3.30 and 3.31 in Section 3.6.
2. Since A is 2 ⇥ 2, so is A2 . Since B is 2 ⇥ 3, the product makes sense. Then






1 2 1 2
7 12
7 12 2
0
1
50
2
2
A =
=
, so that A B =
=
3 5 3 5
18 31
18 31 3
3
4
129
36
93
41
.
106
3. B 2 is not defined, since the number of columns of B is not equal to the number of rows of B. So this
matrix product does not make sense.
4. Since B T is 3 ⇥ 2, A 1 is 2 ⇥ 2, and B is 2 ⇥ 3, this matrix product makes sense as long as A is
invertible. Since det A = 1 · 5 2 · 3 = 1, we have


1
5
2
5
2
A 1=
=
,
3
1
3
1
1
Chapter Review
231
so that
2
2
BT A 1B = 4 0
1
3
3 
5
35
3
4
2
1

2
3
2
3
1 
2
35
3
6
1
1
=4 9
4
17
0
3
0
3
2
1
1
=4 9
4
16
5. Since BB T is always defined (see exercise 34 in section 3.2) and is square, (BB T )
BB T is invertible. We have
2
3


2
3
2
0
1 4
5 2
0
35 =
BB T =
.
3
3
4
2 34
1
4
Then det(BB T ) = 5 · 34
1
=
1
166

34
2
BT B
2
13
6
I =4 9
10
1
makes sense if
1
makes sense if
2
.
5
6. Since B T B is always defined (see exercise 34 in section 3.2) and is square, (B T B)
B T B is invertible. We have
2
3
2
3
2
3 
13
9
10
2
0
1
35
9
125 .
BT B = 4 0
=4 9
3
3
4
1
4
10
12
17
⇥
⇤
To compute the inverse, we row-reduce B T B I :
i
3
5
215 .
41
22̇ = 166, so that
(BB T )
h
3
9
18
9
9
12
10
12
17
1
0
0
0
1
0
3
0
7
05
1
R3
2
13
6
!4 0
10
13 R1
! 0
9
R2 + 13
R1
9
10
1
36
13
66
13
66
13
121
13
9
13
10
13
2
13
6
4 0
R3 + 66
36 R2
! 0
9
10
1
36
13
66
13
0
0
9
13
1
2
Since the matrix on the left has a zero row, B T B is not invertible, so (B T B)
7. The outer product expansion of AAT is

1 ⇥
T
T
T
1
AA = a1 A1 + a2 A2 =
3

⇤
2 ⇥
3 +
2
5

⇤
1
5 =
3
3
0 0
7
1 05
0 1
1
3
0 0
7
1 05 .
11
1
6
does not exist.


3
4 10
5
+
=
9
10 25
13
13
.
34
8. By Theorem 3.9 in Section 3.3, we know that (A 1 ) 1 = A, so we must find the inverse of the given
3
1
matrix. Now, det A 1 = 12 · 4 ( 1)
2 = 2 , so that
"
# "
#
8 2
1 4 1
1
1
A = (A ) =
=
.
1/2 32 12
3 1
2
1
9. Since AX = B = 4 5
3
3
3
05, we have X = A 1 B. To find A
2
2
3
2
1 0
1 1 0 0
1 0 0
6
7
6
1 0 1 05 ! 40 1 0
42 3
0 1
1 0 0 1
0 0 1
1
, we compute
2
1
1
1
2
1
2
1
2
3
3
2
17
.
25
3
2
232
CHAPTER 3. MATRICES
Then
2
32
1
2
1
2
1
2
2
6
X = A 1B = 4 1
1
3
2
17 6
25 4
3
2
1
5
3
3 2
3
0
7 6
05 = 42
2
1
3
9
7
45 .
6
10. Since det A = 1 · 6 2 · 4 = 2 6= 0, A is invertible, so by Theorem 3.12 in Section 3.3, it can be written
as a product of elementary matrices. As in Example 3.29 in Section 3.3, we start by row-reducing A:
A=
So by applying

1
4
2
6
"
1
E1 =
4

1
! 0
R2 4R1
#
0
,
1

1
! 0
2
2
1
2 R2
"
1
E2 =
0
0
1
2
#
2
1

! 1
0
R1 2R2
h
E3 = 1
,
2
0
0
.
1
i
1
in that order to A we get the identity matrix, so that E3 E2 E1 A = I. Since elementary matrices are
invertible, this means that A = (E3 E2 E1 ) 1 = E1 1 E2 1 E3 1 . Since E1 is R2 4R1 , it follows that
E1 1 is R2 + 4R1 . Since E2 is 12 R2 , it follows that E2 1 is 2R2 . Since E3 is R1 2R2 , it follows that
E3 1 is R1 + 2R2 . Thus



1 0 1
0 1 2
1
1
1
A = E1 E2 E 3 =
.
1 1 0
2 0 1
11. To show that (I
(I + A + A2 )(I
A)
1
A) = II
= I + A + A2 , it suffices to show that (I + A + A2 )(I
IA + AI
A2 + A2 I
A3 = I
A+A
A2 + A2
A) = I. But
A3 = I + A3 = I + O = I.
12. Use the multiplier method. We want to reduce A to an upper triangular matrix:
2
1
A = 43
2
3
1 1
1 15
1 1
R3
2
1
! 40
2R1
! 0
R2 3R1
3
1
25
1
1
2
3
R3
2
1
40
3
2 R2
! 0
1
2
0
3
1
25 = U.
2
Then `21 = 3, `31 = 2, and `32 = 32 , so that
2
1
6
L = 43
2
13. We follow the comment after Example 3.48
form of A:
2
3
2
2
4 5 8 5
1
2 4
41
2 2 3 15 R1 $R!
2
4
8 3 2 6
4
2
1
40
R3 +5R2
! 0
R1 R3 2
! 1
R2 3R3 40
!
0
0
1
3
2
3
0
7
05 .
1
in Section 3.5. Start by finding the reduced row echelon
2
4
8
2
5
3
3
8
2
2 2 3
0 1 2
0 0 0
2 2 3
0 1 2
0 0 0
3
2
1
1
2
R2 2R1
55
! 40
0
1
6 R3 4R!
0
0
3
2
1
1
2 2
40
35 1
0 1
3
8 8 R!
0
0 0
3
2
2
0 R1 2R!
1
2
40
05
0
1
0
0
2
1
5
3
3 1
2 35
0 1
0
1
0
3
3 1
2 35
10 2
3
1 0
2 05 .
0 1
Chapter Review
233
Thus a basis for row(A) is given by the nonzero rows of the matrix, or
⇥
⇤ ⇥
⇤ ⇥
2 0
1 0 , 0 0 1 2 0 , 0 0 0
{ 1
0
⇤
1 }.
Note that since the matrix is rank 3, so all three rows are nonzero, the rows of the original matrix also
provide a basis for row(A).
A basis for col(A) is given by the column vectors of the original matrix A corresponding to the leading
1s in the row-reduced matrix, so a basis for col(A) is
8 2 3 2 3 2 39
5
5 =
< 2
415 , 425 , 415 .
:
;
4
3
6
Finally, from the row-reduced form, the solutions to Ax = 0 where
2 3
x1
6x2 7
6 7
7
x=6
6x3 7
4x4 5
x5
can be found by noting that x2 = s and x4 = t are free variables and that x1 = 2s + t, x3 =
x5 = 0. Thus the nullspace of A is
2
3
2 3
2 3
2s + t
2
1
6 s 7
617
6 07
6
7
6 7
6 7
6 2t 7 = s 607 + t 6 27 ,
6
7
6 7
6 7
4 t 5
405
4 15
0
0
0
2t, and
so that a basis for the nullspace is given by
2 3
2
617
6 7
7
{6
607 ,
405
0
2
6
6
6
6
4
3
1
07
7
27
7}.
15
0
14. The row spaces will be the same since by Theorem 3.20 in Section 3.5, if A and B are row-equivalent,
they have the same row spaces. However, the column spaces need not be the same. For example, let
2
3
2
3
1 0 0
1 0 0
40 1 05 = B.
A = 40 1 05
1
1 0 0 R3 R!
0 0 0
Then A and B are row-equivalent, but
02 3 2 3 1
02 3 2 31
1
0
1
0
col(A) = span @405 , 415A but col(B) = span @405 , 415A .
1
0
0
0
15. If A is invertible, then so is AT . But by Theorem 3.27 in Section 3.5, invertible matrices have trivial
null spaces, so that null(A) = null(AT ) = 0. If A is not invertible, then the null spaces need not be
the same. For example, let
2
3
2
3
1 1 1
1 0 0
A = 40 0 05 , so that AT = 41 0 05 .
0 0 0
1 0 0
234
CHAPTER 3. MATRICES
Then AT row reduces to
2
1
40
0
But
0
0
0
3
02 3 2 31
0
0
0
05 , so that null(AT ) = span @415 , 405A .
0
0
1
02
3 2 31
1
1
null(A) = span @4 15 , 4 05A .
0
1
Since every vector in null(AT ) has a zero first component, neither of the basis vectors for null(A) is in
null(AT ), so the spaces cannot be equal.
16. If the rows of A add up to zero, then the rows are linearly dependent, so by Theorem 3.27 in Section
3.5, A is not invertible.
17. Since A has n linearly independent columns, rank(A) = n. Theorem 3.28 of Section 3.5 tells us that
rank(AT A) = rank(A) = n, so AT A is invertible since it is an n ⇥ n matrix. For the case of AAT ,
since m may be greater than n, there is no reason to believe that this matrix must be invertible. One
counterexample is
2
3
2
3
2
3
2
3

1 0
1 0 
1 0 1
1 0 1
1 0 1
1 0 1
A = 40 15 , AT =
, AAT = 40 15
= 40 1 05 ! 40 1 05 .
0 1 0
0 1 0
1 0
1 0
1 0 1
0 0 0
Thus AAT is not invertible, since it row-reduces to a matrix with a zero row.
⇥
⇤
18. The matrix of T is [T ] = T (e1 ) T (e2 ) . From the given data, we compute

✓ 

◆



1 1
1
1 2
1 0
1
1
1
T
=T
+
=
+
=
0
1
4
2 1
2
2 3
2 5

✓ 

◆



1 1
1
1 2
1 0
0
1
1
T
=T
=
=
.
1
1
1
2 1
2
2 3
2 5
Thus
[T ] =

1
4
1
.
1
19. Let R be the rotation and P the projection. The given line has direction vector
Examples 3.58 and 3.59 in Section 3.6,
"
cos 45
[R] =
sin 45
sin 45
cos 45
"
12
1
[P ] = 2
1 + 22 1 · ( 2)
So R followed by P is
[P
R] = [P ][R] =
"
1
5
2
5
2
5
4
5
#
=
"p
1 · ( 2)
( 2)2
# "p
2
p2
2
2
p
2
p2
2
2
#
p
2
p2
2
2
=
2
p2
2
2
#
"
=
#
1
5
2
5
"
2
5
4
5
p
2
10
p
2
5
#

1
. Then from
2
.
p #
3 2
10
p
.
3 2
5
20. Suppose that cv + dT (v) = 0. We want to show that c = d = 0. Since T is linear and T 2 (v) = 0,
applying T to both sides of this equation gives
T (cv + T (dT (v))) = cT (v) + dT (T (v)) = cT (v) + dT 2 (v) = cT (v) = 0.
Since T (v) 6= 0, it follows that c = 0. But then cv + dT (v) = dT (v) = 0. Again since T (v) 6= 0, it
follows that d = 0. Thus c = d = 0 and v and T (v) are linearly independent.
Chapter 4
Eigenvalues and Eigenvectors
4.1
Introduction to Eigenvalues and Eigenvectors
1. Following example 4.1, to see that v

0 3
Av =
3 0
is an eigenvector of A, we show that Av is a multiple of v:




1
0·1+3·1
3
1
=
=
=3
= 3v.
1
3·1+0·1
3
1
Thus v is an eigenvector of A with eigenvalue 3.
2. Following example 4.1, to see that v is an eigenvector of A, we show that Av is a multiple of v:




1 2
3
1 · 3 + 2 · ( 3)
3
Av =
=
=
= v
2 1
3
2 · 3 + 1 · ( 3)
3
Thus v is an eigenvector of A with eigenvalue
1.
3. Following example 4.1, to see that v is an eigenvector of A, we show that Av is a multiple of v:





1 1
1
1 · 1 + 1 · ( 2)
3
1
Av =
=
=
= 3
= 3v.
6 0
2
6 · 1 + 0 · ( 2)
6
2
Thus v is an eigenvector of A with eigenvalue
3.
4. Following example 4.1, to see that v is an eigenvector of A, we show that Av is a multiple of v:




4
2 4
4·4 2·2
12
Av =
=
=
= 3v
5
7 2
5·4 7·2
6
Thus v is an eigenvector of A with eigenvalue 3.
5. Following example 4.1, to see that v is an eigenvector of A, we show that Av is
2
32 3 2
3 2 3
2
6
3 0
0
2
3 · 2 + 0 · ( 1) + 0 · 1
25 4 15 = 40 · 2 + 1 · ( 1) 2 · 15 = 4 35 = 3 4
Av = 40 1
1 0
1
1
1 · 2 + 0 · ( 1) + 1 · 1
3
Thus v is an eigenvector of A with eigenvalue 3.
a multiple of v:
3
2
15 = 3v.
1
6. Following example 4.1, to see that v is an eigenvector of A, we show that Av is a multiple of v:
2
32 3 2
3 2 3
0 1
1
2
0 · ( 2) + 1 · 1 1 · 1
0
15 4 15 = 41 · ( 2) + 1 · 1 + 1 · 15 = 405 = 0v
Av = 41 1
1 2
0
1
1 · ( 2) + 2 · 1 + 0 · 1
0
Thus v is an eigenvector of A with eigenvalue 0.
235
236
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
7. We want to show that
= 3 is an eigenvalue of
A=

2
2
2
1
Following Example 4.2, that means we wish to find a vector v with Av = 3v, i.e. with (A
So we show that (A 3I)v = 0 has nontrivial solutions.



2
2
3 0
1
2
A 3I =
=
.
2
1
0 3
2
4
Row-reducing, we get

1
2

! 1
2
R1
2
4
2
4

1
! 0
R2 2R1
3I)v = 0.
2
.
0

x
It follows that any vector
with x = 2y is an eigenvector, so for example one eigenvector correy

2
sponding to = 3 is
.
1
8. We want to show that
=
1 is an eigenvalue of
A=

2
3
3
2
Following Example 4.2, that means we wish to find a vector v with Av =
(A + I)v = 0. So we show that (A + I)v = 0 has nontrivial solutions.



2 3
1 0
3 3
A+I =
+
=
.
3 2
0 1
3 3
Row-reducing, we get

3
3
3
3
R2

x
It follows that any vector
with x =
y

1
sponding to = 1 is
.
1
9. We want to show that

3
R1
! 0

! 1
0
1
3 R1
3
0
v, i.e. with (A
1
0
y is an eigenvector, so for example one eigenvector corre-
= 1 is an eigenvalue of
A=

0
1
4
5
Following Example 4.2, that means we wish to find a vector v with Av = 1v, i.e. with (A
So we show that (A I)v = 0 has nontrivial solutions.



0 4
1 0
1 4
A I=
=
.
1 5
0 1
1 4
Row-reducing, we get
( I))v =

1
1
4
4
R1
!

1
1
4
4

1
! 0
R2 +R1
I)v = 0.
4
.
0

x
It follows that any vector
with x = 4y is an eigenvector, so for example one eigenvector correy

4
sponding to = 1 is
.
1
4.1. INTRODUCTION TO EIGENVALUES AND EIGENVECTORS
10. We want to show that
=
6 is an eigenvalue of

4
A=
5
237
2
7
Following Example 4.2, that means we wish to find a vector v with Av = 6v, i.e. with (A ( 6I))v =
(A + 6I)v = 0. So we show that (A + 6I)v = 0 has nontrivial solutions.



4
2
6 0
10
2
+
=
.
A + 6I =
5
7
0 6
5
1
Row-reducing, we get
"
10
5
#
2
1

R2
x
It follows that any vector
with x =
y

1
sponding to = 6 is
.
5
11. We want to show that
=
1
2 R1
!
1
5y
"
#
2
10
0
"
! 1
0
1
10 R1
0
1
5
0
#
is an eigenvector, so for example one eigenvector corre-
1 is an eigenvalue of
2
1
A=4 1
2
0
1
0
3
2
15
1
Following Example 4.2, that means we wish to find a vector v with Av = 1v, i.e. with (A + I)v = 0.
So we show that (A + I)v = 0 has nontrivial solutions.
2
3 2
3 2
3
1 0 2
1 0 0
2 0 2
A + I = 4 1 1 15 + 40 1 05 = 4 1 2 15 .
2 0 1
0 0 1
2 0 2
Row-reducing, we get
2
2
4 1
2
0
2
0
3
2
15
2
R3
2
2
4 1
R1
0
!
0
2
0
3
2
15
0
2
1 2
R1 $R2 4
2 0
!
0 0
2
1
2
40
4
0
0
2 3
x
It follows that any vector 4y 5 with x = y =
2z 3
1
corresponding to = 1 is 4 15.
1
12. We want to show that
= 2 is an eigenvalue of
2
3
A = 41
4
3
1
25
0
3
1
45
0
2
! 1
42
0
2
1
1
4 R2 4 0
!
0
R1
2
0
0
2
1
0
3
1
1
25 R2 2R!
0
3
2
2
1 R1 +2R!
1
40
15
0
0
0
1
0
3
1
15
0
z is an eigenvector, so for example one eigenvector
1
1
2
3
1
15
0
238
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
Following example 4.2, that means we wish to find a vector v with Av = 2v, i.e. with (A
So we show that (A 2I)v = 0 has nontrivial solutions.
2
3 2
3 2
3
3 1
1
2 0 0
1
1
1
15 40 2 05 = 41
1
15
A 2I = 41 1
4 2
0
0 0 2
4
2
2
2I)v = 0.
Row-reducing, we have
2
1
41
4
1
1
2
3
1
15
2
R2
R3
2
1
R1
! 40
4R1
! 0
It follows that any vector
eigenvector for
= 2.
3
1
25
2
1
2
2
hxi
y
z
R3
2
1
40
R2
! 0
1
2
0
3
1
25
0
2
1
1
2 R2 4 0
!
0
1
1
0
3
1
15
0
2
! 1
40
0
R1 R2
0
1
0
3
0
15
0
2 3
0
with x = 0 and y = z is an eigenvector. Thus for example 415 is an
1
13. Since A is the reflection F in the y-axis,
 the only vectors that F maps parallel to themselves are vectors
1
parallel to the x-axis (multiples of
), which are reversed (eigenvalue = 1) and vectors parallel
0

0
to the y-axis (multiples of
), which are sent to themselves (eigenvalue = 1). So = ±1 are the
1
eigenvalues of A, with eigenspaces
✓ ◆
✓ ◆
1
0
E 1 = span
E1 = span
.
0
1
F !e1 " ! e1
e!1
F !e!1 " !" e!1
x
F !x"
14. The only vectors that the transformation F corresponding to A maps to themselves are vectors perpendicular to y = x, which are reversed (so correspond to the eigenvalue = 1) and vectors parallel
to y = x, which are sent to
 themselves (so correspond to the eigenvalue = 1. Vectors perpendicular
1
1
to y = x are multiples of
; vectors parallel to y = x are multiples of
, so the eigenspaces are
1
1
E
1
= span
✓
1
1
◆
E1 = span
✓
1
1
◆
.
4.1. INTRODUCTION TO EIGENVALUES AND EIGENVECTORS
239
F !e1 " ! e1
F !e!1 " !" e!1
F !x"
e!1
x
15. If x is an eigenvalue of A, then Ax is a multiple of x:



1 0 x
x
Ax =
=
.
0 0 y
0

x
, is an eigenvector for A with
0
eigenvalue 1. But this projection also transforms vectors parallel to the y-axis
 to the zero vector, since
0
the projection of such a vector on the x-axis is 0. So vectors of the form
are eigenvectors for A
y

x
with eigenvalue 0. Finally, a vector of the form
where neither x nor y is zero is not an eigenvector:
y

x
it is transformed into
, and comparing the x-coordinates shows that would have to be 1, while
0
comparing y-coordinates shows that it would have to be zero. Thus there is no such . So in summary,
✓ ◆
1
= 1 is an eigenvalue with eigenspace E1 = span
0
✓ ◆
0
= 0 is an eigenvalue with eigenspace E0 = span
.
1
Then clearly any vector parallel to the x-axis, i.e., of the form
16. This projection sends a vector to its projection on the given line. Then a vector parallel to the direction
vector of that line will be sent to itself, since its projection on the line is the vector itself. A vector
orthogonal to the direction vector of the line will be "sent
zero vector, since its projection on the
# to the

4
4
line is the zero vector. Since the direction vector is 53 , or
, there are two eigenspaces
3
5
span
✓
4
3
◆
,
= 1;
span
✓
3
4
◆
,
=
1.
If a vector v is neither parallel to nor orthogonal to the direction vector of the line, then the projection
transforms v to a vector parallel to the direction vector, thus not parallel to v. Thus the image of v
cannot be a multiple of v. Hence the eigenvalues and eigenspaces above are the only ones.

x
17. Consider the action of A on a vector x =
. If y = 0, then A multiplies x by 2, so this is an
y
eigenspace for = 2. If x = 0, then A multiplies x by 3, so this is an eigenspace for = 3. If neither x
240
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
nor y is zero, then


x
2x
is mapped to
, which is not a multiple of
y
3y
✓ ◆
✓ ◆
1
0
span
, = 2;
span
,
0
1

x
. So the only eigenspaces are
y
= 3.
18. Since A rotates the plane 90 counterclockwise around the origin, any nonzero vector is taken to
a vector orthogonal to itself. So there are no nonzero vectors that are taken to vectors parallel to
themselves (that is, to multiples of themselves). The only vector taken to a multiple of itself is the
zero vector; this is not an eigenvector since eigenvectors are by definition nonzero. So this matrix has
no eigenvectors and no eigenvalues. Analytically, the matrix of A is


cos ✓
sin ✓
0
1
A=
=
,
sin ✓
cos ✓
1
0
so that

0
Ax =
1
1
0

x
=
y

y
.
x
If this is a multiple of the original vector, then there is a constant a such that x = ay and y = ax.
Substituting ax for y in the first equation gives x = a2 x, which is impossible unless x = 0. But then
the second equation forces y = 0. So again we see that there are no eigenvectors.
19. Since eigenvectors x are those taken to a multiple of themselves, i.e., to a vector parallel to themselves,
by A, we look for unit vectors whose image under A is in the same direction.
✓ It
◆ appears that these
✓ vec◆
1
0
tors are vectors along the x and y-axes, so the eigenspaces are E1 = span
and E2 = span
.
0
1
The length of the resultant vectors in the first eigenspace appears are same length as the original vector, so that 1 = 1. The length of the resultant vectors in the second eigenspace are twice the original
length, so that 2 = 2.
20. Since eigenvectors x are those taken to a multiple of themselves, i.e., to a vector parallel to themselves,
by A, we look for unit vectors whose image under A is in the✓
same◆direction. The only vectors satisfying
1
this criterion are those parallel to the line y = x, i.e., span
, so this is the only eigenspace. The
1

1
terminal point of the line in the direction
appears to be the point (2, 5), so that it has a length of
1
p
p
29. Since the original vector is a unit vector, the eigenvalue is 29.
21. Since eigenvectors x are those taken to a multiple of themselves, i.e., to a vector parallel to themselves,
by A, we look for unit vectors
✓ ◆whose image under A is in the same direction. The vectors parallel to
1
the line y = x, i.e., span
satisfy this criterion, so this is an eigenspace E1 . The vectors parallel
1
✓
◆
1
to the line y = x, i.e. span
, also satisfy this criterion, so this is another eigenspace E2 . For
1
p
E1 , the terminal point of the line along y = x appears to be (2, 2),pso its length is 2 2. Since the
original vector is a unit vector, the corresponding eigenvalue is 1 = 2 2. For E2 , the resultant vectors
are zero length, so 2 = 0.
22. Since eigenvectors x are those taken to a multiple of themselves, i.e., to a vector parallel to themselves,
by A, we look for unit vectors whose image under A is in the same direction. There are none — every
vector is “bent” to the left by the transformation. So this transformation has no eigenvectors.
23. Using the method of Example 4.5, we wish to find solutions to the equation
✓

◆

4
1
0
4
1
0 = det(A
I) = det
= det
2
1
0
2
1
= (4
)(1
)
2 · ( 1) =
2
5 +6=(
2)(
3).
4.1. INTRODUCTION TO EIGENVALUES AND EIGENVECTORS
241
Thus the eigenvalues are 2 and 3.
To find the eigenspace corresponding to
h
A
= 2, we must find the null space of A
"
#
"
#
i
1
2
1 0
1
0
2
!
0 =
2
1 0
0
0 0
2I
so that eigenvectors corresponding to this eigenvalue have the form
1

t
2t ,
or
.
2t
t
A basis for this eigenspace is given by choosing for example t = 1 to get


1
4
A
=
2
2
1
1


1
. As a check,
2



1
4·1 1·2
2
1
=
=
=2
.
2
2·1+1·2
4
2
To find the eigenspace corresponding to
⇥
A
= 3, we must find the null space of A


⇤
1
1 0
1
1 0
0 =
!
2
2 0
0
0 0
3I
so that eigenvectors corresponding to this eigenvalue have the form

t
.
t
A basis for this eigenspace is given by choosing for example t = 1 to get
A


1
4
=
1
2
1
1


4
3
2v
2
3w
v
w
1
2
3
3I.
1
. As a check,
1



1
4·1 1·1
3
1
=
=
=3
1
2·1+1·1
3
1
The plot below shows the e↵ect of A on the eigenvectors


1
1
v=
w=
.
2
1
1
2I.
242
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
24. Using the method of Example 4.5, we wish to find solutions to the equation
0 = det(A
I) = det
Thus the eigenvalues are
✓
2
6

4
0
0
0
◆
= det
= (2

2
4
6
)(
2
4·6=
)
2
24 = (
6)( + 4).
4 and 6.
To find the eigenspace corresponding to
h
=
4, we must find the null space of A + 4I.
"
i
6
0 =
6
A + 4I
4
0
4
0
#
"
!
1
2
3
0
0
#
0
0
so that eigenvectors corresponding to this eigenvalue have the form

2
3t
t
,
or

2t
3t
A basis for this eigenspace is given by choosing for example t = 1 to get
A


2
2
=
3
6
4
0



2
2 · ( 2) + 4 · 3
8
=
=
=
3
6 · ( 2) + 0 · 3
12
To find the eigenspace corresponding to
⇥
A
6I

2
. As a check,
3
4

= 6, we must find the null space of A
⇤
0 =

4
6
4
6

0
1
!
0
0
1
0
2
.
3
6I.
0
0
so that eigenvectors corresponding to this eigenvalue have the form

t
t
A basis for this eigenspace is given by choosing for example t = 1 to get


1
2
A
=
1
6
4
0

1
. As a check,
1



1
2·1+4·1
6
1
=
=
=6
1
6·1+0·1
6
1
The plot below shows the e↵ect of A on the eigenvectors
v=


2
3
w=

1
.
1
4.1. INTRODUCTION TO EIGENVALUES AND EIGENVECTORS
243
6
6w
4
v
2
w
2
!2
4
6
8
!2
!4
!6
!8
!10
!4 v
!12
25. Using the method of Example 4.5, we wish to find solutions to the equation
✓

◆

2 5
0
2
5
0 = det(A
I) = det
= det
0 2
0
0
2
= (2
Thus the only eigenvalue is
)
5·0=(
2)2
= 2.
To find the eigenspace corresponding to
h
A
2I
= 2, we must find the null space of A
"
#
"
#
i
0 5 0
0 1 0
!
0 =
0 0 0
0 0 0
so that eigenvectors corresponding to this eigenvalue have the form

t
.
0
A basis for this eigenspace is given by choosing for example t = 1, to get

1
2
=
0
0
5
2


2I.
1
. As a check,
0



1
2·1+5·0
2
1
=
=
=2
.
0
0·1+2·0
0
0

1
The plot below shows the e↵ect of A on the eigenvector v =
:
0
A

)(2
v
2v
1
2
26. Using the method of Example 4.5, we wish to find solutions to the equation
✓

◆

1 2
0
1
2
0 = det(A
I) = det
= det
2 3
0
2
3
= (1
Since this quadratic has no (real) roots, A has no (real) eigenvalues.
)(3
)+4=
2
4 +7
244
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
27. Using the method of Example 4.5, we wish to find solutions to the equation
0 = det(A
I) = det
✓
1
1
1
1

0
0
◆
= det

1
1
1
1
= (1
)(1
2
1 · ( 1) =
)
2 + 2.
Using the quadratic formula, this quadratic has roots
p
2 ± 22 4 · 2
2 ± 2i
=
=
= 1 ± i.
2
2
Thus the eigenvalues are 1 + i and 1
i.
To find the eigenspace corresponding to

⇥
⇤
i
A (1 + i)I 0 =
1
= 1 + i, we must find the null space of A (1 + i)I.


1
1 0 iR!
1
i 0
1 i 0
1
i 0
1
i 0 R2 +R!
0 0 0
so that eigenvectors corresponding to this eigenvalue have the form

it
.
t
A basis for this eigenspace is given by choosing for example t = 1, to get
To find the eigenspace corresponding to = 1

⇥
⇤
i 1 0
A (1 i)I 0 =
1 i 0

i
.
1
i, we must find the null space of A (1


iR1
i 0
1
i 0
! 1
R2 +R1
1
i 0
0 0
! 0
so that eigenvectors corresponding to this eigenvalue have the form

it
.
t
A basis for this eigenspace is given by choosing for example t = 1, to get

i)I.
i
.
1
28. We wish to find solutions to the equation
0 = det(A
I) = det
✓
2
1
3
0

0
0
To find the eigenspace corresponding to
A
p
(1 + i 2)I

⇤
1
0 =
p
i 2
1
= det

2
3
1
= (2
)(
)
( 3) · 1 =
2
2 + 3.
p
= 1 ± i 2, so these are the eigenvalues.
p
p
= 1 + i 2, we must find the null space of A (1 + i 2)I:
Using the quadratic formula, this has roots
⇥
◆
1

=
3p
i 2
1p
1 i 2
p
2± 4 4·3
2
0
0
R1 $R2
!
1
p
i 2
3
0
0
R2 (1 i
so that eigenvectors corresponding to this eigenvalue have the form

p
(1 + i 2)t
.
t
p

1
2)R1
! 0
p
i 2
1
0
0
,
0
4.1. INTRODUCTION TO EIGENVALUES AND EIGENVECTORS
245

p
1+i 2
. As a check,
1
p
p






p
p
p
p
2
3 1+i 2
2(1 + ip 2) 3 · 1
1 + 2i
1+i 2
p 2 = (1 + i 2) 1 + i 2
A
=
=
=
1
0
1
1
1
(1 + i 2) + 0 · 1
1+i 2
p
p
To find the eigenspace corresponding to = 1 i 2, we must find the null space of A (1 i 2)I.
p

⇥
⇤
p
1+i 2
3p
0 R1 $R2
!
A (1 i 2)I 0 =
1
1+i 2 0
p


p
1p
1+i 2 0
1
1+i 2 0
p
1
0
0
0
1+i 2
3
0 R2 (1+i 2)R!
A basis for this eigenspace is given by choosing for example t = 1, to get
so that eigenvectors corresponding to this eigenvalue have the form

p
(1 i 2)t
.
t

p
i 2
A basis for this eigenspace is given by choosing for example t = 1, to get
. As a check,
1
p
p






p
p
p
p
2
3 1 i 2
2(1 ip 2) 3 · 1
1 2i
2
1 i 2
1
i
2
p
A
=
=
=
= (1 i 2)
1
0
1
1
1
(1 i 2) + 0 · 1
1 i 2
1
29. Using the method of Example 4.5, we wish to find solutions to the equation
✓

◆

1 i
0
1
i
0 = det(A
I) = det
= det
i 1
0
i
1
= (1
)(1
)
i·i=
2
2 + 2.
Using the quadratic formula, this quadratic has roots
p
2 ± 22 4 · 2
2 ± 2i
=
=
= 1 ± i.
2
2
Thus the eigenvalues are 1 + i and 1 i.
To find the eigenspace corresponding to

⇤
⇥
i
A (1 + i)I 0 =
i
= 1 + i, we must find the null space of A (1 + i)I.


1
i 0 iR!
1
1 0
1
1 0
1
i 0
i
i 0 R2 iR!
0
0 0
so that eigenvectors corresponding to this eigenvalue have the form

t
.
t
A basis for this eigenspace is given by choosing for example t = 1, to get
To find the eigenspace corresponding to

⇥
⇤
i
A (1 i)I 0 =
i
=1
i
i

1
.
1
i, we must find the null space of A


iR1
0
1 1 0
! 1 1 0
1
0
i i 0 R2 iR!
0 0 0
so that eigenvectors corresponding to this eigenvalue have the form

t
.
t
A basis for this eigenspace is given by choosing for example t = 1, to get

1
.
1
(1
i)I.
246
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
30. Using the method of Example 4.5, we wish to find solutions to the equation
0 = det(A
I) = det
✓
0
1

1+i
1
i
=(
Thus the eigenvalues are 2 and
0
0
)(1
◆
= det
)

1
1+i
1
i
(1 + i)(1
i) =
2
2=(
2)( + 1)
1.
To find the eigenspace corresponding to = 2, we must find the null space of A 2I.
"
# 1
"
#
"
h
i
1
1
2 R1
2 1+i 0
1
(1 + i) 0
1
!
2
2 (1 + i)
A 2I 0 =
R2 (1 i)R1
1 i
1 0
1 i
1
0
0
! 0
#
0
0
so that eigenvectors corresponding to this eigenvalue have the form
"
#
"
#
1
(1 + i)t
2 (1 + i)t , or, clearing fractions,
.
t
2t
A basis for this eigenspace is given by choosing for example t = 1, to get
To find the eigenspace corresponding to
⇥
A+I
⇤
0 =

=
1
1
i

1+i
.
2
1, we must find the null space of A
1+i
2
0
0

1 1+i
0
! 0
R2 (1 i)R1
so that eigenvectors corresponding to this eigenvalue have the form

(1 + i)t
.
t
A basis for this eigenspace is given by choosing for example t = 1, to get

( I) = A + I.
0
0
1 i
.
1
31. Using the method of Example 4.5, we wish to find solutions in Z3 to the equation
0 = det(A
I) = det
✓
1
1

0
2
0
0
◆
= det

1
0
1
2
= (1
)(2
)
0·1=(
1)(
2).
+ 2 = ( + 2) = (
1).
Thus the eigenvalues are 1 and 2.
32. Using the method of Example 4.5, we wish to find solutions in Z3 to the equation
0 = det(A
I) = det
✓
2
1
= (2
1
2
)(2

0
0
◆
= det
1·1=
)
2

2
1
1
2
4 +4
1=
2
Thus the eigenvalues are 0 and 1.
33. Using the method of Example 4.5, we wish to find solutions in Z5 to the equation
0 = det(A
I) = det
✓
3
4
= (3
Thus the only eigenvalue is 4.
1
0

)(
0
0
)
◆
= det
1·4=

2
3
1
4
3
4=
2
+ 2 + 1 = ( + 1)2 = (
4)2 .
4.1. INTRODUCTION TO EIGENVALUES AND EIGENVECTORS
247
34. Using the method of Example 4.5, we wish to find solutions in Z5 to the equation
0 = det(A
I) = det
✓
1
4

4
0
= (1
)(
0
0
◆
= det
1
4
4
2
4·4=
)

16 =
2
+ 4 + 4 = ( + 2)2 = (
3)2 .
Thus the only eigenvalue is 3.
35. (a) To find the eigenvalues of A, find solutions to the equation
0 = det(A
I) = det
= (a
✓
a
c
b
d
)(d
)

0
0
bc =
◆
= det
2

a
b
c
d
(a + d) + (ad
bc) =
2
tr(A) + det(A).
(b) Using the quadratic formula to find the roots of this quadratic gives
p
tr(A) ± (tr(A))2 4 det(A)
=
2
p
a + d ± a2 + 2ad + d2 4ad + 4bc
=
2
⌘
p
1⇣
2
=
a+d± a
2ad + d2 + 4bc
2
⌘
p
1⇣
=
a + d ± (a d)2 + 4bc .
2
(c) Let
p
1⇣
a + d + (a
2
be the two eigenvalues. Then
1
1
=
+
2
1 2
⌘
d)2 + 4bc ,
2
=
1⇣
a+d
2
p
(a
d)2 + 4bc
⌘
⌘ 1⇣
⌘
p
p
1⇣
a + d + (a d)2 + 4bc +
a+d
(a d)2 + 4bc
2
2
1
= (a + d + a + d)
2
= a + d = tr(A),
✓ ⇣
⌘◆ ✓ 1 ⇣
⌘◆
p
p
1
2
2
=
a+d
(a d) + 4bc
a+d
(a d) + 4bc
2
2
1
=
(a + d)2
(a d)2 + 4bc
4
1
= (a2 + 2ad + d2 a2 + 2ad d2 4bc)
4
1
= (4ad 4bc) = det(A).
4
=
36. A quadratic equation will have two distinct real roots when its discriminant (the quantity under the
square root sign in the quadratic formula) is positive, one real root when it is zero, and no real roots
when it is negative. In Exercise 35, the discriminant is (a d)2 + 4bc.
(a) For A to have to distinct real eigenvalues, we must have (a
d)2 + 4bc > 0.
(b) For A to have a single real eigenvalue, it must be the case that (a d)2 + 4bc = 0, so that
(a d)2 = 4bc. Note that this implies that either a = d and at least one of b or c is zero, or that
a 6= d and either b or c, but not both, is negative.
(c) For A to have no real eigenvalues, we must have (a
c)2 + 4bc < 0.
248
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
37. From Exercise 35, since c = 0, the eigenvalues are
p
1⇣
a + d ± (a
2
⌘ 1
d)2 + 4b · 0 = (a + d ± |a
2
To find the eigenspace corresponding to
⇥
A
d|)
⇢
1
=
(a + d + a
2
d),
1
(a + d + d
2
= a, we must find the null space of A

⇤
0
b
0
aI 0 =
.
0 d a 0
a)
= {a, d}.
aI.
If b = 0 and a = d, then this is the zero matrix, so that any vector is an eigenvector; a basis for the
eigenspace is
⇢

1
0
,
.
0
1
The reason for this is that under these conditions, the matrix A is a multiple of the identity matrix, so
it simply stretches any vector by a factor of a = d. Otherwise, either b 6= 0 or d a 6= 0, so the above
matrix row-reduces to

0 1 0
0 0 0
so that eigenvectors corresponding to this eigenvalue have the form

t
.
0
A basis for this eigenspace is given by choosing for example t = 1, to get




1
a·1+b·0
a
1
A
=
=
=a
.
0
0·1+d·0
0
0
To find the eigenspace corresponding to
⇥
A

1
. As a check,
0
= d, we must find the null space of A

⇤
a d b 0
dI 0 =
.
0
0 0
dI.
If b = 0 and a = d, then this is again the zero matrix, which we dealt with above. Otherwise, either
b 6= 0 or d a 6= 0, so that eigenvectors corresponding to this eigenvalue have the form


bt
bt
, or
.
(a d)t
(d a)t
A basis for this eigenspace is given by choosing for example t = 1, to get
A

b
d

a · b + b · (d
=
a
0 · b + d · (d

b
d
a
. As a check,


a)
bd
b
=
=d
.
a)
d(d a)
d a
38. Using the method of Example 4.5, we wish to find solutions to the equation
0 = det(A
I) = det
✓
a
b
b
a

0
0
◆
= det
= (a

a
b
b
)(a
a
)
4 · ( b) =
2
2a + (a2 + 4b).
4.1. INTRODUCTION TO EIGENVALUES AND EIGENVECTORS
249
Using Exercise 35, the eigenvalues are
p
1⇣
a + a ± (a
2
⌘
1p
a)2 + 4b · ( b) = a ±
4b2 = a ± bi.
2
To find the eigenspace corresponding to
⇥
A
= a + bi, we must find the null space of A
⇤
0 =
(a + bi)I

bi
b
b
bi
(a + bi)I.
0
.
0
If b = 0 then this is the zero matrix, so that any vector is an eigenvector; a basis for the eigenspace is
⇢
1
,
0

0
1
.
The reason for this is that under these conditions, the matrix A is a times the identity matrix, so it
simply stretches any vector by a factor of a. Otherwise, b 6= 0, so we can row-reduce the matrix:

bi
b
b
bi
1
b iR1
0
0
!

1
b
i
bi
0
0

1
! 0
R2 +bR1
i
0
0
.
0
Thus eigenvectors corresponding to this eigenvalue have the form


it
t
=
.
t
it
A basis for this eigenspace is given by choosing for example t = 1, to get

1
A
=
i
To find the eigenspace corresponding to
⇥
A
(a


1
. As a check,
i

a + bi
1
= (a + bi)
.
b + ai
i
=a
bi, we must find the null space of A
⇤
bi)I
0 =

bi
b
b
bi
(a
bi)I.
0
.
0
If b = 0 then this is the zero matrix, which we dealt with above. Otherwise, b 6= 0, so we can row-reduce
the matrix:



1
b iR1
bi b 0
i 0
1
i 0
! 1
.
1
b bi 0
b bi 0 R2 +bR!
0 0 0
Thus eigenvectors corresponding to this eigenvalue have the form

it
.
t
A basis for this eigenspace is given by choosing for example t = 1, to get
A

i
=
1


ai + b
b + ai
=
= (a
bi + a
a bi
bi)


i
.
1
i
. As a check,
1
250
4.2
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
Determinants
1. Expanding along the first row gives
1
5
0
0
1
1
3
1
1 =1
1
2
1
2
0
5
0
1
5
+3
2
0
1
= 1(1 · 2
1
1 · 1)
0 + 3(5 · 1
3
= 1(1 · 2
1
1 · 1)
5(0 · 2
1 · 0) = 16.
Expanding along the first column gives
1
5
0
0
1
1
3
1
1 =1
1
2
1
2
5
0
1
3
0
+0
2
1
3 · 1) + 0 = 16.
2. Expanding along the first row gives
0
2
1
1
3
3
1
3
2 =0
3
0
=0
2
0
2
1
1
1(2 · 0
2
+ ( 1)
0
( 2) · ( 1))
2
1
1(2 · 3
3
3
3 · ( 1)) =
7.
Expanding along the first column gives
0
2
1
1
3
3
1
3
2 =0
3
0
=0
2
0
2
2(1 · 0
1
3
1
1
+ ( 1)
0
3
( 1) · 3)
1(1 · ( 2)
1
2
( 1) · 3) =
7.
3. Expanding along the first row gives
1
1
0
1
0
1
0
0
1 =1
1
1
1
1
( 1)
1
0
1
+0
1
1
0
0
= 1(0·( 1) 1·1)+1( 1·( 1) 1·0)+0 = 0.
1
0
+0
1
1
0
0
= 1(0·( 1) 1·1)+1( 1·( 1) 0·1)+0 = 0.
1
Expanding along the first column gives
1
1
0
1
0
1
0
0
1 =1
1
1
1
1
( 1)
1
1
4. Expanding along the first row gives
1
1
0
1
0
1
0
0
1 =1
1
1
1
1
1
1
0
1
1
+0
1
0
0
= 1(0 · 1
1
1 · 1)
1(1 · 1
1 · 0) + 0 =
2.
0
= 1(0 · 1
1
1 · 1)
1(1 · 1
1 · 0) + 0 =
2.
Expanding along the first column gives
1
1
0
1
0
1
0
0
1 =1
1
1
1
1
1
1
1
0
1
+0
1
0
5. Expanding along the first row gives
1 2
2 3
3 1
3
3
1 =1
1
2
1
2
2
2
3
1
2
+3
2
3
3
= 1(3 · 2
1
1 · 1)
2(2 · 2
1 · 3) + 3(2 · 1
3 · 3) =
18.
4.2. DETERMINANTS
251
Expanding along the first column gives
1 2
2 3
3 1
3
3
1 =1
1
2
1
2
2
2
3
1
2
+3
2
3
3
= 1(3 · 2
1
1 · 1)
2(2 · 2
1 · 3) + 3(2 · 1
3 · 3) =
18.
6. Expanding along the first row gives
1
4
7
2
5
8
3
5
6 =1
8
9
6
9
2
4
7
6
4
+3
9
7
5
= 1(5 · 9
8
6 · 8)
2(4 · 9
6 · 7) + 3(4 · 8
5 · 7) = 0.
3
= 1(5 · 9
6
6 · 8)
4(2 · 9
3 · 8) + 7(2 · 6
3 · 5) = 0.
Expanding along the first column gives
1
4
7
2
5
8
3
5
6 =1
8
9
6
9
4
2
8
3
2
+7
9
5
7. Noting the pair of zeros in the third row, use cofactor expansion along the third row to get
5
1
3
2
1
0
2
2
2 =3
1
0
2
2
5
1
0
2
+0
2
5
1
2
= 3(2 · 2
1
2 · 1)
0 + 0 = 6.
8. Noting the zero in the second row, use cofactor expansion along the second row to get
1
2
3
1
0
2
1
1 =
1
2
1
2
1
1
1
1
3
1
=
2
2(1 · 1
( 1) · ( 2))
1(1 · ( 2)
1 · 3) = 7.
9. Noting the zero in the bottom right, use the third row, since the multipliers there are smaller:
4
2
1
1 3
2 4 =1
1 0
1
2
3
4
( 1)
4
2
3
= 1(1 · 4
4
3 · ( 2)) + 1( 4 · 4
3 · 2) =
12.
10. Noting that the first column has only one nonzero entry, we expand along the first column:
cos ✓
0
0
sin ✓
cos ✓
sin ✓
tan ✓
sin ✓ = cos ✓(cos ✓ · cos ✓
cos ✓
( sin ✓) · sin ✓) = cos ✓(cos2 ✓ + sin2 ✓) = cos ✓.
11. Use the first column:
a
0
a
b
a
0
0
a
b =a
0
b
b
b
+a
b
a
0
= a(a · b
b
b · 0) + a(b · b
0 · a) = a2 b + ab2 .
12. Use the first row (other choices are equally easy):
0
b
0
a
c
e
0
d =
0
a
b d
=
0 0
a(b · 0
d · 0) = 0.
252
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
13. Use the third row, since it contains only one nonzero entry:
1
2
0
1
1 0 3
1
5 2 6
= ( 1)3+2 · 1 2
1 0 0
1
4 2 1
0
2
2
3
6 =
1
1
2
1
0
2
2
3
6 .
1
To compute this determinant, use cofactor expansion along the first row:
1
2
1
0
2
2
3
6 =
1
2
2
1
6
1
2
1
3
2
=
2
1(2 · 1
6 · 2)
3(2 · 2
2 · 1) = 4.
14. Noting that the second column has only one nonzero entry, we expand along that column. The nonzero
entry is in row 3, column 2, so it gets a minus sign, since ( 1)3+2 = 1:
2
1
0
2
0 3
0 2
1 1
0 1
1
2
=
4
3
2
( 1) 1
2
3
2
1
1
2
2 = 1
3
2
3
2
1
1
2 .
3
To evaluate the remaining determinant, we use expansion along the first row:
2
1
2
3
2
1
1
2
2 =2
1
3
2
3
3
= 2(2 · ( 3)
1
2
2
3
2 · 1)
1
2
1
2
1
3(1 · ( 3)
2 · 2)
1(1 · 1
15. Expand along the first row:
0
0
0
g
0
0
d
h
0 a
b c
=
e f
i j
0
a 0
g
0
d
h
b
e .
i
Now expand along the first row again:
0
a 0
g
0
d
h
b
e =
i
ab
0
g
d
=
h
ab(0 · h
dg) = abdg.
16. Using the method of Example 4.9, we get
105
1
4
7
2 3
5 6
8 9
1
4
7
48
2
5
8
45
so that the determinant is
72
84
96
(105 + 48 + 72) + (45 + 84 + 96) = 0.
17. Using the method of Example 4.9,
1
2
3
so that the determinant is
1
0
2
1
1
1
0
2
1
2
3
1
0
2
0
3
(0 + ( 2) + 2) + (0 + 3 + 4) = 7.
2
4
2 · 2) = 8.
4.2. DETERMINANTS
253
18. Using the method of Example 4.9, we get
0
a
0
a
b
a
0
0
b
b
0
a
0
a
b
a
0
a2 b
so that the determinant is
2
0
ab2
0
2
(0 + 0 + 0) + (a b + ab + 0) = a2 b + ab2 .
19. Consider (2) in the text. The three upward-pointing diagonals have product a31 a22 a13 , a32 a23 a11 , and
a33 a21 a12 , while the downward-pointing diagonals have product a11 a22 a33 , a12 a23 a31 , and a13 a21 a32 ,
so the determinant by that method is
det A =
(a31 a22 a13 + a32 a23 a11 + a33 a21 a12 ) + (a11 a22 a33 + a12 a23 a31 + a13 a21 a32 ).
Expanding along the first row gives
det A = a11
a22
a32
a23
a33
= a11 a22 a33
a12
a21
a31
a11 a32 a23
a23
a
+ a13 21
a33
a31
a22
a32
a12 a21 a33 + a12 a23 a31 + a13 a21 a32
a13 a22 a31 ,
which, after rearranging, is the same as the first answer.

a
a12
20. Consider the 2 ⇥ 2 matrix A = 11
. Then
a21 a22
C11 = ( 1)1+1 det[a22 ] = a22 ,
C12 = ( 1)1+2 det[a21 ] =
a21 ,
so that from definition (4)
det A =
2
X
j=1
a1j C1j = a11 C11 + a12 C12 = a11 a12 + a12 · ( a21 ) = a11 a22
a12 a21 ,
which agrees with the definition of a 2 ⇥ 2 determinant.
21. If A is lower triangular, then AT is upper triangular. Since by Theorem 4.10 det A = det AT , if we
prove the result for upper triangular matrices, we will have proven it for lower triangular matrices as
well. So assume A is upper triangular. The basis for the induction is the case n = 1, where A = [a11 ]
and det A = a11 , which is clearly the “product” of the entries on the diagonal. Now assume that the
statement holds for n ⇥ n upper triangular matrices, and let A be an (n + 1) ⇥ (n + 1) upper triangular
matrix. Using cofactor expansion along the last row, we get
det A = ( 1)(n+1)+(n+1) an+1,n+1 An+1,n+1 = an+1,n+1 An+1,n+1 .
But An+1,n+1 is an n ⇥ n upper triangular matrix with diagonal entries a1 , a2 , . . . , an , so applying the
inductive hypothesis we get
det A = an+1,n+1 An+1,n+1 = an+1,n+1 (a1 a2 · · · an ) = a1 a2 · · · an an+1 ,
proving the result for (n + 1) ⇥ (n + 1) matrices. So by induction, the statement is true for all n
22. Row-reducing the matrix in
gives
1
5
0
1.
Exercise 1, and modifying the determinant as required by Theorem 4.3,
0
1
1
3
1
2
1
! 0
0
R2 5R1
0
1
1
3
14
2
R3
1
0
R2
! 0
0
1
0
3
14
16
Note that we did not introduce any factors in front of the determinant since all of the row operations
were of the form Ri + kRj , which does not change the determinant. The final determinant is easy to
evaluate since the matrix is upper triangular; the determinant of the original matrix is 1 · 1 · 16 = 16.
254
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
23. Row-reducing the matrix in Exercise 9, and modifying the determinant as required
gives
4
1 3
1
1 0
1
1 0
1
R 2R1
3
3
2
2 4 R1 $R!
2
2 4 2 !
0
0 4 R2 $R!
0
1
1
1 0
4
1 3 R3 +4R!
0
3 3
0
by Theorem 4.3,
1 0
3 3 .
0 4
There were two row interchanges, each of which introduced a factor of 1 in the determinant. The
other row operations were of the form Ri + kRj , which does not change the determinant. The final
determinant is easy to evaluate since the matrix is upper triangular; the determinant of the original
matrix is 1 · ( 3) · 4 = 12.
24. Row-reducing the matrix in Exercise 13, and modifying the determinant as required by Theorem 4.3,
gives
1
2
0
1
1 0 3
5 2 6
1 0 0
4 2 1
1
! 0
0
R1
! 0
1 0
7 2
1 0
5 2
R2 2R1
R4
R3 7R2
!
R4 5R2
!
1
0
0
0
3
0
0
2
1 0
1 0
0 2
0 2
1
0
0
0
R2 $R3
!
3
0
0
2
R4 R3
!
1
1
7
5
1
0
0
0
0
0
2
2
3
0
0
2
1 0
1 0
0 2
0 0
3
0
.
0
2
There was one row interchange, which introduced a factor of 1 in the determinant. The other row
operations were of the form Ri + kRj , which does not change the determinant. The final determinant
is easy to evaluate since the matrix is upper triangular; the determinant of the original matrix is
(1 · 1 · 2 · ( 2)) = 4.
25. Row-reducing the matrix in Exercise 14, and modifying the determinant as required by Theorem 4.3,
gives
2
1
0
2
0 3
0 2
1 1
0 1
1
2
4
3
R1 $R2
!
1
R2 $R3 0
! 0
0
1
2
0
2
0
0
1
0
0
1
0
0
2
3
1
1
2
1
4
3
2
1
1
3
2
4
5
7
R2 2R1
!
R2 2R1
!
R4
1
0
0
3R3
! 0
1
0
0
0
0
0
1
0
0
1
0
0
2
1
1
3
2
1
1
0
2
5
4
7
2
4
.
5
8
There were two row interchanges, each of which introduced a factor of 1 in the determinant. The
other row operations were of the form Ri + kRj , which does not change the determinant. The final
determinant is easy to evaluate since the matrix is upper triangular; the determinant of the original
matrix is 1 · ( 1) · ( 1) · 8 = 8.
26. Since the third row is a multiple of the first row, the determinant is zero since we can use the row
operation R3 2R1 to zero the third row, and doing so does not change the determinant.
27. This matrix is upper triangular, so its determinant is the product of the diagonal entries, which is
3 · ( 2) · 4 = 24.
28. If we interchange the first and third rows, the result is an upper triangular matrix with diagonal
entries 3, 5, and 1. Since interchanging the rows negates the determinant, the determinant of the
original matrix is (3 · 5 · 1) = 15.
4.2. DETERMINANTS
255
29. The third column is 2 times the first column, so the columns are not linearly independent and thus
the matrix does not have rank 3, so is not invertible. By Theorem 4.6, its determinant is zero.
30. Since the third row is the sum of the first and second rows, the rows are not linearly independent and
thus the matrix does not have rank 3, so is not invertible. By Theorem 4.6, its determinant is zero.
31. Since the first column is the sum of the second and third columns, the columns are not linearly
independent and thus the matrix does not have rank 3, so is not invertible. By Theorem 4.6, its
determinant is zero.
32. Interchanging the second and third rows gives the identity matrix, which has determinant 1. Since
interchanging the rows introduces a minus sign, the determinant of the original matrix is 1.
33. Interchange the first and second rows, and also the third and fourth rows, to give a diagonal matrix with
diagonal entries 3, 2, 1, and 4. Performing two row interchanges does not change the determinant,
so the determinant of the original matrix is 3 · 2 · 1 · 4 = 24.
34. The sum of the first and second rows is equal to the sum of the third and fourth rows, so the rows are
not linearly independent and thus the matrix does not have rank 4, so is not invertible. By Theorem
4.6, its determinant is zero.
35. This matrix results from the original matrix by multiplying the first row by 2, so its determinant is
2 · 4 = 8.
36. This matrix results from the original matrix by multiplying the first column by 2, the second column
by 31 , and the third column by 1. Thus the determinant of this matrix is 2 · 13 · ( 1) = 23 times the
original determinant, so its value is 23 · 4 = 83 .
37. This matrix results from the original matrix by interchanging the first and second rows. This multiplies
the determinant by 1, so the determinant of this matrix is 4.
38. This matrix results from the original matrix by subtracting the third column from the first. This
operation does not change the determinant, so the determinant of this matrix is also 4.
39. This matrix results from the original matrix by exchanging the first and third columns (which multiplies
the determinant by 1) and then multiplying the first column by 2 (which multiplies the determinant
by 2). So the determinant of this matrix is 2 · ( 1) · 4 = 8.
40. To get from the original matrix to this one, row 2 was multiplied by 3, and then twice row 3 was
added to each of rows 1 and 2. The first of these operations multiplies the determinant by 3, while the
additions do not change the determinant. Thus the determinant of this matrix is 3 · 4 = 12.
41. Suppose first that A has a zero row, say the ith row is zero. Using cofactor expansion along that row
gives
n
n
X
X
det A =
aij Cij =
0Cij = 0,
j=1
j=1
so the determinant is zero. Similarly, if A has a zero column, say the j th column, then using cofactor
expansion along that column gives
det A =
n
X
i=1
aij Cij =
n
X
0Cij = 0,
i=1
and again the determinant is zero. Alternatively, the column case can be proved from the row case by
noting that if A has a zero column, then AT has a zero row, and det A = det AT by Theorem 4.10.
256
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
42. Dealing first with rows, we want to show that if A
be the ith row of A. Then
2
3
A1
6 A2
7
6
7
6
7
..
6
7
.
7
B=6
6Ai + kAj 7 ;
6
7
6
7
..
4
5
.
An
Ri +kRj
! B where i 6= j then det B = det A. Let Ai
2
3
A1
6 A2 7
6
7
6 .. 7
6 . 7
7
let C = 6
6kAj 7 .
6
7
6 . 7
.
4 . 5
An
Then A, B, and C are identical except that the ith row of B is the sum of the ith rows of A and C,
so that by part (a), det B = det A + det C. However, since i 6= j, the ith row of C, which is kAj , is a
multiple of the j th row, which is Aj . By Theorem 4.3(d), multiplying the j th row by k multiplies the
determinant of C by k. However, the result is a matrix with two identical rows; both the ith and j th
rows are kAj . So det C = 0 by Theorem 4.3(c). Hence det B = det A as desired.
If instead A
Ci +kCj
! B where i 6= j, then AT
Ri +kRj
! B T , so that by the first part and Theorem 4.10
det A = det AT = det B T = det B.
43. If E interchanges two rows, then det E = 1 by Theorem 4.4. Also, EB is the same as B but with two
rows interchanged, so by Theorem 4.3(b), det(EB) = det B = det(E) det(B). Next, if E results from
multiplying a row by a constant k, then det E = k by Theorem 4.4. But EB is the same as B except that
one row has been multiplied by k, so that by Theorem 4.3(d), det(EB) = k det(B) = det(E) det(B).
Finally, if E results from adding a multiple of one row to another row, then det E = 1 by Theorem 4.4.
But in this case EB is obtained from B by adding a multiple of one row to another, so by Theorem
4.3(f), det(EB) = det(B) = det(E) det(B).
44. Let the row vectors of A be A1 , A2 , . . . , An . Then the row vectors of kA are kA1 , kA2 , . . . , kAn .
Using Theorem 4.3(d) n times,
kA1
A1
A1
A1
kA2
kA2
A2
A2
det(kA) = . = k . = k 2 . = · · · = k n . = k n det A.
..
..
..
..
kAn
kAn
kAn
An
45. By Theorem 4.6, A is invertible if and only if det A 6= 0. To find det A, expand along the first column:
det A = k
k+1
8
1
k
1
k
3
k+1 1
+k
= k((k + 1)(k
1)
3(k + 1)) = k 3
1 · ( 8)) + k( k · 1
4k 2 + 4k = k(k
2)2 .
Thus det A = 0 when k = 0 or k = 2, so A is invertible for all k except k = 0 and k = 2.
46. By Theorem 4.6, A is invertible if and only if det A 6= 0. To find det A, expand along the first row:
k
k2
0
k
2
k
0
2
k =k
k
k
= k(2k
k
k
k
k2 )
k2
0
k(k 3
k
k
0) = k(2k
k2
k 3 ) = k 2 (2
k
k2 ) =
k 2 (k + 2)(k
1).
Since the matrix is invertible precisely when its determinant is nonzero, and since the determinant is
zero when k = 0, k = 1, or k = 2, we see that the matrix is invertible for all values of k except for 0,
1, and 2.
47. By Theorem 4.8, det(AB) = det(A) det(B) = 3 · ( 2) =
6.
4.2. DETERMINANTS
257
48. By Theorem 4.8, det(A2 ) = det(A) det(A) = 3 · 3 = 9.
49. By Theorem 4.9, det(B
1
)=
1
det(B) ,
det(B
1
so that by Theorem 4.8
A) = det(B
1
det(A)
=
det(B)
) det(A) =
3
.
2
50. By Theorem 4.7, det(2A) = 2n det(A) = 3 · 2n .
51. By Theorem 4.10, det(B T ) = det(B), so that by Theorem 4.7, det(3B T ) = 3n det(B T ) =
2 · 3n .
52. By Theorem 4.10, det(AT ) = det(A), so that by Theorem 4.8,
det(AAT ) = det(A) det(AT ) = det(A)2 = 9.
53. By Theorem 4.8, det(AB) = det(A) det(B) = det(B) det(A) = det(BA).
54. If B is invertible, then det(B) 6= 0 and det(B
det(B
1
AB) = det(B
1
(AB)) = det(B
1
1
)=
1
det(B) ,
so that using Theorem 4.8
) det(AB)
= det(B
1
) det(A) det(B) =
1
det(A) det(B) = det(A).
det(B)
55. If A2 = A, then det(A2 ) = det(A). But det(A2 ) = det(A) det(A) = det(A)2 , so that det(A) = det(A)2 .
Thus det(A) is either 0 or 1.
56. We first show that Theorem 4.8 extends to prove that if m 1, then det(Am ) = det(A)m . The case
m = 1 is obvious, and m = 2 is Theorem 4.8. Now assume the statement is true for m = k. Then
using Theorem 4.8 and the truth of the statement for m = k,
det(Ak+1 ) = det(AAk ) = det(A) det(Ak ) = det(A) det(A)k = det(A)k+1 .
So the statement holds for all m
that det(A) = 0.
1. Now, if Am = O, then det(Am ) = det(A)m = det(O) = 0, so
57. The coefficient matrix and constant vector are

1
1
A=
,
1
1
b=

1
.
2
To apply Cramer’s Rule, we compute
det A =
1
1
1
=
1
2
det A1 (b) =
1
2
1
=
1
3
det A2 (b) =
1
1
1
= 1.
2
Then by Cramer’s Rule,
x=
det A1 (b)
3
= ,
det A
2
58. The coefficient matrix and constant vector are

2
1
A=
,
1
3
y=
det A2 (b)
=
det A
b=

5
.
1
1
.
2
258
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
To apply Cramer’s Rule, we compute
det A =
2
1
5
1
det A1 (b) =
det A2 (b) =
1
=7
3
1
= 14
3
2
1
5
=
1
7.
Then by Cramer’s Rule,
det A1 (b)
det A2 (b)
= 2, y =
=
det A
det A
59. The coefficient matrix and constant vector are
2
3
2 3
2 1 3
1
A = 40 1 15 ,
b = 415 .
0 0 1
1
x=
1.
To apply Cramer’s Rule, we compute
2
det A = 0
0
1
1
0
3
1 =2·1·1=2
1
1
det A1 (b) = 1
1
1
1
0
3
1
1 =1
1
1
3
1
+1
1
1
2
det A2 (b) = 0
0
1
1
1
3
1
1 =2
1
1
1
=0
1
2
det A3 (b) = 0
0
1
1
0
1
1
1 =2
0
1
1
= 2.
1
1
=
1
2
Then by Cramer’s Rule,
x=
det A1 (b)
=
det A
1,
y=
60. The coefficient matrix and constant vector
2
1
A = 41
1
det A2 (b)
= 0,
det A
are
1
1
1
To apply Cramer’s Rule, we compute
3
1
15 ,
0
z=
det A3 (b)
= 1.
det A
2 3
1
b = 425 .
3
1
det A = 1
1
1
1
1
1
1
1 =1
1
0
1
1
( 1)
1
1
1
=4
1
1
det A1 (b) = 2
3
1
1
1
1
1
1 =3
1
0
1
1
( 1)
1
2
1
=9
1
1
det A2 (b) = 1
1
1
det A3 (b) = 1
1
1
2
3
1
1
1 =1
2
0
1 1
1 2 =1
1 3
1
1
1
1
3
1
1
1
=
1
2
3
1
1
1
2
1
+1
3
1
3
1
= 2.
1
4.2. DETERMINANTS
259
Then by Cramer’s Rule,
x=
61. We have det A =
1
1
det A1 (b)
9
= ,
det A
4
1
=
1
y=
det A2 (b)
=
det A
3
,
4
det A3 (b)
1
= .
det A
2
z=
2. The four cofactors are
C11 = ( 1)1+1 · ( 1) =
C21 = ( 1)
2+1
·1=
C12 = ( 1)1+2 · 1 =
1,
1,
C22 = ( 1)
2+2
1,
· 1 = 1.
Then
adj A =
"
1
1
62. We have det A =
2
1
#T "
1
1
=
1
1
#
1
, so A
1
1
1
=
adj A =
det A
1
2
"
# "
1
1
= 21
1
2
1
1
1
2
1
2
1
= 7. The four cofactors are
3
C11 = ( 1)1+1 · 3 = 3,
C21 = ( 1)
2+1
· ( 1) = 1,
C12 = ( 1)1+2 · 1 =
C22 = ( 1)
2+2
1,
· 2 = 2.
Then
"
#T "
1
3
=
2
1
3
adj A =
1
#
1
, so A
2
1
1
1
=
adj A =
det A
7
"
# "
1
=
2
3
1
3
7
1
7
1
7
2
7
#
63. We have det A = 2 from Exercise 59. The nine cofactors are
C11 = ( 1)1+1 · 1 = 1,
C21 = ( 1)2+1 · 1 =
C22 = ( 1)2+2 · 2 = 2, C23 = ( 1)2+3 · 0 = 0,
1,
C31 = ( 1)3+1 · ( 2) =
Then
C12 = ( 1)1+2 · 0 = 0, C13 = ( 1)1+3 · 0 = 0,
C32 = ( 1)3+2 · 2 =
2,
2, C33 = ( 1)3+3 · 2 = 2.
2
3T 2
3
0 0
1
1
2
2 05 = 40
2
25 ,
2 2
0
0
2
2
3 2
1
1
1
1
2
2
2
1
16
7 6
=
adj A = 40
2
25 = 4 0
1
det A
2
0
0
2
0
0
1
adj A = 4 1
2
so
A
1
3
1
7
15 .
1
64. We have det A = 4 from Exercise 60. The nine cofactors are
C11 = ( 1)1+1 · 1 = 1,
C21 = ( 1)
2+1
C31 = ( 1)
Then
C12 = ( 1)1+2 · ( 1) = 1, C13 = ( 1)1+3 · ( 2) =
· ( 1) = 1,
3+1
· 2 = 2,
2
C22 = ( 1)
C32 = ( 1)
A
1
3+2
· 1 = 1, C23 = ( 1)
·2=
2+3
2, C33 = ( 1)
3T 2
2
1 1
25 = 4 1 1
0
2 2
2
3 2
1 1
2
1
16
7 6
=
adj A = 4 1 1
25 = 4
det A
4
2 2
0
1
adj A = 41
2
so
2+2
1
1
2
1
4
1
4
1
2
3
2
25 ,
0
1
4
1
4
1
2
· ( 2) = 2,
3+3
3
· 0 = 0.
1
2
17
.
25
0
2,
.
#
.
260
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
65. First, since A
1
=
1
det A
adj A, we have adj A = (det A)A
(adj A) ·
Thus adj A times
1
det A A
1
A = (det A)A
det A
1
·
1
. Then
1
1
A = det A ·
A
det A
det A
1
A = I.
is the identity, so that adj A is invertible and
(adj A)
1
=
1
A.
det A
This proves the first equality. Finally, substitute A 1 for A in the equation adj A = (det A)A
gives adj(A 1 ) = (det A 1 )A = det1 A A, proving the second equality.
66. By Exercise 65, det(adj A) = det((det A)A 1 ); by Theorem 4.7, this is equal to (det A)n det(A
det(A 1 ) = det1 A , so that det(adj A) = (det A)n det1 A = (det A)n 1 .
1
1
; this
). But
67. Note that exchanging Ri 1 and Ri results in a matrix in which row i has been moved “up” one row
and all other rows remain the same. Refer to the diagram at the end of the solution when reading
what follows. Starting with Rs $ Rs 1 , perform s r adjacent interchanges, each one moving the row
that started as row s up one row. At the end of these interchanges, row s will have been moved “up”
to row r, and all other rows will remain in the same order. The original row r is therefore at row r + 1.
We wish to move that row down to row s. Note that exchanging Ri and Ri+1 results in a matrix in
which row i has been moved “down” one row and all other rows remain the same. So starting with
Rr+1 $ Rr+2 , perform s r 1 adjacent interchanges, each one moving the row that started as row
r + 1 “down” one row. At the end of these interchanges, row r + 1 will have been moved down s r 1
rows, so it will be at row s r 1 + r + 1 = s, and all other rows will be as they were. Now, row r + 1
originally started life as row r, so the end result is that row s has been moved to row r, and row r to
row s. We used s r adjacent interchanges to move row s, and another s r 1 adjacent interchanges
to move row r, for a total of 2(s r) 1 adjacent interchanges.
2
3
R1
6 R2 7
6
7
6 .. 7
6 . 7
6
7
6 Rr 1 7
6
7
6 Rr 7
6
7
6Rr+1 7
6
7
Original: 6 . 7
6 .. 7
6
7
6 Rs 1 7
6
7
6 Rs 7
6
7
6Rs+1 7
6
7
6 . 7
4 .. 5
Rn
2
3
2
3
R1
R1
6 R2 7
6 R2 7
6
7
6
7
6 .. 7
6 .. 7
6 . 7
6 . 7
6
7
6
7
6Rr 1 7
6 Rr 1 7
6
7
6
7
6 Rs 7
6 Rs 7
6
7
6
7
6 Rr 7
6Rr+1 7
6
7
6
7
After initial s r exchanges: 6Rr+1 7 After next s r 1 exchanges: 6 . 7 .
6
6 .. 7
7
6 .. 7
6
7
6 . 7
6 Rs 1 7
6
7
6
7
6 Rs 1 7
6 Rr 7
6
7
6
7
6Rs+1 7
6Rs+1 7
6
7
6
7
6 . 7
6 . 7
4 .. 5
4 .. 5
Rn
Rn
4.2. DETERMINANTS
261
68. The proof is very similar to the proof given
be the matrix obtained by moving column j
2
a11 a12 · · ·
6 a21 a22 · · ·
6
A=6 .
..
..
4 ..
.
.
an1
an2
2
a1j
6 a2j
6
B=6 .
4 ..
anj
in the text for the case of expansion along a row. Let B
of matrix A to column 1:
3
a1,j 1 a1j a1,j+1 · · · a1n
a2,j 1 a2j a2,j+1 · · · a2n 7
7
..
..
..
.. 7
..
.
.
.
.
. 5
· · · an,j 1 anj an,j+1 · · · ann
a11
a21
..
.
a12
a22
..
.
···
···
..
.
a1,j
a2,j
..
.
an1
an2
···
an,j
1
a1,j+1
a2,j+1
..
.
···
···
..
.
1
an,j+1
···
1
3
a1n
a2n 7
7
.. 7
. 5
ann
By the discussion in Exercise 67, it requires j 1 adjacent column interchanges to move column j to
column 1, so by Lemma 4.14, det B = ( 1)j 1 det A, so that det A = ( 1)j 1 det B. Then by Lemma
4.13, evaluating det B using cofactor expansion along the first column gives
det A = ( 1)j
1
det B = ( 1)j
1
n
X
bi1 Bi1 = ( 1)j
1
i=1
n
X
aij Bi1 ,
i=1
where Aij and Bij are the cofactors of A and B. It remains only to understand how the cofactors of A
and B are related. Clearly the matrix formed by removing row i and column j of A is the same matrix
as is formed by removing row i and column 1 of B, so the only di↵erence is the sign that we put on
the determinant of that matrix. In A, the sign is ( 1)i+j ; in B, it is ( 1)i+1 . Thus
Bi1 = ( 1)j
1
Aij .
Substituting that in the equation for det A above gives
det A = ( 1)j
1
n
X
aij Bi1 = ( 1)j
1
i=1
n
X
( 1)j
1
aij Aij = ( 1)2j
i=1
2
n
X
i=1
aij Aij =
n
X
aij Aij ,
i=1
proving Laplace expansion along columns.
69. Since P and S are both square, suppose P is n ⇥ n and S is m ⇥ m. Then O is m ⇥ n and Q is n ⇥ m.
Proceed by induction on n. If n = 1, then

p
Q
A = 11
,
O S
and Q is 1 ⇥ n while O is n ⇥ 1. Compute the determinant by cofactor expansion along the first column.
The only nonzero entry in the first column is p11 , so that det A = p11 A11 . Since O has one column
and Q one row, removing the first row and column of A leaves only S, so that A11 = det S and thus
det A = p11 det S = (det P ) det S). This proves the case n = 1 and provides the basis for the induction.
Now suppose that the statement is true when P is k ⇥ k; we want to prove it when P is (k + 1) ⇥ (k + 1).
So let P be a (k + 1) ⇥ (k + 1) matrix. Again compute det A using column 1:
det A =
k+1+m
X
ai1 ( 1)i+1 det Ai1 .
i=1
Now, ai1 = pi1 for i  k + 1, and ai1 = 0 for i > k + 1, so that
det A =
k+1
X
i=1
pi1 ( 1)i+1 det Ai1 .
262
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
Ai1 is the submatrix formed by removing row i and column 1 of A; since i ranges from 1 to k + 1, we
see that this process removes row i and column 1 of P , but does not change S. So we are left with a
matrix

P
Qi
Ai1 = i1
.
O
S
Now Pi1 is k ⇥ k, so we can apply the inductive hypothesis to get
det Ai1 = (det Pi1 )(det S).
Substitute that into the equation for det A above to get
det A =
k+1
X
pi1 ( 1)i+1 det Ai1
i=1
=
k+1
X
pi1 ( 1)i+1 (det Pi1 )(det S)
i=1
k+1
X
= (det S)
pi1 ( 1)
i+1
det Pi1
i=1
!
.
But the sum is just cofactor expansion of P along the first column, so its value is det P , and finally we
get det A = (det S)(det P ) = (det P )(det S) as desired.
70. (a) There are many examples. For instance, let


1 0
1
P =S=
, Q=
0 1
1
Then

A=
0
,
0
2
1
60
Q
=6
40
S
0
P
R
0
1
1
1
1
1
1
0
R=

0
0
1
.
1
3
0
07
7.
05
1
Since the second and third rows of A are identical, det A = 0, but
(det P )(det S)
(det Q)(det R) = 1 · 1
0 · 0 = 1,
and the two are unequal.
(b) We have
BA =
=
=
=




P 1
RP
1
O
I

P
R
Q
S
P 1 P + OR
RP 1 P + IR
I
R+R
I
O
P
RP
P 1 Q + OS
RP 1 Q + IS
1
1
Q
Q+S
P 1Q
.
S RP 1 Q
It is clear that the proof in Exercise 69 can easily be modified to prove the same result for a block
lower triangular matrix. That is, if A, P , and S are square and

P O
A=
,
Q S
Exploration: Geometric Applications of Determinants
263
then det A = (det P )(det S). Applying this to the matrix B gives
1
det B = (det P
)(det I) = det P
1
=
1
.
det P
Further, applying Exercise 69 to BA gives
det(BA) = (det I)(det(S
So finally
det A =
1
RP
Q)) = det( SRP
1
det(BA) = (det P )(det(S
det B
RP
1
1
Q).
Q)).
(c) If P R = RP , then
det A = (det P )(det(S
1
RP
Q)) = det(P (S
= det(P S
RP
1
(P R)P
1
Q))
Q)) = det(P S
RP P
1
Q) = det(P S
Exploration: Geometric Applications of Determinants
1. (a) We have
i
u⇥v = 0
3
j k
1 1 =i
1 2
1
1
1
2
0
3
j
1
0
+k
2
3
1
= 3i + 3j
1
3k.
(b) We have
i
u⇥v = 3
0
j k
1 2 =i
1 1
1
1
2
1
3
0
j
2
3
+k
1
0
1
=
1
3i
3j + 3k.
1
2
2
= 0.
4
(c) We have
u⇥v =
i
1
2
j
2
4
k
3 =i
6
2
4
3
6
j
1
2
k
1
1 =i
2
3
1
3
j
j
1
2
3
+k
6
(d) We have
i
u⇥v = 1
1
1
1
1
1
+k
3
1
1
=i
2
2j + k.
2. First,
i
u · (v ⇥ w) = u · v1
w1
j
v2
w2
k
v3
w3
= (u1 i + u2 j + u3 k) · (i(v2 w3
= (u1 v2 w3
u1 v 3 w 2 )
v3 w 2 )
(u2 v1 w3
j(v1 w3
v3 w1 ) + k(v1 w2
u2 v3 w1 ) + (u3 v1 w2
v2 w1 ))
u3 v2 w1 ).
Computing the determinant, we get
u1
v1
w1
u2
v2
w2
u3
v
v 3 = u1 2
w2
w3
v3
w3
u2
v1
w1
= u1 (v2 w3
v3 w 2 )
= (u1 v2 w3
u1 v 3 w 2 )
and the two expressions are equal.
v3
v
+ u3 1
w3
w1
u2 (v1 w3
(u2 v1 w3
v2
w2
v3 w1 ) + u3 (v1 w2
v2 w1 )
u2 v3 w1 ) + (u3 v1 w2
u3 v2 w1 ),
RQ).
264
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
3. With u, v, and w as in the previous exercise,
(a) We have
i
v ⇥ u = v1
u1
j
v2
u2
k
v3
u3
i
u1
v1
R2 $R3
!=
j
u2
v2
k
u3 =
v3
(u ⇥ v).
(b) We have
i
u ⇥ 0 = u1
0
j
u2
0
k
u3 .
0
This matrix has a zero row, so its determinant is zero by Theorem 4.3. Thus u ⇥ 0 = 0.
(c) We have
i
j
k
u ⇥ 0 = u1 u2 u3 .
u1 u2 u3
This matrix has two equal rows, so its determinant is zero by Theorem 4.3. Thus u ⇥ u = 0.
(d) We have by Theorem 4.3
i
u ⇥ kv = u1
kv1
j
u2
kv2
k
i
u3 = k u1
kv3
v1
j
u2
v2
k
u3 = k(u ⇥ v).
v3
(e) We have by Theorem 4.3
u ⇥ (v + w) =
i
u1
v1 + w 1
j
u2
v2 + w2
k
i
u3
= u1
v3 + w3
v1
j
u2
v2
k
i
u3 + u1
v3
w1
u1
u · (u ⇥ v) = u1
v1
u2
u2
v2
u3
u3
v3
v1
v · (u ⇥ v) = u1
v1
v2
u2
v2
v3
u3 .
v3
j
u2
w2
k
u3 = u ⇥ v + u ⇥ w.
w3
(f ) From Exercise 2 above,
But each of these matrices has two equal rows, so each has zero determinant. Thus
u · (u ⇥ v) = v · (u ⇥ v) = 0.
(g) From Exercise 2 above,
u1
u · (v ⇥ w) = v1
w1
u2
v2
w2
u3
v3 ,
w3
w1
(u ⇥ v) · w = w · (u ⇥ v) = u1
v1
w2
u2
v2
w3
u3 .
v3
The second matrix is obtained from the first by exchanging rows 1 and 2 and then exchanging
rows 1 and 3. Since each row exchange multiplies the determinant by 1, it remains unchanged
so that the two products are equal.
4. The vectors u and v, considered as vectors in R3 , have their third coordinates equal to zero, so the
area of the parallelogram with these edges is
2
3



i
j k
u1 u2
u1 u2
u v1
4
5
A = ku ⇥ vk = det u1 u2 0
= k det
= det
= det 1
.
v1 v2
v1 v2
u2 v 2
v1 v2 0
Exploration: Geometric Applications of Determinants
265
5. The area of the rectangle is (a + c)(b + d). The small rectangles at top left and bottom right each
have area bc. The triangles at top and bottom each have base a and height b, so the area of each is
1
2 ab; altogether they have area ab. Finally, the smaller triangles at left and right each have base d
and height c, so the area of each is 12 cd; altogether they have area cd. So in total, the area in the big
rectangle that is not in the parallelogram is bc + ab + cd, so the area of the parallelogram is
(a + c)(b + d)
2bc ab cd = ab + ad + bc + cd


a
c
For this parallelogram, u =
and v =
, and indeed
b
d
⇥
⇤
det u v = ad bc.
2bc
ab
cd = ad
bc.
As for the absolute value sign, note that the formula A = ku ⇥ vk does not depend on which of the
two sides we choose for u and which for v, but the sign of the determinant does. In our particular
case, when we choose u and v as above, we don’t need the absolute value sign, as the area works out
properly. It is present in the norm formula to represent the fact that we need not make particular
choices for u and v.
6. (a) By Exercise 5, the area is
A = det

2
3
1
4
= |2 · 4
( 1) · 3| = 11.
(b) By Exercise 5, the area is
A = det

3
4
5
5
= |3 · 5
4 · 5| = 5.
7. A parallelepiped of height h with base area a has volume ha. This can be seen using calculus; the
Rh
h
volume is 0 a dz = [az]0 = ah. Alternatively, it can be seen intuitively, since the parallelepiped
consists of “slices” each of area a, and the height of those slices is h. Now, for the parallelepiped in
Figure 4.11, the height is clearly kuk cos ✓. The base is a parallelogram with sides v and w, so its area
is kv ⇥ wk. Thus the volume of the parallelepiped is kuk kv ⇥ wk cos ✓. But
u · (v ⇥ w) = kuk kv ⇥ wk cos ✓,
so that the volume
2 3is u · (v ⇥ w) which, by Exercise 2, is equal to the absolute value of the determinant
u
of the matrix 4 v 5, which is equal to the absolute value of the determinant of the transpose of that
w
⇥
⇤
matrix, or u v w .
8. Compare Figure 4.12 to Figure 4.11. The base of the tetrahedron is a triangle that is one half of the
area of the parallelogram in Figure 4.11. So from the hint, we have
1
(area of the base)(height)
3
1
= (area of the parallelogram)(height)
6
1
= (volume of the parallelepiped)
6
1
= |u · (v ⇥ w)| .
6
V=
9. By problem 4, the area of the parallelogram determined by Au and Av is
⇥
⇤
⇥
⇤
⇥
⇤
TA (P ) = det Au Av = det(A u v ) = (det A)(det u v ) = |det A| (area of P ).
266
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
10. By Problem 7, the volume of the parallelepiped determined by Au, Av, and Aw is
⇥
TA (P ) = det Au
Av
Aw
⇤
⇥
= det(A u
⇤
w )
v
⇥
= (det A)(det u
v
11. (a) The line through (2, 3) and ( 1, 0) is given by
x y
2 3
1 0
⇤
w ) = |det A| (volume of P ).
1
1 .
1
Evaluate by cofactor expansion along the third row:
1
y
3
1
x
+1
1
2
The equation of the line is 3x
y
=
3
1(y
3) + (3x
3y + 3 = 0, or x
2y) = 3x
3y + 3.
y + 1 = 0.
(b) The line through (1, 2) and (4, 3) is given by
x
1
4
y
2
3
1
1 .
1
Evaluate by cofactor expansion along the first row:
x
The equation of the line is
2
3
1
1
y
x + 3y
1
4
1
1
+1
1
4
5 = 0, or x
2
=
3
x + 3y
5.
3y + 5 = 0.
12. Suppose the three points satisfy
ax1 + by1 + c = 0
ax2 + by2 + c = 0
ax3 + by3 + c = 0,
which can be written as the matrix equation
2
x1
XA = 4x2
x3
y1
y2
y3
32 3
1 a
15 4 b 5 = O.
1
c
Then det X 6= 0 if and only if X is invertible, which implies that X 1 XA = X 1 O = O, so that
A = O. Thus A 6= O if and only if det X = O; this says precisely that the three points lie on the line
ax + by + c = 0 with a and b not both zero if and only if det X = 0.
13. Following the discussion at the beginning of the section, since the three points are noncollinear, they
determine a unique plane, say ax + by + cz + d = 0. Since all three points lie on the plane, their
coordinates satisfy this equation, so that
ax1 + by1 + cz1 + d = 0
ax2 + by2 + cz2 + d = 0
ax3 + by3 + cz3 + d = 0.
These three equations together with the equation of the plane then form a system of linear equations
in the variables a, b, c, and d. Since we know that the plane exists, it follows that there are values
Exploration: Geometric Applications of Determinants
267
of a, b, c, and d, not all zero, satisfying these equations.
therefore its coefficient matrix, which is
2
x y z
6 x 1 y 1 z1
6
4 x 2 y 2 z2
x 3 y 3 z3
So the system has a nontrivial solution, and
3
1
17
7
15
1
cannot be invertible, by the Fundamental Theorem of Invertible Matrices. Thus its determinant must
be zero by Theorem 4.6.
To see what happens if the three points are collinear, try row reducing the above coefficient matrix:
2
x
6x1
6
4x2
x3
y
y1
y2
y3
z
z1
z2
z3
3
1
17
7
15
1
If the points are collinear, then
6
6
! 4x2
R2
! x3
R3 R2
R4
2
2
x
x1
y
y1
x1
x1
y2
y3
3
2
x1
x2
y 1 5 = k 4 y2
z1
z2
x3
4 y3
z3
z
z1
y1
y1
z2
z3
z1
z1
3
1
17
7.
05
0
3
x1
y1 5
z1
for some k 6= 0. Then performing R4 kR3 gives a matrix with a zero row. So its determinant is
trivially zero and the system has an infinite number of solutions. This corresponds to the fact that
there are an infinite number of planes passing through the line determined by the three collinear points.
14. Suppose the four points satisfy
ax1 + by1 + cz1 + d = 0
ax2 + by2 + cz2 + d = 0
ax3 + by3 + cz3 + d = 0
ax4 + by4 + cz4 + d = 0,
which can be written as the matrix equation
2
x1
6x2
XA = 6
4x3
x4
y1
y2
y3
y4
z1
z2
z3
z4
32 3
1 a
6 7
17
7 6 b 7 = O.
5
1 4c5
1
d
Then det X 6= 0 if and only if X is invertible, which implies that X 1 XA = X 1 O = O, so that
A = O. Thus A 6= O if and only if det X = O; this says precisely that the four points lie on the plane
ax + by + cz + d = 0 with a, b, and c not all zero if and only if det X = 0.
15. Substituting the values (x, y) = ( 1, 10), (0, 5) and (3, 2) in turn into the equation gives the linear
system in a, b, and c
a
b + c = 10
a
= 5
a + 3b + 9c = 2.
The determinant of the coefficient matrix of this system is
1
X= 1
1
1 1
0 0 =
3 9
1
1
3
1
= 12 6= 0,
9
268
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
Since X is has nonzero determinant, it is invertible by Theorem 4.6, so we have
2 3 2 3
2 3
2 3
a
10
a
10
14 5
4
5
4
5
4
5
b
5
b
5 ,
X
=
)
=X
c
2
c
2
so that the system has a unique solution. To find that solution, we can either invert X or we can
row-reduce the augmented matrix. Using the second method, and writing the second equation first in
the augmented matrix, we have
2
1
41
1
0 0
1 1
3 9
3
5
105
2
R2
R3
2
1
R1
! 40
R1
! 0
0 0
1 1
3 9
2
1
40
0
0
1
0
0
1
12
3
5
55
3
3
1 0
0
5
R2 4
1
55
! 0 1
0 3
9
3 R3
3
2
3
5
1 0
0
5
40 1
55 1
1
55
3
12 12 R!
0 0
1
1
So the system has the unique solution a = 5, b =
is y = 5 4x + x2 .
2
3R2
!
2
1
4
! 0
0
R2 +R3
0
1
0
0
0
1
3
5
45 .
1
4, c = 1 and the parabola through the three points
16. (a) The linear system in a, b, and c is
a+ b+ c=
1
a + 2b + 4c =
4
a + 3b + 9c =
3,
so row-reducing the augmented matrix gives
2
1
41
1
1
2
3
1
4
9
3
1
45
3
R2
R3
2
1 1 1
R1
! 40 1 3
R1
! 0 2 8
2
1 1 1
40 1 3
0 0 1
3
2
3
1
1 1 1
1
40 1 3
55
55 1
3
2
4 R3 2R!
0 0 2
6 2 R!
3 R1 R3 2
3
2
2
1
! 1 1 0
2 R1 R!
1
40
3 40
55 R2 3R!
1 0 145
3
0 0 1
3
0
So the system has the unique solution a =
three points is y = 12 + 14x 3x2 .
(b) The linear system in a, b, and c is
a
12, b = 14, c =
b+ c=
3
a+ b+ c=
1
a + 3b + 9c =
1,
0
1
0
0
0
1
3
12
145 .
3
3 and the parabola through the
so row-reducing the augmented matrix gives
2
1
41
1
1 1
1 1
3 9
3
3
15
1
R2
R3
2
2
1
R1
! 40
R1
! 0
1
40
0
1 1
1 0
0 2
1 1
2 0
4 8
3
3
25
4
2
1
1 1
! 40
1 0
1
0
1 2
4 R3
!
3
2
3
3
1
1 1
3
40
15 1
1 0
15
3
0 2 R!
0
0 1
0
1
2 R2
3
3
15
1
R1 +R2
R3 R2
!
2
R3
! 1
40
0
0
1
0
0
0
1
3
2
15 .
0
So the system has the unique solution a = 2, b = 1, c = 0. Thus the equation of the curve
passing through these points is in fact a line, since c = 0, and its equation is y = 2 + x.
Exploration: Geometric Applications of Determinants
269
17. Computing the determinant, we get
1
1
1
a1
a2
a3
a21
a
a22 = 1 2
a3
a23
a22
a23
= a2 a23
a21
a
+1 1
a23
a2
a1
a3
1
a21
a22
a22 a3
a1 a23 + a21 a3 + a1 a22
a1 a3
a1 a2 + a21 )(a3
a21 a2 .
Now compute the product on the right:
(a2
a1 )(a3
a1 )(a3
a2 ) = (a2 a3
=
=
a2 a23
a2 a23
a22 a3
a22 a3
a1 a23
a1 a23
+ a1 a2 a3
a2 )
a1 a2 a3 + a1 a22 + a21 a3
+ a1 a22 + a21 a3
a21 a2
a21 a2 .
The two expressions are equal. Furthermore, the determinant is nonzero since it is the product of the
three di↵erences a2 a1 , a3 a1 , and a3 a2 , which are all nonzero because we are assuming that the
ai are distinct real numbers.
18. Recall that adding a multiple of one column to another does not change the determinant (see Theorem
4.3). So given the matrix
1
1
1
1
apply in order the column operations C4
1
1
1
1
a1
a2
a3
a4
a1
a1
a1
a1
a21 a21
a22 a2 a1
a23 a3 a1
a24 a4 a1
a1
a2
a3
a4
a21
a22
a23
a24
a1 C 3 , C 3
a31 a31
a32 a22 a1
= (a4
a33 a23 a1
a34 a24 a1
a31
a32
,
a33
a34
a1 C 2 , C 2
a1 )(a3
a1 C1 to get
a1 )(a2
1
1
a1 )
1
1
0
1
1
1
0
a2
a3
a4
0
a22
.
a23
a24
Then expand along the first row, giving
(a4
a1 )(a3
a1 )(a2
1
a1 ) 1
1
a2
a3
a4
a22
a23 .
a24
But we know how to evaluate this determinant from Exercise 17; doing so, we get for the original
determinant
(a4 a1 )(a3 a1 )(a2 a1 )(a3 a2 )(a4 a2 )(a4 a3 )
which is the same as the expression given in the problem statement.
For the second part, we can interpret the matrix above as the coefficient matrix formed by starting
with the cubic y = a + bx + cx2 + dx3 , substituting the four given points for x, and regarding the
resulting equations as a linear system in a, b, c, and d, just as in Exercise 15. Since the determinant
is nonzero if the ai are all distinct, it follows that the system has a unique solution, so that there is a
unique polynomial of degree at most 3 (it may be less than a cubic; for example, see Exercise 16(b))
passing through the four points.
19. For the n ⇥ n determinant given, use induction on n (note that Exercise 18 was essentially an inductive
computation, since it invoked Exercise 17 at the end). We have already proven the cases n = 1, n = 2,
270
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
and n = 3. So assume that the statement holds for n = k, and consider the determinant
1
1
..
.
1
1
apply in order the column operations Ck+1
1
1
..
.
1
1
a1
a2
a1
a1
..
.
ak a1
ak+1 a1
···
···
..
.
···
···
···
···
..
.
a1
a2
..
.
ak
ak+1
ak1 1 ak1 1
ak2 1 a1 ak2 2
..
.
akk 1 a1 akk 2
akk+11 a1 akk+12
···
···
ak1 1
a2k 1
..
.
k 1
ak
akk+11
a1 Ck , C k
ak1
ak2
akk
akk+1
ak1
ak2
..
. .
akk
akk+1
a1 C k
1 , . . . , C2
ak1
a1 ak2 1
k+1
Y
..
=
(aj
.
j=2
k 1
a1 ak
a1 akk+11
a1 C1 to get
1
1
.
a1 ) ..
1
1
0
1
..
.
1
1
···
···
..
.
0
0
ak2 2
ak2 1
···
···
akk 2
akk+12
akk 1
akk+11
..
.
..
.
.
Now expand along the first row, giving
k+1
Y
(aj
a1 )
j=2
1
..
.
1
1
ak2
..
.
···
..
.
···
···
2
akk 2
akk+12
ak2
..
.
1
akk 1
akk+11
.
But the remaining determinant is again a Vandermonde determinant in the k variables a2 , a3 , . . . , ak+1 ,
so we know its value from the inductive hypothesis. Substituting that value, we get for the determinant
of the original matrix
k+1
Y
(aj
a1 )
j=2
Y
(aj
ai ) =
2i<jk+1
Y
(aj
ai )
1i<jk+1
as desired.
For the second part, we can interpret the matrix above as the coefficient matrix formed by starting
with the (n 1)st degree polynomial y = c0 + c1 x + · · · + cn 1 xn 1 , substituting the n given points for
x, and regarding the resulting equations as a linear system in the ci , just as in Exercises 15 and 18.
Since the determinant is nonzero if the x-coordinates of the points are all distinct (being a product of
their di↵erences), it follows that the system has a unique solution, so that there is a unique polynomial
of degree at most n 1 (it may be less; for example, see Exercise 16(b)) passing through the n points.
4.3
Eigenvalues and Eigenvectors of n ⇥ n Matrices
1. (a) The characteristic polynomial of A is
det(A
I) =
1
3
2
6
= (1
)(6
)
( 2) · 3 =
2
7 + 12 = (
(b) The eigenvalues of A are the roots of the characteristic polynomial, which are
3)(
= 3 and
4)
= 4.
(c) To find the eigenspace corresponding to the eigenvalue = 3 we must find the null space of

2 3
A 3I =
2 3
4.3. EIGENVALUES AND EIGENVECTORS OF N ⇥ N MATRICES
271
Row reduce this matrix:
⇥
A
3I
⇤
0 =

2
2

3
2
3
3
0
0
!

2
0

1
0
3
0
0
0

✓ ◆
3t
3
Thus eigenvectors corresponding to = 3 are of the form
, so that E3 = span
.
2t
2
To find the eigenspace corresponding to the eigenvalue = 4 we must find the null space of

3 3
A 4I =
2 2
Row reduce this matrix:
⇥
A
4I
⇤
0 =
3
2
0
0
!
1
0
0
0

t
= 4 are of the form
, so that E4 = span
t
Thus eigenvectors corresponding to
✓
◆
1
1
.
(d) Since = 3 and = 4 are single roots of the characteristic polynomials, the algebraic multiplicity
of each is 1. Since each eigenspace is one-dimensional, the geometric multiplicity of each eigenvalue
is also 1.
2. (a) The characteristic polynomial of A is
det(A
I) =
2
1
= (2
1
)(
2
1 · ( 1) =
)
1)2 .
2 +1=(
(b) The eigenvalues of A are the roots of the characteristic polynomial; this polynomial has as its
only root = 1.
(c) To find the eigenspace corresponding to the eigenvalue we must find the null space of

1
1
A I=
1
1
Row reduce this matrix:
⇥
A
I
⇤
0 =
Thus eigenvectors corresponding to

1
1
1
1
0
0
!
= 1 are of the form


1
0
1
0
0
0
t
, which is span
t
✓
1
1
◆
.
(d) Since
= 1 is a double root of the characteristic polynomial, its algebraic multiplicity is 2.
However, the eigenspace is one-dimensional, so its geometric multiplicity is 1.
3. (a) The characteristic polynomial of A is
1
det(A
I) =
1
0
0
0
1
2
0
= (1
)( 2
)(3
)=
(
1)( + 2)(
3).
3
(b) The eigenvalues of A are the roots of the characteristic polynomial, which are
= 3.
= 1,
=
2, and
(c) To find the eigenspace corresponding to the eigenvalue = 1 we must find the null space of
2
3
0
1 0
3 15
A I = 40
0
0 2
272
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
Row reduce this matrix:
2
0
⇤
0 = 40
0
3
2
3
0
0 1 0 0
A I
05 ! 40 0 1 05
0
0 0 0 0
2 3
02 31
t
1
Thus eigenvectors corresponding to = 1 are of the form 405, so that E1 = span @405A.
0
0
To find the eigenspace corresponding to the eigenvalue = 2 we must find the null space of
2
3
3 1 0
A + 2I = 40 0 35
0 0 5
⇥
1 0
3 1
0 2
Row reduce this matrix:
h
2
3
6
0 = 40
0
i
A + 2I
1
0
0
0
3
5
3
2
0
1
7
6
05 ! 40
0
0
2
1
3
0
0
0
1
0
31
1
Thus eigenvectors corresponding to = 2 are of the form 4 t 5, so that E 2 = span @4 35A.
0
0
To find the eigenspace corresponding to the eigenvalue = 3 we must find the null space of
2
3
2
1 0
5 15
A 3I = 4 0
0
0 0
1
3t
3
3
0
7
05
0
02
Row reduce this matrix:
h
A
3I
2
2
6
0 =4 0
0
i
1 0
5 1
0 0
3
2
0
1
7
6
05 ! 40
0
0
Thus eigenvectors corresponding to = 3 are of the form
2 3
2 3
1
1
10 t
61 7
6 7
4 5 t 5 , or, clearing fractions, 4 2 5 , so that
t
10
1
10
1
5
0
1
0
E
0
2
3
0
7
05
0
02
31
1
B6 7 C
= span @4 2 5A .
10
(d) Since all three eigenvalues are single roots of the characteristic polynomials, the algebraic multiplicity of each is 1. Since each eigenspace is one-dimensional, the geometric multiplicity of each
eigenvalue is also 1.
4. (a) The characteristic polynomial of A is
1
det(A
= (1
= (1
=
0
0
1
I) =
1
1
1
1
)
)(
1
1
1
2
( + 1)(
+1
0
1
1
1
1) + 1( 1 + ) =
1)(
2).
3
+2
2
+
2
4.3. EIGENVALUES AND EIGENVECTORS OF N ⇥ N MATRICES
273
(b) The eigenvalues of A are the roots of the characteristic polynomial, which are
and 3 = 2.
(c) To find the eigenspace corresponding to
=
1,
2
= 1,
1 we must find the null space of
2
3
2 0 1
A + I = 4 0 2 15 .
1 1 1
Row reduce this matrix:
h
1
2
1
2
i 6
0 =6
40
A+I
1
=
0
1
2
1
1
1
3
2
0
1
7
6
7
6
05 ! 40
0
0
2
1
1
2
1
2
0
0
0
3
3
02 3 1
t
1
Thus eigenvectors corresponding to 1 are of the form 4
or 4 t5, which is span @4 15A.
2t
2
t
To find the eigenspace corresponding to 2 = 1 we must find the null space of
2
3
0 0
1
15 .
A I = 40 0
1 1
1
1
2t
1 5
2t ,
2
3
0
7
07
5
0
Row reduce this matrix:
2
0
⇤
0 = 40
1
3
2
3
0
1 1 0 0
A I
05 ! 40 0 1 05
0
0 0 0 0
2 3
02 31
t
1
Thus eigenvectors corresponding to 2 are of the form 4 t5, which is span @4 15A.
0
0
To find the eigenspace corresponding to 3 = 2 we must find the null space of
2
3
1
0
1
1
15 .
A 2I = 4 0
1
1
2
⇥
0
0
1
1
1
1
Row reduce this matrix:
⇥
A
2I
2
3
2
0
1 0
05 ! 40 1
0
0 0
23
t
4
are of the form t5, which is
t
1
0 =4 0
1
⇤
Thus eigenvectors corresponding to
3
0
1
1
1
1
2
3
0
05
0
02 31
1
@
4
15A.
span
1
1
1
0
(d) Since each eigenvalue is a single root of the characteristic polynomial, each has algebraic multiplicity 1. Since each eigenspace is one-dimensional, each eigenvalue also has geometric multiplicity
1.
5. (a) The characteristic polynomial of A is
1
det(A
I) =
2
1
0
0
1
1
1
1
=
3
+
2
=
2
(
1).
274
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
(b) The eigenvalues of A are the roots of the characteristic polynomial, which are
(c) To find the eigenspace corresponding to
Row reduce this matrix:
2
1
⇥
⇤
A 0 =4 1
0
the eigenvalue
Thus eigenvectors corresponding to
= 1.
= 0 we must find the null space of A.
3
2
0
1
05 ! 40
0
0
2 0
1 1
1 1
= 0 and
0
1
0
3
0
05
0
2
1
0
= 0 are of the form
2 3
2 3
2t
2
4 t5 = t 4 15 ,
t
1
02 31
2
so that the eigenspace is E0 = span @4 15A.
1
To find the eigenspace corresponding to the eigenvalue = 1 we must find the null space of
2
3
0
2 0
2 15
A I=4 1
0
1 0
Row reduce this matrix:
h
A
2
0
i
6
0 =4 1
0
I
Thus eigenvectors corresponding to
2 3
t
6 7
405 ,
t
3
2
0
1
7
6
05 ! 40
0
0
2 0
2 1
1 0
0
1
0
1
0
0
= 1 are of the form
3
0
7
05
0
02 3 1
1
B6 7 C
E1 = span @405A .
1
so that
(d) Since = 0 is a double root of the characteristic polynomial, it has algebraic multiplicity 2; since
its eigenspace is one-dimensional, it has geometric multiplicity 1. Since = 1 is a single root of the
characteristic polynomial, it has algebraic multiplicity 1; since its eigenspace is one-dimensional,
it also has geometric multiplicity 1.
6. (a) The characteristic polynomial of A is
1
det(A
I) =
0
3
2
2
3
1
0
1
0
1
1
3
= (1
)
= (1
)( (1 + )(1
= (1
)(
2
+2
)
3
2
1
0
3 · 0) + 2(3 · 0 + 2(1 + ))
1) + 4 + 4 =
3
+
2
+5 +3=
( + 1)2 (
(b) The eigenvalues of A are the roots of the characteristic polynomial, which are
(c) To find the eigenspace corresponding to
=
1 we must find the null space of
2
3
2 0 2
A + I = 4 3 0 35 .
2 0 2
1
=
1
3).
1 and
2
= 3.
4.3. EIGENVALUES AND EIGENVECTORS OF N ⇥ N MATRICES
Row reduce this matrix:
⇥
A+I
2
2
⇤
0 = 43
2
0
0
0
3
2
0
1
05 ! 4 0
0
0
2
3
2
275
0
0
0
1
0
0
3
0
05 .
0
Thus eigenvectors corresponding to 1 are of the form
2 3 2 3 2 3
2 3
2 3
02 3 2 31
s
s
0
1
0
1
0
4 t5 = 4 05 + 4 t5 = s 4 05 + t 415 = span @4 05 , 415A .
s
s
0
1
0
1
0
To find the eigenspace corresponding to
A
Row reduce this matrix:
h
(d)
A
2
= 3 we must find the null space of
2
3
2
0
2
4
35 .
3I = 4 3
2
0
2
2
2
i 6
0 =6
4 3
3I
2
0
2
4
3
0
2
3
2
0
1
7
6
7
6
05 ! 40
0
0
0
1
1
3
2
0
0
Thus eigenvectors corresponding to 2 are of the form
2 3
2 3
2 3
02 31
t
1
2
2
4 3 t5 = t 4 3 5 = s 435 = span @435A .
2
2
2
2
t
1
0
3
7
07
5.
0
1 is a double root of the characteristic polynomial, and its eigenspace has dimension 2, so its
algebraic and geometric multiplicities are both equal to 2. 2 is a single root of the characteristic
polynomial, and its eigenspace has dimension 1, so its algebraic and geometric multiplicities are
both equal to 1.
7. (a) The characteristic polynomial of A is
4
det(A
I) =
0
2
1
1
2
3
0
=
3
+9
2
27 + 27 =
(
3)3 .
2
(b) The eigenvalues of A are the roots of the characteristic polynomial, so the only eigenvalue is
(c) To find the eigenspace corresponding to the eigenvalue
2
1 0
A 3I = 4 2 0
1 0
Row reduce this matrix:
h
A
3I
2
1
i
6
=
0
4 2
1
Thus eigenvectors corresponding to
2 3
2 3
2
t
0
6 7
6 7
6
4 s 5 = s 4 15 + t 4
t
0
0
0
0
1
2
1
= 3 we must find the null space of
3
1
25
1
3
2
0
1
7
6
!
05
40
0
0
0
0
0
1
0
0
3
0
7
05
0
= 3 are of the form
3
02 3 2 3 1
1
0
1
7
B6 7 6 7 C
05 , so that E3 = span @415 , 4 05A .
1
0
1
= 3.
276
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
(d) Since = 3 is a triple root of the characteristic polynomial, it has algebraic multiplicity 3; since
its eigenspace is two-dimensional, it has geometric multiplicity 2.
8. (a) The characteristic polynomial of A is
1
det(A
1
0
1
I) =
1
1
1
1
)
= (2
)((1
(
1
1
= (2
=
1
0
2
2)
)2
2
( 1) · ( 1)) = (2
)(
2
2 )
(b) The eigenvalues of A are the roots of the characteristic polynomial, which are
(c) To find the eigenspace corresponding to 1 = 0 we must find the null space of
2
3
1
1
1
2
05 .
A + 0I = 4 0
1
1
1
Row reduce this matrix:
⇥
A + 0I
2
1
0 =4 0
1
⇤
1
2
1
1
0
1
3
2
0
1
05 ! 40
0
0
0
1
0
1
0
0
Thus eigenvectors corresponding to 1 are of the form
2 3
2 3
02 31
s
1
1
405 = s 405 = span @405A .
s
1
1
To find the eigenspace corresponding to
A
Row reduce this matrix:
⇥
A
2I
1
= 0 and
2
= 2.
3
0
05 .
0
= 2 we must find the null space of
2
3
1
1
1
0
05 .
2I = 4 0
1
1
1
2
2
1
⇤
0 =4 0
1
1
0
1
1
0
1
3
2
0
1
05 ! 40
0
0
1
0
0
1
0
0
3
0
05 .
0
Thus eigenvectors corresponding to 1 are of the form
2
3
2 3
2 3
02 3 2 3 1
s t
1
1
1
1
4 s 5 = s 4 15 + t 4 05 = span @4 15 , 4 05A .
t
0
1
0
1
(d) Since 1 = 0 is a single root of the characteristic polynomial and its eigenspace has dimension
1, its algebraic and geometric multiplicities are both equal to 1. Since 2 = 2 is a double root
of the characteristic polynomial and its eigenspace has dimension 2, its algebraic and geometric
multiplicities are both equal to 2.
9. (a) The characteristic polynomial of A is
3
det(A
I) =
1
1
0
0
0
0
1
0
0
0
0
4
1
1
1
=
4
6
3
+9
2
+4
12 = ( + 1)(
2)2 (
3)
4.3. EIGENVALUES AND EIGENVECTORS OF N ⇥ N MATRICES
277
(b) The eigenvalues of A are the roots of the characteristic polynomial, so the eigenvalues are
= 2, and = 3.
=
(c) To find the eigenspace corresponding to the eigenvalue = 1 we must find the null space of
2
3
4 1 0 0
6 1 2 0 07
7
A+I =6
4 0 0 2 45
0 0 1 2
Row reduce this matrix:
h
A+I
2
i 6
6
0 =6
4
4
1
0
0
1
2
0
0
0
0
2
1
3
2
0
1
7
6
07
60
7 !6
40
05
0
0
4
2
0
0
0
1
0
0
Thus eigenvectors corresponding to = 1 are of the form
2
3
02
0
6 0 7
B6
6
7
B6
4 2t5 , so that E 1 = span @4
t
To find the eigenspace corresponding to the eigenvalue
2
1
1
6 1
1
A 2I = 6
4 0
0
0
0
Row reduce this matrix:
h
A
2I
2
i 6
6
0 =6
4
1
1
0
0
Thus eigenvectors corresponding to
2 3
s
6s7
6 7
6 7,
405
0
1
1
0
0
0
0
1
1
0
0
4
1
h
A
3I
i 6
6
0 =6
4
3
0
07
7
7
05
0
= 2 we must find the null space of
3
0
0
0
07
7
1
45
1
1
1
0
0
0
0
1
0
0
0
0
1
0
= 2 are of the form
02
B6
B6
E2 = span B6
@4
so that
Row reduce this matrix:
0
1
0
0
0
0
2
0
31
0
C
07
7C .
5
2 A
1
3
2
0
1
7
6
07
60
7 !6
5
40
0
0
0
To find the eigenspace corresponding to the eigenvalue
2
0
1
6 1
2
A 3I = 6
4 0
0
0
0
2
0
0
1
0
1
2
0
0
0
0
2
1
0
0
4
2
31
1
C
17
7C
7C .
05 A
0
= 3 we must find the null space of
3
0
0
0
07
7
2
45
1
2
3
2
0
1
60
07
7
6
7 !6
40
05
0
3
0
07
7
7
05
0
0
0
1
0
0
0
0
1
0
0
0
2
0
3
0
07
7
7
05
0
1,
278
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
Thus eigenvectors corresponding to
2 3
0
607
6 7
6 7,
42t5
= 3 are of the form
so that
t
02 31
0
B607C
B6 7C
E3 = span B6 7C .
@425A
1
(d) Since = 3 and = 1 are single roots of the characteristic polynomial, they each have algebraic
multiplicity 1; their eigenspaces are each one-dimensional, so those two eigenvalues have geometric
multiplicity 1 as well. Since = 2 is a double root of the characteristic polynomial, it has algebraic
multiplicity 2; since its eigenspace is one-dimensional, it only has geometric multiplicity 1.
10. (a) The characteristic polynomial of A is
2
det(A
1
0
0
0
I) =
1
4
1
0
0
0
5
1
3
0
2
Since this is a triangular matrix, its determinant is the product of the diagonal entries by Theorem
4.2, so the characteristic polynomial is (2
)2 (1
)(3
).
(b) The eigenvalues of A are the roots of the characteristic polynomial, which are
and 3 = 3.
(c) To find the eigenspace corresponding to
= 1 we must find the null space of
2
3
1 1 1 0
6 0 0 4 57
7
I=6
4 0 0 2 15 .
0 0 0 1
1
A
Row reduce this matrix:
⇥
A
I
2
1
⇤ 60
0 =6
40
0
1
0
0
0
1
4
2
0
Thus eigenvectors corresponding to 1 are
2 3
2
t
6 t7
6
6 7 = t6
4 05
4
0
To find the eigenspace corresponding to
A
Row reduce this matrix:
⇥
A
2I
2
0
⇤ 60
0 =6
40
0
0
5
1
1
3
2
0
1
60
07
7 !6
40
05
0
0
of the form
3
02
1
B6
17
7 = span B6
5
@4
0
0
= 2 we
2
0
60
2I = 6
40
0
2
1
1
0
0
1
4
1
0
0
5
1
0
1
0
0
0
0
1
0
0
3
0
07
7.
05
0
0
0
1
0
31
1
C
17
7C .
5
0 A
0
must find the null space of
3
1 1 0
1 4 57
7.
0 1 15
0 0 0
3
2
0
0
60
07
7 !6
40
05
0
0
1
0
0
0
0
1
0
0
1
1
0
0
3
0
07
7.
05
0
1
= 1,
2
= 2,
4.3. EIGENVALUES AND EIGENVECTORS OF N ⇥ N MATRICES
Thus eigenvectors corresponding to 2 are of the form
2 3
2 3
2 3
02 3
s
1
0
1
6 t7
607
6 17
B607
6 7 = s 6 7 + t 6 7 = span B6 7 ,
4 t5
405
4 15
@405
t
0
1
0
To find the eigenspace corresponding to
= 3 we
2
1
6 0
6
3I = 4
0
0
A
Row reduce this matrix:
⇥
A
3I
2
1
⇤ 6 0
0 =6
4 0
0
3
1
2
0
0
1
4
0
0
0
5
1
1
279
2
31
0
6 17C
6 7C .
4 15A
1
must find the null space of
3
1 1
0
2 4
57
7.
0 0
15
0 0
1
3
2
0
1
60
07
7 !6
40
05
0
0
0
1
0
0
3
0
07
7.
05
0
3 0
2 0
0 1
0 0
Thus eigenvectors corresponding to 1 are of the form
2 3
2 3
02 3 1
3t
3
3
62t7
627
B6 27 C
6 7 = t 6 7 = span B6 7C .
4 t5
415
@4 15 A
0
0
0
(d) Since 1 = 1 is a single root of the characteristic polynomial and its eigenspace has dimension
1, its algebraic and geometric multiplicities are both equal to 1. Since 2 = 2 is a double root
of the characteristic polynomial and its eigenspace has dimension 2, its algebraic and geometric
multiplicities are both equal to 2. Since 3 = 3 is a single root of the characteristic polynomial
and its eigenspace has dimension 1, its algebraic and geometric multiplicities are both equal to 1.
11. (a) The characteristic polynomial of A is
1
det(A
I) =
0
0
1
2
0
0
1
1
1
0
0
0
3
2
1)2 (
=(
3)( + 1)
1
(b) The eigenvalues of A are the roots of the characteristic polynomial, so the eigenvalues are
= 1, and = 3.
=
(c) To find the eigenspace corresponding to the eigenvalue = 1 we must find the null space of
2
3
2 0 0 0
6 0 2 0 07
7
A+I =6
4 1 1 4 05
2 1 2 0
Row reduce this matrix:
⇥
A+I
2
2
⇤ 6 0
0 =6
4 1
2
0
2
1
1
0
0
4
2
0
0
0
0
3
2
0
1
60
07
7 !6
40
05
0
0
0
1
0
0
0
0
1
0
0
0
0
0
3
0
07
7
05
0
1,
280
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
Thus eigenvectors corresponding to = 1 are of the form
2 3
02 31
0
0
6 07
B607C
6 7 , so that E 1 = span B6 7C .
4 05
@405A
t
1
To find the eigenspace corresponding to the eigenvalue
2
0 0 0
6 0 0 0
A I=6
4 1 1 2
2 1 2
= 1 we must find the null space of
3
0
07
7
05
2
Row reduce this matrix:
h
A
I
2
0
0
1
2
i 6
6
0 =6
4
Thus eigenvectors corresponding to
2
2
3t
3
2
6 2s + 2 t7
6
6
3 7 = s6
6
7
6
4
5
4
s
t
0
0
1
1
0
0
2
2
3
2
0
1
7
6
07
60
7 !6
40
05
0
0
0
2
0
0
1
0
0
0
3
0
07
7
7
05
2
3
2
3
0
2
0
0
0
0
0
= 1 are of the form
3
2 23
0
3
6 27
27
7
6 37
7 + t6 7,
4 05
15
0
1
so that
02
B6
B6
E1 = span B6
@4
3 2
0
6
27
7 6
7,6
15 4
0
31
2
C
27
7C
7C .
05A
3
To find the eigenspace corresponding to the eigenvalue = 3 we must find the null space of
2
3
2
0 0
0
6 0
2 0
07
7
A 3I = 6
4 1
1 0
05
2
1 2
4
Row reduce this matrix:
h
A
3I
2
i 6
6
0 =6
4
2
0
1
2
Thus eigenvectors corresponding to
2 3
0
607
6 7
6 7,
42t5
t
0
2
1
1
0
0
0
2
0
0
0
4
3
2
0
1
7
6
07
60
7 !6
40
05
0
0
0
1
0
0
0
0
1
0
0
0
2
0
3
0
07
7
7
05
0
= 3 are of the form
so that
02 31
0
B607C
B6 7C
E3 = span B6 7C .
@425A
1
(d) Since = 3 and = 1 are single roots of the characteristic polynomial, they each have algebraic
multiplicity 1; their eigenspaces are each one-dimensional, so those two eigenvalues have geometric
multiplicity 1 as well. Since = 1 is a double root of the characteristic polynomial, it has algebraic
multiplicity 2; since its eigenspace is two-dimensional, it also has geometric multiplicity 2.
4.3. EIGENVALUES AND EIGENVECTORS OF N ⇥ N MATRICES
281
12. (a) The characteristic polynomial of A is
4
det(A
0
0
0
0
I) =
1
1
4
0
0
1
3
4
=
=
0
0
0
3
0
)
2
= (4
0
1 +(
2
4
6(4
) (
0
1
2
2
(1
2
)(4
4
)
6) = (4
0
0
0
)
1
1
4
0
1
2
)2 (
3)( + 2)
since each of the 3 ⇥ 3 matrices is upper triangular, so that by Theorem 4.2 their determinants
are the product of their diagonal elements.
(b) The eigenvalues of A are the roots of the characteristic polynomial, which are
and 3 = 2.
(c) To find the eigenspace corresponding to
A
Row reduce this matrix:
⇥
A
4I
2
0
⇤ 60
0 =6
40
0
=4
2
0
60
4I = 6
40
0
1
0
0
0
0
1
1
3
0
we must find the null space of
3
0
1
0
0
1
17
7.
0
3
25
0
0
4
0
1
2
4
3
2
0
0
60
07
7 !6
40
05
0
0
Thus eigenvectors corresponding to 1 are of the form
2 3
2 3
2 3
02 3
s
1
0
1
6 t7
6 07
617
B6 07
6 7 = s 6 7 + t 6 7 = span B6 7 ,
405
4 05
405
@4 05
0
0
0
0
To find the eigenspace corresponding to
A
Row reduce this matrix:
⇥
A
3I
2
1
⇤ 60
0 =6
40
0
=3
2
1
60
3I = 6
40
0
0
1
0
0
2
1
1
2
0
0
0
0
0
1
0
0
0
0
1
0
0
3
0
07
7.
05
0
2 31
0
6 17 C
6 7C .
4 05 A
0
we must find the null space of
3
0
1
0
1
1
17
7.
0
2
25
0
0
3
0
1
2
3
3
2
1
0
60
07
7 !6
40
05
0
0
Thus eigenvectors corresponding to 1 are of the form
2
3
2 3
02
t
1
6 2t7
6 27
B6
6
7
6 7
B6
4 t5 = t 4 15 = span @4
t
1
0
1
0
0
31
1
C
27
7C .
5
1 A
1
0
0
1
0
1
2
1
0
3
0
07
7.
05
0
1
= 4,
2
= 3,
282
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
To find the eigenspace corresponding to
2 we must find the null space of
3
6 0 1 0
6 0 6 1 17
7
A + 2I = 6
4 0 0 3 25 .
0 0 3 2
3
=
2
Row reduce this matrix:
h
2
6
i 6
60
0 =6
60
4
0
A + 2I
0
1
0
6
1
1
0
3
2
0
3
2
Thus eigenvectors corresponding to 1 are of the
2
3 2
3
2
1
2s
9t
6 1 7 6
7
6
6 18 t7 6
6
s7
6
7=6
7 = s6
6 2 t7 6 12s7
6
4 3 5 4
5
4
t
18s
3
2
0
1
7
6
6
07
7 ! 60
7
60
05
4
0
0
0
0
1
0
0
1
1
9
1
18
2
3
0
0
0
0
form
3
02
31
2
2
7
B6
7C
B6
7C
17
7 = span B6 17C .
7
B
6
7C
125
@4 125A
18
18
3
7
07
7.
07
5
0
(d) Since 1 = 4 is a double root of the characteristic polynomial and its eigenspace has dimension
2, its algebraic and geometric multiplicities are both equal to 2. Since 2 = 3 is a single root
of the characteristic polynomial and its eigenspace has dimension 1, its algebraic and geometric
multiplicities are both equal to 1. Since 3 = 2 is a single root of the characteristic polynomial
and its eigenspace has dimension 1, its algebraic and geometric multiplicities are both equal to 1.
13. We must show that if A is invertible and Ax = x, then A 1 x = 1 x = 1 x. First note that since A
is invertible, 0 is not an eigenvalue, so that 6= 0 and this statement makes sense. Now,
x = (A
Dividing through by
gives A
1
1
A)x = A
x=
1
1
(Ax) = A
1
( x) = (A
1
x).
x as required.
14. The remark referred to in Section 3.3 tells us that if A is invertible and n > 0 is an integer, then
A n = (A 1 )n = (An ) 1 . Now, part (a) gives us the result in part (c) when n > 0 is an integer. When
n = 0, the statement says that 0 = 1 is an eigenvalue of A0 = I, which is true. So it remains to prove
the statement for n < 0. We use induction on n. For n = 1, this is the statement that if Ax = x,
1
then A 1 x =
x, which is Theorem 4.18(b). Now assume the statement holds for n = k, and
prove it for n = (k + 1). Note that by the remark in Section 3.3,
A
(k+1)
= (Ak+1 )
1
= (Ak A)
1
=A
1
(Ak )
1
=A
1
A
k
k
1
.
Now
A
(k+1)
x=A
1
(A
k
x)
inductive
hypothesis
=
A
1
(
k
x) =
k
A
1
n= 1
x =
x=
(k+1)
x
and we are done.
15. If we can express x as a linear combination of the eigenvectors, it will be easy to computeA10 x since

⇥
⇤ c1
5
we know what A10 does to the eigenvectors. To so express x, we solve the system v1 v2
=
,
c2
1
which is
c1 + c2 = 5
c1 + c2 = 1.
4.3. EIGENVALUES AND EIGENVECTORS OF N ⇥ N MATRICES
283
This has the solution c1 = 2 and c2 = 3, so that x = 2v1 + 3v2 . But by Theorem 4.4(a), v1 and v2
10
are eigenvectors of A10 with eigenvalues 10
1 and 2 . Thus
A10 x = A10 (2v1 + 3v2 ) = 2A10 v1 + 3A10 v2 = 2
Substituting for v1 and v2 gives
"
3072 +
A x=
3072
10
16. As in Exercise 15,
10
1 v1
1
512
1
512
#
+3
✓ ◆k

1
3 · 2k + 2 1
k
A x=2·
v1 + 3 · 2 v2 =
3 · 2k 21
2
k
=
1
v1 + 3072v2 .
512
.
k
As k ! 1, the 21
10
2 v2
k
k
.
terms approach zero, so that Ak x ⇡ 3 · 2k v2 .
17. We want to express x as a linear combination of the eigenvectors, say x = c1 v1 + c2 v2 + c3 v3 , so we
want a solution of the linear system
c1 + c2 + c3 = 2
c2 + c3 = 1
c3 = 2.
Working from the bottom and substituting gives c3 = 2, c2 =
Then
A20 x = A20 (v1
20
= A v1
v2 + 2v3 )
A20 v2 + 2A20 v3
20
1 v1
=
=
1, and c1 = 1, so that x = v1 v2 +2v3 .
✓
2
= 42
20
20
2 v2 + 2 3 v3
2 3
2 3
2 3
◆20 1
✓ ◆20 1
1
1
1
405
415 + 2 · 120 415
3
3
0
0
1
3
2
1 5
320 .
2
18. As in Exercise 17,
Ak x = Ak (v1
k
= A v1
v2 + 2v3 )
Ak v2 + 2Ak v3
k
k
2 v2 + 2 3 v3
2 3
2 3
2 3
✓
◆k 1
✓ ◆k 1
1
1
1
405
415 + 2 · 1k 415
=
3
3
0
0
1
2 1
3
1
± 3k
+2
3k
1
5.
+
2
=4
k
3
2
=
As k ! 1, the
1
3k
k
1 v1
2 3
2
terms approach zero, and Ak x ! 425.
2
284
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
19. (a) Since ( I)T = I T = I, we have
AT
I = AT
( I)T = (A
I)T ,
so that using Theorem 4.10,
det(AT
I)T = det(A
I) = det (A
I).
But the left-hand side is the characteristic polynomial of AT , and the right-hand side is the
characteristic polynomial of A. Thus the two matrices have the same characteristic polynomial
and thus the same eigenvalues.
(b) There are many examples. For instance
A=
has eigenvalues
1
= 1 and

1
1
0
2
= 2, with eigenspaces
✓
◆
✓ ◆
1
0
E1 = span
, E2 = span
.
1
1
2
AT has the same eigenvalues by part (a), but the eigenspaces are
✓ ◆
✓ ◆
1
1
E1 = span
, E2 = span
.
0
1
20. By Theorem 4.18(b), if is an eigenvalue of A, then m is an eigenvalue of Am = O. But the only
eigenvalue of O is zero, since Am takes any vector to zero. Thus m = 0, so that = 0.
21. By Theorem 4.18(a), if is an eigenvalue of A with eigenvector x, then 2 is an eigenvalue of A2 = A
with eigenvector x. Since di↵erent eigenvalues have linearly independent eigenvectors by Theorem 4.20,
we must have 2 = , so that = 0 or = 1.
22. If Av = v, then Av cIv = v cIv, so that (A cI)v = (
that v is an eigenvector of A cI with eigenvalues
c.
c)v. But this equation means exactly
23. (a) The characteristic polynomial of A is
A
3
I =
So the eigenvalues of A are
1
2
= 5 and
A
and for
2,
=
5
2
5I =
=

2
5
we have

5
A + 2I =
5
2
3
2. For
10 = (
1,
we have

1
,
0
2
1
!
5
0

2
1
!
2
0
2
5
0
5)( + 2).
.
It follows that
E5 = span
(b) By Theorem 4.4(b), A
1
✓
1
1
◆
,
E
2
= span
✓
2
5
1
◆
= span
✓
has eigenvalues 15 and 12 , with
✓ ◆
✓
◆
1
2
E1/5 = span
, E 1/2 = span
.
1
5
2
5
◆
.
4.3. EIGENVALUES AND EIGENVECTORS OF N ⇥ N MATRICES
By Exercise 22, A
285
2I has eigenvalues 5 2 = 3 and 2 2 = 4 with
✓ ◆
✓
◆
1
2
E3 = span
, E 4 = span
.
1
5
Also by Exercise 22, A + 2I = A
( 2I) has eigenvalues 5 + 2 = 7 and
E7 = span
✓
◆
1
1
,
E0 = span
✓
2
5
◆
2 + 2 = 0 with
.
24. (a) There are many examples. For instance, let

0
A=
1
1
,
0
B=

1
0
1
.
1
Then the characteristic equation of A is 2 1, so that A has eigenvalues
teristic equation of B is (µ 1)2 , so that B has eigenvalue µ = 1. But
A+B =
has characteristic polynomial
+ µ are 0 and 2.
2

1
1
2
1
1, so it has eigenvalues 1 ±
2
= ±1. The charac-
p
2. The possible values of
(b) Use the same A and B as above. Then
AB =
with characteristic polynomial
+ µ.
(c) If

0
1
1
,
1
p
1 and eigenvalues 12 (1 ± 5), which again does not match
2
and µ both have eigenvector x, then
(A + B)x = Ax + Bx = x + µx = ( + µ)x,
which shows that
+ µ is an eigenvalue of A + B with eigenvector x. For the product, we have
(AB)x = A(Bx) = A(µx) = µAx = µ x = µx,
showing that µ is an eigenvalue of AB with eigenvector x.
25. No, they do not. For example, take B = I2 and let A be any 2 ⇥ 2 rank 2 matrix that has an eigenvalue
0 2
other than 1. For example let A =
, which has eigenvalues 2 and 2. Since A has rank 2, it
2 0
row-reduces to I, so A and I are row-equivalent but do not have the same eigenvalues.
26. By definition, the companion matrix of p(x) = x2
C(p) =

7x + 12 is

12
7
=
0
1
( 7)
1
12
,
0
which has characteristic polynomial
7
12
1
= (7
)(
) + 12 =
2
7 + 12.
286
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
27. By definition, the companion matrix of p(x) = x3 + 3x2 4x + 12 is
2
3 2
3
3
( 4)
12
3 4
12
0
0 5=4 1 0
05 ,
C(p) = 4 1
0
1
0
0 1
0
which has characteristic polynomial
3
4
1
0
12
0 =
3
1
1
=
12
=
3
(( 3
3
C(p) =

28. (a) If p(x) = x2 + ax + b, then
so the characteristic polynomial of C(p) is
a
b
1
12
0
1
2
=( a
3
)(
+4
4
1
)
4)
12.
a
1
b
,
0
)(
)+b=
2
+ a + b.

x
(b) Suppose that is an eigenvalue of C(p), and that a corresponding eigenvector is x = 1 . Then
x2


x1
x
= C(p) 1 means that
x2
x2




x1
a
b x1
ax1 bx2
=
=
,
x2
1
0 x2
x1

so that x1 = x2 . Setting x2 = 1 gives an eigenvector
.
1
29. (a) If p(x) = x3 + ax2 + bx + c, then
2
a
C(p) = 4 1
0
3
c
05 ,
0
b
0
1
so the characteristic polynomial of C(p) is
a
b
1
0
c
0 =
1
=
c
a
1
=
c
0
1
(( a
3
a
2
a
)(
b
b
1
) + b)
c.
2
3
x1
(b) Suppose that is an eigenvalue of C(p), and that a corresponding eigenvector is x = 4x2 5. Then
x3
2 3
2 3
x1
x1
4x2 5 = C(p) 4x2 5 means that
x3
x3
2
3 2
32 3 2
3
x1
a
b
c
x1
ax1 bx2 cx3
4 x2 5 = 4 1
5,
0
05 4 x2 5 = 4
x1
x3
0
1
0
x3
x2
4.3. EIGENVALUES AND EIGENVECTORS OF N ⇥ N MATRICES
287
2
so that x1 = x2 and x2 = x3 . Setting x3 = 1 gives an eigenvector 4
30. Using Exercise 28, the characteristic polynomial of the matrix

a
b
C(p) =
1
0
is
2
2
31. Using Exercise 28, the characteristic polynomial of the matrix
2
3
a
b
c
0
05
C(p) = 4 1
0
1
0
is
(
2
+ a + b. If we want this matrix to have eigenvalues 2 and 5, then
7 + 10, so we can take

7
10
A=
.
1
0
3
+a
2
+ b + c). If we want this matrix to have eigenvalues
(
3
+a
2
+ b + c) =
so we can take
( + 2)(
2
2
A = 41
0
1)(
1
3
5.
+a +b = (
2)(
5) =
2, 1, and 3, then
3
3) =
2
2
2
5 + 6,
3
6
05 .
0
5
0
1
32. (a) Exercises 28 and 29 provide a basis for the induction, proving the cases n = 2 and n = 3. So
assume that the statement holds for n = k; that is, that
ak
ak
1
···
···
..
.
2
1
0
0
0
where p(x) = xk + ak 1 xk
has degree k + 1; then
1
1
0
0
···
···
a1
0
a0
0
0
0
0
= ( 1)k p( ),
1
+ · · · + a1 x + a0 . Now suppose that q(x) = xk+1 + bk xk + · · · + b1 x + b0
C(q) =
bk
1
bk
0
0
0
1
0
0
···
···
..
.
1
···
···
b1
0
b0
0
0
0
0
.
1
Evaluate this determinant by expanding along the last column, giving
1
C(q) = ( 1)k ( b0 ) 0
0
0
1
0
0
···
..
.
···
···
0
0 +(
1
)
bk
1
bk
0
0
0
1
0
0
1
···
···
..
.
···
···
b2
0
b1
0
0
0
0
.
1
The first determinant is of an upper triangular matrix, so it is the product of its diagonal entries,
which is 1. The second determinant is the characteristic polynomial of xk +bk xk 1 +· · ·+b2 x+b1 ,
288
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
which is a k th degree polynomial. So by the inductive hypothesis, this determinant is ( 1)k (
bk k 1 + · · · + b2 + b1 ). So the entire expression becomes
C(q) = ( 1)k ( b0 )
( 1)k (
= ( 1)k+1 b0 + ( 1)k+1 (
= ( 1)
k+1
k+1
(
= ( 1)
k+1
q( ).
k
+ bk
k
+ bk
k+1
k 1
+ bk
k
+ · · · + b2
2
an
0
1
..
.
1
2
+ · · · + b2
2
+ b1 )
+ b1 + b0 )
0
1 x1
an
is an eigenvalue of C(p)
32
3
a0
x1
6
7
0 7
7 6 x2 7
6 .. 7
0 7
76
7
.. 7 6 . 7
5
4
.
xn 1 5
···
1
0
xn
3
· · · a1 xn 1 a0 xn
2 x2
7
x1
7
7
x2
7.
7
..
5
.
xn 1
···
···
···
..
.
But this means that xi = xi+1 for i = 1, 2, . . . , n
2 n
a1
0
0
..
.
1. Setting xn = 1 gives an eigenvector
3
1
6 n 27
6
7
6
7
x = 6 ... 7 .
6
7
4
5
1
33. The characteristic polynomial of A is
cA ( ) = det(A
I) =
1
1
2
Now,
A2 =
= (1
3

1
2
1
3
2
=

)(3
1
8
)+2=
4
,
7
so that
cA (A) = A2 4A + 5I2


1
4
1
=
4
8
7
2
= O.
+
+ · · · + b2 + b1 )
(b) The method is quite similar to Exercises 28(b) and 29(b). Suppose that
with eigenvector
2
3
x1
6 x2 7
6
7
6
7
x = 6 ... 7 .
6
7
4xn 1 5
xn
Then x = C(p)x, so that
2
3 2
x1
an
6 x2 7 6 1
6
7 6
6 .. 7 6 0
6 . 7=6
6
7 6 .
4 xn 1 5 4 ..
xn
0
2
an
6
6
6
=6
6
4
k

1
1
+5
3
0
0
1
2
4 + 5.
4.3. EIGENVALUES AND EIGENVECTORS OF N ⇥ N MATRICES
289
34. The characteristic polynomial of A is
1
cA ( ) = det(A
Now,
2
1
A2 = 4 1
0
so that
32 2
0
2
15 = 41
1
1
1
0
1
cA (A) =
=
2
1
1
0
I) =
1
2
1
1
2
0
0
3
=
+2
2
2
1
A3 = 41
0
3 2
1
1
15 + 41
2
0
3+2+1
2+4+0
3+2+1
33 2
0
3
15 = 43
1
2
1
0
1
1
0
1
0
2
0
3
0
15
1
2
1
2 40
0
2+2+0
3+2+1
3+4+1
4A + 5I = 0, we have A2 = 4A
A3 = A(4A
4
2
5I) = 4A2
2
A = (A )(A ) = (4A
5I)(4A
A3 + 2A2 + A
36. Since
A3 = 2A2 + A
4
5A = 4(4A
1
A
5
=
A
2
= A
1 2
5I, and
5I)
5A = 11A
5I) = 16A
2
20I
40A + 25I = 16(4A
2I) = 2A3 + A2
4A + 5I = 0, we get A(A
1
3
0
05
1
5I)
40A + 25I = 24A
55I.
2I,
A = A(A ) = A(2A2 + A
A
0
1
0
3
3
2
35 ,
3
2I = 0, we have
3
37. Since A2
2.
3
2
3
0
05
2
= O.
35. Since A2
+
1
3
1
15 ,
2
A3 + 2A2 + A 2I3
2
3
2
3 3 2
2 1
43 2 35 + 2 41 2
2 3 3
1 1
3+4+1
=4 3+2+1
2+2+0
0
1
4I) =
4
1
4
I=
A+ I
5
5
5
✓
◆2
1
4
1 2
=
A+ I
=
A
5
5
25
2A = 2(2A2 + A
2I) + A2
5I. Then solving for A
8
16
1
A + I2 =
(4A
25
25
25
1
5I)
2A = 5A2
4I.
yields
8
16
A+ I =
25
25
4
11
A + I.
25
25
38. Since A3 + 2A2 + A 2I = 0, rearranging terms gives A( A2 + 2A + I) = 2I. Then solving for A
yields (using Exercise 36)
A
1
=
A
2
=
=
=
=
1
1 2
1
I( A2 + 2A + I) =
A +A+ I
2
2
2
✓
◆2
1
1
2
A 1 =
A2 + A + I
2
2
1 4
1
1
A
A3 + A2 + A + I 2
4
2
4
1
1
1
2
2
(5A
4I) (2A + A 2I) + A2 + A + I
4
2
4
1 2 5
A + I.
4
4
1
290
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
39. Apply Exercise 69 in Section 4.2 to the matrix A
I:

P
I
Q
cA ( ) = det(A
I) = det
= (det(P
O
S
I
I))(det(S
I)) = cP ( )cS ( ).
(Note that neither O or Q is a↵ected by subtracting an identity matrix, since all the 1’s are along the
diagonal, in either P or S.)
40. Start with the equation given in the problem statement:
I) = ( 1)n (
det(A
This is an equation, so in particular it holds for
0I) = ( 1)n (
det A = det(A
n ).
= 0, and it becomes
2) · · · (
1 )(
2) · · · (
1 )(
n)
= ( 1)n ( 1)n
···
1 2
n
=
1 2
···
n,
proving the first result. For the second result, we will find the coefficients of n 1 on the left and right
sides. On the left side, the easiest method given the tools we have is induction on n. For n = 2,

a
a12
det(A
I) = det 11
= (a11
)(a22
) a12 a21 = 2 (a11 + a22 )
a12 a21 ,
a21
a22
so that the coefficient of n 1 = is tr(A). Claim that in general, for n 1, the coefficient of n 1
in det(A
I) is ( 1)n 1 tr(A). Suppose this statement holds for n = k. Then if A is (k + 1) ⇥ (k + 1),
we have (expanding along the first column of A
I)
det A = (a11
) det(A11
I) +
k+1
X
( 1)i+1 ai1 det(Ai1
I).
i=2
Now, the terms in the sum on the right each consist of a constant times the determinant of a k ⇥ k
matrix derived by removing row i and column 1 from A. Since i > 1, this means we have removed
both the entries a11
I and ai1
I from A, so that there are only k 1 entries in A that contain
. So the sum on the right cannot produce any terms with k . For the other term, note that A11
I
is a k ⇥ k matrix; using the inductive hypothesis we see that the coefficient of k 1 in det(A11
I)
is ( 1)k 1 tr(A11 ). Now
)(( 1)k
det A = (a11
so that the coefficient of
( 1)k a11
( 1)k
1
k
k
+ ( 1)k
1
tr(A11 )
k 1
+ ...) +
in det A is
X
...,
tr(A11 ) = ( 1)k (a11 + tr(A11 )) = ( 1)k (a11 + a22 + · · · + ak+1,k+1 ) = ( 1)k tr(A).
Now look at the product on the right in the original expression
I) = ( 1)n (
det(A
2) · · · (
1 )(
n ).
We can get a constant times n 1 by expanding the right-hand side when exactly n 1 of the factors
use and one uses i for some i; that contribution to the n 1 term in the characteristic polynomial
n 1
is
. Summing over all possible choices for i shows that the coefficient of the n 1 term is
i
( 1)n (
1
2
Since the coefficient of
( 1)n
1
···
n 1
n)
= ( 1)n+1 (
1
+
2
+ ··· +
n)
= ( 1)n
1
(
1
+
2
+ ··· +
on the left must be the same as that on the right, we have
tr(A) = ( 1)n
1
(
1
+
2
+ ··· +
n ),
or
tr(A) =
1
+
2
+ ··· +
n.
n ).
4.4. SIMILARITY AND DIAGONALIZATION
291
41. Suppose that the eigenvalues of A are ↵1 , ↵2 , . . . , ↵n (repetitions included) and for B are
Let the eigenvalues of A + B be 1 , 2 , . . . , n . Then by Exercise 40,
tr(A + B) =
1
+
2
+ ··· +
1,
2, . . . ,
n.
n
tr(A) + tr(B) = ↵1 + · · · + ↵n +
1
+ ··· +
n.
Since tr(A + B) = tr(A) + tr(B) by Exercise 44(a) in Section 3.2, we get
1
+
+ ··· +
2
2
= ↵1 + · · · + ↵n +
1
+ ··· +
n
as desired.
Let µ1 , µ2 , . . . , µn be the eigenvalues of AB. Then again by Exercise 40 we have
µ1 µ2 · · · µn = det(AB) = det(A) det(B) = ↵1 ↵2 · · · ↵n
1 2
···
n
as desired.
Pn
42. Suppose x = i=1 ci vi , where the vi are eigenvectors of A with corresponding eigenvalues i . Then
Theorem 4.18 tells us that vi are eigenvectors of Ak with corresponding eigenvalues ki , so that
k
A x=A
n
X
k
i=1
4.4
ci vi
!
=
n
X
c i Ak v i =
i=1
n
X
k
i vi .
ci
i=1
Similarity and Diagonalization
1. The two characteristic polynomials are
cA ( ) =
cB ( ) =
4
1
3
1
0
1
1
0
= (4
)(1
= (1
)2 =
)
2
2
3=
5 +1
2 + 1.
Since the two characteristic polynomials are unequal, the two matrices are not similar by Theorem
4.22.
2. The two characteristic polynomials are
cA ( ) =
cB ( ) =
2
1
4
6
5
7
3
1
= (2
)(6
)+4=
2
8 + 16
= (3
)(7
)
2
10 + 16.
5=
Since the two characteristic polynomials are unequal, the two matrices are not similar by Theorem
4.22.
3. Both A and B are triangular matrices, so by Theorem 4.15, A has eigenvalues 2 and 4, while B has
eigenvalues 1 and 4. By Theorem 4.22, the two matrices are not similar.
292
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
4. The two characteristic polynomials are
1
2
0
0
cA ( ) =
1
1
1
1
1
= (1
)
= (1
)((1
2
cB ( ) =
0
1
1
1
)2
1
0
2
0
1
2
1
2
)
= (1
)((2
0
0
1) = (1
1
1
)(
2
2 )=
3
+3
2
2
3
+4
1
0
1
= (1
2
1
)(1
)
2) = (1
)(
2
3 )=
2
3 .
Since the two characteristic polynomials are unequal, the two matrices are not similar by Theorem
4.22.
5. By Theorem 4.23, the columns of P are the eigenvectors of A corresponding to the eigenvalues, which
are the diagonal entries of D. Thus
✓ ◆
1
=
4
with
eigenspace
span
1
1
✓ ◆
1
span
.
2 = 3 with eigenspace
2
6. By Theorem 4.23, the columns of P are the eigenvectors of A corresponding to the eigenvalues, which
are the diagonal entries of D. Thus
02 31
3
span @415A
1 = 2 with eigenspace
2
02 31
1
span @4 15A
2 = 0 with eigenspace
0
02 31
0
1 with eigenspace span @4 15A .
3 =
1
7. By Theorem 4.23, the columns of P are the eigenvectors of A corresponding to the eigenvalues, which
are the diagonal entries of D. Thus
02 31
3
@
4
25A
=
6
with
eigenspace
span
1
3
02 3 2 31
0
1
@
4
5
4
1
05 A .
=
2
with
eigenspace
span
,
2
1
1
8. To determine whether A is diagonalizable, we first compute its eigenvalues:
det(A
I) =
5
2
2
5
= (5
)2
4=
2
10 + 21 = (
3)(
7).
4.4. SIMILARITY AND DIAGONALIZATION
293
Since A is a 2 ⇥ 2 matrix with 2 distinct eigenvalues, it is diagonalizable by Theorem 4.25. To find the
diagonalizing matrix P , we find the eigenvectors of A. To find the eigenspace corresponding to 1 = 3
we must find the null space of

2 2
A 3I =
.
2 2
Row reduce this matrix:
⇥
A
2
0 =
2
3I
Thus eigenvectors corresponding to

7I
Thus eigenvectors corresponding to
P =
"

!
1
0
1
0
0
.
0
= 7 we must find the null space of

2
2
A 7I =
.
2
2
⇤
0 =
1

Thus
0
0
2
Row reduce this matrix:
A
2
2
are of the form

✓
◆
t
1
1
=t
= span
.
t
1
1
1
To find the eigenspace corresponding to
⇥

⇤

2
2
2
2
0
0
!

1
0
1
0
0
.
0
are of the form

✓ ◆
t
1
1
=t
= span
.
t
1
1
#
1
1
1
1
"
1
2
1
2
,
so that
P
1
=
"
1
2
1
2
1
2
1
2
#
and so by Theorem 4.23,
1
P
AP =
1
2
1
2
#"
5
2
2
5
#"
#
1
1
1
1
=
"
#
0
3
0
7
.
9. To determine whether A is diagonalizable, we first compute its eigenvalues:
det(A
I) =
3
4
1
1
=( 3
)(1
)
4 · ( 1) =
2
+ 2 + 1 = ( + 1)2 .
So the only eigenvalue is = 1, with algebraic multiplicity 2. To see if A is diagonalizable, we must see
if its geometric multiplicity is also 2 by finding its eigenvectors. To find the eigenspace corresponding
to = 1, we must find the null space of

2 4
A+I =
1 2
by row-reducing it:
⇥
A+I
⇤
0 =

2
1
4
2
0
0
!

1
0
2
0
0
.
0
Thus eigenvectors corresponding to 1 are of the form


2t
2
, so a basis for the eigenspace is
.
t
1
Thus = 1 has geometric multiplicity 1, which is less than the algebraic multiplicity, so that A is
not diagonalizable by Theorem 4.27.
294
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
10. To determine whether A is diagonalizable, we first compute its eigenvalues:
3
det(A
1
0
0
I) =
0
1
3
0
)3
= (3
3
since the matrix is a triangular matrix. Thus the only eigenvalue is = 3, with algebraic multiplicity
3. To see if A is diagonalizable, we must compute the eigenspace
corresponding
to this eigenvalue. To
⇥
⇤
find the eigenspace corresponding to = 3 we row-reduce A 3I 0 :
⇥
A
3I
2
0
0 = 40
0
⇤
1
0
0
0
1
0
3
0
05 .
0
But this matrix is already row-reduced, and thus the eigenspace is
2 3
2 3
02 31
s
1
1
405 = s 405 = span @405A .
0
0
0
Thus the geometric multiplicity is only 1 while the algebraic multiplicity is 3. So by Theorem 4.27,
the matrix is not diagonalizable.
11. To determine whether A is diagonalizable, we first compute its eigenvalues:
1
det(A
I) =
0
0
1
1
1
1
=
3
+2
2
+
2=
( + 1)(
1)(
2).
1
Since A is a 3 ⇥ 3 matrix with 3 distinct eigenvalues, it is diagonalizable by Theorem 4.25. To find
the diagonalizing matrix P , we find the eigenvectors of A. To find the eigenspace corresponding to
= 1, we must find the null space of
2
3
2 0 1
A + I = 4 0 2 15
1 1 1
by row-reducing it:
h
A+I
Thus eigenvectors corresponding to
2
6
4
2
2
i
6
0 = 40
1
0
2
1
1
1
1
3
2
0
1
7
6
05 ! 4 0
0
0
are of the form
3
1
1
2t
1 7
2 t5
t
2
1
2
1
2
0
3
0
7
05
0
3
t
6 7
or, clearing fractions 4 t5 .
2t
3
1
Thus a basis for this eigenspace is 4 15.
2
To find the eigenspace corresponding to
0
1
0
2
= 1, we must find the null space of
2
3
0 0
1
15
A I = 40 0
1 1
1
4.4. SIMILARITY AND DIAGONALIZATION
by row-reducing it:
⇥
A
295
2
0
⇤
0 = 40
1
I
0
0
1
3
2
0
1
05 ! 40
0
0
1
1
1
1
0
0
0
1
0
3
0
05
0
Thus eigenvectors corresponding to 1 are of the form
2 3
2 3
t
1
4 t5 , so a basis for the eigenspace is 4 15 .
0
0
To find the eigenspace corresponding to
A
by row-reducing it:
⇥
A
= 2, we must find
2
1
0
1
2I = 4 0
1
1
2
1
0 =4 0
1
⇤
2I
0
1
1
1
1
2
the null space of
3
1
15
2
3
2
0
1
05 ! 40
0
0
0
1
0
1
1
0
Thus eigenvectors corresponding to 1 are of the form
23
2 3
t
1
4t5 , so a basis for the eigenspace is 415 .
t
1
Then a diagonalizing matrix P is the
2
1
6
P =4 1
2
and so by Theorem 4.23,
P
1
2
1
6
1
2
1
3
6
AP = 4
1
6
1
2
1
3
3
0
05
0
matrix whose columns are the eigenvectors of A:
3
2
3
1
1
1
1 1
6
6
3
7
6
7
1
1 15 , so that P 1 = 4 12
05
2
1
1
1
0 1
3
3
3
1
3
32
1
76
05 40
1
1
3
0
1
1
32
1
1
76
15 4 1
0
2
3 2
1 1
1
7 6
1 15 = 4 0
0 1
0
12. To determine whether A is diagonalizable, we first compute its eigenvalues:
1
det(A
I) =
0
2
3
0
1
2
0
1
= (1
)
2
1
0
1
So the eigenvalues are 1 = 1 with algebraic multiplicity 2, and
⇥ 2=
To find the eigenspace corresponding to 1 = 1 we row-reduce A I
3
2
2
1 0 0 0
1 0 0
⇥
⇤
A I 0 = 42 2 1 05 ! 40 1 1
3 0 1 0
0 0 0
Thus the eigenspace is
2
3
2 3
02 31
0
0
0
4 t5 = t 4 15 = span @4 15A .
t
1
1
= (1
0
1
0
3
0
7
05 .
2
)2 (2
).
2 with
algebraic multiplicity 1.
⇤
0 :
3
0
05 .
0
Thus the geometric multiplicity of this eigenvalue is only 1 while the algebraic multiplicity is 2. So by
Theorem 4.27, the matrix is not diagonalizable.
296
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
13. To determine whether A is diagonalizable, we first compute its eigenvalues:
1
2
1
1
3
=
So A has eigenvalues 0 and 1. Considering = 0, we row-reduce
2
3
2
1 2 1 0
1 0
1
4 1 0 1 05 ! 40 1
1
1 1 0 0
0 0
0
A
det(A
I) =
1
1
so that the eigenspace is
Since
=
2
2
(
1).
1
02
31
1
span @4 15A .
1
3
0
05 ,
0
0I = A:
= 0 has algebraic multiplicity 2 but geometric multiplicity 1, the matrix is not diagonalizable.
14. To determine whether A is diagonalizable, we first compute its eigenvalues:
2
det(A
I) =
0
0
0
0
0
2
3
0
0
2
1
0
3
0
= (1
)(2
)(3
)2
1
since the matrix is upper triangular. So the eigenvalues are 1 = 1 with algebraic multiplicity 1,
2 = 2 with algebraic multiplicity 1, and
⇥ 3 = 3 with
⇤ algebraic multiplicity 2. To find the eigenspace
corresponding to 3 = 3 we row-reduce A 3I 0 :
2
3
2
3
1 0 0
2 0
1 0 0 0 0
6
7
⇥
⇤ 6 0 0 2
1 07
7 ! 60 0 1 0 07 .
A 3I 0 = 6
4 0 0 0
5
4
0 0
0 0 0 1 05
0 0 0
2 0
0 0 0 0 0
Thus the eigenspace is
2 3
2 3
02 31
0
0
0
6t7
617
B617C
6 7 = t 6 7 = span B6 7C .
405
405
@405A
0
0
0
Thus the geometric multiplicity of this eigenvalue is only 1 while the algebraic multiplicity is 2. So by
Theorem 4.27, the matrix is not diagonalizable.
15. Since A is triangular, we can read o↵ its eigenvalues, which are 1 = 2 and 2 = 2, each of which
has algebraic multiplicity 2. Computing eigenspaces, we have
8 2 3 2 39
2
3
2
3
0 0
0
4 0
0 0 1 0 0
1
0 >
>
>
< 6 7 6 7>
=
7
60 0 0 1 07
⇥
⇤ 60 0
0
0
0
0
7!6
7 , so E2 has basis 6 7 , 617
A 2I 0 = 6
40 0
5
4
5
4
5
4
5
4
0 0
0 0 0 0 0
0
0 >
>
>
>
:
;
0 0
0
4 0
0 0 0 0 0
0
0
8 2 3 2 39
2
3
2
3
4 0 0 4 0
1 0 0 1 0
1
0 >
>
>
< 6 7 6 7>
=
60 1 0 0 07
⇥
⇤ 6 0 4 0 0 07
0
7
6
7
6 7 6 07
A + 2I 0 6
40 0 0 0 05 ! 40 0 0 0 05 , so E 2 has basis >4 05 , 415> .
>
>
:
;
0 0 0 0 0
0 0 0 0 0
1
0
Since each eigenvalue also has geometric
2
1 0
60 1
P =6
40 0
0 0
multiplicity 2, the matrix is diagonalizable, and
3
2
3
1 0
2 0
0
0
6
0 07
0
07
7 , D = 60 2
7
5
4
0 1
0 0
2
05
1 0
0 0
0
2
4.4. SIMILARITY AND DIAGONALIZATION
provides a solution to
2
1
6
0
P 1 AP = 6
40
0
0
1
0
0
32
1
2
60
07
76
15 40
0
0
0
0
0
1
0
2
0
0
297
32
4 1
6
07
7 60
05 40
2 0
0
0
2
0
3 2
1 0
2
60
0 07
7=6
0 15 40
1 0
0
0
1
0
0
16. As in Example 4.29, we must find a matrix P such that

4
1
1
P AP = P
3
0
2
0
0
3
0
07
7 = D.
05
2
0
0
2
0
6
P
5
is diagonal. To do this, we find the eigenvalues and a basis for the eigenspaces of A. The eigenvalues
of A are the roots of its characteristic polynomial
det(A
I) =
4
6
3
=( 4
5
)(5
) + 18 =
2
This polynomial factors as (
2)( + 1), so the eigenvalues are 1 = 1 and
associated with 1 can be found by finding the null space of A + I:


⇥
⇤
3 6 0
1
2 0
A+I 0 =
!
3 6 0
0
0 0
Thus a basis for this eigenspace is given by
The eigenspace associated with
⇥
A
2
.
1

1
.
1
can be found by finding the null space of A


⇤
6 6 0
1
1 0
2I 0 =
!
.
3 3 0
0
0 0
2
Thus a basis for this eigenspace is given by
Set


1
P =
1
2
,
1
1
so that P
=

1
1
2
1

2.
2
= 2. The eigenspace
2I:
2
1
Then
D=P
1
AP =

2
0
0
, so that A5 = P D5 P
1
1
=

1
1
32
0
0
1

1
1
2
=
1

34
33
66
65
17. As in Example 4.29, we must first diagonalize A. To do this, we find the eigenvalues and a basis for
the eigenspaces of A. The eigenvalues of A are the roots of its characteristic polynomial
det(A
so the
⇥ eigenvalues
⇤ are
and A 2I 0 :
⇥
1
I) =
⇥
A
1
A + 3I
2I
6
=( 1
1
=
3 and

⇤
2
0 =
1

⇤
3
0 =
1
2
6
3
)(
)
6=
2
+
6 = ( + 3)(
2),
⇥
= 2. To compute the eigenspaces, we row-reduce A + 3I

⇢
0
1 3 0
3
!
, so E 3 has basis
0
0 0 0
1

⇢
6 0
1
2 0
2
!
so E2 has basis
2 0
0
0 0
1
.
⇤
0
298
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
Thus
P =
satisfy P
1
"
#
2
1
3
1
P
1
=
"
1
5
1
5
2
5
3
5
#
D=
"
3
0
0
2
#
AP = D, so that
"
#10
6
= (P DP 1 )10
0
"
#"
3 2
=
1 1
"
#"
3 2 (
=
1 1
"
35,839
=
11,605
1
1
= P D10 P =1
3
0
0
2
#10 "
3)10
0
1
5
1
5
0
210
#
69,630
.
24,234
#10 "
2
5
3
5
#
1
5
1
5
2
5
3
5
#
18. As in Example 4.29, we must first diagonalize A. To do this, we find the eigenvalues and a basis for
the eigenspaces of A. The eigenvalues of A are the roots of its characteristic polynomial
det(A
so
are
⇥ the eigenvalues
⇤
A 5I 0 :
⇥
⇥
4
I) =
A
⇤
0 =
⇤
0 =
5I
Set
P =
"


= (4
2
= 1 and
1
I
A
3
1
2
3
1
1
1
3
3
#
3
1
1
,
A
6
= PD
P
1

0
1
!
0
0

0
1
!
0
0
1
0
3
0
1
so that P
Then
6
)
2
3=
6 +5=(
5)(
1),
⇥
= 5. To compute the eigenspaces, we row-reduce A
3
1
1
)(2
=
"
#"
3 1
1
1
1
0
=
⇢
0
1
, so that E1 has basis
0
1
⇢
0
3
, so that E5 has basis
0
1
"
0
6
5
1
4
1
4
#"
3
4
1
4,
#
1
4
1
4
,
3
4
1
4
D=
#
=
"
"
1
0
3907
56
3906
56
#
0
5
I
⇤
0 and
,
.
.
11718
56
11719
56
#
.
19. As in Example 4.29, we must first diagonalize A. To do this, we find the eigenvalues and a basis for
the eigenspaces of A. The eigenvalues of A are the roots of its characteristic polynomial
det(A
I) =
so the
⇥ eigenvalues
⇤ are
and A + I 0 :
⇥
A
3I
⇥
A+I
1
3
1
= 3 and
⇤
0 =

3
1

⇤
1
0 =
1
=(
2
2
3
1
3
3
=
)(2
)
3=
2
2
3=(
3)( + 1),
⇥
1. To compute the eigenspaces, we row-reduce A

⇢
0
1
1 0
1
!
, so that E3 has basis
0
0 0 0
1

⇢
0
1 3 0
3
!
, so that E 1 has basis
0
0 0 0
1
,
.
3I
⇤
0
4.4. SIMILARITY AND DIAGONALIZATION
Set
P =
"
#
1
3
1
1
,
so that P
Then
k
k
1
A = PD P
"
=
299
=
3k +3( 1)k
4
3k ( 1)k
4
"
1
1
1
=
"
1
4
1
4
3
4
1
4,
#"
3 3k
1
0
3k+1 3( 1)k
4
3k+1 +( 1)k
4
#
,
D=
0
#
( 1)k
#"
1
4
1
4
"
#
0
3
0
3
4
1
4
1
.
#
.
20. As in Example 4.29, we must first diagonalize A. To do this, we find the eigenvalues and a basis for
the eigenspaces of A. The eigenvalues of A are the roots of its characteristic polynomial
2
det(A
so
are
⇥ the eigenvalues
⇤
A 5I 0 :
h
⇥
A
A
= 0 and
2
i 62
0 =6
42
2
2
2
1
1
=5
Set
1
2
2
1
6
P =6
4 2
0
Then
so that
1
2
1
0
1
1
3
7
17
5,
1
6
=6
4 2
0
32
1 0
76
6
0 17
5 40
1 1 0
1
3
=
2
(
5),
⇥
= 5. To compute the eigenspaces we row-reduce A
so that P
2
0
D = P 1 AP = 40
0
1
2
2
3
2
3
0
1 12 1 0
7
6
7
6
7
1 2 07
5 ! 40 0 0 05 , so
2 1 2 0
0 0 0 0
2
3
2
3
1
2 0
1 0
1
⇤
0 =4 2
4
2 05 ! 4 0 1
1
2
1
3 0
0 0
0
5I
A8 = P D 8 P
1
1
2
2
I) =
0
0
0
32
0
76
6
07
54
58
1
5
2
5
2
5
1
8 2 3 2 39
>
1 >
> 1
>
<
6 7 6 7=
7 , 6 07
E0 has basis 6
2
4 5 4 5>
>
>
>
: 0
1 ;
82 39
3
0
< 1 =
05 , so E5 has basis 415 .
:
;
0
1
2
6
=6
4
0
0
0
2
5
1
5
1
5
0I
1
5
2
5
2
5
3
0
05 ,
5
3
1
57
37
55
2
5
2
5
1
5
1
5
2
⇤
0 and
3
1
57
37 .
55
2
5
156250
6
=6
4156250
156250
78125
78125
78125
156250
3
7
1562507
5.
156250
21. As in Example 4.29, we must first diagonalize A. To do this, we find the eigenvalues and a basis for
the eigenspaces of A. Since A is upper triangular, its eigenvalues ⇥are the diagonal
are
⇤
⇥ entries, which
⇤
A
+
I
0
A
I
0
=
1
and
=
1.
To
compute
the
eigenspaces,
we
row-reduce
and
:
1
2
8 2 3 2 39
2
3
2
3
>
2 1 1 0
1 12 12 0
1 >
< 1
=
h
i
6
7
6
7
6 7 6 7
A + I 0 = 40 0 0 05 ! 40 0 0 05 , so E 1 has basis 4 25 , 4 05 ,
>
>
:
;
0 0 0 0
0 0 0 0
0
2
8 2 39
2
3
2
3
0
1
1 0
0 1 0 0
< 1 =
⇥
⇤
A I 0 = 40
2
0 05 ! 40 0 1 05 , so E1 has basis 405 .
:
;
0
0
2 0
0 0 0 0
0
300
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
Set
2
3
1
7
2 07
5,
0 0
1
1
6
P =6
4 0
2
Then
1
so that P
2
A2015 = P D2015 P
1
2
1
2
2
0
1
6
6
=4 0
2
2
1
6
6
= 40
1
2
2
0
1
1
0
0
1
1
2
1
2
6
=6
40
3
1
27
07
5.
1
2
32
32
1 ( 1)2015
0
0
0
76
76
2015
7
6
7
6
05 4
0
( 1)
0 5 40
2015
0
0
0
1
1
3
32
32
1
1
0 0 0 0 12
76
76
7
1
6
6
7
07
1 07
54 0
5 40 2 05
0
0
0 1 1 12 12
3
1
7
07
5 = A.
1
6
=6
4 0
0
3
0 0
1 05 ,
0 1
1
D = P 1 AP = 4 0
0
so that
2
0
0
1
2
1
2
3
1
27
07
5
1
2
1
22. As in Example 4.29, we must first diagonalize A. To do this, we find the eigenvalues and a basis for
the eigenspaces of A. The eigenvalues of A are the roots of its characteristic polynomial
2
det(A
⇥
Set
A
A
I
3I
= (1
)((2
1
6
P =6
41
1
2
1
2
1
)2
= 1 and
2
1 0
⇤
0 = 41 0
1 0
2
1
⇤
0 =4 1
1
2
0
2
)
1
1
1
1
= (1
Thus⇥ the eigenvalues
are
⇤
and A 3I 0 .
⇥
0
1
1
I) =
2
1) = (1
)(
2
4 + 3) =
1)2 (
(
3).
⇥
= 3. To compute the eigenspaces, we row-reduce A
82 3 2 39
3
2
3
0
1 0 1 0
0 =
< 1
05 ! 40 0 0 05 , so E1 has basis 4 05 , 415
:
;
0
0 0 0 0
1
0
82 39
3
2
3
0
1 0
1 0
1 0
< 1 =
2
1 05 ! 4 0 1
1 05 , so E3 has basis 415 .
:
;
0
1 0
0 0
0 0
1
1
1
1
3
0
7
0 17
5,
1 0
1
so that P
1
2
6
=6
4
1
2
1
2
1
2
0
0
1
3
1
27
17 ,
25
1
2
2
3
6
D=6
40
0
0
1
0
3
0
7
07
5,
1
I
⇤
0
4.4. SIMILARITY AND DIAGONALIZATION
301
so that
Ak = P D k P
1
2
1
6
=6
41
2
0
1
1
1
1
6
=6
41
2
32
0 3k
76
6
17
54 0
0
0
32
0 3k
76
6
17
54 0
0
0
1
0
1
1
1 k
(3 + 1)
62
1
6
= 4 2 (3k 1)
1 k
1)
2 (3
0
1
0
0
0
32
1
2
1
2
1
2
76
6
07
54
1k
k
0
1
32
0 0
76
6
1 07
54
0 1
3
1 k
1)
2 (3
7
1 k
1)7
5
2 (3
1 k
(3
+
1)
2
1
2
1
2
1
2
3
1
27
17
25
1
2
0
0
1
3
1
27
17
25
1
2
0
0
1
23. As in Example 4.29, we must first diagonalize A. To do this, we find the eigenvalues and a basis for
the eigenspaces of A. The eigenvalues of A are the roots of its characteristic polynomial
1
det(A
I) =
1
2
0
1
2
2
1
Thus
the eigenvalues
are⇤ 1 =⇥1, 2 = 2, ⇤and 2 =
⇥
⇤ ⇥
A I 0 , A 2I 0 , and A + 3I 0 .
2
3
2
0
1 0 0
1 0
⇥
⇤
A I 0 = 42
3 2 05 ! 40 1
0
1 0 0
0 0
2
3
2
1
1
0 0
1
⇥
⇤
A 2I 0 = 4 2
4
2 05 ! 4 0
0
1
1 0
0
2
3
2
4 1 0 0
1 0
⇥
⇤
A + 3I 0 = 42 1 2 05 ! 40 1
0 1 4 0
0 0
Set
2
so that
1
6
P =4 0
1
Ak = P D k P
1
1
1
1
2
3
1
7
45 ,
1
1
6
=4 0
1
2
1
1
1
1
10 (5
1
so that P
32
1 1
76
45 4 0
1 0
=
(
1)(
2)( + 3).
1
0
2k
0
2
6
=4
3. To compute the eigenspaces, we row-reduce
82
3
0
<
05 , so E1 has basis 4
:
0
3
0
1 0
1
1 05 , so E2 has basis
0
0 0
82
3
1 0
<
4 05 , so E3 has basis 4
:
0 0
1
0
0
1
2
2
5
1
10
0
1
5
1
5
32
0
76
05 4
( 3)k
1 k
5 (2
1
5 (4(
1 k
5 (2
+ ( 3)k + 2k+2 )
6
2
k
=4
2k )
5 (( 3)
1
k
k+2
)
10 ( 5 + ( 3) + 2
1
2
2
5
1
10
3
1
2
27
,
55
1
10
0
1
5
1
5
( 3)k )
3)k + 2k )
( 3)k )
2
1
6
D = 40
0
I) =
1
1
0
k
=(
1)(
3
0
7
05 ,
3
3
1
2
27
55
1
10
3
5 + ( 3)k + 2k+2 )
7
2
k
2k )
5.
5 (( 3)
1
k
k+2
)
10 (5 + ( 3) + 2
1
10 (
24. The characteristic polynomial of A is
det(A
0
2
0
39
1 =
05
;
1
82 39
< 1 =
415
:
;
1
9
3
1 =
45 .
;
1
k),
302
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
so that if k 6= 1, then A is a 2 ⇥ 2 matrix with two distinct eigenvalues, so it is diagonalizable. If k = 1,
then

0 1
A I=
,
0 0

1
which gives a one-dimensional eigenspace with basis
. Since = 1 has algebraic multiplicity 2 but
0
geometric multiplicity 1, A is not diagonalizable. So A is diagonalizable exactly when k 6= 1.
25. The characteristic polynomial of A is
det(A
I) =
1
k
0
The only eigenvalue is 1, and then
A
I=

0
0
1)2 .
=(
1
k
.
0
If k = 0, then
 A is diagonal to start with, while if k 6= 0, then we get a one-dimensional eigenspace
1
with basis
. Since = 1 then has algebraic multiplicity 2 but geometric multiplicity 1, A is not
0
diagonalizable. So A is diagonalizable exactly when k = 0.
26. The characteristic polynomial of A is
det(A
I) =
k
1
1
p
= (
k)
2
1=
k
1,
2
so that A has the eigenvalues
= k± 2k +4 . Since k 2 + 4 is never zero, and always positive, the
eigenvalues of A are real and distinct, so A is diagonalizable for all k.
27. Since A is triangular, the only eigenvalue is 1, with algebraic multiplicity 3. Then
2
3
0 0 k
A I = 40 0 0 5 .
0 0 0
If k = 0, then
to start with, while if k 6= 0, then we get a one-dimensional eigenspace
82 A
3 is2diagonal
39
0 =
< 1
with basis 405 , 415 . Since = 1 then has algebraic multiplicity 3 but geometric multiplicity 2,
:
;
0
0
A is not diagonalizable. So A is diagonalizable exactly when k = 0.
28. Since A is triangular, the eigenvalues are 1, with
plicity 1. Then
2
0 k
A I = 40 1
0 0
algebraic multiplicity 2, and 2, with algebraic multi3
2
0
0
05 ! 4 0
0
0
1
0
0
3
0
05
0
82 3 2 39
0 =
< 1
regardless of the value of k, so = 1 has a two-dimensional eigenspace with basis 405 , 405 , i.e.,
:
;
0
1
with geometric multiplicity 2. Since this is equal to the algebraic multiplicity of , we see that A is
diagonalizable for all k.
29. The characteristic polynomial of A is
1
det(A
I) =
1
1
1
k
k
1
1
k
=
3
+2
2
+k
2
=
2
(
(k + 2)).
4.4. SIMILARITY AND DIAGONALIZATION
If k =
2, then A has only the eigenvalue
2
1
A = 41
1
303
0, with algebraic multiplicity 3, and then
3
2
3
1
2
1 1
2
1
25 ! 4 0 0
05 ,
1
2
0 0
0
so the eigenspace is two-dimensional and therefore the geometric multiplicity is not equal to the algebraic multiplicity, so that A is not diagonalizable. If k 6= 2, then 0 has algebraic multiplicity 2, and
then
2
3
2
3
1 1
k
1 1
k
k 5 ! 40 0
05 ,
A = 41 1
1 1
k
0 0
0
and the eigenspace is again two-dimensional, so the geometric multiplicity equals the algebraic multiplicity and A is diagonalizable. Thus A is diagonalizable exactly when k 6= 2.
30. We are given that A ⇠ B, so that B = P 1 AP for some invertible matrix P , and that B ⇠ C, so that
C = Q 1 BQ for some invertible matrix Q. But then
1
C=Q
1
so that C is R
1
BQ = Q
P
1
1
AP Q = (P Q)
AP Q,
AR for the invertible matrix R = P Q.
31. A is invertible if and only if det A 6= 0. But since A and B are similar, det A = det B by part (a) of
this theorem. Thus A is invertible if and only if det B 6= 0, which happens if and only if B is invertible.
32. By Exercise 61 in Section 3.5, if U is invertible then rank A = rank(U A) = rank(AU ). Now, since A
and B are similar, we have B = P 1 AP for some invertible matrix P , so that
1
rank B = rank(P
(AP )) = rank(AP ) = rank A,
using the result above twice.
33. Suppose that B = P 1 AP where P is invertible. This can be rewritten as BP
an eigenvalue of A with eigenvector x. Then
B(P
1
x) = (BP
1
1
)x = (P
A)x = P
1
1
(Ax) = P
1
=P
1
( x) = (P
1
A. Let
be
x).
Thus P 1 x is an eigenvector for B, corresponding to . So every eigenvalue of A is also an eigenvalue
of B. By writing the similarity equation as A = Q 1 BQ, we see in exactly the same way that every
eigenvalue of B is an eigenvalue of A. So A and B have the same eigenvalues. (Note that we have also
proven a relationship between the eigenvectors corresponding to a given eigenvalue).
34. Suppose that A ⇠ B, say A = P
m > 0, then
Am = P
But A
m
=P
1
1
BP
m
m
1
=P
|
B P means that A
35. From part (f), Am ⇠ B m for m
BP . Then if m = 0, we have Am = B m = I, so that A0 ⇠ B 0 . If
m
1
BP · P
1
1
BP ·{z
··P
m times
m
⇠B .
BP · P
1
m
1
= P
1
B
m
1
P
1
since A m and B m are similar because m > 0. But P 1 B m P
=P
P 1 B m P . Thus Am ⇠ B m for m < 0, so they are similar for all integers m.
1
. Then
BA = BABB
showing that AB ⇠ BA.
B m P.
0. If m < 0, then
Am = A
36. Let P = B
1
BP} = P
1
= B(AB)B
1
=P
1
(AB)P,
1
(B
m
)
1
P
1
1
=
304
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
37. Exercise 45 in Section 3.2 shows that if A and B are n ⇥ n, then tr(AB) = tr(BA). But then if
B = P 1 AP , we have
1
tr B = tr(P
AP ) = tr((P
38. A is triangular, so it has eigenvalues
det(B
1
I) =
1
)2
= (1
1
So the eigenvalues of B are also
1
A)) = tr((P P
2
4=
1 and 3, and thus

1
A⇠B⇠
0
2
3 = ( + 1)(
1
4
◆
3I shows that E 1 = span
and E3 = span
✓
◆
✓ ◆
1
1
3I shows that E 1 = span
and E3 = span
. Thus
1
1
D=Q
1
D=R
1
"
#"
3
0
#"
1
1
2
1
2
2
1
4
1
4
0
AQ =
1
"
1
2
1
2
BR =
3).
0
= D.
3
✓
Reducing A + I and A
and B
)A) = tr A.
1 and 3. For B, we have
2
2
1
A)P ) = tr(P (P
✓
1
0
◆
. Reducing B + I
#"
#
1
1 1
1
4 0
#"
#
2
1 1
,
1
1 1
so that finally
B = RDR
1
= RQ
1
AQR
Thus
P = QR
is an invertible matrix such that B = P
1
1
=
1
= (QR

1
2
1
1
)
A(QR
1
).
0
2
AP .
39. We have
det(A
I) =
det(B
I) =
5
3
4
= (5
2
1
1
6
)( 2
=( 1
4
)(4
So A and B both have eigenvalues 1 and 2, so that

1
A⇠B⇠
0
)+6=
2
3 +2=(
3 +2=(
1)(
1)(
2)
2).
0
= D.
2
✓ ◆
✓ ◆
3
1
2I shows that E1 = span
and E2 = span
. Reducing B
4
1
✓ ◆
✓ ◆
1
1
2I shows that E1 = span
and E3 = span
. Thus
2
3
Reducing A
B
2
) + 12 =
I and A
D=Q
1
AQ =
D=R
1
BR =


1
4
1
3
3
2
1
1


5
4
3
2
1
6
1
4


3
4
1
1
1
2
1
,
3
I and
4.4. SIMILARITY AND DIAGONALIZATION
305
so that finally
B = RDR
1
= RQ
1
AQR
1
1

Thus
P = QR
1
is an invertible matrix such that B = P
3
I) =
So B also has eigenvalues
Reducing A + 2I, A
E
2
Reducing B + 2I, B
E
Therefore
2
2
A⇠B⇠4 0
0
( + 2)(
2
B = RDR
1
32
1 0
1
76
2 15 4 4
0 1
0
32
32
3
3 2
5 1
2
76
76
15 40
15 41 2
1
2 2
4 1
2
0
1
2
1
2
1
2
1
2
= RQ
1
0
AQR
Thus
1
= (QR
2
1
=4 1
3
1
P = QR
1
02 3 1
1
E2 = span @425A .
1
32
1
4
is an invertible matrix such that B = P
2),
02 31
1
E2 = span @405A .
0
1
2
12
17 6
0
5
4
3
1
0
4
1
4
6
D = R 1 BR = 4 1
1)(
3
0
05 = D.
2
0
1
0
I, and B 2I gives
02 3 1
02 31
1
1
= span @405A , E1 = span @4 15A ,
1
0
0
6
D = Q 1 AQ = 40
1
2
so that finally
=
4
2I gives
31
02 3 1
1
1
= span @4 45A , E1 = span @4 15A ,
0
3
2
).
2
3
5
1
2
2, 1, and 2, so that
2
I, and A
02
1
A(QR
2, 1, and 2. For B, we have
2
1
2
7
10
1
)
AP .
40. A is upper triangular, so it has eigenvalues
det(B
=
1
= (QR
AP .
1
)
1
3
1 1
7
1 05
3 0
3
1 1
7
1 25 ,
0 1
A(QR
1
).
3
0
55
3
0
2
0
41. We have
1
det(A
I) =
0
1
2
0
3
det(B
I) =
2
1
1
3)
1
2
6
4
( + 1)2 (
=
0
0
5
4
1
=
( + 1)2 (
3),
306
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
so that A and B both have eigenvalues
Reducing A + I and A
3I gives
E
Reducing B + I and B
1
1 (of algebraic
2
1
0
1
A⇠B⇠4 0
0
0
02
3 2 31
0
1
= span @4 05 , 415A ,
1
0
02 31
2
E3 = span @415A .
2
3I gives
E
1
Therefore
D=Q
D=R
so that finally
3 2 31
1
0
= span @4 15 , 405A ,
0
1
1
2
1
4
1
4
6
AQ = 4
1
1
02
2
2
B = RDR
1
1
2
1
2
1
2
32
0
3
76
15 4 6
0
4
1
1
= RQ
AQR
Thus
P = QR
is an invertible matrix such that B = P
1
31
1
E3 = span @4 35A .
2
32
3
2
6
BR = 4 1
02
1
1
2
17 6
4 5 41
1
2
4
0
1
0
1
AP .
2
6
=4
multiplicity 2) and 3. Therefore
3
0
05 = D.
3
1
32
0 2
1
76
1 15 4 0
0 1
1
32
2
0
76
5
05 4
4
1
= (QR
1
2
3
2
5
2
1
1
)
0
1
0
1
1
0
A(QR
3
2
7
15
2
0
0
1
1
3
1
7
35 ,
2
).
3
0
7
15
0
1
2
3
2
3
2
42. Suppose that B = P 1 AP . This gives P B = AP , so that (P B)T = (AP )T . Expanding yields
B T P T = P T AT , so that B T ⇠ AT and thus AT ⇠ B T .
43. Suppose that P
1
AP = D, where D is diagonal. Then
D = DT = (P
Now let Q = (P T )
1
1
AP )T = P T AT (P T )
; the equation above becomes D = Q
1
1
.
AT Q, so that AT is diagonalizable.
44. If A is invertible, it has no zero eigenvalues. So if it is diagonalizable, the diagonal matrix D is also
invertible since none of the diagonal entries are zero. Then D 1 is also diagonal, consisting of the
inverses of each of the diagonal entries of D. Suppose that D = P 1 AP . Then
D
1
= P
1
AP
1
=P
1
A
1
P
1
1
=P
1
A
1
P,
showing that A 1 is also diagonalizable (and, incidentally, showing again that the eigenvalues of A
are the inverses of the eigenvalues of A, by looking at the entries of D 1 .)
1
45. Suppose that A is diagonalizable, and that is its only eigenvalue. Then if D = P 1 AP is diagonal,
all of its diagonal entries must equal , so that D = I. Then I = P 1 AP implies that P ( I)P 1 =
P P 1 AP P 1 = A. However, P ( I)P 1 = P IP 1 = I. Thus A = I.
4.4. SIMILARITY AND DIAGONALIZATION
307
46. Since A and B each have n distinct eigenvalues, they are both diagonalizable by Theorem 4.25. Recall
that diagonal matrices commute: if D and E are both diagonal, then DE = ED. Now, if A and B have
the same eigenvectors, then since the diagonalizing matrix P has columns equal to the eigenvectors,
we see that for the same invertible matrix P , both DA = P 1 AP and DB = P 1 BP are diagonal.
But then
AB = (P DA P
1
)(P DB P
1
) = P DA DB P
1
1
= P DB DA P
= (P DB P
1
)(P DA P
1
) = BA.
Conversely, if AB = BA, let ↵1 , ↵2 , . . . , ↵n be the distinct eigenvalues of A, and pi the corresponding
eigenvectors, which are also distinct since they are linearly independent. Then
A(Bpi ) = (AB)pi = (BA)pi = B(Api ) = B(↵i Pi ) = ↵i Bpi .
Thus Bpi is also an eigenvector of A corresponding to ↵i , so it must be a multiple of pi . That is,
Bpi = i pi for all i, so that the eigenvectors of B are also pi . So every eigenvector of A is an
eigenvector of B. Since B also has n distinct eigenvectors, the set of eigenvectors must be the same.
47. If A ⇠ B, then A and B have the same characteristic polynomial, so the algebraic multiplicities of
their eigenvalues are the same.
48. From the solution to Exercise 33, we know that if B = P 1 AP and x is an eigenvector of A with
eigenvalue , then P 1 x is an eigenvalue of B with eigenvalue . Therefore, if is an eigenvalue of A
with multiplicity m, its eigenspace has a basis consisting of m distinct vectors v1 , v2 , . . . , vm , so that
the vectors P 1 v1 , P 1 v2 , . . . , P 1 vm are in the eigenspace of B corresponding to . These vectors
are linearly independent since P 1 is an invertible linear transformation. So each of the n eigenvectors
of A corresponds to a distinct eigenvector of B; since each matrix has n eigenvectors, the reverse is true
as well. Therefore the m eigenvalues of B found above are all of the eigenvalues of B corresponding to
, so that the dimension of the eigenspace of for B is m as well. So the geometric multiplicities of
the eigenvalues of A and B are the same.
49. If D = P 1 AP and D is diagonal, then every entry of D is either 0 or 1, since every eigenvalue of A
is either 0 or 1 by assumption. Since D2 is computed by squaring each diagonal entry, we see that
D2 = D. But then P DP 1 = A, and
A2 = (P DP
1 2
) = (P DP
1
)(P DP
1
1
)=P
D2 P = P
1
DP = A,
so that A is idempotent.
50. Suppose that A is nilpotent and diagonalizable, so that D = P
Am = O for some m > 1. Then
Dm = (P
1
AP )m = P
1
Am P = P
1
1
AP (so that A = P DP
1
) and
OP = O.
Since D is diagonal, Dm is found by taking the mth power of each diagonal entry. If this is the zero
matrix, then all the diagonal entries must be zero, so that D = O. But then A = P OP 1 = O, so that
A is the zero matrix.
51. (a) Suppose we have three vectors v1 , v2 , and v3 with
Av1 = v1 ,
Av2 = v2 ,
Av3 = v3 .
Then all three vectors must lie in the eigenspace E1 corresponding to = 1, since all three vectors
are multiplied by 1 when A is applied. However, since the algebraic multiplicity of = 1 is 2,
the geometric multiplicity is no greater than 2, by Lemma 4.26. But any three vectors in a twodimensional space must be linearly independent, so that v1 , v2 , and v3 are linearly dependent.
(b) By Theorem 4.27, A is diagonalizable if and only if the algebraic multiplicity of each eigenvalue
equals its geometric multiplicity. So the dimensions of the eigenspaces must be dim E 1 = 1,
dim E1 = 2, dim E2 = 3, since the algebraic multiplicity of 1 is 1, that of 1 is 2, and that of 2 is
3.
308
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
52. (a) The characteristic polynomial of A is
det(A
I) =
a
b
c
d
=
Therefore the eigenvalues are
p
a + d ± (a + d)2
=
2
2
(a + d) + (ad
4(ad
bc)
=
bc) =
a+d±
p
(a
2
2
tr(A) + det(A).
d)2 + 4bc
.
So if (a d)2 + 4bc > 0, then the two eigenvalues of A are real and distinct and therefore A is
diagonalizable. If (a d)2 + 4bc < 0, then A has no real eigenvalues and so is not diagonalizable.
(b) If (a d)2 + 4bc = 0, then A has one real eigenvalue, and A may or may not be diagonalizable
depending on the geometric multiplicity of that eigenvalue. For example, the zero matrix O has
(a d)2 + 4bc = 0, and it is already a diagonal matrix, so is diagonalizable. However,

1
1
A=
1
1
d)2 + 4bc = 0; from the formula above,
✓ ◆ its only real eigenvalue is = 0, and it is
1
easy to see by row-reducing A that E0 = span
. So the geometric multiplicity of = 0 is
1
only 1 and thus A is not diagonalizable, by Theorem 4.27.
satisfies (a
4.5
Iterative Methods for Computing Eigenvalues
1. (a) For the given value of x5 , we estimate a dominant eigenvector of A to be
"
# "
#
1
1
.
11,109 ⇡
2.500
4443
The corresponding dominant eigenvalue is computed by finding
"
#"
# "
#
1 2
4443
26,661
x6 = Ax5 =
=
,
5 4 11,109
66,651
and
26,661
4443
(b) det(A
⇡ 6.001,
I) =
66,651
11,109
1
⇡ 6.000.
2
5
4
= (1
)(4
)
10 =
2
5
6 = ( + 1)(
eigenvalue is 6.
2. (a) For the given value of x5 , we estimate a dominant eigenvector of A to be
"
# "
#
1
1
⇡
.
3904
0.500
7811
The corresponding dominant eigenvalue is computed by finding
"
#"
# "
#
7
4
7811
39,061
x6 = Ax5 =
=
,
3
1
3904
19,529
and
39,061
7811
⇡ 5.001,
19,529
3904
⇡ 5.002.
6), so the dominant
4.5. ITERATIVE METHODS FOR COMPUTING EIGENVALUES
(b) det(A
I) =
7
4
3
1
dominant eigenvalue is 5.
= (7
)( 1
) + 12 =
309
2
6 +5 = (
1)(
5), so the
3. (a) For the given value of x5 , we estimate a dominant eigenvector of A to be
" # "
#
1
1
⇡
.
89
0.618
144
The corresponding dominant eigenvalue is computed by finding
"
#"
# "
#
2 1 144
377
x6 = Ax5 =
=
,
1 1
89
233
and
377
144
(b) det(A
⇡ 2.618,
I) =
233
89
2
⇡ 2.618.
1
= (2
)(1
) 1=
1
1
p
So the dominant eigenvalue is 12 (3 + 5) ⇡ 2.618.
2
3 + 1, which has roots
= 12 (3 ±
p
5).
4. (a) For the given value of x5 , we estimate a dominant eigenvector of A to be
"
# "
#
1
1
⇡
.
239.500
3.951
60.625
The corresponding dominant eigenvalue is computed by finding
"
#"
# "
#
1.5 0.5
60.625
210.688
x6 = Ax5 =
=
,
2.0 3.0 239.500
839.75
and
210.688
60.625
(b) det(A
= 1 and
⇡ 3.475,
839.75
239.500
⇡ 3.506.
1.5
0.5
= (1.5
)(3.0
)
2.0
3.0
= 3.5. So the dominant eigenvalue is 3.5.
I) =
1.0 =
2
4.5 + 3.5, which has roots
5. (a) The entry of largest magnitude in x5 is m5 = 11.001, so that
"
# "
#
3.667
0.333
1
11.001 =
y5 =
x5 =
.
11.001
1
1
So the dominant eigenvalue is about 11 and y5 approximates the dominant eigenvector.
(b) We have
"
#"
# "
#
2
3
0.333
3.666
Ay5 =
=
3 10
1
10.999
"
# "
#
0.333
3.663
=
,
1 y5 = 11.001
1
11.001
so we have indeed approximated an eigenvalue and an eigenvector of A.
6. (a) The entry of largest magnitude in x10 is m10 = 5.530, so that
"
# "
#
1
1
1
y10 =
x10 = 1.470 ⇡
.
5.530
0.266
5.530
So the dominant eigenvalue is about 5.53 and y10 approximates the dominant eigenvector.
310
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
(b) We have
Ay10
1 y10
"
#"
# "
#
2
1
5.532
=
2 0.266
1.468
"
# "
#
1
5.530
= 5.530
=
,
0.266
1.471
5
=
2
so we have indeed approximated an eigenvalue and an eigenvector of A.
7. (a) The entry of largest magnitude in x8 is m8 = 10, so that
2
3 2
3
1
1
1
6
7 6
7
y8 =
x8 = 4 0.001
⇡ 40.00015 .
10.0 5
10
1
1
So the dominant eigenvalue is about 10
(b) We have
2
4
6
Ay8 = 4 1
6
2
and y8 approximates the dominant eigenvector.
32
3 2
3
6
1
10
76
7 6
7
15 40.00015 = 40.00035
4
1
10
3 2
3
1
10
6
7 6
7
y
=
10
=
0.0001
0.001
4
5
4
5,
1 8
1
10
0
3
0
so we have indeed approximated an eigenvalue and an eigenvector of A, though the approximation
in the middle component is not great.
8. (a) The entry of largest magnitude in x10 is m10 = 3.415, so that
2
3 2
3
1
1
1
6
7 6
7
y10 =
x10 = 4 2.914
⇡ 4 0.853 5 .
3.415 5
3.415
1.207
0.353
3.415
So the dominant eigenvalue is about 3.415 and y10 approximates the dominant eigenvector.
(b) We have
2
32
3 2
3
1
2
2
1
3.412
6
76
7 6
7
Ay10 = 41
1
35 4 0.853 5 = 4 2.912 5
0
1
1
0.353
1.206
2
3 2
3
1
3.415
6
7 6
7
1 y10 = 3.415 4 0.853 5 = 4 2.913 5 ,
0.353
1.205
so we have indeed approximated an eigenvalue and an eigenvector of A.
9. With the given A and x0 , we get the values below:
k
xk
yk
mk


0
1
1
1
1
1


1
26
8
1.000
0.308
26

2
17.692
5.923

1.000
0.335
17.692

3
18.017
6.004

1.000
0.333
18.017

4
17.999
6.000

1.000
0.333
17.999

5
18.000
6.000

1.000
0.333
18
4.5. ITERATIVE METHODS FOR COMPUTING EIGENVALUES
311

We conclude that the dominant eigenvalue of A is 18, with eigenvector
3
.
1
10. With the given A and x0 , we get the values below:
k
xk
yk


mk
0
1
0

1
0

1

6
8

0.75
1.00
1
8
2
3

8.500
8.000
9.765
9.882

1.000
0.941
0.988
1.000
8.5
4

9.929
9.905

9.882
5

9.990
9.995

1.000
0.998
9.929
6

9.997
9.996

1.000
1.000
9.995
1.000
1.000
9.997
We conclude that the dominant eigenvalue of A is 10 (not 10, since at each step the sign of each
1
component of yk is reversed) , with eigenvector
.
1
11. With the given A and x0 , we get the values below:
k
xk
yk


mk
0
1
0

1
0

1

7
2
1.000
0.286
1

7
2
7.571
2.857
1.000
0.377


7.571
3
7.755
3.132
1.000
0.404


7.755
4

7.808
3.212

1.000
0.411
7.808
5

7.823
3.234

1.000
0.412
7.823
6
7.827
3.240
1.000
0.414
7.827

1
We conclude that the dominant eigenvalue of A is about 7.83, with eigenvector about
. The
0.414

p
1
true values are = 5 + 2 2 with eigenvector p
.
2 1
12. With the given A and x0 , we get the values below:
k
xk
yk
mk


0

1
0

1
0
1
1
3.5
1.5
1.000
0.429


3.5
2
4.143
1.286
1.000
0.310


4.143
3
3.966
1.345
1.000
0.339
3.966


4

4.009
1.330

1.000
0.332
4.009
5

3.998
1.334

1.000
0.334
3.998
We conclude that the dominant eigenvalue of A is 4, with eigenvector

6
4.001
1.333
1.000
0.333
4.001
3
.
1
13. With the given A and x0 , we get the values below:
k
xk
yk
mk
0
2 3
1
415
1
2 3
1
415
1
1
1
3
21
4155
13
2
3
1.000
40.7145
0.619
2
21
2
3
16.810
412.2385
10.714
2
3
1.000
40.7285
0.637
2
16.81
3
3
17.011
412.3715
10.824
2
3
1.000
40.7275
0.636
2
17.011
4
3
16.999
412.3635
10.818
2
3
1.000
40.7275
0.636
2
16.999
5
3
17.000
412.2645
10.818
2
3
1.000
40.7275
0.636
2
17
312
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
2
3
1
We conclude that the dominant eigenvalue of A is 17, with corresponding eigenvector about 40.7275.
0.636
02 3 2 31
1
0
In fact the eigenspace corresponding to this eigenvalue is span @425 , 4 25A, and we have found one
0
1
eigenvector in this space.
14. With the given A and x0 , we get the values below:
k
0
2 3
1
4 15
1
2 3
1
4 15
1
xk
yk
mk
1
2 3
4
455
4
2 3
0.8
41.05
0.8
1
5
2
3
3.4
44.65
3.4
2
3
0.739
41.0005
0.739
2
4.6
3
3
3.217
44.4785
3.217
2
3
0.718
41.0005
0.718
2
4.478
4
3
3.155
44.4375
3.155
2
3
0.711
41.0005
0.711
5
3
3.133
44.4225
3.133
2
3
0.709
41.0005
0.709
2
6
3
3.126
44.4175
3.126
2
3
0.708
41.0005
0.708
2
4.437
2
4.422
4.417
2
3
0.708
We conclude that the dominant eigenvalue of A is ⇡ 4.4, with corresponding eigenvector ⇡ 41.0005.
0.708
2p 3
2
p
6 2 7
In fact the dominant eigenvalue is 3 + 2 with eigenspace spanned by 4 p1 5.
2
2
2 3
1
15. With the given matrix, and using x0 = 415, we get
1
k
xk
yk
mk
0
2 3
1
4 15
1
2 3
1
4 15
1
1
1
2 3
8
425
4
2
3
1
40.255
0.50
8
2
2
3
3
2
5.75
40.505
2.25
2
3
1
40.095
0.39
3
5.26
40.175
1.87
2
3
1
40.035
0.36
xk
yk
mk
0
2 3
2
415
1
2 3
2
415
1
2
5.26
4
3
5.10
40.075
1.74
2
3
1
40.015
0.34
1
3
24
425
6
2
3
1
40.085
0.25
2
24
5.10
2
5
3
5.04
40.035
1.70
2
3
1
40.015
0.34
2
3
11.0
4 1.55
2.5
2
3
1
4 0.14 5
0.23
2
11
5.04
2
6
3
5.02
40.015
1.68
2
3
1
4 0 5
0.33
3
3
14.18
4 2.455
7.91
2
3
1
4 0.17 5
0.56
2
14.18
5.02
2
7
3
5.01
4 0 5
1.67
2
3
1
4 0 5
0.33
4
3
16.38
4 3.125
11.65
2
3
1
4 0.19 5
0.71
2
16.38
5.01
2
8
3
5
4 0 5
1.67
2
3
1
4 0 5
0.33
5
2 3
3
We conclude that the dominant eigenvalue of A is 5, with corresponding eigenvector 405.
1
2 3
1
16. With the given matrix, and using x0 = 415, we get
1
k
5.75
2
5
3
16.38
4 3.125
11.65
2
3
1
4 0.20 5
0.77
2
17.41
2
9
3
5
4 0 5
1.67
2
3
1
4 0 5
0.33
5
4.5. ITERATIVE METHODS FOR COMPUTING EIGENVALUES
k
xk
yk
mk
6
3
17.80
4 3.545
14.05
2
3
1
4 0.20 5
0.79
2
17.8
7
3
17.93
4 3.585
14.28
2
3
1
4 0.2 5
0.8
8
3
17.98
4 3.595
14.36
2
3
1
4 0.2 5
0.8
2
9
3
17.99
4 3.605
14.39
2
3
1
4 0.2 5
0.8
2
17.93
313
2
17.98
17.99
10
3
18.00
4 3.605
14.40
2
3
1
4 0.2 5
0.8
2
18
11
3
18.00
4 3.605
14.40
2
3
1
4 0.2 5
0.8
2
18
2
3
1
We conclude that the dominant eigenvalue of A is 18, with corresponding eigenvector 4 0.2 5, or
0.8
2 3
5
4 15 .
4
17. With the given A and x0 , we get the values below:
k

xk
R(xk )
0

1
0
7
1

7
2
7.755
2
7.571
2.857

7.823
3
7.755
3.132
7.828
The Rayleigh quotient converges to the dominant eigenvalue, to three decimal places, after three
iterations.
18. With the given A and x0 , we get the values below:
k

xk
R(xk )
0

1
0
3.5
1

3.5
1.5
3.966
2
4.143
1.286

3
3.966
1.345
3.998
4
The Rayleigh quotient converges to the dominant eigenvalue after three iterations.
19. With the given A and x0 , we get the values below:
k
xk
R(xk )
0
2 3
1
415
1
2
16.333
1
3
2
21
4155
13
16.998
2
3
16.810
412.2385
10.714
17
The Rayleigh quotient converges to the dominant eigenvalue after two iterations.
20. With the given A and x0 , we get the values below:
k
xk
R(xk )
0
2 3
1
4 15
1
4.333
1
2 3
4
455
4
4.404
2
2
3
3.4
44.65
3.4
4.413
2
3
3
3.217
44.4785
3.217
4.414
The Rayleigh quotient converges to the dominant eigenvalue, to three decimal places, after three
iterations.
314
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
21. With the given matrix, and using x0 =
k
xk

0
1
1
yk

mk
4.5
1
1


1

5
4

1
0.8
4.49
2


4.8
3.2

1
0.67
4.46
1
, we get
1
3
4

4.67
2.67

4.57
2.29

1
0.5
1
0.57
4.43

4.4
5
4.50
2.00
1
0.44


4.37
6
4.44
1.78
1
0.40
4.34


7

4.40
1.60

1
0.36
4.32
8
4.36
1.45
1
0.33
4.3
The characteristic polynomial is
det(A
I) =

4
1
0
4)2 ,
=(
4
so that A has the double eigenvalue 4. To find the corresponding eigenspace,
⇥

A

⇤
0
0 =
0
4I
1
0
0
,
0
1
. The values mk appear to be converging to 4, but very slowly;
0
the vectors yk may also be converging to the eigenvector, but very slowly if at all.
so that the eigenspace has basis
22. With the given matrix, and using x0 =
k
xk
yk
mk


0
1
1
1
1
2


1
4
0
1
0
3


2
2.8
4

2.67
1.33

1
0.33
1
, we get
1
3

3
1

2.50
1.50

1
0.5
2.6
5

2.40
1.60

1
0.6
2.47
1
0.67


2.38
6
2.33
1.67
1
0.71
7

2.29
1.71

2.32
1
0.75
2.28


8
2.25
1.70
1
0.78
2.25
The characteristic polynomial is
det(A
I) =

3
1
1
1
=
2
4 +4=(
2)2 ,
so that A has the double eigenvalue 2. To find the corresponding eigenspace,
⇥
A
2I

⇤
0 =

1
1
1
1

0
1
!
0
0
1
0
0
0
1
. The values mk appear to be converging to 2, but very slowly;
1
the vectors yk may also be converging to the eigenvector, but very slowly.
so that the eigenspace has basis
4.5. ITERATIVE METHODS FOR COMPUTING EIGENVALUES
315
2 3
1
23. With the given matrix, and using x0 = 415, we get
1
k
xk
yk
mk
0
2 3
1
415
1
2 3
1
415
1
3.33
1
2 3
5
445
1
2 3
1
40.85
0.2
4.05
2
2
3
4.2
43.25
0.2
2
3
1
40.765
0.05
4.03
3
2
3
4.05
43.055
0.05
2
3
1
40.755
0.01
4.01
4
2
3
4.01
43.015
0.01
2
3
1
40.755
0
4
The matrix is upper triangular, so that A has an
eigenspace for = 4,
2
0 0
⇥
⇤
A 4I 0 = 40 0
0 0
5
2 3
4
435
0
2
3
1
40.755
0
4
6
2 3
4
4 35
0
2
3
1
40.755
0
4
7
2 3
4
435
0
2
3
1
40.755
0
4
8
2 3
4
435
0
2
3
1
40.755
0
4
eigenvalue 1 and a double eigenvalue 4. To find the
1
0
3
3
2
0
0
05 ! 40
0
0
0
0
0
3
0
05
0
1
0
0
8 2 3 2 39
0 =
< 1
so that the eigenspace has basis 405 , 415 . The values mk converge to 4, while yk converge
:
;
0
0
2
3
1
apparently to 40.755, which is in the eigenspace.
0
2 3
1
24. With the given matrix, and using x0 = 415, we get
1
k
xk
yk
mk
0
2 3
1
4 15
1
2 3
1
4 15
1
3.67
1
2 3
0
465
5
2
3
0
4 1 5
0.83
5.49
2
2
3
0
45.835
4.17
2
3
0
4 1 5
0.71
5.47
2
3
3
0
45.715
3.57
2
3
0
4 1 5
0.62
5.45
The matrix is upper triangular, so that A has
eigenspace for = 5,
2
5
⇥
⇤
A 5I 0 = 4 0
0
2
4
3
0
45.625
3.12
2
3
0
4 1 5
0.56
5.42
2
5
3
2
0
45.565
2.78
2
3
0
4 1 5
0.50
6
3
0
45.505
2.50
2
3
0
4 1 5
0.45
5.4
5.38
2
7
3
0
45.455
2.27
2
3
0
4 1 5
0.42
5.36
2
8
3
0
45.425
2.08
2
3
0
4 1 5
0.38
5.34
an eigenvalue 0 and a double eigenvalue 5. To find the
0
0
0
0
1
0
3
2
0
1
05 ! 40
0
0
0
0
0
0
1
0
3
0
05
0
2 3
0
so that the eigenspace has basis 415. The values mk appear to converge slowly to 5, while yk also
0
2 3
0
appear to converge slowly to 415.
0
316
CHAPTER 4. EIGENVALUES AND EIGENVECTORS

25. With the given matrix, and using x0 =
k

xk

yk
mk
1
, we get
1
0

1
1

1
1
1
1
I) =


1
1

1
0
1
2

1
0

1
1
1
3

1
0

1
0
1
4
1
1
1
1
1
The characteristic polynomial is
det(A
1
2
1
=
1
2
+ 1,
so that A has no real eigenvalues, so the power method cannot succeed.

1
26. With the given matrix, and using x0 =
, we get
1
k
xk
yk


mk
0
1
1
1
1


1
1

3
3

1
1
3
2

3
3

1
1
3
3
3
3
1
1
3
The characteristic polynomial is
det(A
I) =

2
1
2
2
=
5
7 + 12 = (
3)(
4)
so
 that the power method converges to a non-dominant eigenvalue. This happened because we chose
1
as the initial value, which is in fact an eigenvector for = 3:
1




2 1 1
3
1
=
=3
.
2 5 1
3
1
27. With the given matrix, and using x0 =
k
xk
yk
mk
0
2 3
1
4 15
1
2 3
1
4 15
1
1
1
2 3
3
445
3
2
3
0.75
4 1 5
0.75
4
2
3
2.5
44.05
2.5
2
3
0.62
4 1 5
0.62
2
4

1
, we get
1
3
3
2.25
4 4 5
2.25
2
3
0.56
4 1 5
0.56
2
4
4
3
2.12
4 4 5
2.12
2
3
0.53
4 1 5
0.53
2
4
5
3
2.06
4 4 5
2.06
2
3
0.52
4 1 5
0.52
2
4
6
3
2.03
4 4 5
2.03
2
3
0.51
4 1 5
0.51
2
4
7
3
2.02
4 4 5
2.02
2 3
0.5
415
0.5
2
4
The characteristic polynomial is
det(A
2
I) = 4
5
1
0
7
7
0
4
1
5
3
5 = ( + 12)(
4)(
2)
8
3
2.01
4 4 5
2.01
2 3
0.5
415
0.5
2
4
4.5. ITERATIVE METHODS FOR COMPUTING EIGENVALUES
317
so
that the power method converges to a non-dominant eigenvalue. This happened because we chose
2 3
1
415 as the initial value. This vector is orthogonal to the eigenvector corresponding to the dominant
1
2 3
1
eigenvalue = 12, which is 4 05:
1
2
32 3 2
3
2 3
5 1
7
1
12
1
4 0 4
05 4 05 = 4 05 = 12 4 05 .
7 1
5
1
12
1

1
28. With the given matrix, and using x0 =
, we get
1
k
xk
yk
mk
0
2 3
1
415
1
2 3
1
415
1
1
1
2 3
0
425
1
2 3
0
415
0.5
2
2
3
1
4 1 5
0.5
2
3
1
4 1 5
0.5
0.6
1.44
3
2
3
2
4
3
0.8
40.85
1.8
2
3
0.44
40.445
1
2
4 0 5
2.5
2 3
0.8
405
1
1.49
1
5
2
3
0
40.895
1
2
3
0
40.895
1
0.5
6
2
3
2
0.89
4 0.895
0.11
2
3
1
4 1 5
0.12
7
3
2
4 0 5
1.88
2
3
1
4 0 5
0.94
0.88
1.5
2
8
3
1
4 1 5
1.94
2
3
0.52
40.525
1
1
Neither mk nor yk shows clear signs of converging to anything. The characteristic polynomial is
2
3
1
1
0
1
0 5=(
det(A
I) = 4 1
1)( 2 2x + 2)
1
1
1
so that A has eigenvalues 1 and 1 ± i. The absolute value of each of the complex eigenvectors is 2, so
1 is not the eigenvalue of largest magnitude, and the power method cannot succeed.
29. In Exercise 9, we found that
= 18. To find

4
A 18I =
5
1
k
xk
yk
mk
Thus
2
1
=
19, so that
2
30. In Exercise 10, we found that
1
=
=


0
1
1
1
1


2,
1
8
10
0.8
1
1
10
we apply the power method to

12
1
with x0 =
.
15
1


2
A + 10I =
4
8
4
8
19
2
1
= 12, so that
2
3
15.2
19.0
0.8
1
19
1 is the second eigenvalue of A.
10, so we apply the power method to
with x0 =

1
:
1
xk
yk
mk
Thus

0.8
1
k


15.2
19.0


0
1
1
1
1
1
= 2 is the second eigenvalue of A.


1
8
16
0.5
1
16


2
6
12
0.5
1
12


3
6
12
0.5
1
12
318
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
31. In Exercise 13, we found that
= 17, so we apply the power method to
2
3
2 3
8
4
8
1
2
45 with x0 = 415 .
17I = 4 4
8
4
8
1
1
A
k
0
2 3
1
415
1
2 3
1
415
1
xk
yk
mk
Thus
2
1
=
18, so that
32. In Exercise 14, we found that
A
k
xk
yk
mk
Thus
2
1
⇡
0
2 3
1
4 15
1
2 3
1
4 15
1
⇡ 4.4, so we
2
1.4
4.4I = 4 1
0
1
0.6
2.81, so that
2
2
3
18
4 95
18
2 3
1
40.55
1
3
2
3
18
4 95
18
2 3
1
40.55
1
18
18
1 is the second eigenvalue of A.
3
0.4
4 0.65
0.4
2
3
0.667
4 1 5
0.667
1
2
4
1
2
3
4
4 25
4
2 3
1
40.55
1
1
=
2
1
2
2
2
apply the power method to
3
2 3
1
0
1
1.4
1 5 with x0 = 415 .
1
1.4
1
3
3
2
1.933
4 2.7335
1.933
2
3
0.707
4 1 5
0.707
3
1.990
4 2.8155
1.990
2
3
0.707
4 1 5
0.707
2.733
4
2
3
3
1.990
4 2.8145
1.990
2
3
0.707
4 1 5
0.707
1.990
4 2.8145
1.990
2
3
0.707
4 1 5
0.707
2.815
5
2
2.814
2.814
⇡ 1.59 is the second eigenvalue of A (in fact,
33. With A and x0 as in Exercise 9, proceed to calculate xk by solving Axk = yk
the inverse power method. We get
k

xk

yk
mk
0

1
1

1
1
1
1

0.5
0.5

1
1
0.5
2

0.833
1.056

0.789
1
1.056
3

0.798
0.997

0.801
1
0.997
4

0.8
1

0.8
1
1
We conclude that the eigenvalue of A of smallest magnitude is
=
xk
yk
mk


0
1
1
1
1
1


1
0.1
0.4
0.25
1
0.4


2
0.225
0.400
0.562
1
0.4


3
0.256
0.525
0.488
1
0.525


4
0.249
0.495
0.502
1
0.495
We conclude that the eigenvalue of A of smallest magnitude is
1
0.5
p
2).
at each step to apply
0.8
1
0.8
1
1
1.
34. With A and x0 as in Exercise 10, proceed to calculate xk by solving Axk = yk
the inverse power method. We get
k
=3
5
1
1
1
2

5
0.250
0.501

0.5
1
0.501
= 2.
1

at each step to apply
6
0.25
0.50

0.5
1
0.5
4.5. ITERATIVE METHODS FOR COMPUTING EIGENVALUES
319
35. With A as in Exercise 7, and x0 as given, proceed to calculate xk by solving Axk = yk
to apply the inverse power method. We get
k
2
0
3
1
4 15
1
2 3
1
4 15
1
xk
yk
mk
1
2
3
0.5
4 0 5
0.5
2 3
1
4 05
1
1
0.5
2
3
0.500
4 0.3335
0.500
2
3
1
4 0.6675
1
3
3
0.500
4 0.1115
0.500
2
3
1
4 0.2225
1
2
2
0.5
0.5
4
3
0.500
4 0.2595
0.500
2
3
1
4 0.5195
1
5
3
0.50
4 0.165
0.50
2
3
1
4 0.3215
1
2
2
0.5
We conclude that the eigenvalue of A of smallest magnitude is
1
0.5
=
xk
yk
mk
0
2 3
1
415
1
2 3
1
415
1
1
2
1
3
0.286
40.1435
0.286
2 3
1
40.55
1
0.286
2
2
3
0.357
4 0.0715
0.357
2
3
1
4 0.25
1
0.357
3
2
3
2
0.457
4 0.3715
0.457
2
3
1
4 0.8125
1
4
3
0.545
4 0.6345
0.545
2
3
0.859
4 1 5
0.859
0.457
0.634
We conclude that the eigenvalue of A of smallest magnitude is ⇡
p
is 3
2.
2
6
2.
5
1
3
0.511
4 0.6745
0.511
2
3
0.758
4 1 5
0.758
1
0.645
at each step
0.5
36. With A and x0 as in Exercise 14, proceed to calculate xk by solving Axk = yk
the inverse power method. We get
k
1
0.674
at each step to apply
2
6
3
0.468
4 0.6455
0.468
2
3
0.725
4 1 5
0.725
0.645
⇡ 1.55. In fact the eigenvalue
37. With ↵ = 0, we are just using the inverse power method, so the solution to Exercise 33 provides the
eigenvalue closest to zero.

1
38. Since ↵ = 0, there is no shift required, so this is the inverse power method. Take x0 = y0 =
, and
1
proceed to calculate xk by solving Axk = yk 1 at each step to apply the inverse power method. We
get
k
0
1
2
3
4
5
6







1
0.125
0.417
0.306
0.340
0.332
0.334
xk
1
0.375
0.750
1.083
0.981
1.005
0.999







1
0.333
0.556
0.282
0.346
0.33
0.334
yk
1
1
1
1
1
1
1
mk
1
0.375
0.75
1.083
0.981
1.005
0.999
We conclude that the eigenvalue of A closest to 0 (i.e., the eigenvalue smallest in magnitude) is
39. We first shift A:
A
2
1
5I = 4 1
6
0
2
0
1
1
=
1.
3
6
15 .
1
2 3
1
Now take x0 = y0 = 415, and proceed to calculate xk by solving Axk = yk
1
1
at each step to apply
320
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
the inverse power method. We get
k
0
2 3
1
4 15
1
2 3
1
4 15
1
xk
yk
mk
1
2
3
0.2
4 0.55
0.2
2
3
0.4
4 1 5
0.4
1
2
2
2
0.5
0.5
2
11
A + 2I = 4 4
8
3
3
0.32
4 0.505
0.32
2
3
0.064
4 1 5
0.064
0.08
4 0.505
0.08
2
3
0.16
4 1 5
0.16
0.5
1
0.5
We conclude that the eigenvalue of A closest to 5 is 5 +
40. We first shift A:
3
= 3.
3
7
45 .
11
4
17
4
2 3
1
Now take x0 = y0 = 415, and proceed to calculate xk by solving Axk = yk
1
the inverse power method. We get
k
xk
yk
mk
0
2 3
1
4 15
1
2 3
1
4 15
1
1
2
3
0.158
4 0.1585
0.263
2
3
0.6
4 0.6 5
1
1
2
2
3
2
0.832
4 0.4325
0.853
2
3
0.975
4 0.506 5
1
0.263
41. The companion matrix of p(x) = x2 + 2x
3
0.990
4 0.4965
0.991
2
3
0.999
4 0.5 5
1
0.853
We conclude that the eigenvalue of A closest to
3
0.991
2 is
2+
1
1
=
4
2
1
3
2
at each step to apply
5
3
1
40.55
1
2 3
1
40.55
1
0.999
4 0.5005
1.000
2 3
1
40.55
1
1
1
1.
2 is

2
,
0

1
so we apply the inverse power method with x0 = y0 =
, calculating xk by solving Axk = yk
1
each step. We get
C(p) =
k
xk
yk
mk


0
1
1
1
1
1


1
1
1.5
0.667
1
1.5
So the root of p(x) closest to 0 is ⇡
42. The companion matrix of p(x) = x2
2

1
1.333

0.75
1
1.333
1
1.366
x


2
1
3
1
1.375
0.727
1


1.375
4
1
1.364
0.733
1
1.364


5
1
1.367
0.732
1
C(p) =

1
1
3
,
0

1.367
⇡ 0.732. The exact value is
3 is

1+
6
1
1.366
0.732
1
1.366
p
3.
1
at
4.5. ITERATIVE METHODS FOR COMPUTING EIGENVALUES
321
so we apply the inverse power method to
C(p)
with x0 = y0 =

2I =

1
1
1
, calculating xk by solving Axk = yk
1
k

xk

yk
mk
0

1
1

1
1
1
1

5
2

1
0.4
5
2

3.2
1.4

1
0.438
3.2
43. The companion matrix of p(x) = x3


1
0.434
at each step. We get
1
3.312
1
3.303
So the root of p(x) closest to 2 is ⇡ 2 +
3
3.312
1.438
3
2
4

3.302
1.434

1
0.434
1
0.434


6
3.303
1.434
1
0.434
3.303
p
⇡ 2.303. The exact value is 12 (1 + 13).
2x2 + 1 is
2
2
C(p) = 41
0
3.302
5
3.303
1.434
3.303
3
1
05 ,
0
0
0
1
2 3
1
so we apply the inverse power method to C(p). If we first start with y0 = 415, we find that this is an
1
eigenvector of C(p) with
eigenvalue
1,
but
this
may
not
be
the
closest
root
to 0, so we start instead
2 3
1
with with x0 = y0 = 405, calculating xk by solving Axk = yk 1 at each step. We get
1
k
xk
yk
mk
0
2 3
1
405
1
2 3
1
405
1
1
k
xk
yk
mk
2
1
3
0
4 15
1
2 3
0
4 15
1
1
2
6
2
3
1
4 15
2
2
3
0.5
4 0.55
1
2
3
2
3
7
3
0.615
4 1 5
1.615
2
3
0.381
4 0.6195
1
1.615
2x2 + 1 is
2
5
C(p) = 41
0
2
4
2
3
2
1.667
8
3
0.619
4 1 5
1.619
2
3
0.382
4 0.6185
1
1.619
1
1.618
⇡=
1
0
1
3
1
05 ,
0
2
5
3
0.6
4 1 5
1.6
2
3
0.375
4 0.6255
1
0.667
4 1 5
1.667
2
3
0.4
4 0.65
1
1.6
We conclude that the root of p(x) closest to zero is
44. The companion matrix of p(x) = x3
3
0.5
4 1 5
0.5
2
3
0.333
4 0.6675
1
2
0.625
4 1 5
1.625
2
3
0.385
4 0.6155
1
1.625
2
1.6
9
3
0.618
4 1 5
1.618
2
3
0.382
4 0.6185
1
1.618
0.618. The actual root is 12 (1
p
5).
322
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
so we apply the inverse power method to C(p)
solving Axk = yk
k
xk
yk
mk
0
2 3
1
4 15
1
2 3
1
4 15
1
1
1
at each step. We get
1
3
2.333
4 0.6675
0.333
2
3
1
40.2865
0.143
2
2.333
2
3
3.762
4 0.8105
0.190
2
3
1
40.2155
0.051
2
3.762
2 3
1
5I starting with x0 = y0 = 415, calculating xk by
1
3
3
3.909
4 0.8255
0.175
2
3
1
40.2115
0.045
2
3.909
We conclude that the root of p(x) closest to zero is 5 +
4
3
3.918
4 0.8265
0.174
2
3
1
40.2115
0.044
2
3.918
1
3.919
5
3
3.919
4 0.8265
0.174
2
3
1
40.2115
0.044
2
3.919
6
3
3.919
4 0.8265
0.174
2
3
1
40.2115
0.044
2
3.919
⇡= 4.745.
45. First, A ↵I is invertible, since ↵ is an eigenvalue of A precisely when A ↵I is not invertible. Let x be
an eigenvector for . Then Ax = x, so that Ax ↵x = x ↵x, and therefore (A ↵I)x = (
↵)x.
Now left multiply both sides of this equation by (A ↵I) 1 , to get
x = (A
↵I)
1
(
But the last equation says exactly that
↵)x, so that
1
↵
x = (A
↵ is an eigenvalue of (A
1
↵I)
↵I)
1
x.
with eigenvector x.
46. If is dominant, then | | > | | for all other eigenvalues . But this means that the algebraic multiplicity
of is 1, since it appears only once in the list of eigenvalues listed with multiplicity, so its geometric
multiplicity is 1 and thus its eigenspace is one-dimensional.
47. Using rows, the Gerschgorin disks are:
Center 1,
Center 5,
radius 1 + 0 = 1,
1 1
radius + = 1,
2 2
radius 1 + 0 = 1.
Center 1,
radius
Center 4,
Using columns, they are
Center 4,
Center 5,
1
3
+1= ,
2
2
radius 1 + 0 = 1,
1
1
radius 0 + = :
2
2
From this diagram, we see that A has a real eigenvalue between 0 and 2.
4.5. ITERATIVE METHODS FOR COMPUTING EIGENVALUES
323
48. Using rows, the Gerschgorin disks are:
Center 2,
Center 2i,
Center
2i,
radius | i| + 0 = 1,
radius |1| + |1 + i| = 1 +
p
2,
radius 0 + 1 = 1.
Using columns, they are
Center 2,
Center 2i,
Center
2i,
radius 1 + 0 = 1,
radius | i| + 1 = 2,
radius 0 + |1 + i| =
p
2:
From this diagram, we conclude that A has an eigenvalue within one unit of
2i.
49. Using rows, the Gerschgorin disks are:
Center 4
Center
3i,
1 + i,
Center 5 + 6i,
Center
5 5i,
radius |i| + 2 + 2 = 5,
radius |i| + 0 + 0 = 1,
radius |1 + i| + | i| + |2i| = 3 +
radius 1 + | 2i| + |2i| = 5.
p
Using columns, they are
Center 4
3i,
Center
1 + i,
Center 5 + 6i,
Center
5
5i,
radius |i| + |1 + i| + 1 = 2 +
radius |i| + | i| + | 2i| = 4,
radius 2 + 0 + |2i| = 4
radius | 2| + 0 + |2i| = 4.
p
2,
2
324
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
From this diagram, we conclude that A has an eigenvalue within 1 unit of
1 + i.
50. Using rows, the Gerschgorin disks are:
Center 2,
Center 4,
Center 6,
Center 8,
1
,
2
1
radius +
4
1
radius +
6
1
radius .
8
1
1
= ,
4
2
1
2
=
6
3
1
,
4
1
radius +
2
1
radius +
4
1
radius .
6
1
1
= ,
6
3
1
3
=
8
8
radius
Using columns, they are
Center 2,
Center 4,
Center 6,
Center 8,
radius
Since the circles are all disjoint, we conclude that A has four real eigenvalues, near 2, 4, 6, and 8.
51. Let A be a strictly diagonally dominant square matrix. Then inP
particular, none of the diagonal entries
of A can be zero, since each row has the property that |aii | > j6=i |aij |. Therefore each Gerschgorin
disk is centered at a nonzero point, and its radius is less than the absolute value of its center, so the disk
does not contain the origin. Since every eigenvalue is contained in some Gerschgorin disk, it follows
that 0 is not an eigenvalue, so that the matrix is invertible by Theorem 4.17.
4.5. ITERATIVE METHODS FOR COMPUTING EIGENVALUES
325
52. Let be any eigenvalue of A. Then by Theorem 4.29, is contained in the Gerschgorin disk for some
row i, so that
X
|
aii | 
|aij | .
j6=i
By the Triangle Inequality, | |  |
| ||
aii | + |aii |, so that
aii | + |aii |  |aii | +
X
j6=i
|aij | =
n
X
j=1
|aij |  kAk ,
since kAk is the maximum of the possible row sums.
53. A stochastic matrix is a square matrix whose entries are probability vectors, which means that the
sum of the entries in any column of A is 1, so the sum of the entries of any row in AT is 1, so that
AT = 1. By Exercise 52, any eigenvalue of AT must satisfy | |  AT = 1. But by Exercise 19
in Section 4.3, the eigenvalues of A and of AT are the same, so that if is any eigenvalue of A, then
| |  1.
54. Using rows, the Gerschgorin disks are:
Center 0,
radius 1,
Center 5,
radius 2,
1 1
radius + = 1
2 2
3
radius .
4
Center 3,
Center 7,
Using columns, they are
Center 0,
radius 2 +
Center 5,
radius 1,
3
radius
4
1
radius .
2
Center 3,
Center 7,
1
5
= ,
2
2
From the row diagram, we see that there is exactly one eigenvalue in the circle of radius 1 around 0, so
it must be real and therefore there is exactly one eigenvalue in the interval [ 1, 1]. From the column
diagram,
conclude that there is exactly one eigenvalue in the interval [4, 6] and exactly
⇥ we15similarly
⇤
one in 13
2 , 2 .
Since the matrix has real entries, the fourth eigenvalue must be real as well. From the row diagram,
since the⇥ circle⇤ around zero contains exactly one eigenvalue, the fourth eigenvalue must be in the
interval 2, 7 34 . But it cannot be larger than 3 34 since then, from the column diagram, it would either
326
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
not be⇥in a disk
⇤ or would be in an isolated disk already containing an eigenvalue. So it must lie in the
range 2, 3 34 .
In summary, the four eigenvalues are each in a di↵erent one of the four intervals
[ 1, 1],
The actual values are ⇡
4.6
[2, 3.75],
[4, 6],
[6.5, 7.5].
0.372, ⇡ 2.908, ⇡ 5.372, and ⇡ 7.092.
Applications and the Perron-Frobenius Theorem

0
1
regular.
1. Since
1
0
2
=

1
0

0
0
, it follows that any power of
1
1
1
contains zeroes. So this matrix is not
0
2. Since the given matrix is upper triangular, any of its powers is upper triangular as well, so will contain
a zero. Thus this matrix is not regular.
3. Since
"
#2 "
7
1
= 92
0
9
1
3
2
3
1
3
2
3
#
,
the square of the given matrix has all positive entries, so this matrix is regular.
4. We have
2
1
2
61
42
0
2
1
2
61
42
0
2
1
2
61
42
0
32 2
1
0 1
4
7
61
0 05 = 4 4
1
1 0
2
33 2
1
0 1
2
7
61
0 05 = 4 2
1 0
0
34 2
1
0 1
2
7
61
0 05 = 4 2
1 0
0
1
0
0
0
0
1
0
0
1
3
1
2
17
25
0
32
1
1
4
7 61
05 4 4
1
0
2
32
5
1
8
7 61
05 4 8
1
0
4
1
0
0
1
2
1
2
0
3
1
2
17
25
=
2
=
2
0
3
1
4
17
45
1
2
5
8
61
48
1
4
3
1
2
1
2
0
9
16
65
4 16
1
8
1
4
1
4
1
2
1
4
17
45
1
2
3
5
8
17
85
1
4
so that the fourth power of this matrix contains all positive entries, so this matrix is regular.
5. If A and B are any 3 ⇥ 3 matrices with a12 = a32 = b12 = b32 = 0, then
(AB)12 = a11 b12 + a12 b22 + a13 b32 = 0 + 0 + 0 = 0,
(AB)32 = a31 b12 + a32 b22 + a33 b32 = 0 + 0 + 0 = 0,
so that AB has the same property. Thus any power of the given matrix A has this property, so no
power of A consists of positive entries, so A is not regular.
6. Since the third row of the matrix is all zero, the third row of any power will also be all zero, so this
matrix is not regular.
7. The transition matrix P has characteristic equation
det(P
I) =
1
3
1
6
2
3
5
6
=
1
(
6
1)(6
1),
so its eigenvalues are 1 = 1 and 2 = 16 . Thus P has two distinct eigenvalues, so is diagonalizable.
Now, this matrix is regular, so the remark at the end of the proof of Theorem 4.33 shows that the
4.6. APPLICATIONS AND THE PERRON-FROBENIUS THEOREM
327
vector x which forms the columns of L is a probability vector parallel to an eigenvector for
compute this, we row-reduce P I:
"
#
"
#
h
i
2
1
1
0
1
0
3
6
4
!
P I 0 =
2
1
0
0
0 0
3
6
= 1. To
so that the corresponding eigenspace is
E1 = span
Thus the columns of L are of the form

✓
1
4
◆
.
t
where t + 4t = 5t = 1, so that t = 15 . Therefore
4t
"
#
L=
1
5
4
5
1
5
4
5
.
8. The transition matrix P has characteristic equation
1
2
det(P
I) =
1
3
1
2
0
1
6
1
3
1
2
1
6
1
2
=
1
(
36
1)(36
2
18 + 1).
So one eigenvalue is 1; the roots of the quadratic are distinct, so that P has three distinct eigenvalues,
so is diagonalizable. Now, this matrix is regular (its square has all positive entries), so the remark
at the end of the proof of Theorem 4.33 shows that the vector x which forms the columns of L is a
probability vector parallel to an eigenvector for = 1. To compute this, we row-reduce P I:
2
3
2
3
1
1
1
7
0
1 0
0
h
i
2
3
6
3
6
7
6
7
1
1
P I 0 = 4 12
05 ! 40 1
3 05
2
3
1
1
0
0
0 0
0 0
6
2
so that the corresponding eigenspace is
02 31
7
E1 = span @495A .
3
2 3
7t
Thus the columns of L are of the form 49t5 where 7t + 9t + 3t = 19t = 1, so that t =
3t
2
3
7 7 7
1 4
9 9 95 .
L=
19
3 3 3
1
19 .
Therefore
9. The transition matrix P has characteristic equation
det(P
I) =
0.2
0.6
0.2
0.3
0.1
0.6
0.4
0.4
0.2
=
(
1)(
2
+ 0.5 + 0.08)
So one eigenvalue is 1; the roots of the quadratic are distinct (and complex), so that P has three
distinct eigenvalues, so is diagonalizable. Now, this matrix is regular, so the remark at the end of the
328
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
proof of Theorem 4.33 shows that the vector x which forms the columns of L
parallel to an eigenvector for = 1. To compute this, we row-reduce P I:
2
3
2
0.8
0.3
0.4 0
1 0
0.889
h
i
6
7
6
P I 0 = 4 0.6
0.9
0.4 05 ! 40 1
1.037
0.2
0.6
0.8 0
0 0
0
is a probability vector
so that the corresponding eigenspace is
3
0
7
05
0
02
31
0.889
E1 = span @41.0375A .
1
2
3
0.889t
Thus the columns of L are of the form 41.037t5 where 0.889t + 1.037t + t = 2.926t = 1, so that
1t
1
t = 2.926 . Therefore
2
3
0.304 0.304 0.304
L = 40.354 0.354 0.3545 .
0.341 0.341 0.341
(If the matrix
2 3 is converted to fractions and the computation done that way, the resulting eigenvector
24
will be 4285, which must then be converted to a unit vector.)
27
10. Suppose that the sequence of iterates xk approach both u and v, and choose ✏ > 0. Choose k such
that |xk u| < 2✏ and |xk v| < 2✏ . Then
|u
so that |u
v| = |(u
xk ) + (xk
v)|  |u
xk | + |xk
v| < ✏,
v| is less than any positive ✏. So it must be zero; that is, we must have u = v.
11. The characteristic polynomial is
det(L
I) =
2
0.5
=
2
1 = ( + 1)(
1),
so the positive eigenvalue is 1. To find a corresponding eigenvector, row-reduce


⇥
⇤
1
2 0
1
2 0
L I 0 =
!
.
0.5
1 0
0
0 0
✓ ◆
2
Thus E1 = span
.
1
12. The characteristic polynomial is
det(L
I) =
1
1.5
0.5
=
2
0.75 = (
1.5)( + 0.5),
so the positive eigenvalue is 1.5. To find a corresponding eigenvector, row-reduce


⇥
⇤
0.5
1.5 0
1
3 0
L 1.5I 0 =
!
.
0.5
1.5 0
0
0 0
✓ ◆
3
Thus E1.5 = span
.
1
4.6. APPLICATIONS AND THE PERRON-FROBENIUS THEOREM
329
13. The characteristic polynomial is
7
I) = 0.5
0
det(L
4
0
3
=
+ 3.5 + 1 =
0.5
1
(
2
2)(2
2
+ 4 + 1),
so 2 is a positive eigenvalue. The other two roots are
4±
4
p
8
,
both of which are negative. To find an eigenvector for 2, row-reduce
02
⇥
L
2I
31
16
Thus E2 = span @4 4 5A.
1
2
2
7
⇤
0 = 40.5
2
0 0.5
4
0
2
3
2
0
1
05 ! 4 0
0
0
0
1
0
16
4
0
5
2
x3 + x2 + x + =
3
3
1
(x
3
3
0
05 .
0
14. The characteristic polynomial is
1
det(L
I) =
5
1
3
3
0
=
2
3
0
so 2 is a positive eigenvalue. The other two eigenvalues are complex. To
row-reduce
2
3
2
1
5
3 0
1 0
18
h
i
6 1
7
6
L 2I 0 = 4 3
2
0 05 ! 40 1
3
2
0
2
0
0
0
0
3
02 31
18
Thus E2 = span @4 3 5A.
1
2)(3x2 + 3x + 1)
find an eigenvector for 2,
3
0
7
05 .
0
15. Since the eigenvector corresponding to the positive eigenvalue 1 for the Leslie matrix represents a
stable ratio of population classes, one that proportion is reached, if 1 > 1, the population will grow
without bound; if 1 = 1, it will be stable, and if 1 < 1 it will decrease and eventually die out.
16. From the Leslie matrix in Equation (3), we have
b1
det(L
For n = 1, this is just b1
I) =
b2
b3
0
s1
0
0
..
.
s2
0
..
.
s3
..
.
0
0
0
···
···
···
···
..
.
bn 1
0
0
0
..
.
···
sn
1
, which is equal to
cL ( ) = ( 1)1 (
1
b1
0
) = b1
.
bn
0
0
0 .
..
.
330
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
This forms the basis for the induction. Now assume the statement holds for n
n. Expand the above determinant along the last column, giving
det(L
I) = ( 1)
n+1
bn
s1
0
0
..
.
s2
0
..
.
s3
..
.
0
0
0
···
···
···
..
.
0
b1
0
0
0
..
.
···
+(
sn
)
1
b2
1, and we prove it for
b3
0
s1
0
0
..
.
s2
0
..
.
s3
..
.
0
0
0
···
···
···
···
..
.
bn 2
0
0
0
..
.
···
sn
bn 1
0
0
0 .
..
.
0
2
The first matrix is upper triangular, so its determinant is the product of its diagonal entries; for the
second, we use the inductive hypothesis. So
det(L
I) = ( 1)n+1 bn (s1 s2 · · · sn
( 1)
n 1
= ( 1)
b1
2 s1 s2
· · · sn
(
bn
n+1
1)
n 1
bn (s1 s2 · · · sn
n
+ ( 1) (
n
1)
bn
n
= ( 1) (
n
b1
2 s1 s2
n 1
bn
n 1
b1
b2 s 1
2 s1 s2
n 2
· · · sn
n 2
· · · sn
2
3
bn
3
b2 s 1
3
n 3
b2 s 1
1 s1 s2
n 2
2
bn
···
bn
···
· · · sn
···
1 s1 s2
1 s1 s2
2)
· · · sn
· · · sn
2
)
bn s 1 s 2 · · · s n
2
1 ),
proving the result.
17. P is a diagonal matrix, so inverting it consists of inverting each of the diagonal entries. Now, multiplying
any matrix on the left by a diagonal matrix multiplies each row by the corresponding diagonal entry,
and multiplying on the right by a diagonal matrix multiplies each column by the corresponding diagonal
entry, so
2
3
b1 b2
b3 · · ·
bn 1
bn
6
7
0
···
0
07
61 0
6
7
60 1
0
···
0
07
s1
6
7
1
P L = 60 0
1
···
0
07
6
7
s1 s2
6.
..
..
..
.. 7
..
6.
7
.
.
.
.
.5
4.
1
0 0
0
· · · s1 s2 ···s
0
n 2
and then
2
b1
6
61
6
60
6
1
P LP = 6
60
6.
6.
4.
0
b2 s 1
0
1
0
..
.
0
b3 s 1 s 2
0
0
1
..
.
0
···
···
···
···
..
.
bn
1 s1 s2
0
0
0
..
.
1
···
· · · sn
2
bn s 1 s 2 · · · s n
0
0
0
..
.
0
But this is the companion matrix of the polynomial
p(x) = xn
b1 x n
1
b2 s 1 x n
2
···
bn
1 s1 s2
· · · sn
2x
1
3
7
7
7
7
7
7.
7
7
7
5
bn s 1 s 2 · · · s n
1,
and by Exercise 32 in Section 4.3, the characteristic polynomial of P 1 LP is ( 1)n p( ). Finally, the
characteristic polynomials of L and P 1 LP are the same since they have the same eigenvalues, so the
characteristic polynomial of L is
det(L
I) = det(P
= ( 1)n (
1
I) = ( 1)n p( )
LP
n
b1
n 1
b2 s 1
n 2
···
bn
1 s1 s2
· · · sn
2
bn s 1 s 2 · · · s n
1)
4.6. APPLICATIONS AND THE PERRON-FROBENIUS THEOREM
331
18. By Exercise 46 in Section 4.4, the eigenvectors of P 1 LP are of the form P 1 x where x is an eigenvalue
of L, and by Exercise 32 in Section 4.3, an eigenvector of P 1 LP for 1 is
2 n 13
1
So an eigenvector of L corresponding to
2 n
6 n1 2 7
6 n 37
6
7
6 1 7
6 . 7.
6 .. 7
6
7
4 1 5
1
is
3
2
1
1
3
n 1
1
s1 n1 2
s1 s2 n1 3
1
n 27
1
7
n 37
1
7
6
6
7
6
6
7
6
6
7
6
6
7
P6 . 7=6
7.
.
..
6 .. 7 6
7
6
7 6
7
4 1 5 4 s1 s2 · · · sn 2 1 5
1
s1 s2 · · · sn 2 sn 1
Finally, we can scale this vector by dividing each entry by
2
1
6
s
/
1
1
6
6
s1 s2 / 21
6
6
..
6
.
6
4 s1 s2 · · · sn 2 / n1 2
s1 s2 · · · sn 2 sn 1 / n1
19. Using a CAS, we find the positive eigenvalue of
2
1
L = 40.7
0
1
0
0.5
n 1
1
to get
3
1
7
7
7
7
7.
7
7
5
3
3
05
0
to be ⇡ 1.7456, so this is the steady state growth rate of the population with this Leslie matrix.
From Exercise 18, a corresponding eigenvector is
2
3 2
3
1
1
4
5 ⇡ 40.4015 .
0.7/1.7456
0.7 · 0.5/(1.7456)2
0.115
So the long-term proportions of the age classes are in the ratios 1 : 0.401 : 0.115.
20. Using a CAS, we find the positive eigenvalue of
2
0
60.5
L=6
40
0
1
0
0.7
0
3
2 5
0 07
7
0 05
0.3 0
to be ⇡ 1.2023, so this is the steady state growth rate of the population with this Leslie matrix.
From Exercise 18, a corresponding eigenvector is
2
3 2
3
1
1
6
7 60.4167
0.5/1.2023
6
7 6
7
4 0.5 · 0.7/(1.2023)2 5 ⇡ 40.2425
0.5 · 0.7 · 0.3/(1.2023)3
0.060
So the long-term proportions of the age classes are in the ratios 1 : 0.416 : 0.242 : 0.06.
332
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
21. Using a CAS, we find the positive eigenvalue
2
0 0.4
60.3 0
6
6 0 0.7
6
0
L=6
60
60
0
6
40
0
0
0
of
1.8
0
0
0.9
0
0
0
1.8 1.8 1.6
0
0
0
0
0
0
0
0
0
0.9 0
0
0 0.9 0
0
0 0.6
3
0.6
07
7
07
7
07
7.
07
7
05
0
to be ⇡ 1.0924, so this is the steady state growth rate of the population with this Leslie matrix.
From Exercise 18, a corresponding eigenvector is
2
3 2
3
1
1
7 60.2757
6
0.3/1.0924
6
7 6
7
2
6
7 60.1767
0.3 · 0.7/(1.0924)
6
7
6
7
6
7 ⇡ 60.1457 .
0.3 · 0.7 · 0.9/(1.0924)3
6
7
6
7
6
7 60.1197
0.3 · 0.7 · 0.9 · 0.9/(1.0924)4
6
7
6
7
4 0.3 · 0.7 · 0.9 · 0.9 · 0.9/(1.0924)5 5 40.0985
0.3 · 0.7 · 0.9 · 0.9 · 0.9 · 0.6/(1.0924)6
0.054
The entries of this vector give the long-term proportions of the age classes.
22. (a) The table gives us 13 separate age classes, each of duration two years; from the table, the Leslie
matrix of the system is
2
3
0
0.02 0.70 1.53 1.67 1.65 1.56 1.45 1.22 0.91 0.70 0.22 0
60.91
0
0
0
0
0
0
0
0
0
0
0
07
6
7
6 0
0.88
0
0
0
0
0
0
0
0
0
0
07
6
7
6 0
0
0.85
0
0
0
0
0
0
0
0
0
07
6
7
6 0
0
0
0.80
0
0
0
0
0
0
0
0
07
6
7
6 0
0
0
0
0.74
0
0
0
0
0
0
0
07
6
7
0
0
0
0
0.67
0
0
0
0
0
0
07
L=6
6 0
7.
6 0
7
0
0
0
0
0
0.59
0
0
0
0
0
0
6
7
6 0
0
0
0
0
0
0
0.49
0
0
0
0
07
6
7
6 0
0
0
0
0
0
0
0
0.38
0
0
0
07
6
7
6 0
0
0
0
0
0
0
0
0
0.27
0
0
07
6
7
4 0
0
0
0
0
0
0
0
0
0
0.17
0
05
0
0
0
0
0
0
0
0
0
0
0
0.15 0
Using a CAS, this matrix has positive eigenvalue ⇡ 1.333, with a corresponding eigenvector
(normalized to add to 100)
2
3
36.1
6 24.65 7
6
7
6 16.27 7
6
7
6 10.38 7
6
7
6 6.23 7
6
7
6 3.46 7
6
7
6 1.74 7 .
6
7
6 0.77 7
6
7
6 0.28 7
6
7
6 0.08 7
6
7
6 0.016 7
6
7
4 0.0021 5
0.00023
4.6. APPLICATIONS AND THE PERRON-FROBENIUS THEOREM
333
(b) In the long run, the percentage of seals in each age group is given by the entries in the eigenvector
in part (a), and the growth rate will be about 1.333.
23. (a) Each term in r is the number of daughters born to a female in a particular age class times the
probability of a given individual surviving to be in that age class, so it represents the expected
number of daughters produced by a given individual while she is in that age class. So the sum
of the terms gives the expected number of daughters ever born to a particular female over her
lifetime.
(b) Define g( ) as suggested. Then g( ) = 1 if and only if
n
g( ) = b1
n
b1
n 1
+ b2 s 1
n 1
b2 s 1
n 2
+ · · · + bn s 1 s 2 · · · s n
n 2
···
bn s 1 s 2 · · · s n
1
1
=
n
= 0.
)
But from Exercise 16, this is true if and only if is a zero of the characteristic polynomial of the
Leslie matrix, indicating that is an eigenvalue of L. So r = g(1) = 1 if and only if = 1 is an
eigenvalue of L, as desired.
(c) For given values of bi and si , it is clear that g( ) is a decreasing function of . Also, r = g(1) and
g( 1 ) = 1. So if 1 > 1, then r = g(1) > g( 1 ) = 1 so that r > 1. On the other hand, if 1 < 1,
then r = g(1) < g( 1 ) = 1, so that r < 1. So we have shown that if the population is decreasing,
then r < 1; if the population is increasing, then r > 1, and (part (b)) if the population is stable,
then r = 1. Since this exhausts the possibilities, we have proven the converses as well. Thus r < 1
if and only if the population is decreasing and r > 1 if and only if the population is increasing.
Note that this corresponds to our intuition, since if the net reproduction rate is less than one, we
would expect the population in the next generation to be smaller.
24. If 1 is the unique positive eigenvalue, then 1 x = Lx for the corresponding eigenvector x. If h is the
sustainable harvest ratio, then (1 h)Lx = x, which means that Lx = 1 1 h x. Thus 1 = 1 1 h , so that
1
h=1
.
1
25. (a) Exercise 21 found the positive eigenvalue of L to be
1
1 1.0924
⇡ 0.0846.
1
⇡ 1.0924, so from Exercise 24, h = 1
1
1
⇡
(b) Reducing the initial population levels by 0.0846 gives
2
x00 = (1
Then
2
0
60.3
6
60
6
0
0
x1 = Lx0 = 6
60
60
6
40
0
0.4
0
0.7
0
0
0
0
3 2
3
10
9.15
6 2 7 6 1.831 7
6 7 6
7
6 8 7 6 7.323 7
6 7 6
7
7 6
7
0.0846) 6
6 5 7 = 6 4.577 7 .
6127 610.9857
6 7 6
7
405 4 0 5
1
0.915
1.8 1.8 1.8 1.6
0
0
0
0
0
0
0
0
0.9 0
0
0
0 0.9 0
0
0
0 0.9 0
0
0
0 0.6
32
3 2
3
0.6
9.15
42.47
6
7 6
7
07
7 6 1.831 7 6 2.75 7
7
6
7
6
0 7 6 7.323 7 6 1.28 7
7
6
7 6
7
07
7 6 4.577 7 = 6 6.59 7 .
7
6
7
6
0 7 610.9857 6 4.12 7
7
0 5 4 0 5 4 9.89 5
0
0.915
0
The population has still grown substantially, because we did not start with a stable population
334
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
vector. If we repeat with the stable population ratios computed in Exercise 21, we get
2
3 2
3
1
0.915
60.2757 60.2527
6
7 6
7
60.1767 60.1617
6
7
6
7
7 6
7
x00 = (1 0.0846) 6
60.1457 = 60.1337 ,
60.1197 60.1097
6
7 6
7
40.0985 40.0905
0.054
0.049
and
2
0
60.3
6
60
6
x01 = Lx00 = 6
60
60
6
40
0
0.4
0
0.7
0
0
0
0
1.8 1.8 1.8 1.6
0
0
0
0
0
0
0
0
0.9 0
0
0
0 0.9 0
0
0
0 0.9 0
0
0
0 0.6
which is essentially x0 .
26. Since
1
⇡ 1.333, we have h = 1
1
=1
1
1
1.333
32
3 2
3
0.6 0.915
0.999
6
7 6
7
07
7 60.2527 60.2757
60.1617 60.1767
07
76
7 6
7
6
7 6
7
07
7 60.1337 = 60.1457 ,
60.1097 60.1197
07
76
7 6
7
0 5 40.0905 40.0985
0
0.049
0.053
⇡ 0.25, which is about one quarter of the population.
27. Suppose 1 = r1 (cos 0 + i sin 0); then r1 = 1 > 0 since 1 is a positive eigenvalue. Let
i sin ✓) be some other eigenvalue. Then g( ) = g( 1 ) = 1. Computing g( ) gives
b1
b2 s 1
bn s 1 s 2 · · · s n 1
+
+ ··· + n
r(cos ✓ + i sin ✓) r2 (cos ✓ + i sin ✓)2
r (cos ✓ + i sin ✓)n
b1
b2 s 1
b2 s 1 s 2 · · · s n 1
= (cos ✓ i sin ✓) + 2 (cos ✓ i sin ✓)2 + · · · +
(cos ✓
r
r
rn
= r(cos ✓ +
1 = g( ) =
where we get to the final equation by multiplying each fraction by
identity (cos ✓ + i sin ✓)(cos ✓
1 = g( ) =
b1
(cos ✓
r
(cos ✓ i sin ✓)k
(cos ✓ i sin ✓)k
= 1 and using the
i sin ✓) = cos2 ✓ + sin2 ✓ = 1. Now apply DeMoivre’s theorem to get
i sin ✓) +
b2 s 1
(cos 2✓
r2
i sin 2✓) + · · · +
b2 s 1 s 2 · · · s n
rn
1
(cos n✓
Since g( ) = 1, the imaginary parts must all cancel, so we are left with
1 = g( ) =
b1
b2 s 1
b2 s 1 s 2 · · · s n
cos ✓ + 2 cos 2✓ + · · · +
r
r
rn
1
cos n✓.
But cos k✓  1 for all k, and bi , si , and r are all nonnegative, so
1 = g( ) 
Since g(
1)
b1
b2 s 1
b2 s 1 s 2 · · · s n
+ 2 + ··· +
r
r
rn
1
.
= 1 as well, we get
b1
b2 s 1
b2 s 1 s 2 · · · s n
+ 2 + ··· +
r1
r1
r1n
Then since r1 > 0, we get r1
i sin ✓)n ,
1

b1
b2 s 1
b2 s 1 s 2 · · · s n
+ 2 + ··· +
r
r
rn
r.
28. The characteristic polynomial of A is
det(A
I) =
2
0
1
1
=(
2)(
1),
1
.
i sin n✓).
4.6. APPLICATIONS AND THE PERRON-FROBENIUS THEOREM
so that the Perron root is
so that

1
335
= 2. To find its Perron eigenvector,


⇥
⇤
0
0 0
1
1
A 2I 0 =
!
1
1 0
0
0
1
is an eigenvector for
1
1
0
,
0
= 2. The entries of the Perron eigenvector must sum to 1, however,
" #
1
2
1
2
so dividing through by 2 gives for the Perron eigenvector
.
29. The characteristic polynomial of A is
det(A
so that the Perron root is
so that

1
I) =
3
= (1
2
)(
)
2
6=
6=(
= 3. To find its Perron eigenvector,
"
#
"
h
i
2
3 0
1
!
A 3I 0 =
2
3 0
0
3)( + 2),
1
3
is an eigenvector for
2
1
3
2
0
#
0
,
0
= 3. The entries of the Perron eigenvector must sum to 1, however,
" #
3
5
2
5
so dividing through by 5 gives for the Perron eigenvector
.
30. The characteristic polynomial of A is
1
det(A
so that the Perron root is
h
1
1
I) =
= 2. To find its
2
2
i
6
2I 0 = 4 1
1
1
A
2 3
1
so that 415 is an eigenvector for
1
1
1
1
( + 1)2 (
=
2),
1
Perron eigenvector,
3
2
1
1 0
1 0
7
6
2
1 05 ! 4 0 1
1
2 0
0 0
1
1
0
3
0
7
05 ,
0
= 2. The entries of the Perron eigenvector must sum to 1, however,
2 3
1
3
6 7
so dividing through by 3 gives for the Perron eigenvector 4 13 5.
1
3
31. The characteristic polynomial of A is
2
det(A
so that the Perron root is
h
A
I) =
1
1
1
= 3. To find its
2
1
i
6
3I 0 = 4 1
1
1
1
0
1
0
=
(
1)(
3),
1
Perron eigenvector,
3
2
1
1 0
1 0
7
6
2
0 05 ! 4 0 1
0
2 0
0 0
2
1
0
3
0
7
05 ,
0
336
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
2 3
2
so that 415 is an eigenvector for
1
1
= 3. The entries of the Perron eigenvector must sum to 1, however,
2 3
1
2
6 7
so dividing through by 4 gives for the Perron eigenvector 4 14 5.
1
4
32. Here n = 4, and
(I + A)n
2
1
60
=6
40
1
1
so A is irreducible.
33. Here n = 4, and
(I + A)n
so A is reducible. Interchanging
required form, so that
2
0 0
60 1
6
40 0
1 0
1
0
1
1
0
2
1
60
=6
41
1
33 2
0
1
63
17
7 =6
41
05
1
3
1
0
1
0
0
1
0
1
3
1
3
1
33 2
0
4
66
17
7 =6
44
05
1
6
1
1
1
0
3
1
37
7 > O,
35
1
3
1
1
3
0
4
0
4
4
6
4
6
3
0
47
7,
05
4
rows 1 and 4 and then columns 1 and 4 produces a matrix in the
3 2
1
0
60
07
7A6
05 40
0
1
0
0
1
0
0
1
0
0
3 2
1
0
61
07
7=6
05 4 0
0
0
0
0
1
0
1
0
0
0
3
1
07
7.
15
0
0
1
0
1
34. Here n = 5, and
(I + A)n
1
2
1
60
6
=6
61
40
1
1
1
0
0
0
0
1
2
1
0
0
0
0
2
0
so that A is reducible. Exchanging rows 1 and
matrix in the required form, so that
2
3 2
0 0 0 1 0
0 0
6 0 1 0 0 07 6 0 1
6
7 6
6 0 0 1 0 07 A 6 0 0
6
7 6
4 1 0 0 0 05 4 1 0
0 0 0 0 1
0 0
and B is a 1 ⇥ 1 matrix while D is 4 ⇥ 4.
35. Here n = 5, and
(I + A)n
so that A is irreducible.
1
2
1
60
6
=6
61
40
0
1
1
0
0
0
0
0
1
1
0
0
0
0
1
1
34 2
0
11
622
17
7
6
6
17
7 = 628
5
423
0
1
6
6
11
16
7
6
11
17
23
33
5
0
0
0
16
0
3
11
177
7
227
7,
185
6
4 and then exchanging columns 1 and 4 produces a
0
0
1
0
0
1
0
0
0
0
3 2
0
1
60
07
7 6
6
07
7 = 60
5
40
0
1
0
34 2
0
1
61
17
7
6
6
17
7 = 65
5
46
0
2
5
4
1
6
4
1
0
0
0
1
0
1
5
6
5
11
1
1
1
0
0
0
0
1
0
1
3
0
17
7
17
7,
05
0
3
5 11
11 167
7
12 217
7,
6 125
16 22
4.6. APPLICATIONS AND THE PERRON-FROBENIUS THEOREM
337
36. (a) Let G be a graph and A its adjacency matrix. Let G0 be the graph obtained from G by adding
an edge from each vertex to itself if no such edge exists in G. Then G0 is connected if and only if
G is, since every vertex is always connected to itself. The adjacency matrix A0 of G results from
A by replacing every 0 on the diagonal by a 1. Now consider the matrix A + I. This matrix is
identical to A0 except that some diagonal entries of A0 equal to 1 may be equal to 2 in A + I (if
the vertex in question had an edge from it to itself in G). But since all entries in A + I and A0
are nonnegative, (A0 )n 1 > O if and only (A + I)n 1 > O — whether those diagonal entries are
1 or 2 does not a↵ect whether a particular entry in the power is zero or not. So G is connected
if and only if G0 is connected, which is the case if and only if (A0 )n 1 of its adjacency matrix is
> O, so that there is a path between any two vertices of length n 1 (note that any shorter path
can be made into a path of length n 1 by adding traverses of the edge from the vertex to itself).
But (A0 )n 1 > O then means that (A + I)n 1 > O, so that A is irreducible.
(b) Since all the graphs in Section 4.0 are connected, they all have irreducible adjacency matrices.
The graph in Figure 4.1 has a primitive adjacency matrix, since
2
32 2
3
0 1 1 1
3 2 2 2
6 1 0 1 17
6 2 3 2 27
7
6
7
A2 = 6
41 1 0 15 = 42 2 3 25 > O.
1 1 1 0
2 2 2 3
So does the graph in Figure 4.2:
2
0 1 0 0 1 0 1 0
61 0 1 0 0 0 0 1
6
60 1 0 1 0 0 0 0
6
60 0 1 0 1 0 0 0
6
61 0 0 1 0 1 0 0
4
A =6
60 0 0 0 1 0 0 1
6
61 0 0 0 0 0 0 0
6
60 1 0 0 0 1 0 0
6
40 0 1 0 0 1 1 0
0 0 0 1 0 0 1 1
0
0
1
0
0
1
1
0
0
0
34 2
0
15
64
07
7
6
69
07
7
6
69
17
6
7
6
07
7 =64
69
07
7
6
64
17
7
6
69
17
7
6
49
05
0
9
4
15
4
9
9
9
9
4
9
9
9
4
15
4
9
9
9
9
4
9
The graph in Figure 4.3 has a primitive adjacency matrix:
2
34 2
0 1 0 0 1
6 1 4
6 1 0 1 0 07
61 6 1
6
7
6
7
6
A4 = 6
6 0 1 0 1 07 = 6 4 1 6
4 0 0 1 0 15
44 4 1
1 0 0 1 0
1 4 4
However, the adjacency matrix for the graph is
2
0 0
60 0
6
60 0
A=6
61 1
6
41 1
1 1
9
9
4
15
4
9
9
9
9
4
Figure 4.4 is
3
0 1 1 1
0 1 1 17
7
0 1 1 17
7,
1 0 0 07
7
1 0 0 05
1 0 0 0
4
4
1
6
1
4
9
9
4
15
4
9
9
9
9
9
9
9
9
4
15
9
4
4
9
4
9
9
9
9
9
15
9
4
4
9
4
9
9
9
4
9
15
9
4
9
9
4
9
9
4
4
9
15
9
3
9
97
7
97
7
47
7
97
7>O
97
7
47
7
47
7
95
15
3
1
47
7
47
7 > O.
15
6
and any power of A will always have a 3 ⇥ 3 block of zeroes in either the upper left or upper right,
so A is not primitive.
37. (a) Suppose G is bipartite with disjoint vertex sets U and V and A is its adjacency matrix. Recall
that (Ak )ij gives the number of paths between vertices i and j. Also, from Exercise 80 in Section
3.7, if u 2 U and v 2 V , then any path from u to itself will have even length, and any path from
338
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
u to v will have odd length, since in any path, each successive edge moves you from U to V and
back. So odd powers of A will have zero entries along the diagonal, while even powers of A will
have zero entries in any cell corresponding to a vertex in U and a vertex in V . Since every power
of A has a zero entry, A is not primitive.
(b) Number the vertices so that v1 , v2 , . . . , vk are in U and vk+1 , vk+2 , . . . , vn are in V . Then the
adjacency matrix of G is

O B
A=
.
BT O

v
Write the eigenvector v corresponding to as 1 , where v1 is a k-element column vector and
v2
v2 is an (n k)-element column vector. Then

⇥
⇤
v1
T
Av = v () Bv2 B v1 =
,
v2
so that Bv2 = v1 and B T v1 = v2 . Then




Bv2
v1
v1
v1
=
=
=
A
.
v2
v2
v2
B T v1
This shows that
is an eigenvalue of A with corresponding eigenvector

v1
.
v2
38. (a) Since each vertex of G connects to k other vertices, we see that each column of the adjacency
matrix A of G sums to k. Thus A = kP , where P is a matrix each of whose columns sums to
1. So P is a transition matrix, and by Theorem 4.30, 1 is an eigenvalue of P , so that k is an
eigenvalue of kP = A.
(b) If A is primitive, then Ar > O for some r, and therefore P r > O as well. Therefore P is regular,
so by Theorem 4.31, every other eigenvalue of P has absolute value strictly less than 1, so that
every other eigenvalue of A = kP has absolute value strictly less than k.
39. This exercise is left to the reader after doing the exploration suggested in Section 4.0.
40. (a) Let A = [aij ]. Then cA = [caij ], so that
|cA| = [|caij |] = [|c| · |aij |] = |c| [|aij |] = |c| · |A| .
(b) Using the Triangle Inequality,
|A + B| = [|aij + bij |]  [|aij | + |bij |] = [|aij |] + [|bij |] = |A| + |B| .
(c) Ax is the n ⇥ 1 matrix whose ith entry is
ai1 x1 + ai2 x2 + · · · + ain xn .
Thus the i
th
entry of |Ax| is
|(Ax)i | = |ai1 x1 + ai2 x2 + · · · + ain xn |
 |ai1 x1 | + |ai2 x2 | + · · · + |ain xn |
= |ai1 | |x1 | + |ai2 | |x2 | + · · · + |ain | |xn | .
But this is just the ith entry of |A| |x|.
(d) The ij entry of |AB| is
|AB|ij = |ai1 b1j + ai2 b2j + · · · + ain bnj |
 |ai1 b1j | + |ai2 b2j | + · · · + |ain bnj |
= |ai1 | |b1j | + |ai2 | |b2j | + · · · + |ain | |bnj | ,
which is the ij entry of |A| |B|.
4.6. APPLICATIONS AND THE PERRON-FROBENIUS THEOREM
339
41. A matrix is reducible if, after permuting rows and columns according to the same permutation, it can
be written

B C
,
O D
where B and D are square and O is the zero matrix. If A is 2 ⇥ 2, then B, C, O, and D must all be
1 ⇥ 1 matrices (which are therefore just the entries of A). If A is in the above form after applying no
permutations, leaving A alone, then a21 = 0. If A is in this form after applying the only other possible
permutation — exchanging the two rows and the two columns — then the resulting matrix is

a22 a21
,
a12 a11
so that a12 = 0. This proves the result.
42. (a) If 1 and v1 are the Perron eigenvalue and eigenvector respectively, then by Theorem 4.37, 1 > 0
and v1 > 0. By Exercise 22 in Section 4.3, 1 1 is an eigenvalue, with eigenvector v1 , for
A I, and therefore 1
A, again with eigenvector v1 . Since I A is
1 is an eigenvalue for I
invertible, Theorem 4.18 says that 1 1 1 is an eigenvalue for (I A) 1 , again with eigenvector v1 .
But (I A) 1 0 and v1 0, so that
1
1
1
v1 = (I
1
A)
v1
0.
It follows that 1 1 1
0 and thus 1 < 1. Since we also have
(b) Since 0 < 1 < 1, v1 > 1 v1 = Av1 .
1
> 0, the result follows: 0 <
1
< 1.
43. x0 = 1, x1 = 2x0 = 2, x2 = 2x1 = 4, x3 = 2x2 = 8, x4 = 2x3 = 16, x5 = 2x4 = 32.
44. a1 = 128, a2 =
a1
2
= 64, a3 =
45. y0 = 0, y1 = 1, y2 = y1
a2
2
= 32, a4 =
y0 = 1, y3 = y2
a3
2
= 16, a5 =
a4
2
y1 = 0, y4 = y3
= 8.
y2 =
1, y5 = y4
y3 =
1.
46. b0 = 1, b1 = 1, b2 = 2b1 + b0 = 3, b3 = 2b2 + b1 = 7, b4 = 2b3 + b2 = 17, b5 = 2b4 + b3 = 41.
47. The recurrence is xn 3xn 1 4xn 2 = 0, so the characteristic equation is 2 3 4 = ( 4)( +1) = 0.
Thus the system has eigenvalues 1 and 4, so xn = c1 ( 1)n + c2 · 4n , where c1 and c2 are constants.
Using the initial conditions, we have
0 = x0 = c1 ( 1)0 + c2 · 40 = c1 + c2 ,
5 = x1 = c1 ( 1)1 + c2 · 41 =
c1 + 4c2 .
Add the two equations to solve for c2 , then substitute back in to get c1 =
xn = ( 1)n + 4n = 4n + ( 1)n 1 .
1 and c2 = 1, so that
48. The recurrence is xn 4xn 1 +3xn 2 = 0, so the characteristic equation is 2 4 +3 = ( 3)( 1) = 0.
Thus the system has eigenvalues 1 and 3, so xn = c1 · 1n + c2 · 3n , where c1 and c2 are constants. Using
the initial conditions, we have
0 = x 0 = c 1 · 10 + c 2 · 30 = c 1 + c 2 ,
1 = x1 = c1 · 11 + c2 · 31 = c1 + 3c2 .
Subtract the two equations to solve for c2 , then substitute back in to get c1 =
xn = 12 · 1n + 12 · 3n = 12 (3n 1).
1
2
and c2 = 12 , so that
49. The recurrence is yn 4yn 1 + 4yn 2 = 0, so the characteristic equation is 2 4 + 4 = (
2)2 = 0.
n
n
Thus the system has the double eigenvalue 2, so yn = c1 · 2 + c2 · n2 , where c1 and c2 are constants.
Using the initial conditions, we have
1 = y1 = c1 · 21 + c2 · 1 · 21 = 2c1 + 2c2 ,
6 = y2 = c1 · 22 + c2 · 2 · 22 = 4c1 + 8c2 .
Solving the system gives c1 =
1
2
and c2 = 1, so that yn =
1
2
· 2n + n · 2n = n2n
2n
1
.
340
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
50. The recurrence is an an 1 + 14 an 2 = 0, so the characteristic equation is 14 (4 2 4 +1) = 14 (2 1)2 =
0. Thus the system has the double eigenvalue 12 = 2 1 , so an = c1 · 2 n + c2 · n2 n , where c1 and c2
are constants. Using the initial conditions, we have
4 = a0 = c1 · 2
0
+ c2 · 0 · 2
0
1 = a1 = c1 · 2
1
+ c2 · 1 · 2
1
Solving the system gives c1 = 4 and c2 =
= c1 ,
1
1
= c1 + c2 .
2
2
2, so that an = 4 · 2
n
2n · 2
n
= 22
n
n21
n
.
51. The recurrence is bn 2bn 1 2bn 2 = 0, so the characteristic
equation is 2 2
2 = 0. Using
p
the quadraticpformula, this equation
has
roots
=
1
±
3,
which
are
therefore
the
eigenvalues.
So
p n
bn = c1 · (1 + 3)n + c2 · (1
3) , where c1 and c2 are constants. Using the initial conditions, we have
p
p 0
0 = b0 = c1 · (1 + 3)0 + c2 · (1
3) = c1 + c2 ,
p 1
p 1
p
1 = b1 = c1 · (1 + 3) + c2 · (1
3) = c1 + c2 + (c1 c2 ) 3.
p
From the first equation, c1 + c2 = 0, so the second becomes 1 = (c1 c2 ) 3, so that c1
p
p
3
Since c2 = c1 , we get c1 = 63 and c2 =
6 , so that the recurrence is
p
p
p ⇣
p n
p n
p
p n⌘
3
3
3
bn =
(1 + 3)
(1
3) =
(1 + 3)n (1
3) .
6
6
6
c2 =
p
3
3 .
52. The recurrence is yn yn 1 + yn 2 = 0, so the characteristic
equation is 2
+ 1 = 0. Using the
p
1
quadratic formula, this equation has roots = 2 (1 ± 3 i), which are therefore the eigenvalues. So
p
p
n
n
yn = c1 · 21 (1 + 3 i) +c2 · 12 (1
3 i) , where c1 and c2 are constants. Using the initial conditions,
we have
✓
◆0
✓
◆0
p
p
1
1
0 = y0 = c 1 ·
(1 + 3 i) + c2 ·
(1
3 i) = c1 + c2 ,
2
2
✓
◆1
✓
◆1
p
p
p
p
1
1
1
1
1 = y1 = c 1 ·
(1 + 3 i) + c2 ·
(1
3 i) = c1 (1 + 3 i) + c2 (1
3 i).
2
2
2
2
p
p
From the first equation, c1 + c2 = 0, so the second becomes 2 = (c1 c2 ) 3 i, so that c1 c2 = 2 3 3 i.
p
p
3
3
Since c2 = c1 , we get c1 =
3 i and c2 = 6 , so that the recurrence is
p ✓
◆n p ✓
◆n
p
p
3 1
3 1
yn =
i
(1 + 3 i)
+
i
(1
3 i) .
3
2
3
2
53. Working with the matrix A from the Theorem, since it has distinct eigenvalues
diagonalized, so that there is some invertible P such that


e f
0
1
P =
,
P AP = D = 1
.
g h
0
2
First consider the case where one of the eigenvalues, say
of the form

a 0
A=
,
1 0
2,
1
6=
2,
it can be
is zero. In this case the matrix must be
so that the other eigenvalue is a. In this case, the recurrence relation is xn = axn 1 + 0xn 2 = axn 1 ,
so if we are given xj and xj 1 as initial conditions, we have xn = an j xj . So take c1 = xj a j and
c2 = 0; then xn = c1 n1 + c2 n2 as desired. Now suppose that neither eigenvalue is zero. Then

 k


1
1
e f
0
h
f
eh k1 f g k2
ef k1 + ef k2
1
Ak = P D k P 1 =
=
.
k
k
k
g e
0
f g k1 + eh k2
eh f g g h
eh f g gh 1 gh 2
2
4.6. APPLICATIONS AND THE PERRON-FROBENIUS THEOREM
341
If we are given two terms of the sequence, say xj 1 and xj , then for any n we have

xn
xn =
= An j x j
xn 1


1
eh n1 j f g n2 j
ef n1 j + ef n2 j
xj
=
eh f g gh n1 j gh n2 j
f g n1 j + eh n2 j xj 1
2⇣
⌘
⇣
⌘
3
eh n1 j f g n2 j xj +
ef n1 j + ef n2 j xj 1
1
4⇣
⌘
⇣
⌘
5
=
eh f g
gh n1 j gh n2 j xj +
f g n1 j + eh n2 j xj 1
2⇣
⌘
⇣
⌘
(ehxj ef xj 1 ) 1 j n1 + ( f gxj + ef xj 1 ) 2 j
1
4⇣
⌘
⇣
⌘
=
eh f g
(ghxj f gxj 1 ) 1 j n1 + ( ghxj + ehxj 1 ) 2 j
n
2
n
2
Looking at the first row, we see that for
c1 =
ehxj ef xj
eh f g
1
j
1
,
f gxj + ef xj
eh f g
c2 =
n
1
c1 and c2 are independent of n, and we have xn = c1
n
2,
+ c2
1
j
2
3
5.
,
as required.
54. (a) First suppose that the eigenvalues are distinct, i.e., that 1 6= 2 . Then Exercise 53 gives one
formula for c1 and c2 involving the diagonalizing matrix P . But since we now know that c1 and c2
exist, we can derive a simpler formula. We know the solutions are of the form xn = c1 n1 + c2 n2 ,
valid for all n. Since x0 = r and x1 = s, we get the following linear system in c1 and c2 :
0
1 c1
1
1 c1
+
+
0
2 c2
1
2 c2
=r
or
c1 +
=s
1 c1
+
c2 = r
2 c2
= s.
Solving this system, for example by setting up the augmented matrix and row-reducing, gives a
unique solution
s
s
1r
1r
c1 = r
,
c2 =
.
2
1
2
1
So the system has a unique solution in terms of x0 = r, x1 = s, 1 , and 2 . Note that the system
has a unique solution (and these expressions make sense) since 1 6= 2 .
Next suppose that the eigenvalues are the same, say 1 = 2 = . Then the solutions are of
the form xn = c1 n + c2 n n . Note that we must have 6= 0, since otherwise the characteristic
polynomial of the recurrence matrix is 2 , so the recurrence relation is xn = 0. With x0 = r and
x1 = s, we get the following linear system in c1 and c2 :
0
1
c1 + 0 ·
c1 + 1 ·
0
c2 = r
1
c2 = s
or
c1
=r
c1 + c2 = s.
Thus c1 = r, and substituting into the second equation gives (since 6= 0) c2 = s r . So in this
case as well we can find the scalars c1 and c2 uniquely. Finally, note that 0 is an eigenvalue of the
recurrence matrix only when b = 0 in the recurrence relation xn = axn 1 + bxn 2 .
(b) From part (a), the general solution to the recurrence is
✓
◆
s
1r
n
n
xn = c 1 1 + c 2 2 = r
2
Substituting x0 = r = 0 and x1 = s = 1 gives
✓
◆
✓
◆
1
1
1·0
1·0
n
xn = 0
+
1
2
1
2
1
n
2
1
n
1
1
=
2
+
(
1
✓
s
1r
2
n
1
+
1
n
2)
◆
n
2.
1
=
1
(
2
n
1
n
2 ).
342
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
55. (a) For n = 1, since f0 = 0 and f1 = f2 = 1, the equation holds:


1 1
f f1
A1 = A =
= 2
.
1 0
f1 f0
Now assume that the equation is true for n = k, and consider Ak+1 :




f
fk
1 1
f
+ fk fk+1
f
Ak+1 = Ak A = k+1
= k+1
= k+2
fk
fk 1 1 0
fk + fk 1
fk
fk+1
as desired.
(b) Since det A =
fk+1
fk
n
1, it follows that det(An ) = (det A) = ( 1)n . But from part (a),
det(An ) = fn+2 fn
2
fn+1
= ( 1)n .
(c) If we look at the 5 ⇥ 13 “rectangle” as being composed of two “triangles”, we see that in fact they
are not triangles: the “diagonal” of the “rectangle” is not straight because the 8 ⇥ 3 triangular
5
pieces are not similar to the entire triangles, since 13
6= 38 . So there is empty space along the
diagonal of the figure that adds up to the missing 1 square unit.
56. (a) t1 = 1 (the only way is using the 1 ⇥ 1 rectangle); t2 = 3 (two 1 ⇥ 1 rectangles or either of the
1 ⇥ 2 rectangles); t3 = 5; t4 = 11; t5 = 21. For t0 , we can think of that as the number of ways to
tile an empty rectangle; there is only one way — using no rectangles.
(b) The rightmost rectangle in a tiling of a 1 ⇥ n rectangle is either the 1 ⇥ 1 or one of the 1 ⇥ 2
rectangles. So the number of tilings of a 1 ⇥ n rectangle is the number of tilings of a 1 ⇥ (n 1)
rectangle (adding the 1 ⇥ 1) plus twice the number of tilings of a 1 ⇥ (n 2) rectangle (adding
either 1 ⇥ 2). So the recurrence relation is
tn = tn
(c) The characteristic equation is
find the constants, we solve
2
1
+ 2tn
2=(
2.
2)( + 1), so the eigenvalues are
t1 = 1 = c1 ( 1)1 + c2 · 21 =
2
1 and 2. To
c1 + 2c2
2
t2 = 3 = c1 ( 1) + c2 · 2 = c1 + 4c2 .
This gives c1 =
1
3
and c2 = 23 , so that
tn =
1
2
1
( 1)n + · 2n =
( 1)n + 2n+1 .
3
3
3
Evaluating for various values of t gives
1
( 1)0 + 2 = 1,
3
1
t2 =
( 1)2 + 23 = 3,
3
1
t4 =
( 1)4 + 25 = 11,
3
1
( 1)1 + 22 = 1,
3
1
t3 =
( 1)3 + 24 = 5,
3
1
t5 =
( 1)5 + 26 = 21.
3
t0 =
t1 =
These agree with part (a).
57. (a) d1 = 1, d2 = 2, d3 = 3, d4 = 5, d5 = 8. As in Exercise 56, d0 can be interpreted as the number of
ways to cover a 2 ⇥ 0 rectangle, which is one — use no dominos.
(b) The right end of any tiling of a 2 ⇥ n rectangle either consists of a domino laid vertically, or of
two dominos each laid horizontally. Removing those leaves either a 2 ⇥ (n 1) or a 2 ⇥ (n 2)
rectangle. So the number of tilings of a 2 ⇥ n rectangle is equal to the number of tilings of a
2 ⇥ (n 1) rectangle plus the number of tilings of a 2 ⇥ (n 2) rectangle, or
dn = dn
1
+ dn
2.
4.6. APPLICATIONS AND THE PERRON-FROBENIUS THEOREM
343
(c) We recognize the recurrence as the recurrence for the Fibonacci sequence; the initial values shift
it by 1, so we get
p !n+1
p !n+1
1
1
1+ 5
1
5
p
dn = p
.
2
2
5
5
Computing values using this formula gives the same values as in part (a), including the value for
d0 .
58. Row-reducing gives
h
h
A
p
1+ 5
2 I
A
1
p
5
2
"
i
1
0 =
p
1+ 5
2
I
So the eigenvectors are
v1 =
"
p
1
5
2
1
p #
1+ 5
2
"
and v2 =
1
"
p
1
5
2
1
#
p
1
1
p
!
1+ 5
0
2
#
"
0p
1
!
1
5
0
2
1
"
i
1
0 =
#
0
5
2
0
#
p #
1+ 5
2
.
0
.
Then
2
lim
xk
k
1
k!1
= lim
k!1
p
1 p5
1+ 5
2
6
= lim 4
k!1
Since
p ⌘k
1+ 5
2
6
4 ⇣ p ⌘k 1
1+
5
1
p
2
5
p1
5
⇣
p1
5
p1
5
·
p1
5
p1
5
⇣ p ⌘k
1+ 5
2
⇣
p1
5
2p
1+ 5
p1
5
·
⇣
⇣
1
p
5
1
2
p
5
2
⌘k
⌘k
3
7
15
3
p ⌘k
1 p5
7
1+ 5
p ⇣
p ⌘k 5 .
1
5 1 p5
2
1+ 5
⇡ 0.381 < 1, the terms involving k th powers of this fraction go to zero as k ! 1, leaving
us with (in the limit)
"
p1
5
p 2 p
5(1+ 5)
#
2
p
=p
5(1 + 5)
"
p #
1+ 5
2
1
=p
p
2
5
5
p v1 =
v1 ,
10
5(1 + 5)
which is a constant times v1 .
59. The coefficient matrix has characteristic polynomial

1
3
= (1
)(2
) 6=
2
2
so that the eigenvalues are
reduction:
⇥
A
⇥
4I
A+I
Then by Theorem 4.40,
1
= 4 and
⇤
0 =


⇤
2
0 =
2
2
3
2
=
2

0
1
!
0
0

0
1 32
!
0
0 0
x = C1 e4t
4=(
4)( + 1),
1. The corresponding eigenvectors are found by row-
3
2
3
3
3

1
+ C2 e
1
1
0
0
0
t
0
0
)

3
.
2

1
v1 =
1

3
v2 =
.
2
)
344
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
Substitute t = 0 to determine the constants:




0
1
3
C1 + 3C2
x(0) =
= C1 e0
+ C2 e0
=
.
5
1
2
C1 2C2
These two equations give C1 = 3 and C2 =
x = 3e4t

1
1
e
1, so that the solution is

t
x(t) = 3e4t 3e t
3
, or
2
y(t) = 3e4t + 2e t .
60. The coefficient matrix has characteristic polynomial

2
1
= (2
)2 1 =
1
2
so that the eigenvalues are
reduction:
⇥
⇥
A
A
I
3I
Then by Theorem 4.40,
1
= 1 and
⇤
0 =
⇤
0 =


2
1
1
1
1
1
1
1
1
2
4 +3=(
1)(
3),
= 3. The corresponding eigenvectors are found by row
0
1
1 0
!
0
0
0 0

0
1 1 0
!
0
0 0 0
x = C1 e
t

1
+ C2 e3t
1

)
)

1
v1 =
1

1
v2 =
.
1
1
.
1
Substitute t = 0 to determine the constants:




1
1
C1 + C2
0 1
0
x(0) =
= C1 e
+ C2 e
=
.
1
1
1
C1 C2
These two equations give C1 = 1 and C2 = 0, so that the solution is
x=e
t

1
+ 0 · e3t
1

x(t) = et
1
, or
1
y(t) = et .
61. The coefficient matrix has characteristic polynomial

1
1
= (1
)( 1
) 1=
1
1
so that the eigenvalues are
row-reduction:

⇥
⇤
p
1
A
2I 0 =
⇥
A+
p
2I
1
=
p
2 and
2
1
p
⇤
1+ 2
0 =
1

p
1
2
1p
2
1p
1+ 2
Then by Theorem 4.40,
x = C1 e
p
2t

=
p
2
p
2=(
2 )( +
p
2 ),
2. The corresponding eigenvectors are found by

0
1
!
0
0

0
1
!
0
0
p
1+ 2
+ C2 e
1
1
0
1+
0
p
2t
p
2
p
0
0
)
2
0
0
)

1
p
1
2

p
1+ 2
1

p
1
2
v2 =
.
1
v1 =
.
Substitute t = 0 to determine the constants:




p
p
p
1
1+ 2
1
2
C1 (1 + 2) + C2 (1
x(0) =
= C1 e0
+ C2 e0
=
0
1
1
C1 + C2
p
2)
.
4.6. APPLICATIONS AND THE PERRON-FROBENIUS THEOREM
345
p
The second equation
gives C1 + C2 = 0; substituting
into pthe first equation gives (C1 C2 ) 2 = 1, so
p
p
2
that C1 C2 = 22 . It follows that C1 = 42 and C2 =
4 , so that the solution is
p
p
2 + 2 p2 t 2
2 p2 t
p
p


p
p
x
(t)
=
e
+
e
p
p
1
2 2t 1 + 2
2
1
2
4
x=
e
e 2t
, or
p 4⇣
⌘
p
1
1
4
4
2 p2 t
x2 (t) =
e
e 2t .
4
62. The coefficient matrix has characteristic polynomial

1
1
= (1
)2 + 1 =
1
1
so that the eigenvalues are
row-reduction:
⇥
⇥
= 1 + i and
1
A
(1 + i)I
A
(1
⇤
0 =


⇤
i
0 =
1
i)I
Then by Theorem 4.40,
2
i
1
=1
1
i
1
i
y = C1 e(1+i)t
2
2 + 2,
i. The corresponding eigenvectors are found by


0
1
i 0
i
!
) v1 =
0
0
0 0
1


0
1 i 0
i
!
) v2 =
.
0
0 0 0
1

i
+ C2 e(1
1
i)t

i
.
1
Substitute t = 0 to determine the constants:




1
i
(C1 C2 )i
0 i
0
y(0) =
= C1 e
+ C2 e
=
.
1
1
1
C1 + C2
Solving gives C1 =
1 i
2 ,
C1 =
or
i
so that the solution is

1 i (1+i)t i
1 + i (1
y=
e
+
e
1
2
2
i)t

i
,
1
1 + i (1 i)t
i + 1 (1+i)t 1 i (1
e
· ( i) =
e
+
e
2
2
2
1 i (1+i)t 1 + i (1 i)t
y2 (t) =
e
+
e
.
2
2
y1 (t) =
1
1+i
2 ,
2
e(1+i)t i +
Now substitute cos t + i sin t for eit , giving
i + 1 (1+i)t 1 i (1 i)t
e
+
e
2
2
i+1 t
1 i t
=
e (cos t + i sin t) +
e (cos t
2
2◆
✓
i 1
i 1 t
= et cos t +
+
e sin t
2
2
y1 (t) =
= et (cos t sin t)
1 i (1+i)t 1 + i (1 i)t
y2 (t) =
e
+
e
2
2
1 i t
1+i t
=
e (cos t + i sin t) +
e (cos t
2
✓
◆2
1+i 1 i t
= et cos t +
+
e sin t
2
2
= et (cos t + sin t).
i sin t)
i sin t)
i)t
346
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
63. The coefficient matrix has characteristic polynomial
2
3
1
1
3
4 1
1 5=
+
1
1
so that the eigenvalues are
row-reduction:
⇥
A
⇥
⇥
A
0I
I
A+I
Then by Theorem 4.40,
1
= 0,
2
0 1
⇤
0 = 41 0
1 1
2
1
⇤
0 =4 1
1
2
1 1
⇤
0 = 41 1
1 1
2
= 1, and
1
1
0
1
1
1
1
1
1
3
=
=
(1
)(1 + ),
1. The corresponding eigenvectors are found by
3
2
3
1 0
1 0
0
05 ! 40 1
1 05
0
0 0
0 0
3
2
1 0
1 0
0
1 05 ! 4 0 1
1
1 0
0 0
0
3
2
3
0
1 1 0 0
05 ! 40 0 1 05
0
0 0 0 0
2
3
2 3
1
0
x = C1 e0t 4 15 + C2 et 415 + C3 e
1
1
2
3
1
) v1 = 4 15
1
3
2 3
0
0
05 ) v2 = 415
0
1
2 3
1
) v 2 = 4 15 .
0
2
3
1
t4
15 .
0
Substitute t = 0 to determine the constants:
2 3
2 3
2 3
2 3 2
3
1
1
0
1
C1 C3
x(0) = 4 05 = C1 e0 4 15 + C2 e0 415 + C3 e0 4 15 = 4C1 + C2 + C3 5
1
1
1
0
C1 + C2
Solving gives C1 =
or
2, C3 = 1, C2 = 1, so that the solution is
2 3
2 3
2 3
1
0
1
x = 2 4 15 + et 415 + e t 4 15
1
1
0
x(t) = 2
e
t
y(t) =
2 + et + e
z(t) =
2 + et .
64. The coefficient matrix has characteristic polynomial
2
3
1
0
3
4 1
2
1 5=
3
0
(
t
4)( + 2)2
so that the eigenvalues are 1 = 4 and 2 = 2 (with algebraic multiplicity 2). The corresponding
eigenvectors are found by row-reduction:
2
3
2
3
2 3
1 0
1 0
3
0
3 0
3
⇥
⇤
1
5 ) v1 = 415
A 4I 0 = 4 1
6
1 05 ! 4 0 1
0
3
3
0
3 0
3
0 0
0 0
2
3
2
3
2 3
2 3
3 0 3 0
1 0 1 0
0
1
⇥
⇤
A + 2I 0 = 41 0 1 05 ! 40 0 0 05 ) v2 = 415 , v3 = 4 05 .
3 0 3 0
0 0 0 0
0
1
4.6. APPLICATIONS AND THE PERRON-FROBENIUS THEOREM
347
Then by Theorem 4.40,
2 3
3
x = C1 e4t 415 + C2 e
3
2 3
0
2t 4 5
1 + C3 e
0
2
3
1
2t 4
05 .
1
Substitute t = 0 to determine the constants:
2 3
2 3
2 3
2 3 2
3
2
3
0
1
3C1 C3
x(0) = 435 = C1 e0 415 + C2 e0 415 + C3 e0 4 05 = 4 C1 + C2 5
4
3
0
1
3C1 + C3
Solving gives C1 = 1, C2 = 2, C3 = 1, so that the solution is
2 3
2 3
3
0
x = e4t 415 + 2e 2t 415 + e
3
0
or
x(t) = 3e4t
e
2t
y(t) = e4t + 2e
2t
z(t) = 3e4t + e
2t
65. (a) The coefficient matrix has characteristic polynomial

1.2
0.2
= (1.2
)(1.5
) 0.04 =
0.2
1.5
so that the eigenvalues are
row-reduction:
⇥
⇥
A
1.1I
A
1.6I
Then by Theorem 4.40,
1
= 1.1 and

⇤
0.1
0 =
0.2

⇤
0.4
0 =
0.2
2
x = C1 e
3
1
2t 4
05
1
.
2
2.7 + 1.76 = (

0
1
!
0
0

0.2 0
1
!
0.1 0
0

2
0
0
0
)
v1 =
1
2
0
0
)
v2 =
0
2
+ C2 e1.6t
1

1
.
2
Substitute t = 0 to determine the constants:




400
2
1
2C1 C2
x(0) =
= C1 e0
+ C2 e0
=
.
500
1
2
C1 + 2C2
Solving gives C1 = 260, C2 = 120, so that the solution is


1
1.1t 2
1.6t
x = 260e
+ 120e
1
2
or
x(t) = 520e1.1t
120e1.6t
y(t) = 260e1.1t + 240e1.6t .
A plot of the populations over time is
1.1)(
1.6)
= 1.6. The corresponding eigenvectors are found by
0.2
0.4
1.1t
2

2
1

1
.
2
348
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
Population
15 000
X
10 000
5000
Y
0
1.0
0.5
2.0
1.5
2.5
3.0
Time !days"
From the graph, the population of Y dies out after about three days.
(b) Substituting a and b for 400 and 500 in the initial value computation above gives




a
2
1
2C1 C2
x(0) =
= C1 e0
+ C2 e0
=
.
b
1
2
C1 + 2C2
Solving gives C1 = 15 (2a + b), C2 = 15 ( a + 2b), so that the solution is


1
1
1
1.1t 2
1.6t
x = (2a + b)e
+ ( a + 2b)e
1
2
5
5
or
2
1
x(t) = (2a + b)e1.1t
( a + 2b)e1.6t
5
5
1
2
y(t) = (2a + b)e1.1t + ( a + 2b)e1.6t .
5
5
In the long run, the e1.6t term will dominate, so if a > 2b, then a + 2b < 0, so the coefficient
of e1.6t in x(t) is positive and in y(t) is negative, so that Y will die out and X will dominate.
If a < 2b, then a + 2b > 0, so the reverse will happen — X will die out and Y will dominate.
Finally, if a = 2b, then the e1.6t term disappears in the solutions, and both populations will grow
at a rate proportional to e1.1t , though one will grow twice as fast as the other.
66. The coefficient matrix has characteristic polynomial

0.8
0.4
= ( 0.8
)( 0.2
0.4
0.2
)
0.16 =
2
+
= ( + 1)
so that the eigenvalues are 1 = 0 and 2 = 1. The corresponding eigenvectors are found by rowreduction:



1
⇥
⇤
0.8
0.4 0
1
1
0
2
A 0I 0 =
!
) v1 =
0.4
0.2 0
2
0
0 0



⇥
⇤
0.2 0.4 0
1 2 0
2
A+I 0 =
!
) v2 =
.
0.4 0.8 0
0 0 0
1
Then by Theorem 4.40,
x = C1 e
0t

1
+ C2 e
2
t

2
.
1
Substitute t = 0 to determine the constants:




15
2
C1 2C2
0 1
0
x(0) =
= C1 e
+ C2 e
=
.
10
2
1
2C1 + C2
Solving gives C1 = 7, C2 =
or
4, so that the solution is

1
x=7
4e
2
t
x(t) = 7 + 8e
y(t) = 14

2
1
t
4e t .
As t ! 1, the exponential terms go to zero, and the populations stabilize at 7 of X and 14 of Y .
4.6. APPLICATIONS AND THE PERRON-FROBENIUS THEOREM
349
67. We want to convert x0 = Ax + b into an equation u0 = Au. If we could write b = Ab0 , then letting
u = x + b0 would do the trick. To find b0 , we invert A:
"
#"
# "
#
1
1
30
10
0
1
2
2
=
.
b =A b= 1
1
10
20
2
2

10
Then let u = x +
, so that
20
✓

◆0

✓

◆
10
10
10
u0 = x +
= x0 = Ax + b = Ax + AA 1
=A x+A 1
= Au.
20
20
20
So now we want to solve this equation. A has characteristic polynomial

1
1
= (1
)2 + 1 = 2 2 + 2,
1
1
so the eigenvalues are 1 + i and 1
⇥
⇥
A
(1 + i)I
A
(1
i)I
Then by Theorem 4.40,
i. Eigenvectors are found


⇤
i
1 0
1 i
0 =
!
1
i 0
0 0


⇤
i 1 0
1
i
0 =
!
1 i 0
0
0
u = C1 e
(1+i)t

1
+ C2 e(1
i
by row-reducing:
i)t
0
0
)
v1 =
0
0
)
v2 =



1
i
i
.
1
i
.
1
Substitute t = 0 to determine the constants:





10
10
C1 + iC2
0 1
0 i
u(0) = x(0) +
=
= C1 e
+ C2 e
=
.
20
10
i
1
iC1 + C2
Solving gives C1 = C2 = 5
u = (5
= (5
= et
= et


5i, so that the solution is


1
i
5i)e(1+i)t
+ (5 5i)e(1 i)t
i
1

1
5i)et (cos t + i sin t)
+ (5 5i)et (cos t
i
i sin t)

i
1
(5 5i) cos t + (5 + 5i) sin t + (5 + 5i) cos t + (5 5i) sin t
(5 + 5i) cos t + ( 5 + 5i) sin t + (5 5i) cos t + ( 5 5i) sin t
10 cos t + 10 sin t
.
10 cos t 10 sin t
Rewriting in terms of the original variables gives




10
10
t 10 cos t + 10 sin t
t 10 cos t + 10 sin t
x=e
=e
+
.
10 cos t 10 sin t
20
10 cos t 10 sin t
20
A plot of the populations is below. Both populations die out; species Y after about 1.2 time units and
species X after about 2.4.
Population
60
X
50
40
30
20
Y
10
0
0.0
0.5
1.0
1.5
2.0
2.5
3.0
Time !days"
350
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
68. We want to convert x0 = Ax + b into an equation u0 = Au. If we could write b = Ab0 , then letting
u = x + b0 would do the trick. To find b0 , we invert A:
"
#" # "
#
1
1
0
20
0
1
2
2
b =A b= 1
=
.
1
40
20
2
2

20
Then let u = x +
, so that
20
✓

◆0

✓

◆
20
0
0
0
0
1
1
u = x+
= x = Ax + b = Ax + AA
=A x+A
= Au.
20
40
40
So now we want to solve this equation. A has characteristic polynomial

1
1
=( 1
)2 + 1 = 2 + 2 + 2,
1
1
so the eigenvalues are
⇥
⇥
1 + i and
A
( 1 + i)I
A
( 1
Then by Theorem 4.40,
i)I
1
i. Eigenvectors are found by row-reducing:



⇤
i
1 0
1 i 0
1
0 =
!
) v1 =
1
i 0
0 0 0
i



⇤
i 1 0
1
i 0
i
0 =
!
) v2 =
.
1 i 0
0
0 0
1
u = C1 e
( 1+i)t

1
+ C2 e(
i
1 i)t

i
.
1
Substitute t = 0 to determine the constants:





20
10
1
i
C1 + iC2
u(0) = x(0) +
=
= C1 e0
+ C2 e0
=
.
20
10
i
1
iC1 + C2
Solving gives C1 =
5
5i and C2 = 5 + 5i, so that the solution is


1
i
u = ( 5 5i)e( 1+i)t
+ (5 + 5i)e( 1 i)t
i
1


1
i
= ( 5 5i)e t (cos t + i sin t)
+ (5 + 5i)e t (cos t i sin t)
i
1

( 5 5i) cos t + (5 5i) sin t + ( 5 + 5i) cos t + (5 + 5i) sin t
=e t
(5 5i) cos t + (5 + 5i) sin t + (5 + 5i) cos t + (5 5i) sin t

10 cos t + 10 sin t
=e t
.
10 cos t + 10 sin t
Rewriting in terms of the original variables gives


10 cos t + 10 sin t
20
t
x=e
=e
10 cos t + 10 sin t
20
t


10 cos t + 10 sin t
20
+
.
10 cos t + 10 sin t
20
A plot of the populations is below; both populations stabilize at 20 as the e
Population
30
Y
25
20
X
15
10
5
0
0
1
2
3
4
Time !days"
5
t
term goes to zero.
4.6. APPLICATIONS AND THE PERRON-FROBENIUS THEOREM
351
69. (a) Making the change of variables y = x0 and z = x, the original equation x00 + ax0 + bx = 0 becomes
y 0 + ay + bz = 0; since z = x, we have z 0 = x0 = y. So we get the system
y0 =
ay
0
z =
(b) The coefficient matrix of the system in (a) is

whose characteristic polynomial is ( a
70. Make the substitutions y1 = x(n
1)
x(n) + an
1x
y
a
1
)(
, y2 = x(n
2)
bz
.
b
,
0
)+b=
, . . . , yn
(n 1)
2
+ a + b.
= x0 , yn = x. Then the original equation
1
+ · · · + a1 x0 + a0 = 0
can be written as
y10 + an
1 y1
+ an
and for i = 2, 3, . . . , n, we have yi0 = yi
y10 = an
y20
y30
1 y1
2 y2
1.
an
+ · · · + a 2 yn
2
+ a 1 yn
1
+ a0 yn = 0,
So we get the n equations
···
2 y2
= y1
a 2 yn
2
a 1 yn
1
a 0 yn
= y2
..
.
yn0 = yn
whose coefficient matrix is
1
2
6
6
6
6
6
4
an
1
0
..
.
an
0
1
..
.
1
0
0
This is the companion matrix of p( ).
2
···
···
···
..
.
a1
0
0
..
.
···
1
3
a0
0 7
7
0 7
7.
.. 7
. 5
0
71. From Exercise 69, we know that the characteristic polynomial is
and 3. Therefore the general solution is x(t) = C1 e2t + C2 e3t .
2
5 + 6, so the eigenvalues are 2
72. From Exercise 69, we know that the characteristic polynomial is 2 + 4 + 3, so the eigenvalues are
and 3. Therefore the general solution is x(t) = C1 e t + C2 e 3t .
1
73. 
From Exercise
59, the matrix of the system has eigenvalues 4 and 1 with corresponding eigenvectors

1
3
and
. So by Theorem 4.41, the system x0 = Ax has solution x = eAt x(0). In this case
1
2

0
x(0) =
, we get
5
"
#"
#"
# 1 "
#
2 4t
3
3 4t
3
4t
t
t
e
+
e
e
e
1
3
e
0
1
3
eAT = P (et )D P 1 =
= 52 4t 52 t 53 4t 52 t .
1
2
0 e t 1
2
5e
5e
5e + 5e
Thus the solution is
"
2 4t
5e
2 4t
5e
+ 35 e
2
5e
t
t
3 4t
5e
3 4t
5e
which matches the result from Exercise 59.
+
3
t
5e
2
t
5e
#" # "
#
0
3e4t 3e t
=
,
5
3e4t + 2e t
352
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
74. From
Exercise
60, the matrix of the system has eigenvalues 1 and 3 with corresponding eigenvectors


1
1
and
. So by Theorem 4.41, the system x0 = Ax has solution x = eAt x(0). In this case
1
1

1
x(0) =
, we get
1
e
AT
t D
= P (e ) P
1
Thus the solution is
"
#"
1 et
1 0
1
=
1
"
1 t
2e
1 t
2e
0
e3t
+ 12 e3t
1 3t
2e
1 t
2e
1 t
2e
#"
1
1
+
#
1
1
1 3t
2e
1 3t
2e
1
=
"
1 t
2e
1 t
2e
+ 12 e3t
1 3t
2e
1 t
2e
1 t
2e
+
1 3t
2e
1 3t
2e
#
.
#" # " #
1
et
= t ,
1
e
which matches the result from Exercise 60.
75. From
of the system has eigenvalues 0, 1, and 1 with corresponding eigenvectors
2 3Exercise
2 3 63, 2the matrix
3
1
0
1
4 15, 415, and 4 15. So by Theorem 4.41, the system x0 = Ax has solution x = eAt x(0). In this
1
1
0
2 3
1
case x(0) = 4 05, and we get
1
eAT = P (et )D P
Thus the solution is
2
1
4 1+e t
1 + et
1
2
1
=4 1
1
2
0
1
1
1
=4 1+e t
1 + et
1 e t
1 + e t + et
1 + et
which matches the result from Exercise 63.
32
1
1
15 40
0 0
32
0
1
0 54 1
e t
1
0
et
0
0
1
1
3
1+e t
1 e t 5.
1
1 e t
1 + e t + et
1 + et
3
1
15
0
1
32 3 2
3
1+e t
1
2 e t
1 e t 5 4 05 = 4 2 + e t + e t 5 ,
1
1
2 + et
76. From Exercise 64, the matrix of 2the3 system
82has
3 eigenvalues
2 39 4 and 2 (with algebraic multiplicity 2)
0
1 =
3
<
with corresponding eigenvectors 415 and 415 , 4 05 . So by Theorem 4.41, the system x0 = Ax
:
;
3
02 3
1
2
has solution x = eAt x(0). In this case x(0) = 435, and we get
4
eAT = P (et )D P
1
2
3
6
= 41
3
2
=
0
1
0
1 4t
2e
6 1 4t
46e
1 4t
2e
+
32
1 e4t
76
05 4 0
1
0
1
2t
2e
1
2t
6e
1
2t
2e
e
0
e
2t
0
0
2t
0
32
0
3
76
0 5 41
e 2t
3
1 4t
2e
1 4t
6e
1 4t
2e
+
3
0
1
0
1
2t
2e
1
2t 7 .
5
6e
1
2t
e
2
3
1
7
05
1
1
4.6. APPLICATIONS AND THE PERRON-FROBENIUS THEOREM
Thus the solution is
2
1 4t
2e
6 1 4t
46e
1 4t
2e
+ 12 e
1
6e
1
2e
2t
2t
2t
2t
e
32 3
1 4t
2e
1 4t
6e
1 4t
2e
0
0
+
1
2t
2
2e
1
2t 7 6 7
e
35
5
4
6
1
2t
4
2e
which matches the result from Exercise 64.

2 1
77. (a) With A =
, we have
0 3



1
3
9
x1 = A
=
, x2 = Ax1 =
,
1
3
9
353
2
3e4t e
6 4t
= 4e + 2e
3e4t + e
2t
3
2t 7 ,
2t
x3 = Ax2 =
5

27
.
27
25
20
15
10
5
(b) With A =

10
5
2
0
1
, we have
3


1
2
x1 = A
=
,
0
0
20
15

4
x2 = Ax1 =
,
0
25
x3 = Ax2 =

8
.
0
1.0
0.5
2
4
6
8
!0.5
!1.0
(c) Since the matrix is upper triangular, we see that its eigenvalues are 2 and 3; since these are both
greater than 1 in magnitude, the origin is a repeller.
(d) Some other trajectories are
10
5
5
!5
!5
!10
10
354
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
78. (a) With A =

0.5
0
0.5
, we have
0.5


1
0
x1 = A
=
,
1
0.5
x2 = Ax1 =

0.25
,
0.25
x3 = Ax2 =

0.25
.
0.125
1.0
0.8
0.6
0.4
0.2
0.2
!0.2
(b) With A =

0.5
0
0.4
0.6
0.8
1.0
0.5
, we have
0.5
x1 = A


1
0.5
=
,
0
0
x2 = Ax1 =

0.25
,
0
x3 = Ax2 =

0.125
.
0
0.1
0.2
0.4
0.6
0.8
1.0
1.2
1.4
!0.1
(c) Since the matrix is upper triangular, we see that its only eigenvalue is 0.5; since this is less than
1 in magnitude, the origin is an attractor.
(d) Some other trajectories are
1.0
0.5
!1.0
0.5
!0.5
1.0
!0.5
!1.0
!1.5
!2.0
79. (a) With A =

2
1
1
, we have
2
x1 = A
so that

1
is a fixed point:
1


1
1
=
,
1
1
1.5
2.0
4.6. APPLICATIONS AND THE PERRON-FROBENIUS THEOREM
355
2.0
1.5
1.0
0.5
1.0
0.5
(b) With A =

2
1
2.0
1.5
1
, we have
2

1
x1 = A
=
0

2
,
1
2
x2 = Ax1 =
4
6

5
,
4
8
10
x3 = Ax2 =
12

14
.
13
14
!2
!4
!6
!8
!10
!12
(c) The characteristic polynomial of the matrix is

2
1
det(A
I) =
=
1
2
2
4 +3=(
3)(
1).
Thus one eigenvalue is greater than 1 in magnitude and the other is equal to 1. So the origin is
not a repeller, an attractor, or a saddle point; there are no nonzero values of x0 for which the
trajectory of x0 approaches the origin.
(d) Some other trajectories are
10
5
5
!5
10
!5
!10
80. (a) With A =

4
1
2
, we have
3
x1 = A

1
=
1

2
,
2
x2 = Ax1 =

4
,
4
x3 = Ax2 =

8
.
8
356
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
4
!4, 4"
2
!1, 1"
!8
!6
!4
2
!2
!!2, !2"
4
!2
!4
!6
(b) With A =

!!8, !8"
4
1
!8
2
, we have
3


1
4
x1 = A
=
,
0
1
x2 = Ax1 =

18
,
7
x3 = Ax2 =

86
.
39
40
!!86, 39"
30
20
10
!!4, 1"
!80
!60
!40
!1, 0"
20
!20
!18, !7"
(c) The characteristic polynomial of the matrix is

4
2
det(A
I) =
1
3
Since both eigenvalues, 5 and
(d) Some other trajectories are
=
2
+ 7 + 10 = ( + 5)( + 2).
2, are greater than 1 in magnitude, the origin is a repeller.
x3
x2 20
30
x0 !!"1, 1"
50
20
100
150
x1
"20
10
"40
x0 !!2, 1"
"80
"60
"40
"20
"60
20
x2
x1
x3
"80
x3
x3
50
100
40
30
20
50
10
x1
x1
x0 !!2, "1"
"250
"200
"150
"100
"50
50
x2
100
"60
"40
"20
x0 !!"1, "2" "10
x2
4.6. APPLICATIONS AND THE PERRON-FROBENIUS THEOREM
81. (a) With A =

1.5
1
357
1
, we have
0


1
x1 = A
=
1
2
,
2

4
x2 = Ax1 =
,
4
1.0
x3 = Ax2 =

8
.
8

86
.
39
!1, 1"
0.5
1.0
0.5
2.0
1.5
3.0
2.5
!1.75, !0.5"
!0.5
!1.0
!0.5, !1"
!1.5
!3.125, !1.75"
(b) With A =

1.5
1
1
, we have
0
x1 = A

1
=
0

4
,
1
!1, 0"
1
x2 = Ax1 =
2
3

18
,
7
4
x3 = Ax2 =
5
6
!0.5
!1.5, !1"
!1.0
!3.25, !1.5"
!1.5
!2.0
!2.5
!3.0
!6.375, !3.25"
(c) The characteristic polynomial of the matrix is
det(A
I) =

1.5
1
1
=
2
1.5 + 1 = (
2)( + 0.5)
Since one eigenvalue is greater than 1 in magnitude and the other is less, the origin is a saddle
point. Eigenvectors for = 0.5 are attracted to the origin, and other vectors are repelled.
358
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
(d) Some other trajectories are
2.0
5
x3
x0 !!1, 2"
1.5
4
1.0
3
0.5
x2
x2
2
"0.4
0.2
"0.2
0.4
0.6
0.8
1.0
x0 !!"1, 1"
x3
x1
1
x1
"0.5
"1.0
"8
"6
"4
"2
1.0
10
5
x0 !!2, "1"
15
x1
0.5
x3
x1
"2
"1.0
"0.8
"0.6
"0.4
x2
"4
0.2
"0.2
x2
"0.5
"1.0
"6
"1.5
82. (a) With A =

0.1
0.5

"2.0
0.9
, we have
0.5
x1 = A
so
x0 !!"1, "2"
x3
"8


1
1
=
,
1
1
1
is a fixed point - it is an eigenvector for the eigenvalue
1
= 1.
2.0
1.5
1.0
!1, 1"
0.5
0.5
(b) With A =

0.1
0.5
1.0
1.5
2.0
0.9
, we have
0.5
x1 = A


1
0.1
=
,
0
0.5
x2 = Ax1 =

0.46
,
0.30
x3 = Ax2 =

0.316
.
0.38
0.4
4.6. APPLICATIONS AND THE PERRON-FROBENIUS THEOREM
359
0.5
!0.1, 0.5"
0.4
!0.316, 0.38"
0.3
!0.46, 0.3"
0.2
0.1
!1, 0"
0.2
0.4
0.6
0.8
1.0
(c) The characteristic polynomial of the matrix is
det(A
I) =

0.1
0.5
0.9
0.5
=
2
0.6
0.45 = (
1)( + 0.4).
Since one eigenvalue is less than 1 in magnitude and the other is equal to 1, the origin is neither
an attractor, a repeller, nor a saddle point.
(d) Some other trajectories are
1.5
5
x0 !!"9, 5"
x1
4
x3
x2
3
1.0
x0 !!2, 1"
2
x2
1
0.5
"8
"6
"4
2
"2
x3
"1
0.0
0.5
"2.0
"1.5
1.0
1.5
"1.0
"0.5
x1
2.0
"2
x1
0.5
"0.5
"0.2
"0.5
x3
"0.4
x2
"1.0
"0.6
x1
x3
x2
"1.5
"0.8
x0 !!"1, "2"
"2.0
83. (a) With A =

0.2
0.2
x0 !!1, "1"
"1.0
0.4
, we have
0.8
x1 = A


1
0.6
=
,
1
0.6
x2 = Ax1 =

0.36
,
0.36
x2 = Ax2 =

0.216
.
0.216
1.0
360
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
1.0
0.8
0.6
0.4
0.4
(b) With A =

0.2
0.2
0.6
0.8
1.0
0.4
, we have
0.8
x1 = A

1
=
0

0.2
,
0.2
x2 = Ax1 =

0.04
,
0.20

0.088
.
0.152
0.4)(
0.6).
x3 = Ax2 =
0.2
0.1
0.2
!0.2
0.4
0.6
0.8
1.0
!0.1
!0.2
!0.3
(c) The characteristic polynomial of the matrix is
det(A
I) =

0.2
0.4
0.8
0.2
=
2
+ 0.24 = (
Since both eigenvalues are smaller than 1 in magnitude, the origin is an attractor.
(d) Some other trajectories are
4
2
!2
1
!1
2
!2
!4
84. (a) With A =

0
1.2
1.5
, we have
3.6
x1 = A

1
=
1

1.5
,
4.8
x2 = Ax1 =

7.2
,
15.48
x2 = Ax2 =

23.22
.
47.088
4.6. APPLICATIONS AND THE PERRON-FROBENIUS THEOREM
361
40
30
20
10
(b) With A =

!20
0
1.2
!10
!15
!5
1.5
, we have
3.6
x1 = A


1
0
=
,
0
1.2
x2 = Ax1 =

1.8
,
4.32
x3 = Ax2 =

6.48
.
13.392
12
10
8
6
4
2
!6
!4
!5
!3
!2
(c) The characteristic polynomial of the matrix is

1.5
det(A
I) =
=
1.2 3.6
2
1
!1
3.6 + 1.8 = (
3)(
0.6).
Since one eigenvalue is smaller than 1 in magnitude and the other is greater, the origin is a saddle
point; the origin attracts some initial points and repels others.
(d) Some other trajectories are
6
4
2
!4
2
!2
4
!2
!4
p
a 2 + b2 =
#"
0 cos ✓
r sin ✓
#
85. With the given matrix A, we have r =
"
r
A=
0
p
⇡
4,
2 and ✓ =
sin ✓
=
cos ✓
"p
2
0
so that
#"
0
p
2
p1
2
p1
2
p1
2
p1
2
#
.
362
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
Since r = | | =
p
2 > 1, the origin is a spiral repeller:
2.0
1.5
1.0
0.5
!4
!3
!2
1
!1
p
86. With the given matrix A, we have r = a2 + b2 = 0.5 and ✓ = 3⇡
4 , so that
"
#"
# "
#"
#
r 0 cos ✓
sin ✓
0.5 0
0 1
.
A=
=
0 r sin ✓
cos ✓
0 0.5
1 0
Since r = | | = 0.5 < 1, the origin is a spiral attractor:
1.0
0.5
0.2
!0.2
0.4
0.6
0.8
1.0
1.2
!0.5
p
87. With the given matrix A, we have r = a2 + b2 = 2 and ✓ = 5⇡
3 , so that
"
#"
# "
#"
p #
3
1
r 0 cos ✓
sin ✓
2 0
2p
2
A=
=
.
3
1
0 r sin ✓
cos ✓
0 2
2
2
Since r = | | = 2 > 1, the origin is a spiral repeller:
!8
!6
!4
2
!2
!2
!4
!6
!8
88. With the given matrix A, we have r =
"
r
A=
0
0
r
p
a 2 + b2 =
#"
cos ✓
sin ✓
r⇣
p
# "
sin ✓
1
=
cos ✓
0
3
2
⌘2
+
#"
0
1
1 2
2
p
3
2
1
2
= 1 and ✓ =
1
p2
3
2
#
.
5⇡
6 ,
so that
4.6. APPLICATIONS AND THE PERRON-FROBENIUS THEOREM
363
Since r = | | = 1, the origin is the orbital center:
1.0
x3
x0
0.5
!1.0
1.0
0.5
!0.5
x1
x2
!0.5
!1.0
89. The characteristic polynomial of A is
det(A
I) =

0.1
0.1
0.2
0.3
=
2
0.4 + 0.05;
= 0.2 ± 0.1i for the eigenvalues. Row-reducing gives

⇥
⇤
1
1
eigenvector. Thus P = Re x Im x =
, and so
1
0




0
1 0.1
0.2
1
1
0.2
0.1
1
C = P AP =
=
.
1
1 0.1
0.3
1
0
0.1
0.2
p
Since | | = 0.05 < 1, the origin is a spiral attractor. The first six points are



1
0.1
0.09
x0 =
, x1 =
, x2 =
,
1
0.4
0.11



0.031
0.0079
0.00161
x3 =
, x4 =
, x5 =
.
0.024
0.0041
0.00044
solving this quadratic gives
1.0
0.8
0.6
0.4
0.2
0.2
0.4
0.6
0.8
1.0
90. The characteristic polynomial of A is
det(A
I) =

2
1
2
=
2
2 + 2;

1 i
for an
1
364
CHAPTER 4. EIGENVALUES AND EIGENVECTORS

1 i
= 1 ± i for the eigenvalues. Row-reducing gives
for an eigen2

⇥
⇤
1
1
vector. Thus P = Re x Im x =
, and so
2
0
"
#"
#"
# "
#
1
0
2
1
1
1
1
1
2
C = P 1 AP =
=
.
1
1
2 0
2
0
1 1
2
solving this quadratic gives
Since | | =
p
2 > 1, the origin is a spiral repeller. The first six points are





1
3
4
2
4
x0 =
, x1 =
, x2 =
, x3 =
, x4 =
,
1
2
6
8
4
x5 =

12
.
8
5
!10
!5
!5
91. The characteristic polynomial of A is
det(A
solving this quadratic gives
=
1
2
I) =
±
p
3
2

1
1
⇥
P = Re x
C=P
1
AP =
"
0
p2
3
2
+ 1;
i for the eigenvalues. Row-reducing gives
eigenvector. Thus
and so
=
1
1
p1
3
⇤
Im x =
#"

1
1
0
3
2
1
0
p
3
2
1
2
1
0
#
=
"
1.0
0.5
0.5
!0.5
!0.5
!1.0
p
1
,
1
p2
3
2
Since | | = 1, the origin is an orbital center. The first six points are





1
0
1
1
0
x0 =
, x1 =
, x2 =
, x3 =
, x4 =
,
1
1
0
1
1
!1.0
1
2
p
1
2
#"
1

1.0
p
3
2
1
2
#
.
x5 =

1
.
0
3
2 i
for an
Chapter Review
365
92. The characteristic polynomial of A is
det(A
solving this quadratic gives
=
I) =
p
3
2
±
1
2

p 1
3
1
⇥
P = Re x
C=P
1
AP =
"
0
p
2
3 + 1;
i for the eigenvalues. Row-reducing gives
eigenvector. Thus
and so
p
2
=
⇤
Im x =
1
#"
3
0
p
1

p
3
2
1
2
1
#"
1
3
p
3
2
1
1
2i
for an
,
0
p

3
2
1
2
1
0
#
=
"p
3
2
1
2
1
p2
3
2
#
.
Since | | = 1, the origin is an orbital center. The first six points are
p



1p
1
1 p3
x0 =
, x1 =
, x2 =
,
1
1+ 3
2+ 3
p
p



p
2 p3
2 p3
1
3
x3 =
, x4 =
, x5 =
.
1
2+ 3
1+ 3
3
2
1
!3
!2
1
!1
2
3
!1
!2
!3
Chapter Review
1. (a) False. What is true is that if A is an n ⇥ n matrix, then det( A) = ( 1)n det A. This is because
multiplying any row by 1 multiplies the determinant by 1, and A multiplies every row by
1. See Theorem 4.7.
(b) True, since det(AB) = det(A) det(B) = det(B) det(A) = det(BA). See Theorem 4.8.
(c) False. It depends on what order the columns are in. Swapping any two columns of A multiplies
the determinant by 1. See Theorem 4.4.
(d) False. While det AT = det A by Theorem 4.10, det A 1 = det1 A by Theorem 4.9. So unless
det A = ±1, the statement is false.

0 1
(e) False. For example, the matrix
has a characteristic polynomial of 0, so 0 is its only
0 0
eigenvalue.
(f ) False. For example, the identity matrix has only
for every i.
= 1 as an eigenvalue, but ei is an eigenvector
(g) True. See Theorem 4.25 in Section 4.4.
(h) False. Any diagonal matrix with two identical entries on the diagonal (such as the identity matrix)
provides a counterexample.
366
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
(i) False. If A and B are similar, then they have the same eigenvalues, but not necessarily the same
eigenvectors. Exercise 48 in Section 4.4 shows that if B = P 1 AP and v is an eigenvector of A,
then P 1 v is an eigenvector for B for the same eigenvalue.
(j) False. For example, if A is invertible, then its reduced row echelon form is the identity matrix,
which has only the eigenvalue 1. So if A has an eigenvalue other than 1, it cannot be similar to
the identity matrix.
2. (a) For example, expand along the first row:
1
3
7
3
5
9
5
5
7 =1
9
11
7
11
3
= 1(5 · 11
3
7
7
3
+5
11
7
7 · 9)
= 0.
5
9
3(3 · 11
7 · 7) + 5(3 · 9
5 · 7)
(b) Apply row operations to diagonalize the matrix:
2
1
43
7
3
5
9
3
5
75
11
2
1
! 40
7R1
! 0
R2 3R1
R3
3
4
12
3
5
85
24
R3
2
1
40
3R2
! 0
3
5
85 .
0
3
4
0
Then det A is the product of the diagonal entries, which is zero.
3. To get from the first matrix to the second:
• Multiply the first column by 3; this multiplies the determinant by 3.
• Multiply the second column by 2; this multiplies the determinant by 2.
• Subtract 4 times the third column from the second; this does not change the determinant.
• Interchange the first and second rows; this multiplies the determinant by
So the determinant of the resulting matrix is 3 · 3 · 2 · ( 1) =
4. (a) We have
det C = det (AB)
1
=
1.
18.
1
1
1
=
=
det(AB)
det A det B
2·
1
4
=
2.
(b) We have
det C = det A2 B(3AT ) = det A · det A · det B · det(3AT )
=
det(3AT ) =
34 det(AT ) =
34 det A =
162.
5. For any n ⇥ n matrix, det A = det(AT ), and det( A) = ( 1)n det A. So if A is skew-symmetric, then
AT = A and
det A = det(AT ) = det( A) = ( 1)n det A = det A
since n is odd. But det A =
det A means that det A = 0.
6. Compute the determinant by expanding along the first row:
1
1
2
1
2
1
1 k =1
4
2
4 k
k
k2
= 1(k 2
= 2k
2
( 1)
1
2
4k) + 1(k 2
6k + 4 = 2(k
So the determinant is zero when k = 1 or k = 2.
k
1
+2
k2
2
1
4
2k) + 2(4
2)
2)(k
1).
Chapter Review
367
7. Since

3
Ax =
4

1
3


1
5
1
=
=5
,
2
10
2
we see that x is indeed an eigenvector, with corresponding eigenvalue 5.
8. Since
2
13
Ax = 4 5
10
32 3 2
3 2 3
2 3
45
3
3 · 13 1 · ( 60) + 2 · ( 45)
9
3
155 4 15 = 4 3 · ( 5) 1 · 18 + 2 · 15 5 = 4 35 = 3 4 15 ,
32
2
3 · 10 1 · ( 40) + 2 · ( 32)
6
2
60
18
40
we see that x is indeed an eigenvector, with corresponding eigenvalue 3.
9. (a) The characteristic polynomial of A is
5
det(A
6
3
0
I) =
0
=( 2
=
3
3
4
2
5
)
( + 2)(
2
6
3
4
+
2) =
( + 2)( + 2)(
1).
(b) The eigenvalues of A are the roots of the characteristic polynomial, which are
(with algebraic multiplicity 2).
(c) We row-reduce A
⇥
A
I
2
6
⇤
0 =4 3
0
6
3
0
2
1
! 40
0
2
R1 +R2
! 1
40
R3 +3R1
! 0
A + 2I
= 1 and
2
=
2
I for each eigenvalue :
R2 +6R1
⇥
1
2
3
⇤
0 =4 3
0
3
3
3
1
0
0
1
0
0
6
6
0
Thus the eigenspaces are
02
3
0
05
0
2
3
R1 $R2 4
6
!
0
3
1 0
3 05
1 0
3
0 0
1 05
0 0
3
3
0
31
1
E1 = span @4 15A ,
0
3
0
05
0
E
2
1
! 40
0
R2 $R3
2
3
R2 +R2 4
0
!
0
2
3
6
0
1
0
0
3
3
3
1
1
3
6 3
0 0
0 0
3
0
05
0
1
3 R1
!
1
3 R3
!
3
0
05
0
3
0
05
0
2
1
4 6
0
2
! 1
40
0
1
3 R1
2
0
0
1
6
0
1
3
1
1
0
0
3
0
05 .
0
82
39
02 3 2 31
1
2
< 2s + t =
= 4 s 5 = span @4 15 , 405A .
:
;
t
0
1
3
0
05
0
(d) Even though A does not have three distinct eigenvalues, it is diagonalizable since the eigenvalue
of algebraic multiplicity 2, which is 2 = 2, also has geometric multiplicity 2. A diagonalizing
matrix is a matrix P whose columns are the eigenvectors above, so
2
3
1
2 1
1 05 ,
P =4 1
0
0 1
368
CHAPTER 4. EIGENVALUES AND EIGENVECTORS
and
2
1
P 1 AP = 4 1
0
2
1
0
32
1
5
15 4 3
1
0
32
3
1
35 4 1
2
0
6
4
0
3 2
2 1
1
1 05 = 40
0 1
0
0
2
0
3
0
05 = D.
2
10. Since A is diagonalizable, A ⇠ D where D is a diagonal matrix whose diagonal entries are its eigenvalues. Since A ⇠ D, we know that A and D have the same eigenvalues, so the diagonal entries of
D are 2, 3, and 4 and thus det D = 24. But A ⇠ D also implies that det D = det(P 1 AP ) =
det(P 1 ) det A det P = det A, so that det A = 24.
11. Since
we know by Theorem 4.19 that

✓ ◆
1
3
A 5
=5
7
2

5


3
1
=5
7
1
1
1
2( 1)
2
5


1
,
1



1
160
2
162
=
+
=
.
1
160
2
158
12. Since A is diagonalizable, it has n linearly
independent eigenvectors, which therefore form a basis for
P
Rn . So if x is any vector, then x =
ci vi for some constants ci , so that by Theorem 4.19, and using
the Triangle Inequality,
!
n
X
X
X
n
n
n
kA xk = A
ci vi
=
ci ni vi 
|ci | | i | kvi k .
i=1
n
Then | i | ! 0 as n ! 1 since | i | < 1, so each term of this sum goes to zero, so that kAn xk ! 0 as
n ! 1. Since this holds for all vectors x, it follows that An ! O as n ! 1.
13. The two characteristic polynomials are
pA =
pB =
4
2
3
1
3
2
2
2
= (4
)(1
)
6=
2
5
2
= (2
)(2
)
6=
2
4
2.
Since the characteristic polynomials are di↵erent, the matrices are not similar.
14. Since both matrices are diagonal, we can read o↵ their eigenvalues; both matrices have eigenvalues 2
and 3. To convert A to B, we want to reverse the rows, and reverse the columns, so choose

0 1
P =P 1=
.
1 0
Then
P
1
AP =

0
1
1
0

2
0
15. These matrices are not similar. If there were, then
2
a
P = 4d
g
such that P
1
AP = B, or AP = P B. But
2
a+d b+e
AP = 4d + g e + h
g
h

0
3
0
1

1
3
=
0
0
0
= B.
2
there would be some 3 ⇥ 3 matrix
3
b c
e f5
h i
3
c+f
f + i5
i
2
a
P B = 4d
g
3
a+b c
d + e f5 .
g+h i
Chapter Review
369
If these matrices are to be equal, then comparing entries in the last row and column gives f = i = g = 0.
The upper left entries also give d = 0. The two middle entries are e + h and d + e = e, so that h = 0 as
well. Thus the entire bottom row of P is zero, so P cannot be invertible and A and B are not similar.
Note that this is a counterexample to the converse of Theorem 4.22. These two matrices satisfy parts
(a) through (g) of the conclusion, yet they are not similar.
16. The characteristic polynomial of A is
det(A
I) =
(a) A has eigenvalues 3 and
1 if
2
k
= (2
1
2
2
k=(
)(
)
2
k=
2
k.
3)( + 1), so when k = 3.
(b) A has an eigenvalue with algebraic multiplicity 2 when its characteristic polynomial has a double
root, which occurs when its discriminant is 0. But its discriminant is ( 2)2 4 · 1 · ( k) = 4 + 4k,
so we must have k = 1.
(c) A has no real eigenvalues when the discriminant 4 + 4k of its characteristic polynomial is negative,
so for k < 1.
17. If
is an eigenvalue of A with eigenvector x, then Ax = x, so that
A3 x = A(A(A(x))) = A(A( x)) = A(
If A3 = A, then Ax =
3
x, so that
=
3
and then
= 0 or
2
x) =
3
x.
= ±1.
18. If A has two identical rows, then Theorem 4.3(c) says that det A = 0, so that A is not invertible. But
then by Theorem 4.17, 0 must be an eigenvalue of A.
19. If x is an eigenvalue of A with eigenvalue 3, then
(A2
5A + 2I)x = A2 x
5Ax + 2Ix = 9x
This shows that x is also an eigenvector of A2
5 · 3x + 2x =
5A + 2I, with eigenvalue
4x.
4.
20. Suppose that B = P 1 AP where P is invertible. This can be rewritten as BP
an eigenvalue of A with eigenvector x. Then
B(P
Thus P
1
1
x) = (BP
1
)x = (P
1
A)x = P
x is an eigenvector for B, corresponding to .
1
(Ax) = P
1
( x) = (P
1
=P
1
x).
1
A. Let
be
Chapter 5
Orthogonality
5.1
Orthogonality in Rn
1. Since
v1 · v2 = ( 3) · 2 + 1 · 4 + 2 · 1 = 0,
v1 · v3 = ( 3) · 1 + 1 · ( 1) + 2 · 2 = 0,
v2 · v3 = 2 · 1 + 4 · ( 1) + 1 · 2 = 0,
this set of vectors is orthogonal.
2. Since
v1 · v2 = 4 · ( 1) + 2 · 2
v1 · v3 = 4 · 2 + 2 · 1
v2 · v3 =
5 · 0 = 0,
5 · 2 = 0,
1 · 2 + 2 · 1 + 0 · 2 = 0,
this set of vectors is orthogonal.
3. Since v1 · v2 = 3 · ( 1) + 1 · 2 1 · 1 = 2 6= 0, this set of vectors is not orthogonal. There is no need
to check the other two dot products; all three must be zero for the set to be orthogonal.
4. Since v1 · v3 = 5 · 3 + 3 · 1 + 1 · ( 1) = 17 6= 0, this set of vectors is not orthogonal. There is no need
to check the other two dot products; all three must be zero for the set to be orthogonal.
5. Since
v1 · v2 = 2 · ( 2) + 3 · 1
1 · ( 1) + 4 · 0 = 0,
v1 · v3 = 2 · ( 4) + 3 · ( 6)
v2 · v3 =
2 · ( 4) + 1 · ( 6)
1 · 2 + 4 · 7 = 0,
1 · 2 + 0 · 7 = 0,
this set of vectors is orthogonal.
6. Since v2 · v4 = 1 · 0 + 0 · ( 1) + 1 · 1 + 2 · 1 = 3 6= 0, this set of vectors is not orthogonal. There is no
need to check the other five dot products (they are all zero); all dot products must be zero for the set
to be orthogonal.
7. Since the vectors are orthogonal:


4
1
·
=4·1
2
2
371
2 · 2 = 0,
372
CHAPTER 5. ORTHOGONALITY
they are linearly independent by Theorem 5.1. Since there are two vectors, they form an orthogonal
basis for R2 . By Theorem 5.2, we have
v1 · w
v2 · w
4·1
v1 +
v2 =
v1 · v1
v2 · v2
4·4
w=
2 · ( 3)
1 · 1 + 2 · ( 3)
1
v1 +
v2 = v1
2 · ( 2)
1·1+2·2
2
so that
[w]B =

1
2
1
v2 ,
.
8. Since the vectors are orthogonal:

1
·
3

6
= 1 · ( 6) + 3 · 2 = 0,
2
they are linearly independent by Theorem 5.1. Since there are two vectors, they form an orthogonal
basis for R2 . By Theorem 5.2, we have
w=
v1 · w
v2 · w
1·1+3·1
v1 +
v2 =
v1 +
v1 · v1
v2 · v2
1·1+3·3
so that
[w]B =
"
2
5
1
10
6·1+2·1
2
v2 = v1
6 · ( 6) + 2 · 2
5
#
1
v2 ,
10
.
9. Since the vectors are orthogonal:
2
3 2 3
1
1
4 05 · 425 = 1 · 1 + 0 · 2 1 · 1 = 0
1
1
2 3 2 3
1
1
4 05 · 4 15 = 1 · 1 + 0 · ( 1) 1 · 1 = 0
1
1
2 3 2 3
1
1
425 · 4 15 = 1 · 1 + 2 · ( 1) + 1 · 1 = 0,
1
1
they are linearly independent by Theorem 5.1. Since there are three vectors, they form an orthogonal
basis for R3 . By Theorem 5.2, we have
v1 · w
v2 · w
v3 · w
v1 +
v2 +
v3
v1 · v1
v2 · v2
v3 · v3
1·1+0·1 1·1
1·1+2·1+1·1
1·1 1·1+1·1
=
v1 +
v2 +
v3
1 · 1 + 0 · 0 1 · ( 1)
1·1+2·2+1·1
1 · 1 1 · ( 1) + 1 · 1
2
1
= v2 + v3 ,
3
3
w=
so that
2 3
0
627
[w]B = 4 3 5 .
1
3
5.1. ORTHOGONALITY IN RN
373
10. Since the vectors are orthogonal:
2 3 2
1
415 · 4
1
2 3 2
1
415 · 4
1
2 3 2
1
4 15 · 4
0
3
1
15 = 1 · 1 + 1 · ( 1) + 1 · 0 = 0
0
3
1
15 = 1 · 1 + 1 · 1 + 1 · ( 2) = 0
2
3
1
15 = 1 · 1 1 · 1 + 0 · ( 2) = 0,
2
they are linearly independent by Theorem 5.1. Since there are three vectors, they form an orthogonal
basis for R3 . By Theorem 5.2, we have
v1 · w
v2 · w
v3 · w
v1 +
v2 +
v3
v1 · v1
v2 · v2
v3 · v3
1·1+1·2+1·3
1·1 1·2+0·3
1·1+1·2 2·3
=
v1 +
v2 +
v3
1·1+1·1+1·1
1 · 1 1 · ( 1) + 0 · 0
1 · 1 + 1 · 1 2 · ( 2)
1
1
= 2v1
v2
v3 ,
2
2
w=
so that
11. Since kv1 k =
mal.
2
q
6
[w]B = 4
3 2
5
+
4 2
5
= 1 and kv2 k =
q
4 2
5
2
3
17
.
25
1
2
+
3 2
5
= 1, the given set of vectors is orthonor-
q
q
1 2
1 2
1 2
1 2
p1 6= 1 and kv2 k =
12. Since kv1 k =
+
=
+
= p12 6= 1, the given set of vectors
2
2
2
2
2
is not orthonormal. To normalize the vectors, we divide each by its norm to get
(
" #
" #) (" # "
#)
1
1
p1
p1
1
1
2
2 ,
2
p 2 , p
=
.
1
p1
p1
1/ 2 12
1/ 2
2
2
2
13. Taking norms, we get
s✓ ◆
✓ ◆ 2 ✓ ◆2
2
1
2
2
kv1 k =
+
+
=1
3
3
3
s✓ ◆
p
✓
◆2
2
2
1
5
kv2 k =
+
+ 02 =
3
3
3
s
p
✓ ◆2
5
3 5
2
2
kv3 k = 1 + 2 +
=
.
2
2
So the first vector is already normalized; we must normalize the other two by dividing by their norms,
to get
82 3
2 3
2 39 8 2 3 2 2 3 2 2 39
2
p
p
> 13
>
1 > > 1
<
1 6 31 7
1 6 7= <6 32 7 6 15 7 6 3 4 5 7=
627
p
p
p
p
,
,
=
,
,
2
435
4 35
4 5
435 4
5 4 3 55 .
5
>
> >
>
5/3
3 5/2
: 2
: 2
;
5 ;
5
p
0
0
3
2
3
3 5
374
CHAPTER 5. ORTHOGONALITY
14. Taking norms, we get
s✓ ◆
✓ ◆2 ✓
◆ 2 ✓ ◆2
2
1
1
1
1
kv1 k =
+
+
+
=1
2
2
2
2
s
p
✓ ◆2 ✓ ◆ 2 ✓ ◆2
1
2
1
6
2
2
kv2 k = 0 +
+
+
+0 =
3
3
3
3
s✓ ◆
p
✓
◆2 ✓ ◆ 2 ✓
◆2
2
1
1
1
1
3
kv3 k =
+
+
+
=
.
2
6
6
6
3
So the first vector is already normalized; we must normalize the other two
to get
82 1 3
2 3
2 1 39 8 2 1 3 2 3
>
0
0
>
>
2
2 >
2
>
>
>
> >
>
6 7
6
6
<6 1 7
<
6
17
1 7=
1 7 6 p1 7
1 637
1 6 67
6 27
6 7
67
,
6 17 , p
627 , p
6 1 7 = 6 21 7 , 6
p2 7
>4 2 5
>
435
4 6 5>
4 25 6
6/3
3/3
>
>
>
4
65
>
>
>
:
>
1
1
1 ;
1
:
1
2
3
6
p
2
6
15. Taking norms, we get
by dividing by their norms,
2
6
6
6
6
4
p
39
3 >
>
p2 7>
>
3 7=
6
7
p
.
3 7>
>
p6 5>
;
3 >
6
s✓ ◆
✓ ◆2 ✓
◆2 ✓ ◆ 2
2
1
1
1
1
kv1 k =
+
+
+
=1
2
2
2
2
v
u
p !2 ✓
◆2 ✓
◆2
u
6
1
1
t
2
p
kv2 k = 0 +
+ p
+
=1
3
6
6
v
u p !2
p !2
p !2
p !2
u
3
3
3
3
kv3 k = t
+
+
+
=1
2
6
6
6
s
✓
◆2 ✓
◆2
1
1
2
2
kv4 k = 0 + 0 + p
+ p
= 1.
2
2
So these vectors are already orthonormal.
16. The columns of the matrix are orthogonal:


0
1
·
= 0 · ( 1) + 1 · 0 = 0.
1
0
Further, they are orthonormal since


0
0
·
= 0 · 0 + 1 · 1 = 1,
1
1

1
·
0

1
=
0
1 · ( 1) + 0 · 0 = 1.
Thus the matrix is an orthogonal matrix, so by Theorem 5.5, the inverse of this matrix is its transpose,
or

0 1
.
1 0
17. The columns of the matrix are orthogonal:
"
p1
2
p1
2
# "
·
p1
2
p1
2
#
1
1
=p ·p
2
2
1
1
p · p = 0.
2
2
5.1. ORTHOGONALITY IN RN
375
Further, they are orthonormal since
"
# "
p1
2
p1
2
"
·
p1
2
p1
2
# "
·
p1
2
p1
2
p1
2
p1
2
#
#
1
1
=p ·p
2
2
1
p ·
2
✓
1
p
2
◆
=1
1
1
1
1
= p · p + p · p = 1.
2
2
2
2
Thus the matrix is an orthogonal matrix, so by Theorem 5.5, the inverse of this matrix is its transpose,
or
"
#
p1
2
p1
2
p1
2
p1
2
.
18. This is not an orthogonal matrix since none of the columns have norm 1:
q1 · q1 =
1 1 1
1
+ + = ,
9 9 9
3
q2 · q2 =
1 1
1
+ +0= ,
4 4
2
19. This matrix Q is orthogonal by Theorem 5.4 since QQT = I:
2
32
cos ✓ sin ✓
cos ✓ sin ✓
cos ✓
sin2 ✓
cos ✓
QQT = 4 cos2 ✓
sin ✓
cos ✓ sin ✓5 4
sin ✓
0
cos ✓
sin2 ✓
2
2
2
2
4
3
sin ✓ cos ✓+cos ✓ sin3 ✓
2
2
1
= 40
0
cos4 ✓+sin2 ✓+cos2 ✓ sin2 ✓
2
2
sin ✓ cos ✓ sin ✓ cos ✓
0
1
0
3
0
05 .
1
Thus by Theorem 5.5 Q
1
cos2 ✓
sin ✓
cos ✓ sin ✓
cos ✓ sin ✓ sin ✓ cos ✓+cos ✓ sin3 ✓
cos ✓ sin ✓+cos ✓+sin ✓
= 4cos3 ✓ sin ✓
q3 · q3 =
2
sin ✓ cos ✓ cos ✓ sin ✓
1
1
4
6
+
+
=
.
25 25 25
25
3
sin ✓
0 5
cos ✓
cos ✓ sin2 ✓ cos ✓ sin2 ✓
3
cos2 ✓ sin ✓ cos2 ✓ sin ✓ 5
sin2 ✓+cos2 ✓
= QT .
20. The columns are all orthogonal since each dot product consists of two terms of
The columns are orthonormal since the norm of each column is
s✓
◆2 ✓
◆2 ✓
◆2 ✓
◆2
1
1
1
1
±
+ ±
+ ±
+ ±
= 1.
2
2
2
2
1
2
· 12 and two of
So the matrix is orthonormal, so by Theorem 5.5 its inverse is equal to its transpose,
2 1
3
1
1
1
6
6
6
4
2
1
2
1
2
1
2
2
1
2
1
2
1
2
2
1
2
1
2
1
2
2
17
27 .
17
5
2
1
2
21. The dot product of the first and fourth columns is
2 3 2 1 3
p
1
67
6 7 6
p1 7
607 6
1
6
67 = 1 · p
6 7·6
6 0,
=
607 6 p1 7
7
6
4 5 4
65
p1
0
2
so the matrix is not orthonormal.
1
2
· 12 .
376
CHAPTER 5. ORTHOGONALITY
1
22. By Theorem 5.5, Q
Q 1 = QT , so that
is orthogonal if and only if Q
1 T
Q
and Q
1
1
= QT
1 T
= Q
= Q
1
1
1
1
. But since Q is orthogonal,
,
is orthogonal.
23. Since det Q = det QT , we have
1 = det(I) = det(QQ
1
) = det(QQT ) = (det Q)(det QT ) = (det Q)2 .
Since (det Q)2 = 1, we must have det Q = ±1.
24. By Theorem 5.5, it suffices to show that (Q1 Q2 ) 1 = (Q1 Q2 )T . But using the fact that Q1 1 = QT1
and Q2 1 = QT2 , we have
(Q1 Q2 ) 1 = Q2 1 Q1 1 = QT2 QT1 = (Q1 Q2 )T .
Thus Q1 Q2 is orthogonal.
25. By Theorem 3.17, if P is a permutation matrix, then P
1
= P T . So by Theorem 5.5, P is orthogonal.
26. Reordering the rows of Q is the same as multiplying it on the left by a permutation matrix P . By the
previous exercise, P is orthogonal, and then by Exercise 24, P Q is also orthogonal. So reordering the
rows of an orthogonal matrix results in an orthogonal matrix.
27. Let ✓ be the angle between x and y, and the angle between Qx and Qy. By Theorem 5.6, kQxk = kxk,
kQyk = kyk, and Qx · Qy = x · y, so that
cos ✓ =
x·y
Qx · Qy
=
= cos .
kxk kyk
kQxk kQyk
Since 0  ✓,
 ⇡, the fact that their cosines are equal implies that ✓ = .

a c
28. (a) Suppose that Q =
is an orthogonal matrix. Then
b d
T
T
QQ = Q Q = I
)

a
c
b
d

a
b
 2
c
a + b2
=
d
ac + bd

ac + bd
1
=
c 2 + d2
0
0
.
1
So a2 + b2 = 1 and c2 + d2 = 1, and thus both columns of Q are unit vectors. Also, ac + bd = 0,
which implies that ac = bd and therefore a2 c2 = b2 d2 . But c2 = 1 d2 and b2 = 1 a2 ;
substituting gives
a2 (1
d2 ) = (1
a2 )d2
)
a2
a2 d2 = d2
a2 d2
)
If a = d = 0, then b = c = ±1, so we get one of the matrices




0 1
0
1
0 1
0
,
,
,
1 0
1
0
1 0
1
a2 = d2
)
a = ±d.
1
,
0
all of which are of one of the required forms. Otherwise, if a = d 6= 0, then c2 = 1 d2 = 1 a2 = b2 ,
so that b = ±c and we get

a ±b
,
b a
and we must choose the minus sign in order that the columns be orthogonal. So the result is

a
b
,
b
a
5.1. ORTHOGONALITY IN RN
377
which is also of the required form. Finally, if a = d 6= 0, then again b2 = c2 and we get

a ±b
;
b
a
this time we must choose the plus sign for the columns to be orthogonal, resulting in

a
b
,
b
a
which is again of the required form.

a
(b) Since
is a unit vector, we have a2 +b2 = 1, so that (a, b) is a point on the unit circle. Therefore
b
there is some ✓ with (a, b) = (cos ✓, sin ✓). Substituting into the forms from part (a) gives the
desired matrices.
(c) Let

cos ✓
Q=
sin ✓
sin ✓
,
cos ✓
R=

cos ✓
sin ✓
sin ✓
.
cos ✓
From Example 3.58 in Section 3.6, Q is the standard matrix of a rotation through angle ✓. From
Exercise 26(c) in Section 3.6, R is a reflection through the line ` that makes an angle of ✓2 with
the x-axis, which is the line y = tan ✓2 x (or x = 0 if ✓ = ⇡).
(d) From part (c), we have det Q = cos2 ✓ + sin2 ✓ = 1 and det R = cos2 ✓ sin2 ✓ = 1.
⇣
⌘
p1 · p1
29. Since det Q = p12 · p12
= 1, this matrix represents a rotation. To find the angle, we have
2
2
cos ✓ =
p1
2
1
2
30. Since det Q =
have cos ✓ =
31. Since det Q =
p
p1 ,
2
and sin ✓ =
1
2
1
2
·
so that ✓ = ⇡4 = 45 .
⇣ p ⌘
p
3
3
·
= 1, this matrix represents a rotation. To find the angle, we
2
2
1
2
·
1
2
p
3
2 , so that
⇣p ⌘
p
3
3
·
=
2
2
and sin ✓ =
✓=
4⇡
3
= 240 .
1, this matrix represents a reflection. If cos ✓ =
1
2
and
sin ✓ = 23 , then ✓ = 2⇡
, so that by part (d) of the previous exercise the reflection occurs through the
p3
line y = tan ⇡3 x = 3 x.
4
4
32. Since det Q = 35 · 35
1, this matrix represents a reflection. Let ✓ be the (third
5 ·
5 =
3
quadrant) angle with cos ✓ = 5 and sin ✓ = 45 ; then by part (d) of the previous exercise, the
reflection occurs through the line y = tan ✓2 x.
33. (a) Since A and B are orthogonal matrices, we know that AT = A
A(AT + B T )B = A A
1
+B
1
B = AA
1
B + AB
1
1
and B T = B
1
, so that
B = B + A = A + B.
(b) Since A and B are orthogonal matrices, we have det A = ±1 and det B = ±1. If det A+det B = 0,
then one of them must be 1 and the other must be 1, so that (det A)(det B) = 1. But by part
(a),
det(A + B) = det A(AT + B T )B = (det A) det(AT + B T ) (det B)
= (det A) det((A + B)T ) (det B) = (det A)(det B) (det(A + B)) =
Since det(A+ B) =
det(A + B).
det(A+ B), we conclude that det(A+ B) = 0 so that A+ B is not invertible.
34. It suffices to show that QQT = I. Now,
"
#"
x1
yT ⌘
x1
T
⇣
QQ =
1
t
y I
y
1 x1 yy
I
T
⇣ y ⌘
1
1 x1
yyt
#
.
378
CHAPTER 5. ORTHOGONALITY
Also, x21 + yyT = x21 + x22 + · · · + x2n = 1, since x is a unit vector. Rearranging gives yyT = 1 x21 .
Further, we have
✓
✓
◆
◆
✓
◆
1
1
T
T
T
y
I
yy
=y
yT (1 x1 )2 = yT (1 + x1 )yT
1 x1
1 x1
✓
✓
I
✓
1
1
x1
◆
I
◆✓
yyT
I
✓
1
1
✓
x1
◆
1
1
x1
yyT
◆
◆
yyT
=
y=y
◆
=
x1 y T
✓
1
x1 y
= I2
=I
=I
=I
1
2
✓
x1
1
1
◆
(1
◆
x21 )y = y
yyT +
✓
1
(1 + x1 )y
◆2
x1
1 x1
2(1 x1 ) T
1 x21
yy
+
yyT
(1 x1 )2
(1 x1 )2
1 2x1 + x21 T
yy
(1 x1 )2
yyT yyT
yyT .
Then evaluating QQT gives
2
⇣
⇣
⌘
⌘
3
1
2
T
T
T
T
x
+
y
y
x
y
+
y
I
yy
1
⇣ 1⇣
⌘
⌘
⇣
⇣
⌘
⌘1⇣ x1 ⇣
⌘
⌘5
QQT = 4
1
1
1
T
T
T
T
x1 y + I
yy
y
yy
+
I
yy
I
yy
1 x1
1 x1
1 x1

1
x1 y T x1 y T
=
x1 y x1 y yyT + (I yyT )

1 O
=
O I
= I.
Thus Q is orthogonal.
35. Let Q be orthogonal and upper triangular. Then the columns qi of Q are orthonormal; that is,
(
n
X
0 i 6= j
qi · qj =
qki qkj =
.
1 i=j
k=1
Since Q is upper triangular, we know that qij = 0 when i > j. We prove by induction on the rows that
each row has only one nonzero entry, along the diagonal. Note that q11 6= 0, since all other entries in
the first column are zero and the norm of the first column vector must be 1. Now the dot product of
q1 with qj for j > 1 is
n
X
q1 · qj =
qk1 qkj = q11 q1j .
k=1
Since these columns must be orthogonal, this product is zero, so that q1j = 0 for j > 1 and therefore
q11 is the only nonzero entry in the first row. Now assume that q11 , q22 , . . . , qii are the only nonzero
entries in their rows, and consider qi+1,i+1 . It is the only nonzero entry in its column, since entries
in lower-numbered rows are zero by the inductive hypothesis and entries in higher-numbered rows are
zero since Q is upper triangular. Therefore, qi+1,i+1 6= 0 since the (i + 1)st column must have norm 1.
Now consider the dot product of qi+1 with qj for j > i + 1:
qi+1 · qj =
n
X
k=1
qk,i+1 qkj = qi+1,i+1 qi+1,j .
5.2. ORTHOGONAL COMPLEMENTS AND ORTHOGONAL PROJECTIONS
379
Since these columns must be orthogonal, this product is zero, so that qi+1,j = 0 for j > i + 1. So all
entries to the right of the (i + 1)st column are zero. But also entries to the left of the (i + 1)st column
are zero since Q is upper triangular. Thus qi+1,i+1 is the only nonzero entry in its row. So inductively,
qkk is the only nonzero entry in the k th row for k = 1, 2, . . . , n, so that Q is diagonal.
36. If n > m, then the maximum rank of A is m. By the Rank Theorem (Theorem 3.26, in Section 3.5),
we know that rank(A) + nullity(A) = n; since n > m, we have nullity(A) = n m > 0. So there is
some nonzero vector x such that Ax = 0. But then kxk =
6 0 while kAxk = k0k = 0.
37. (a) Since the {vi } are a basis, we may choose ↵i ,
x=
n
X
i
such that
↵k vk ,
y=
k=1
n
X
k vk .
k=1
Because {vi } form an orthonormal basis, we know that vi · vi = 1 and vi · vj = 0 for i 6= j. Then
!
!
n
n
n
n
n
X
X
X
X
X
x·y =
↵k v k ·
=
↵i j v i · vj =
↵i i vi · v i =
↵i i .
k vk
k=1
i,j=1
k=1
i=1
Pn
But x · vi = k=1 ↵k vk · vi = ↵i and similarly y · vi =
displayed equation above is
n
X
i=1
↵i
i
=
n
X
i=1
i.
i=1
Therefore the sum at the end of the
(x · vi )(y · vi ) = (x · v1 )(y · v1 ) + (x · v2 )(y · v2 ) + · · · + (x · vn )(y · vn ),
proving Parseval’s Identity.
(b) Since [x]B = [↵k ] and [y]B = [
k ],
Parseval’s Identity says that
x·y =
n
X
↵k
k=1
k
= [x]B · [y]B .
In other words, the dot product is invariant under a change of orthonormal basis.
5.2
Orthogonal Complements and Orthogonal Projections
1. W is the column space of A =
matrix is already row-reduced:

⇥
1
, so that W ? is the null space of AT = 1
2
⇥
1
2
⇤
0 ,
⇤
2 . The augmented
so that a basis for the null space, and thus for W ? , is given by

2
.
1
2. W is the column space of A =
gives

⇥
4
4
, so that W ? is the null space of AT =
3
h
i
h
i
3
4 3 0 ! 1
0 ,
4
so that a basis for the null space, and thus for W ? , is given by
" #
" #
3
3
4 ,
or
.
1
4
⇤
3 . Row-reducing
380
CHAPTER 5. ORTHOGONALITY
3. W consists of
8 2 3 2 39
3
2 3
2 3
x
1
0
0 =
< 1
4 y 5 = x 405 + y 415 , so that 405 , 415 is a basis for W .
:
;
x+y
1
1
1
1
2
Then W is the column space of
2
1
A = 40
1
3
0
15 ,
1
so that W ? is the null space of AT . The augmented matrix is already row-reduced:

1 0 1 0
0 1 1 0
so that the null space, which is W ? , is the set
2 3
2 3
02 31
t
1
1
4 t5 = t 4 15 = span @4 15A .
t
1
1
This vector forms a basis for W ? .
4. W consists of
8 2 3 2 39
3
2 3
2 3
x
1
0
0 =
< 1
42x + 3z 5 = x 425 + y 435 , so that 425 , 435 is a basis for W .
:
;
z
0
1
0
1
2
Then W is the column space of
2
1
A = 42
0
3
0
35 ,
1
so that W ? is the null space of AT . Row-reduce the augmented matrix:
"
#
"
#
2
1 2 0 0
1 0
0
3
!
1
0 3 1 0
0 1
0
3
so that the null space, which is W ? , is the set
2
3
2 3
02 31
2
2
t
2
3
3
6 1 7
6 7
B6 7C
4 3 t5 = t 4 13 5 = span @4 15A .
t
1
3
This vector forms a basis for W ? .
2
3
1
⇥
5. W is the column space of A = 4 15, so that W ? is the null space of AT = 1
3
augmented matrix is already row-reduced:
⇥
⇤
1
1 3 0 ,
so that the null space, which is W ? , is the set
2
3
2 3
2 3
02 3 2 3 1
1
3
s 3t
1
3
4 s 5 = s 415 + t 4 05 = span @415 , 4 05A
t
0
1
0
1
These two vectors form a basis for W ? .
1
⇤
3 . The
5.2. ORTHOGONAL COMPLEMENTS AND ORTHOGONAL PROJECTIONS
2
6
6. W is the column space of A = 4
gives
3
1
2
17
,
25
so that W ? is the null space of AT =
2
⇥1
2
1
2
⇤
⇥
0 ! 1
2
1
381
⇥1
1
2
2
⇤
2 . Row-reducing
⇤
0 ,
4
so the null space, which is W ? , is the set
2
3
2 3
2 3
02 3 2 3 1
s 4t
1
4
1
4
4 s 5 = s 415 + t 4 05 = span @415 , 4 05A
t
0
1
0
1
These two vectors form a basis for W ? .
⇥
⇤
7. To find the required bases, we row-reduce A 0 :
2
3
2
1
1
3 0
1
6 5
7
60
2
1
0
6
7!6
4 0
40
1
2 05
1
1
1 0
0
0
1
0
0
1
2
0
0
3
0
07
7.
05
0
Therefore, a basis for the row space of A is the nonzero rows of the reduced matrix, or
⇥
⇤
⇥
⇤
2 .
u1 = 1 0 1 ,
u2 = 0 1
The null space null(A) is the set of solutions of the augmented matrix, which are given by
2 3
s
4 2s5 .
s
Thus a basis for null(A) is
2
3
1
v = 4 25 .
1
Every vector in row(A) is orthogonal to every vector in null(A):
u1 · v = 1 · ( 1) + 0 · 2 + 1 · 1 = 0,
8. To find the required bases, we
2
1
6 2
6
4 2
3
⇥
row-reduce A
1
0
2
1
1 0
2 4
2 0
3 4
2
4
1
5
u2 · v = 0 · ( 1) + 1 · 2
⇤
0 :
3
2
0
1
6
07
7 ! 60
40
05
0
0
0
1
0
0
1
0
0
0
2 0
2 0
0 1
0 0
2 · 1 = 0.
3
0
07
7.
05
0
Thus a basis for the row space of A is the nonzero rows of the reduced matrix, or
⇥
⇤
⇥
⇤
⇥
1
2 0 ,
u1 = 1 0
u2 = 0 1 0 2 0 ,
u3 = 0 0 0
The null space null(A) is the set of solutions of the augmented matrix,
2
3
2 3
2 3
02 3 2
s + 2t
1
2
1
6 2t 7
607
6 27
B607 6
6
7
6 7
6 7
B6 7 6
6 s 7 = s 617 + t 6 07 = span B617 , 6
6
7
6 7
6 7
B6 7 6
4 t 5
405
4 15
@405 4
0
0
0
0
0
⇤
1 .
which are given by
31
2
C
27
7C
7
07 C
C.
15 A
0
382
CHAPTER 5. ORTHOGONALITY
Thus a basis for null(A) is
2 3
1
6 07
6 7
7
v1 = 6
6 17 ,
4 05
0
2
6
6
v2 = 6
6
4
3
2
27
7
07
7.
15
0
Every vector in row(A) is orthogonal to every vector in null(A):
u1 · v1 = 1 · 1 + 0 · 0
1·1
2 · 0 + 0 · 0 = 0,
u2 · v1 = 0 · 1 + 1 · 0 + 0 · 1 + 2 · 0 + 0 · 0 = 0,
u3 · v1 = 0 · 1 + 0 · 0 + 0 · 1 + 0 · 0 + 1 · 0 = 0,
u1 · v2 = 1 · 2 + 0 · ( 2)
1·0
2 · 1 + 0 · 0 = 0,
u2 · v2 = 0 · 2 + 1 · ( 2) + 0 · 0 + 2 · 1 + 0 · 0 = 0,
u3 · v2 = 0 · 2 + 0 · ( 2) + 0 · 0 + 0 · 1 + 1 · 0 = 0.
9. Using the row reduction from Exercise 7, we see that a basis
2 3
2
1
6 57
6
7
u1 = 6
u2 = 6
4 05 ,
4
1
⇥
⇤
To find the null space of AT , we must row-reduce AT 0 :
2
3
2
1 5
0
1 0
1 0
6
7
6
1
1 05 ! 40 1
4 1 2
3 1
2
1 0
0 0
for the column space of A is
3
1
27
7.
15
1
5
7
1
7
3
7
2
7
0
0
3
0
7
05 .
0
The null space null(AT ) is the set of solutions of the augmented matrix, which are given by
2 5
3
2 53
2 33
02 3 2 31
3
5
3
7s
7t
7
7
6 1 s + 2 t7
6 17
6 27
B6 17 6 27C
6 7
7 6 7C
7 7 = s 6 7 7 + t 6 7 7 = span B6
6
7
6 7
6 7
B6 7 , 6 7C .
4
5
4 15
4 05
@4 75 4 05A
s
t
Thus a basis for null(AT ) is
0
2
3
5
6 17
7
v1 = 6
4 75 ,
0
1
0
7
2
3
3
6 27
7
v2 = 6
4 05 .
7
Every vector in col(A) is orthogonal to every vector in null(AT ):
u1 · v1 = 1 · 5 + 5 · ( 1) + 0 · 7
u2 · v1 =
1 · 5 + 2 · ( 1) + 1 · 7
1 · 0 = 0,
1 · 0 = 0,
u1 · v2 = 1 · ( 3) + 5 · 2 + 0 · 0
u2 · v2 =
1 · ( 3) + 2 · 2 + 1 · 0
1 · 7 = 0,
1 · 7 = 0.
10. Using the row reduction from Exercise 8, we see that a basis for the column space of A is
2 3
2 3
2 3
1
1
2
6 27
6 07
647
7
7
7
u1 = 6
u2 = 6
u3 = 6
4 25 ,
4 25 ,
415
3
1
5
⇥
⇤
To find the null space of AT , we must row-reduce AT 0 :
2
3
2
3
1
2
2
3 0
1 0 0
1 0
6 1
60 1 0
0
2
1 07
1 07
6
7
6
7
6 1
7
6
2
2
3 07 ! 60 0 1
1 07
6
7.
4 0
40 0 0
4
0
4 05
0 05
2
4
1
5 0
0 0 0
0 0
5.2. ORTHOGONAL COMPLEMENTS AND ORTHOGONAL PROJECTIONS
383
The null space null(AT ) is the set of solutions of the augmented matrix, which are given by
2 3
2 3
s
1
6 s7
6 17
6 7 = s6 7.
4s5
4 15
s
1
Thus a basis for null(AT ) is
2
3
1
6 17
7
v1 = 6
4 15 .
1
Every vector in col(A) is orthogonal to every vector in null(AT ):
u1 · v1 = 1 · ( 1)
2 · ( 1) + 2 · 1
3 · 1 = 0,
u2 · v1 = 1 · ( 1) + 0 · ( 1) + 2 · 1
u3 · v1 = 2 · ( 1) + 4 · ( 1) + 1 · 1 + 5 · 1 = 0.
11. Let
⇥
A = w1
w2
⇤
2
2
=4 1
2
1 · 1 = 0,
3
4
05 .
1
Then the subspace W spanned by w1 and w2 is the column space of A, so that by Theorem 5.10,
W ? = (col(A))? = null(AT ). To compute null(AT ), we row-reduce
"
#
"
#
h
i
1
2
1
2
0
1
0
0
4
!
.
AT 0 =
5
4 0
1 0
0 1
0
2
Thus null(AT ) is the set of solutions of this augmented matrix, which is
2
3
2 3
02 31
1
1
1
4t
4
6 5 7
6 57
B6 7C
=
t
=
span
4 2t 5
4 25
@4 105A
t
1
4
so this vector forms a basis for W ? .
12. Let
⇥
A = w1
2
1
⇤ 6 1
w2 = 6
4 3
2
3
0
17
7.
25
1
Then the subspace W spanned by w1 and w2 is the column space of A, so that by Theorem 5.10,
W ? = (col(A))? = null(AT ). To compute null(AT ), we row-reduce


⇥ T
⇤
1
1
3
2 0
1 0
1
1 0
A
0 =
!
.
0
1
2
1 0
0 1
2
1 0
Thus null(AT ) is the set of solutions of this augmented matrix, which is
2
3
2 3
2 3
02 3 2
s+t
1
1
1
6 2s t 7
6 27
6 17
B6 27 6
6
7
6 7
6 7
B6 7 6
4 s 5 = s 4 15 + t 4 05 = span @4 15 , 4
t
0
1
0
so these two vectors are a basis for W ? .
31
1
C
17
7C ,
5
0 A
1
384
CHAPTER 5. ORTHOGONALITY
13. Let
⇥
A = w1
w2
2
3
1 2
2 57
7.
3 65
2 1
2
⇤ 6 1
w3 = 6
4 6
3
Then the subspace W spanned by w1 , w2 , and w3 is the column space of A,
W ? = (col(A))? = null(AT ). To compute null(AT ), we row-reduce
2
3
2
4
2
1
6
3 0
1 0 3
h
i
3
6
7
6
1
AT 0 = 4 1
2
3
2 05 ! 4 0 1 0
3
2
5
6
1 0
0 0 0
0
so that by Theorem 5.10,
3
0
7
05 .
0
Thus null(AT ) is the set of solutions of this augmented matrix, which is
2
3
2 3
2 43
02 3 2 31
3s 43 t
3
3
4
3
6
7
6
7
6
B
6
7
6
7C
1
17
6
7
6 07
6 7
B6 07 6 17C
3t
6
7 = s 6 7 + t 6 3 7 = span B6 7 , 6 7C ,
4
5
4 15
4 05
@4 15 4 05A
s
t
0
1
0
3
so these two vectors form a basis for W ? .
14. Let
2
6
⇤ 6
v2 = 6
6
4
⇥
A = v1
4
6
1
1
1
1
2
0
1
3
3
2
27
7
27
7.
15
2
Then the subspace W spanned by w1 , w2 , and w3 is the column space of A,
W ? = (col(A))? = null(AT ). To compute null(AT ), we row-reduce
2
3
2
4
6
1
1
1
0
1 0 0
2
h
i 6
7
6
3
T
6
7
6
=
!
A
0
0
1
3 05
41 2
40 1 0
2
2 2
2
1
2 0
0 0 1
0
Thus null(AT ) is the set of solutions of this augmented matrix,
2
3
2 3
2
2s 7t
2
6 3
7
6 37
6
6 s + 5t7
6 7
6
6 2
7
6 27
6
6
7
6 7
6
6
7 = s 6 07 + t 6
t
6
7
6 7
6
6
7
6 17
6
s
4
5
4 5
4
t
0
Clearing fractions, we get
2
6
6
6
6
4
as a basis for W ? .
15. We have
projW (v) =
✓
u1 · v
u1 · u1
◆
3
4
37
7
07
7,
25
0
2
6
6
6
6
4
3
7
57
7
17
7
05
1
so that by Theorem 5.10,
7
5
1
which is
3
7
7
57
7
7
17
7
07
5
1
1 · 7 + 1 · ( 4)
3
u1 =
u1 = u1 =
1·1+1·1
2
" #
3
2
3
2
0
3
7
07
5.
0
5.2. ORTHOGONAL COMPLEMENTS AND ORTHOGONAL PROJECTIONS
385
16. We have
projW (v) =
✓
=
=
=
◆
✓
◆
u1 · v
u2 · v
u1 +
u2
u1 · u1
u2 · u2
1 · 3 + 1 · 1 + 1 · ( 2)
1·3
u1 +
1·1+1·1+1·1
1·1
2
u1 + u2
3
2 3
2 3 2 3
5
1
1
37
24 5 4 5 6
17
1 +
1 =6
4
35
3
1
0
2
1 · 1 + 0 · ( 2)
u2
1 · ( 1) + 0 · 0
3
17. We have
◆
✓
◆
u1 · v
u2 · v
u1 +
u2
u1 · u1
u2 · u2
2·1 2·2+1·3
1·1+2·2+4·3
=
u1 +
u2
2 · 2 2 · ( 2) + 1 · 1
1 · ( 1) + 1 · 1 + 4 · 4
1
13
= u1 + u2
9
2 3 182
3 2 3
projW (v) =
✓
6
=6
4
2
97
27
95
1
9
6
+6
4
13
18 7
13 7
18 5
52
18
6
=6
4
1
27
1 7.
25
3
18. We have
projW (v) =
=
=
=
◆
✓
◆
✓
◆
u1 · v
u2 · v
u3 · v
u1 +
u2 +
u3
u1 · u1
u2 · u2
u3 · u3
1 · 3 + 1 · ( 2) + 0 · 4 + 0 · ( 3)
1 · 3 1 · ( 2)
u1 +
1·1+1·1+0·0+0·0
1 · 1 1 · ( 1)
0 · 3 + 0 · ( 2) + 1 · 4 + 1 · ( 3)
u3
0·0+0·0+1·1+1·1
1
1
1
u1
u2 + u3
2 2 3 2 2 23
2 3 2 3
1
1
0
0
7 1 6 17 1 607 617
16
1
6 7
6 7 + 6 7 = 6 7.
2 405 2 4 15 2 415 415
0
1
1
0
✓
19. Since W is spanned by the vector u1 =
projW (v) =
✓
u1 · v
u1 · u1

◆
1 · 4 + 1 · ( 3)
u2 +
1 · ( 1) + 1 · 1
1
, we have
3
"

2 1
=
5 3
1 · 2 + 3 · ( 2)
u1 =
u1 =
1·1+3·3
The component of v orthogonal to W is thus
projW ? (v) = v
projW (v) =
"
12
5
4
5
#
.
2
5
6
5
#
.
386
CHAPTER 5. ORTHOGONALITY
2 3
1
20. Since W is spanned by the vector u1 = 415, we have
1
projW (v) =
✓
u1 · v
u1 · u1
◆
2 3
2 3
4
1
37
1 · 3 + 1 · 2 + 1 · ( 1)
44 5 6
47
1 =6
u1 =
u1 =
4 3 5.
1·1+1·1+1·1
3
1
4
3
The component of v orthogonal to W is thus
projW ? (v) = v
21. Since W is spanned by
we have
2 3
1
u1 = 425
1
2
6
projW (v) = 6
4
and
3
5
37
27 .
35
7
3
2
3
1
u2 = 4 15 ,
1
✓
◆
✓
◆
u1 · v
u2 · v
projW (v) =
u1 +
u2
u1 · u1
u2 · u2
1 · 4 + 2 · ( 2) + 1 · 3
1·4
=
u1 +
1·1+2·2+1·1
1·1
1
= u1 + 3u2
2 3 2 3 2 3
2
1
7
3
2
2
6 7 6 7 6 7
= 4 1 5 + 4 35 = 4 25 .
1
7
3
2
2
1 · ( 2) + 1 · 3
u2
1 · ( 1) + 1 · 1
The component of v orthogonal to W is thus
projW ? (v) = v
22. Since W is spanned by
we have
2 3
1
617
6
u1 = 4 7
15
0
1
2
3
6 7
projW (v) = 4 05 .
and
1
2
2
3
1
6 07
7
u2 = 6
4 15 ,
1
◆
✓
◆
u1 · v
u2 · v
u1 +
u2
u1 · u1
u2 · u2
1 · 2 + 1 · ( 1) + 1 · 5 + 0 · 6
1 · 2 + 0 · ( 1) 1 · 5 + 1 · 6
=
u1 +
u2
1·1+1·1+1·1+0·0
1 · 1 + 0 · 0 1 · ( 1) + 1 · 1
= 2u1 + u2
2 3 2 3 2 3
1
1
3
617 6 07 627
7 6 7 6 7
= 26
415 + 4 15 = 415 .
0
1
1
projW (v) =
✓
2
5.2. ORTHOGONAL COMPLEMENTS AND ORTHOGONAL PROJECTIONS
387
The component of v orthogonal to W is thus
2
3
1
6 37
7
projW (v) = 6
4 45 .
5
projW ? (v) = v
23. W ? is defined to be the set of all vectors in Rn that are orthogonal to every vector in W . So if
w 2 W \ W ? , then w · w = 0. But this means that w = 0. Thus 0 is the only element of W \ W ? .
24. If v 2 W ? , then since v is orthogonal to every vector in W , it is certainly orthogonal to wi , so that
v · wi = 0 for i = 1, 2, . . . , k. For the converse, suppose w 2 W ; then since W = span(w1 , . . . , wk ),
Pk
there are scalars ai such that w = i=1 ai wi . Then
v·w =v·
k
X
ai wi =
i=1
k
X
i=1
ai v · wi = 0
since v is orthogonal to each of the wi . But this means that v is orthogonal to every vector in W , so
that v 2 W ? .
2 3
2 3
1
0
25. No, it is not true. For example, let W ⇢ R3 be the xy-plane, spanned by e1 = 405 and e2 = 415.
0
0
02 31
0
Then W ? = span(e3 ) = span @405A. Now let w = e1 and w0 = e2 ; then w · w0 = 0 but w0 2
/ W ?.
1
26. Yes, this is true. Let V = span(vk+1 , . . . , vn ); we want to show that V = W ? . P
By Theorem 5.9(d),
n
all we must show is that for every v 2 V , vi · v = 0 for i = 1, 2, . . . , k. But v = j=k+1 cj vj , so that
if i  k, since the vi are an orthogonal basis for Rn we get
vi ·
n
X
cj v j =
j=k+1
n
X
j=k+1
cj vj · vi = 0.
27. Let {uk } be an orthonormal basis for W . If
◆
n ✓
n
X
X
uk · x
x = projW (x) =
uk =
(uk · x)uk ,
uk · uk
k=1
k=1
then
Pn x is in the span of the uk , which is W . For the converse, suppose that x 2 W . Then x =
k=1 ak uk . Since the uk are an orthogonal basis for W , we get
x · ui =
Therefore
x=
n
X
n
X
k=1
ak u k · u i = ai u i · u i = ai .
ak u k =
k=1
n
X
k=1
(x · uk )uk = projW (x).
28. Let {uk } be an orthonormal basis for W . Then
◆
n ✓
n
X
X
uk · x
projW (x) =
uk =
(uk · x)uk .
uk · uk
k=1
k=1
So if projW (x) = 0, then uk · x = 0 for all k since the uk are linearly independent. But then by
Theorem 4.9(d), x 2 W ? , so that x is orthogonal to W . Conversely, if x is orthogonal to W , then
uk · x = 0 for all k, so the sum above is zero and thus projW (x) = 0.
388
CHAPTER 5. ORTHOGONALITY
29. If uk is an orthonormal basis for W , then
projW (x) =
n
X
k=1
(uk · x)uk 2 W.
By Exercise 27, x 2 W if and only if projW (x) = x. Applying this to projW (x) 2 W gives
projW (projW (x)) = projW (x)
as desired.
30. (a) Let W = span(S) = span{v1 , . . . , vk }, and let T = {b1 , . . . , bj } be an orthonormal basis for
W ? . By Theorem 5.13, dim W + dim W ? = k + j = n. Therefore S [ T is a basis for Rn .
Further, it is an orthonormal basis: any two distinct members of S are orthogonal since S is an
orthonormal set; similarly, any two distinct members of T are orthogonal. Also, vi · bj = 0 since
one of the vectors is in W and the other is in W ? . Finally, vi · vi = bj · bj = 1 since both sets
are orthonormal sets.
Choose x 2 Rn ; then there are ↵r , s such that
x=
k
X
↵r v r +
r=1
j
X
s bs .
s=1
Then using orthonormality of S [ T , we have
!
j
k
X
X
x · vi =
↵r v r +
· vi = ↵i and x · bi =
s bs
r=1
s=1
k
X
↵r vr +
r=1
j
X
s bs
s=1
!
· bi =
i.
But then (again using orthonormality)
2
kxk = x · x
=
k
X
↵r vr +
r=1
=
k
X
r=1
=
k
X
r=1
k
X
r=1
↵r2 +
j
X
s=1
j
X
s bs
!
k
X
↵r v r +
r=1
j
X
s bs
s=1
!
2
s
s=1
2
|x · vr | +
j
X
s=1
|x · bs |
2
2
|x · vr | ,
as required.
(b) Looking at the string of equalities and inequalities above, we see that equality holds if and only
Pj
2
if
s=1 |x · bs | = 0. But this means that x · bs = 0 for all s, which in turn means that
x 2 span{vj } = W .
5.3
The Gram-Schmidt Process and the QR Factorization
1. First obtain an orthogonal basis:
" #
1
v1 = x1 =
1
v2 = x2
" #
1
projv1 (x2 )v1 =
2
" # " #
1
1·1+1·2 1
=
1·1+1·1 1
2
" #
3
2
3
2
=
"
1
2
1
2
#
.
5.3. THE GRAM-SCHMIDT PROCESS AND THE QR FACTORIZATION
389
Normalizing each of the vectors gives
" # " #
p1
v1
1 1
q1 =
=p
= 12 ,
kv1 k
p
2 1
2
v2
1
q2 =
= p
kv2 k
1/ 2
"
1
2
1
2
#
=
"
p1
2
p1
2
#
.
2. First obtain an orthogonal basis:
v1 = x1 =

3
3


3·3 3·1
3
3
projv1 (x2 )v1 =
3
1
3 · 3 3 · ( 3)




1
3
3
1
2
=
=
.
3
1
1
2
3
v2 = x2

3
=
1
Normalizing each of the vectors gives
v1
1
q1 =
=p
kv1 k
18
"
#
3
3
=
"
p1
2
p1
2
#
,
" # " #
p1
v2
1 2
q2 =
=p
= 12
kv2 k
p
8 2
2
3. First obtain an orthogonal basis:
2
3
1
v 1 = x 1 = 4 15
1
v2 = x2
2 3
0
projv1 (x2 )v1 = 435
3
1·0 1·3
1 · 1 1 · ( 1)
v3 = x3 projv1 (x3 )v1 projv2 (x3 )v2
2 3
2 3
3
1
1
·
3
1
·
2
1
·
4
4 15
= 425
1 · 1 1 · ( 1) 1 · ( 1)
4
1
2 3 2 3 2 3 2 3
2
1
0
2
1·3
4 15 = 435 + 4 25 = 415
1 · ( 1)
1
3
2
1
2 3 2 3 2 3
2
3
1
2·3+1·2+1·44 5 4 5 4 5
1 = 2 +
1
2·2+1·1+1·1
1
4
1
Normalizing each of the vectors gives
2
3 2
3
p1
1
37
6
v1
1 6 7
7 = 6 p1 7 ,
q1 =
=p 6
1
kv1 k
3 4 5 4 13 5
p
1
3
2 3 2 3
2
p
2
7 6 16 7
v2
1 6
6
7
6
q2 =
= p 415 = 4 p6 7
5,
kv2 k
6
1
p
1
6
2 3 2
3
0
0
6 1 7
v3
1 6 7
p 7.
q3 =
=p 6
=6
17
4
5
4
25
kv3 k
2
p1
1
2
2 3 2 3
2
0
2 415 = 4 15 .
1
1
390
CHAPTER 5. ORTHOGONALITY
4. First obtain an orthogonal basis:
2 3
1
v1 = x1 = 415
1
2 3
1
projv1 (x2 )v1 = 415
0
v2 = x2
2 3
1
1·1+1·1+1·0
v1 = 415
1·1+1·1+1·1
0
v3 = x3 projv1 (x3 )v1 projv2 (x3 )v2
2 3
1
1
1
1·1+1·0+1·0
3 ·1+ 3 ·0
= 405
v1
1 1
1 1
1·1+1·1+1·1
3 · 3 + 3 · 3
0
2 3
2 3
2 3
1
1
1
6 7 1 6 7 1 6 37
7
6 7
6 17
=6
405 3 415 2 4 3 5
2
0
1
3
2 3
6
=6
4
2
3
2
3
·
·0
2
2 3
1
24 5 6
1 =6
4
3
1
3
1
37
17
35
2
3
v2
2
3
1
27
1 7.
25
0
Normalizing each of the vectors gives
2 3 2 3
p1
1
7 6 13 7
v1
1 6
q1 =
=p 6
17 = 6 p 7 ,
kv1 k
3 4 5 4 13 5
p
1
3
q3 =
q2 =
2
v3
1 6
= p 6
kv3 k
1/ 2 4
2
v2
1 6
6
=p
kv2 k
2/3 4
3
1
27
17
25
0
2
6
=6
4
3
p1
27
p1 7
25
3
1
37
17
25
2
3
2
6
=6
4
3
p1
67
p1 7 ,
65
p2
6
0
5. Using Gram-Schmidt, we get
2 3
1
6 7
v1 = x1 = 415
0
v2 = x2
2 3
3
6 7
projv1 (x2 )v1 = 445
2
2 3 2 3
1
3
1·3+1·4+0·26 7 6 7
415 = 445
1·1+1·1+0·0
0
2
2 3 2 3
1
1
7 6 7 6 21 7
415 = 4 2 5 .
2
0
2
5.3. THE GRAM-SCHMIDT PROCESS AND THE QR FACTORIZATION
391
6. Using Gram-Schmidt, we get
2
3
2
6 17
7
v1 = x1 = 6
4 15
2
v2 = x2
2
2
6
projv1 (x2 )v1 = 6
4
3
3
6 17
7
=6
4 05
4
2
3
2
6
7 6
36
6 17 = 6
2 4 15 6
4
2
2
3
3
17
7
05
4
3
0
17
7
27
37
25
1
2·3
2·2
1 · ( 1) + 1 · 0 + 2 · 4
v1
1 · ( 1) + 1 · 1 + 2 · 2
v3 = x3 projv1 (x3 )v1 projv2 (x3 )v2
2 3
1
617
2·1 1·1+1·1+2·1
7
=6
415 2 · 2 1 · ( 1) + 1 · 1 + 2 · 2 v1
1
2 3 2 3
1
2 3
2 3
0
1
2
6 1 7 6 57 7
617 2 6 17
6 27 657
7
6 7
6 7 6 7
=6
415 5 4 15 + 0 6 3 7 = 6 3 7.
4 25 455
1
2
1
1
5
0 · 1 + 12 · 1
0 · 0 + 12 · 12
3
2
3
2
·1+1·1
v2
3
·
2 +1·1
7. First compute the projection of v on each of the vectors v1 and v2 from Exercise 5:
2 3 2 3
1
0
v1 · v
1 · 4 + 1 · ( 4) + 0 · 3 6 7 6 7
projv1 v =
v1 =
=
415 405
v1 · v1
1·1+1·1+0·0
0
0
2 3
projv2 v =
v2 · v
v2 =
v2 · v2
1
2
·4+
1 2
2
1
2
· ( 4) + 2 · 3 6
4
2
+ 12 + 22
Then the component of v orthogonal to both v1 and v2 is
2
3
4
6 7
4 45
3
2
6
4
3
2
9
27
95
8
9
2
6
=4
so that
2
6
v=4
3
38
9
38 7
9 5
19
9
3
38
9
38 7
,
9 5
19
9
4
+ v2 .
9
1
2
17
25
2
=
2
46
4
9
3
1
2
17
25
2
2
6
=4
3
2
9
27
.
95
8
9
392
CHAPTER 5. ORTHOGONALITY
8. First compute the projection of v on each of the vectors v1 and v2 from Exercise 6:
2 3
2 3 2 23
2
2
5
7 1 6 17 6 1 7
1
v1 · v
2·1 1·4+1·0+2·26
6 7
6 7 6 57
projv1 v =
v1 =
6 7 = 6 7 = 6 17
v1 · v1
22 + 12 + 12 + 22
4 15 5 4 15 4 5 5
2
2
5
3
2 3 2
0
0
6
6
17
17
27 = 8 6 27 = 6
6 37 6
37
5 7 4 25 4
2
2
projv2 v =
projv3 v =
v3 · v
v3 =
v3 · v3
1
5 ·1+
1 2
+
5
7
5 ·4
7 2
5
Then the component of v orthogonal to both
2 23
2 3
5
4
6 17
6 7 6 57
4 45 6 1 7
4 55
3
2
5
so that
3
2 ·
3 2
2
0 · 1 + f rac12 · 4
v2 · v
v2 =
2
v2 · v2
02 + 12 +
2
6
6
v=6
4
3
1
12
1 7
84 7
1 7
5
28
5
84
+
+
2
0+1·26
6
6
2
4
+1
213
5
677
3
1
·
0
+
·
2
6
7
5
5
657
3 2
1 2 435
+ 5
5
5
1
5
1
=
31
12
v1 and v2 is
2
3 2 31 3 2
0
60
6 4 7 6 217 7 6
6 7 7 6 60 7 6
6 12 7 6 93 7 = 6
4 7 5 4 60 5 4
8
7
31
60
213
5
677
657
637
455
1
5
=
1
2
3
31
60
6 217 7
6 60 7
6 93 7 .
4 60 5
31
60
3
0
47
77
12 7
5
7
8
7
3
1
12
1 7
84 7 ,
1 7
5
28
5
84
1
8
31
+ v1 + v2 + v3 .
5
7
12
9. To find the column space of the matrix, row-reduce it:
2
3
2
0 1 1
1
41 0 15 ! 40
1 1 0
0
0
1
0
3
0
05 .
1
It follows that, since there are 1s in all three columns, the columns of the original matrix are linearly
independent, so they form a basis for the column space. We must orthogonalize those three vectors.
2 3
0
6 7
v1 = x1 = 415
1
2 3
2 3 2 3
2 3 2 3
1
0
1
0
1
6 7 0 · 1 + 1 · 0 + 1 · 1 6 7 6 7 1 6 7 6 17
v2 = x2 projv1 (x2 )v1 = 405
415 = 405
4 15 = 4 2 5
0·0+1·1+1·1
2
1
1
1
1
1
2
v3 = x3 projv1 (x3 )v1 projv2 (x3 )v2
2 3
2 3
1
0
1·1
6 7 0·1+1·1+1·06 7
= 4 15
415
0·0+1·1+1·1
1·1
0
1
2 3
2 3
2 3 2 3
2
1
0
1
3
6 7 1 6 7 1 6 17 6 27
= 4 15
415
4 25 = 4 35 .
2
3
1
2
0
1
2
3
1
2
1
2
·
·1+
1
2
1
2
·0
+ 12 ·
1
2
2
6
4
1
3
17
25
1
2
5.3. THE GRAM-SCHMIDT PROCESS AND THE QR FACTORIZATION
10. To find the column space of the matrix, row-reduce it:
2
3
2
1
1 1
1
6 1
7
60
1
2
6
7!6
4 1
40
1 05
1
5 1
0
0
1
0
0
393
3
0
07
7.
15
0
It follows that, since there are 1s in all three columns, the columns of the original matrix are linearly
independent, so they form a basis for the column space. We must orthogonalize those three vectors.
2 3
1
6 17
7
v1 = x1 = 6
4 15
1
2 3
2 3 2 3 2 3 2 3
1
1
1
1
0
6 17 1 · 1 + 1 · ( 1) 1 · 1 + 1 · 5 6 17 6 17 6 17 6 27
7
6 7 6 7 6 7 6 7
v2 = x2 projv1 (x2 )v1 = 6
4 15 1 · 1 + 1 · 1 1 · ( 1) + 1 · 1 4 15 = 4 15 4 15 = 4 25
5
1
5
1
4
v3 = x3 projv1 (x3 )v1 projv2 (x3 )v2
2 3
2
1
627
1·1+1·2 1·0+1·1 6
7
6
=6
405 1 · 1 + 1 · 1 1 · ( 1) + 1 · 1 4
1
2 3 2 3
2 3 2 3
1
1
0
0
627 6 17
6 27 617
7 6 7
6 7 6 7
=6
405 4 15 + 0 4 25 = 415 .
1
1
4
0
3
1
17
7
15
1
2
3
0
7
0·1 2·2+2·0+4·1 6
6 27
0 · 0 2 · ( 2) + 2 · 2 + 4 · 4 4 25
4
2 3
2 3
0
0
11. The given vector, x1 , together with x2 = 415 and x3 = 405 form a basis for R3 ; we want to
0
1
orthogonalize that basis.
2 3
3
6 7
v1 = x1 = 415
5
2 3
2 3 2 3
2 3 2
3
3
0
3
0
3
35
1 6 7 6 34 7
6 7 3·0+1·1+5·06 7 6 7
v2 = x2 projv1 (x2 )v1 = 415
415 = 415
415 = 4 35 5
2
2
2
3 +1 +5
35
1
0
5
0
5
7
12. Let
v3 = x3 projv1 (x3 )v1 projv2 (x3 )v2
2 3
2 3
0
3
3
6 7 3·0+1·0+5·16 7
35 · 0 +
= 405
1
4
5
3 2
32 + 12 + 52
+
35
1
5
2 3
2 3
2
3 2
3
3
15
0
3
35
34
5 6 34 7 6
6 7 16 7
7
= 405
4 15 +
4 35 5 = 4 05 .
7
34
1
9
1
5
7
34
2
3
1
6 27
7
v1 = 6
4 15 ,
0
34
35 · 0
34 2
35
1
7
+
2 3
1
607
7
v2 = 6
415 .
3
2
·1 6
4
1 2
7
3
3
35
34 7
35 5
1
7
394
CHAPTER 5. ORTHOGONALITY
Note that v1 · v2 = 0. Now let
2 3
0
607
7
x3 = 6
415 ,
0
2 3
0
607
7
x4 = 6
405 .
1
To see that v1 , v2 , x3 , and x4 form a basis for R4 , row-reduce the matrix that has these vectors as its
columns:
2
3
2
3
1 1 0 0
1 0 0 0
6 2 0 0 07
6 0 1 0 07
6
7
6
7
4 1 1 1 05 ! 40 0 1 05 .
0 3 0 1
0 0 0 1
Since there are 1s in all four columns, the column vectors are linearly independent, so they form a
basis for R4 . Next we need to orthogonalize that basis. Since v1 and v2 are already orthogonal, we
start with x3 :
v3 = x3
✓
v1 · x3
v1 · v1
◆
v1
✓
2 3
2 3
0
1
✓
◆
607 1 6 27
v2 · x3
7
6 7
v2 = 6
415 + 6 4 15
v2 · v2
0
0
◆
✓
◆
✓
◆
v1 · x4
v2 · x4
v3 · x4
v4 = x4
v1
v2
v3
v1 · v1
v2 · v2
v3 · v3
2 3
2 3
2 3
2
3 2
3
5
12
0
1
1
66
49
6 7
6 7
6 7
6 1 7 6 6 7
607
6 27
7 6
7
07
3 6
18 6
6
7
6
7
6
7
6 3 7 = 6 49 7.
= 6 7 + 06 7
+
7
6
6
7
6
7
49
405
4 15 11 415 49 4 66 5 4 0 5
3
4
1
0
3
11
49
2
2 3
1
6
7 6
1 6
6 07 = 6
11 415 6
4
3
3
5
66
1 7
7
3 7
49 7
66 5
3
11
2 3
0
13. To start, we let x3 = 405 be the third column of the matrix. Then the determinant of this matrix
1
is (for example, expanding along the second row) p16 6= 0, so the resulting matrix is invertible and
therefore the three column vectors form a basis for R3 . The first two vectors are orthogonal and each
is of unit length, so we must orthogonalize the third vector as well:
v3 = x3
✓
q1 · x3
q1 · q1
◆
q1
✓
q2 · x3
q2 · q2
◆
q2
2 3
2 1 3
2 1 3
p
p
0
3
2
p1
p1
6 7
6
7
6 p1 7
2
3
= 405 ⇣ ⌘2
05 ⇣ ⌘ 2 ⇣ ⌘ 2 ⇣ ⌘ 2 4 3 5
⇣
⌘2 4
p1
p1
p1
+ 02 +
+ p13 + p13
p1
p1
1
2
2
3
2
3
2 3
2 1 3
2 1 3 2 3
1
p
p
0
2
1 6
1 6 13 7 6 61 7
6 7
7
= 405 + p 4
05 p 4 p3 5 = 4 3 5 .
2
3 1
1
1
p
p
1
6
2
3
Then v3 is orthogonal to the other two columns of the matrix, but we must normalize v3 to convert it
to a unit vector.
s✓ ◆
✓
◆2 ✓ ◆2
2
1
1
1
1
kv3 k =
+
+
=p ,
6
3
6
6
5.3. THE GRAM-SCHMIDT PROCESS AND THE QR FACTORIZATION
395
so that
q3 =
2
1
6
p v3 = 4
1/ 6
3
p1
6
7
p2 5 ,
6
p1
6
so that the completed matrix Q is
2
6
Q=4
p1
2
p1
3
p1
3
p1
3
0
p1
2
3
p1
6
7
p2 5 .
6
p1
6
2 3
2 3
1
2
617
6 17
7
6 7
14. To simplify the calculations, we will start with the multiples v1 = 6
415 and v2 = 4 05 of the given
1
3
basis vectors (column vectors of Q). We will find an orthogonal basis containing those two vectors and
then normalize all four at the very end. Note that v1 and v2 are already orthogonal. If we choose
2 3
2 3
0
0
607
6 07
7
6 7
x3 = 6
415 and x4 = 405 ,
0
1
⇥
⇤
then calculating the determinant of v1 v2 x3 x4 , for example by expanding along the last column,
shows that the matrix is invertible, so that these four vectors form a basis for R4 . We now need to
orthogonalize the last two vectors:
p3 = x3
2 3
0
607
6 7
=6 7
415
0
2 3
0
607
6 7
=6 7
415
0
✓
v1 · x3
v1 · v1
◆
v1
✓
v2 · x3
v2 · v2
◆
v2
2 3
1
6
1 · 0 + 1 · 0 + 1 · 1 + 1 · 0 6 17
7
6 7
1 · 1 + 1 · 1 + 1 · 1 + 1 · 1 4 15
1
2 3
2 3
1
1
7 1 6 17
1
16
6 7
6 7
6 7 = 6 7.
4 415 4 4 35
1
1
2
2·0+1·0+0·1 3·0 6
6
6
2 · 2 + 1 · 1 + 0 · 0 3 · ( 3) 4
3
2
17
7
7
05
3
396
CHAPTER 5. ORTHOGONALITY
2
6
6
Now let v3 = 4p3 = 6
4
v4 = x4
2 3
0
607
6 7
=6 7
405
3
1
17
7
7, again avoiding fractions. Next,
35
1
✓
v1 · x4
v1 · v1
1
✓
v2 · x4
v2 · v2
◆
✓
v2
2 3
1
6
1 · 0 + 1 · 0 + 1 · 0 + 1 · 1 6 17
7
6 7
1 · 1 + 1 · 1 + 1 · 1 + 1 · 1 4 15
v3 · x4
v3 · v3
◆
v3
2
2·0+1·0+0·0 3·1 6
6
6
2 · 2 + 1 · 1 + 0 · 0 3 · ( 3) 4
2
6
1·0 1·0+3·0 1·1
6
6
1 · ( 1) 1 · ( 1) + 3 · 3 1 · ( 1) 4
1
6
6
=6
4
v1
1
2 3
0
607
6 7
=6 7
405
2
◆
3
2 3
2 3
2 3
1
2
1
7
6 17
6 17
1
16
3
1
6 7
6 7
6 7
6 7+
6 7+
6 7
4 415 14 4 05 12 4 35
1
3
1
3
1
17
7
7
35
1
3
2
17
7
7
05
3
2
21
5 7
42 7 .
7
05
1
42
We now normalize v1 through v4 ; note that we already have normalized forms for the first two.
p
p
kv3 k = 12 + 12 + 32 + 12 = 2 3
s✓ ◆
✓
◆2
✓ ◆2
2
2
5
1
1
kv4 k =
+
+ 02 +
=p ,
21
42
42
42
so that the matrix is
Q=
15. From Exercise 9, kv1 k =
p
2, kv2 k =
h
Q = q1
Therefore
2
0
6
R = QT A = 4 p26
p1
3
21
p
6
2 ,
p
3
p6
3
p6
3
2p
3
6
p2
14
p1
14
2
61
62
6
61
42
1
2
0
p3
14
and kv3 k =
q2
p1
2
p1
6
p1
3
i
2
p
2 3
3 ,
0 7
5
p1
42
so that
p2
6
p1
6
p1
6
0
6
q3 = 4 p12
p1
2
32
p1
0
2
76
p1 5 41
6
p1
1
3
3
p4
42
7
p5 7
42 7 .
1
0
1
3
p1
3
7
p1 5 .
3
p1
3
3 2p
2
1
7 6
15 = 4 0
0
0
p1
2
p3
6
0
3
p1
2
7
p1 5 .
6
p2
3
5.3. THE GRAM-SCHMIDT PROCESS AND THE QR FACTORIZATION
p
p
16. From Exercise 10, kv1 k = 2, kv2 k = 2 6, and kv3 k = 2, so that
h
Q = q1
Therefore
2
1
62
1
p2
6
6
p1
2
R = QT A = 6
40
0
17. We have
2
2
3
61
43
2
3
R = QT A =
18. We have
T
R=Q A=
q2
"
p1
6
p1
3
q3
2
6
6
=6
6
4
i
3
2
1
p2
6
6
p1
2
1
6
p2 7 6
67 6
3 56
1
3
2
3
2
3
2
3
27 6
35 4
1
3
p2
6
0
0
p
3
0
7
p1 7
27 .
7
p1 5
2
0
3
2
1 1
2
7
6
1 27
7 = 60
4
1 07
5
0
5 1
1
1
2
1
2
0
6
p6
6
p6
6
3
4 1
1
32
p1
6
p1
3
1
2
1
2
1
2
1
2
397
2
p
2 6
0
3 2
2
3 9
7 6
15 = 4 0 6
1
0 0
8
7
2
2
1
#
0 6
6 2
6
1
p
4 1
3
0
3
3
"p
47
6
7
7=
15
0
2
3
7
.
0 7
p 5
2
3
1
3
27
35
7
3
p #
2 6
p
.
3
1
19. If A is already orthogonal, let Q = A and R = I, which is upper triangular; then A = QR.
20. First, if A is invertible, then A has linearly independent columns, so that by Theorem 5.16, A = QR
where R is invertible. Since R is upper triangular, its determinant is the product of its diagonal entries,
so that R invertible implies that all of its diagonal entries are nonzero. For the converse, if R is upper
triangular with all diagonal entries nonzero, then it has nonzero determinant; since Q is orthogonal, it
is invertible so has nonzero determinant as well. But then det A = det(QR) = (det Q)(det R) 6= 0, so
that A is invertible. Note that when this occurs,
A
1
= (QR)
1
=R
1
Q
1
=R
1
QT .
⇥
⇤
21. From Exercise 20, A 1 = R 1 QT . To compute R 1 , we row-reduce R I using Exercise 15:
2p
3
2
p
p
p 3
6
3
2 p12 p12 1 0 0
1 0 0 22
p6
p6 7
6
7
6
6
37 .
6 0 p3
6
p1
0 1 07
0
4
5 ! 40 1 0
3
6
6
p6 5
3
0 0 1
0
0
0
0 p23 0 0 1
2
Thus
2p
2
2
A
1
=R
22. From Exercise 20, A
1
6
QT = 6
4 0
0
1
p
6
p6
6
3
0
p
32
= R 1 QT . To compute R 1 ,
2
3
2
3 9 13 1 0 0
1
6
7
6
2
4 0 6 3 0 1 05 ! 4 0
0 0 73 0 0 1
0
3
p1
2
p1
6
p1
3
3
0
p6 7 6
3 7 6 p2
p6 5 4 6
3
p1
2
3
p1
27
p1 7
65
p1
3
6
=6
4
⇥
we row-reduce R
0
1
0
0
0
1
1
3
0
0
1
2
1
6
0
2
1
2
1
2
1
2
1
2
1
2
1
2
3
1
27
17 .
25
1
2
⇤
I using Exercise 17:
3
2
21
1 7
.
21 5
3
7
398
CHAPTER 5. ORTHOGONALITY
Thus
A
1
1
=R
2
1
3
32
1
2
1
6
6
QT = 4 0
0
2
2
21
3
7
6
1
1
5
4
21
3
3
2
7
3
0
3
1
3
2
3
2
3
2
3
27
35
1
3
=
2
5
42
61
4 42
2
7
3
2
7
1
7
2
7
11
21
2 7
.
21 5
1
7
23. If A = QR is a QR-factorization of A, then we want to show that Rx = 0 implies that x = 0 so that
R is invertible by part (c) of the Fundamental Theorem. Note that since Q is orthogonal, so is QT , so
that by Theorem 5.6(b), QT x = kxk for any vector x. Then
(QT A)x = QT (Ax) = kAxk = 0 ) Ax = 0.
Rx = 0 ) kRxk = 0 )
P
If ai are the columns of A, then Ax = 0 means that
ai xi = 0. But this means that all of the xi are
zero since the columns of A are linearly independent. Thus x = 0. By part (c) of the Fundamental
Theorem, we conclude that R is invertible.
24. Recall that row(A) = row(B) if and only if there is an invertible matrix M such that A = M B. Then
A = QR implies that AT = RT QT ; since A has linearly independent columns, R is invertible, so that
RT is as well. Thus row(AT ) = row(QT ), so that col(A) = col(Q).
Exploration: The Modified QR Process
1. I
T
2uu = I

d ⇥
2 1 d1
d2
d2 = I
2. (a) From Exercise 1,
Q=
(b) Since kx
yk =
"
p
1

⇤

2d1 d2
1 2d21
=
2
2d2
2d1 d2
# "
9
2d1 d2
1 2 · 25
=
1 2d22
2 · 35 · 45
2d21
2d1 d2
(5
2d21
2d1 d2
1
2d1 d2
.
2d22
# "
2 · 53 · 45
=
1 2 · 16
25
7
25
24
25
24
25
7
25
⌘3
"
#
.
p
7)2 = 2 5, we have
1)2 + (5
1
u = p (x
2 5
y) =
"
4
p
2 5
2
p
2 5
#
=
"
p2
5
p1
5
#
.
Then from Exercise 1,
Q=
3. (a) QT = (I
"
1
2d21
2d1 d2
2uuT )T = I
2d1 d2
1
2d22
#
2 uuT
2
=4
T
2·
⇣
·
4
5
2 uT
T
1
2·
=I
p2
5
p1
5
p2
5
2·
⌘
·
2·
1
uT = I
⇣
p1
5
1
5
5=
3
5
4
5
4
5
3
5
#
.
2uuT = Q, so that Q is symmetric.
(b) Start with
QQT = QQ = (I
2uuT )(I
2uuT ) = I
4uuT +
2uuT
2
=I
4uuT + 4uuT uuT .
Now, uT u = u · u = 1 since u is a unit vector, so the equation above becomes
QQT = I
Thus QT = Q
1
4uuT + 4uuT uuT = I
4uuT + 4uuT = I.
and it follows that Q is orthogonal.
(c) This follows from parts (a) and (b): Q2 = QQ = QQT = QQ
1
= I.
Exploration: The Modified QR Process
399
4. First suppose that v is in the span of u; say v = cu. Then again using the fact that uT u = u · u = 1,
2uuT )(cu) = c(u
Qv = Q(cu) = (I
If v · u = uT v = 0, then
2uuT )v = v
Qv = (I
5. First normalize the given vector to get
Then
2
6
uuT = 4
1
6
1
6
1
3
1
6
1
6
1
3
2uuT u) = c(u
2
6
u=6
4
3
1
3
17
,
35
2
3
2u) =
cu =
v.
2uuT v = v.
3
p1
67
p1 7 .
65
p2
6
2
2
3
1
3
2
3
6
2uuT = 4
so that Q = I
1
3
2
3
2
3
3
2
3
27
.
35
1
3
For Exercise 3, clearly Q is symmetric, and each column has norm 1. A short computation shows that
each pair of columns is orthogonal. Therefore Q is orthogonal. Finally,
2
32
3
6
Q =4
2
2
3
1
3
2
3
1
3
2
3
2
3
For Exercise 4, first suppose that v = cu. Then
2
6
Qv = Q(cu) = cQu = c 6
4
If v · u = uT v = 0, then
h
so that x1
then
p1
6
p1
6
2
3
1
3
2
3
1
3
2
3
2
3
2
3
27 6
35 4
1
3
2
3
1
3
2
3
32
2
37 6
27 6
35 4
1
3
1
3
2
3
2
3
2
3
27
35
1
3
3
p1
67
p1 7
65
p2
6
2 3
i x1
p2
4x2 5 = p1 x1
6
6
x3
2
6
= c6
4
= I.
3
p1
67
p1 7
65
p2
6
=
cu =
v.
1
2
p x2 + p x3 = 0,
6
6
2 3
2 3
1
0
x2 + 2x3 = 0 and therefore x2 = x1 + 2x3 . Thus if v · u = 0, then v = s 415 + t 425. But
0
1
0 2 3
2 31
1
0
B 6 7
6 7C
Qv = Q @s 415 + t 425A
0
1
2 3
2 3
1
0
6 7
6 7
= sQ 415 + tQ 425
0
1
2
32 3
2
2
1
2
1
3
3
3
6
6
27 6 7
= s 4 13 23
+ t4
3 5 415
2
2
1
0
3
3
3
2 3
2 3
1
0
6 7
6 7
= s 415 + t 425 = v.
0
1
2
3
1
3
2
3
1
3
2
3
2
3
32 3
2
0
3
27 6 7
3 5 425
1
1
3
400
CHAPTER 5. ORTHOGONALITY
6. x
y is a multiple of u, so that Q(x
y) =
(x + y) · u =
(x
y) by Exercise 4. Next,
1
kx
yk
(x + y) · (x
y).
But by Exercise 61 in Section 1.2, this dot product is zero because kxk = kyk. Thus x+y is orthogonal
to u, so that Q(x + y) = x + y. So we have
Q(x
y) =
(x
y)
Q(x + y) = x + y
)
Qx
Qy =
Qx + Qy = x + y
7. We have
u=
so that
2
6
uuT = 4
Then
1
3
1
3
1
3
q
yk
3
1
3
1
3
1
3
Qx =
8. Since kyk =
1
kx
2
x+y
1
3
17
35
1
3
1
3
62
43
2
3
(x
2
3
2
1
y) = p 4 25 ,
2 3
2
2uuT =
) Q=I
2
3
1
3
2
3
) 2Qx = 2y ) Qx = y.
32 3
2
1
3
27 6 7
3 5 425
1
2
3
2
1
3
62
43
2
3
3
2
3
1
3
2
3
2
3
27
.
35
1
3
2 3
3
6 7
= 405 = y.
0
2
kxk = kxk, if ai is the ith column vector of A, we can apply Exercise 6 to get
⇥
Q 1 A = Q 1 a1
Q 1 a2
...
⇤ ⇥
Q1 an = Q1 x
Q1 a2
...
since the last m 1 entries in y are all zero. Here A1 is (m
the right-hand ⇤ is 1 ⇥ (n 1).
⇤
Q1 an
⇥
= y Q1 a2
1) ⇥ (n
...
Q1 an =
so that Q2 is orthogonal as well. Further,
1 0
Q2 Q1 A =
0 P2

2

⇤
⇤ ⇤
⇤
⇤
=
= 40
0 A1
0 P2 A 1
0
3
⇤ ⇤
⇤ ⇤ 5.
0 A2
10. Using induction, suppose that
Qk Qk
2
⇤
4
0
·
·
·
Q
A
=
1
1
0
⇤
⇤
0
3
⇤
⇤ 5.
Ak

⇤ ⇤
,
0 A1
1), the left-hand ⇤ is 1 ⇥ 1, and
9. Since P2 is a Householder matrix, it is orthogonal, so that


1 0
1 0
Q2 QT2 =
= I,
0 P2 0 P2T

⇤
Repeat Exercise 8 on the matrix Ak , giving a Householder matrix Pk+1 such that

⇤
⇤
Pk+1 Ak =
.
0 Ak+1
Exploration: The Modified QR Process
401

1
0
. Then using the fact that Pk+1 is orthogonal, we again conclude that Qk+1 is
0 Pk+1
orthogonal, and that
2
3
⇤ ⇤
⇤
⇤ 5.
Qk+1 Qk Qk 1 · · · Q1 A = 40 ⇤
0 0 Ak+1
Let Qk+1 =
The result then follows by induction. The resulting matrix R = Qm
by construction.
11. If Q = Q1 Q2 · · · Qm
1,
· · · Q2 Q1 A is upper triangular
then Q is orthogonal since each Qi is orthogonal, and
QT = (Q1 Q2 · · · Qm
Thus QT A = Qm
1
1)
T
= QTm
· · · QT2 QT1 = Qm
1
· · · Q2 Q1 .
1
· · · Q2 Q1 A = R, so that QQT A = A = QR, since QQT = I.

p
3
32 + ( 4)2 = 5
12. (a) We start with x =
, so that y =
. Then
4
0
"
#
"
#
1
p1
5
p2
5
u=
Then
Q1 = I
T
2uu =
"
1
5
2
5
) uuT =
3
5
4
5
4
5
3
5
#
2
5
4
5
.
"
5
) Q1 A =
0
#
1
= R,
2
3
9
and A = QT1 R.
2 3
2 3
1
3
(b) We start with x = 425, so that y = 405. Then
2
0
3
2
p1
37
p1 7
35
p1
3
6
u=6
4
Thus
Q1 = I
2uuT =
Then
2
2
1
3
1
3
1
3
6
) uuT = 6
4
1
3
62
43
2
3
3
2
3
1
3
2
3
"
2
3
27
35
1
3
4
A1 =
3
3
1
3
1
3
1
3
1
3
2
1
37
17 .
35
1
3
3
6
) Q1 A = 40
0
1
3
4
3
#
5 1
4 3
3 1
3
8
3
17
.
35
4
3
,
so that in the next iteration
x=
" #
4
)
3
y=
Then
T
uu =
So
P2 = I
2uuT =
"
4
5
3
5
3
5
4
5
#
" #
5
"
0
)
u=
1
10
3
10
3
10
9
10
) Q2 =
"
#
"
p1
10
p3
10
#
.
.
1 0
0 P2
#
2
1
6
= 40
0
0
4
5
3
5
0
3
37
.
55
4
5
402
CHAPTER 5. ORTHOGONALITY
Finally,
2
1
6
Q2 Q1 A = 40
0
0
0
32
3
37 6
5 5 40
4
0
5
4
5
3
5
and Q1 Q2 R = A.
3
8
3
17
35
4
3
5 1
4 3
3 1
2
3
6
= 40
0
3
8
3
16 7
15 5
13
15
5 1
5 3
0 1
= R,
Exploration: Approximating Eigenvalues with the QR Algorithm
1. We have
1
A1 = RQ = (Q
Q)RQ = Q
1
(QR)Q = Q
1
AQ,
so that A1 ⇠ A. Thus by Theorem 4.22(e), A1 and A have the same eigenvalues.
2. Using Gram-Schmidt, we have
" #
1
v1 =
1
" #
0
v2 =
3
" # " #
0
1·0+1·3 1
=
1·1+1·1 1
3
Normalizing these vectors gives
"
Q=
p1
2
p1
2
p1
2
p1
2
Thus
A1 = RQ =
"p
#
" # "
3 1
=
2 1
T
, so that R = Q A =
p #"
3 2
p1
2
2
p
3 2
p1
2
2
2
0
p1
2
p1
2
#
"p
=
#
.
p #
3 2
2
p
.
3 2
2
2
0
"
3
2
3
2
5
2
3
2
1
2
3
2
#
.
Then
det(A
I) =
det(A1
1
I) =
0
1
= (1
3
5
2
1
2
3
2
3
2
=
✓
)(3
5
2
)
◆✓
◆
3
2
3
=
4
2
4 +3=(
1)(
3).
So A and A1 do indeed have the same eigenvalues.
3. The proof is by induction on k. We know that A1 ⇠ A. Now assume An
An = R n
1 Qn 1
so that An ⇠ An
1
= (Qn
1
1 Qn 1 )Rn 1 Qn 1
= Qn
1
1
⇠ A. Then
1 (Qn 1 Rn 1 )Qn 1
= Q n 1 1 An
1 Qn 1 ,
⇠ A.
4. Converting A1 = Q1 R1 to

A1 = Q 1 R 1 ⇡

A2 = Q 2 R 2 ⇡

A3 = Q 3 R 3 ⇡

A4 = Q 4 R 4 ⇡
decimals gives, using technology


0.86 0.51 2.92
1.20
3.12 0.47
) A2 = R 1 Q 1 ⇡
0.51 0.86
0
1.04
0.53 0.88


0.99 0.17
3.16
0.32
3.06
0.84
) A3 = R 2 Q 2 ⇡
0.17 0.99
0
0.95
0.16
0.93


1.00
0.05
3.06 0.79
3.02 0.94
) A4 = R 3 Q 3 ⇡
0.05
1.00
0 0.97
0.05 0.97


1.00 0.02
3.02
0.92
3.00
0.98
) A5 = R 4 Q 4 ⇡
.
0.02 1.00
0
0.99
0.02 0.99
The Ak are approaching an upper triangular matrix U .
Exploration: Approximating Eigenvalues with the QR Algorithm
403
5. Since Ak ⇠ A for all k, the eigenvalues of A and Ak are the same, so in the limit, the upper triangular
matrix U will have eigenvalues equal to those of A; but the eigenvalues of a triangular matrix are its
diagonal entries. Thus the diagonal entries of U will be the eigenvalues of A.
6. (a) Follow the same process as in Exercise 4.



0.71
0.71 2.83 2.83
4.02
0.00
A = QR ⇡
) A1 = RQ ⇡
0.71
0.71
0 1.41
1.00
1.00



0.97
0.24
4.14
0.24
3.96
1.23
A1 = Q 1 R 1 ⇡
) A2 = R 1 Q 1 ⇡
0.24
0.97
0
0.97
0.23
0.94



1.00
0.06
3.97
1.17
4.04
0.93
A2 = Q 2 R 2 ⇡
) A3 = R 2 Q 2 ⇡
0.06
1.00
0
1.01
0.06
1.01



1.00
0.01
4.04
0.94
4.03
0.98
A3 = Q 3 R 3 ⇡
) A4 = R 3 Q 3 ⇡
0.01
1.00
0
1.00
0.01
1.00



1.00 0.00
4.03
0.98
4.03
0.98
A4 = Q 4 R 4 ⇡
) A5 = R 4 Q 4 ⇡
.
0.00 1.00
0
1.00
0.00
1.00
The eigenvalues of A are approximately 4.03 and 1.00.
(b) Follow the same process as in Exercise 4.



0.45
0.89 2.24 1.34
2.20
1.40
A = QR ⇡
) A1 = RQ ⇡
0.89
0.45
0 0.45
0.40
0.20



0.98
0.18
2.24
1.34
2.44
0.92
A1 = Q 1 R 1 ⇡
) A2 = R 1 Q 1 ⇡
0.18
0.98
0
0.45
0.08
0.44



1.00
0.03
2.44
0.93
2.41
1.01
A2 = Q 2 R 2 ⇡
) A3 = R 2 Q 2 ⇡
0.03
1.00
0
0.41
0.01
0.41



1 0
2.41
1.01
2.42
1
A3 = Q 3 R 3 ⇡
) A4 = R 3 Q 3 ⇡
0 1
0
0.42
0
0.42



1.00 0.00
2.42
1
2.41
1
A4 = Q 4 R 4 ⇡
) A5 = R 4 Q 4 ⇡
.
0.00 1.00
0
0.41
0
0.41
p
The eigenvalues of A are approximately 2.41 and 0.41 (the true values are 1 ± 2).
(c) Follow the same process as in Exercise 4.
2
32
3
0.24
0.06
0.97
4.24 0.47
0.94
0.97
05 4 0 1.94
1.265 )
A = QR ⇡ 4 0.24
0.94
0.23
0.24
0
0
0.73
2
3
2.00
0
3.89
0.315
A1 = RQ ⇡ 4 0.73 2.18
0.69 0.17
0.18
2
32
3
0.89
0.33
0.30
2.24
0.76
3.32
0.94
0.075 4
0
2.05
1.575 )
A1 = Q1 R1 ⇡ 4 0.33
0.31
0.03
0.95
0
0
1.31
2
3
3.27
0.13
2.44
1.98
1.645
A2 = R1 Q1 ⇡ 4 0.18
0.40
0.04
1.25
2
32
3
0.99
0.05 0.12
3.3
0.03
2.47
1.00 0.015 4
0
1.99
1.85 )
A2 = Q2 R2 ⇡ 4 0.06
0.12
0.02 0.99
0
0
1.31
2
3
2.96
0.16
2.86
1.95
1.815
A3 = R2 Q2 ⇡ 4 0.33
0.11
0.02
0.91
404
CHAPTER 5. ORTHOGONALITY
2
32
3
0.11 0.04
2.98
0.06
2.61
0.99 0.015 4
0
1.95
2.105 )
0.01 1.00
0
0
1.03
2
3
3.07
0.30
2.49
1.96
2.095
A4 = R3 Q3 ⇡ 4 0.14
0.04
0.01
1.03
2
32
3
1
0.04 0.01
3.07
0.21
2.41
1
05 4
0
1.97
2.205 )
A4 = Q4 R4 ⇡ 40.04
0.01
0
1
0
0
0.99
2
3
3.03 0.34
2.45
2.215 .
A5 = R4 Q4 ⇡ 4 0.12 1.96
0.01
0
0.99
0.99
A3 = Q3 R3 ⇡ 4 0.11
0.04
The eigenvalues of A are approximately 3.03, 1.96, and
(d) Follow the same process as
2
0.45
0
A = QR ⇡ 4
0.89

1.00
A1 = Q 1 R 1 ⇡
0.00
in Exercise 4.
3
0.72 
2.24
0.605
0
0.36

0.00 3.00
1.00
0
3.13
3.35
0.99. (The true values are 3, 2, and
2.24
0
1.07
2.00
) A1 = RQ ⇡
) A2 = R 1 Q 1 ⇡


3.00
0
3.00
0
1).
1.07
2.00
1.07
.
2.00
Since Q1 is the identity matrix, we get Q2 = I and R2 = R1 , so that successive Ai are equal to
A1 . Thus the eigenvalues of A are approximately 3 and 2 (in fact, they are 3, 2, and 0).
7. Following the process in Exercise 4,
A = QR =
"
A1 = Q 1 R 1 =
"
p2
5
p1
5
p2
5
p1
5
p1
5
p2
5
# "p
p1
5
p2
5
p8
5
p1
5
5
0
#"
p1
5
0
#
) A1 = RQ =
p8
5
p
#
5
"
2
5
1
5
) A2 = R 1 Q 1 ⇡
"
21
5
2
5
#
2
1
#
3
2
= A.
Here A2 = A (Q1 = Q 1 and R1 = R 1 ); the algorithm fails because the eigenvalues of A, which are
±1, do not have distinct absolute values.
8. Set
B = A + 0.9I =

2.9
1
3
.
1.1
Then


0.95 0.33
3.07
B = QR ⇡
0.33 0.95
0


1 0.01
1.86
B 1 = Q 1 R1 ⇡
0.01
1
0


1.00
0
1.9
B 2 = Q 2 R2 ⇡
0 1.00
0
3.19
0.06
4.02
0.1
4
0.1
) B1 = RQ ⇡


1.86
0.02
1.9
0

1.9
) B3 = R2 Q 2 ⇡
0
) B 2 = R1 Q 1 ⇡
4.02
0.06
4
0.1
4
.
0.1
Further iterations will simply reverse the sign of the upper right entry of Bi . The eigenvalues of
B are approximately 1.9 and 0.1, so the eigenvalues of A are approximately 1.9 0.9 = 1.0 and
0.1 0.9 = 1.0. These are correct (see Exercise 7).
5.4. ORTHOGONAL DIAGONALIZATION OF SYMMETRIC MATRICES
405
9. We prove the first equality using induction on k. For k = 1, it says that Q0 A1 = AQ0 , or QA1 = AQ.
Now, A1 = RQ and A = QR, so that QA1 = QRQ = (QR)Q = AQ. This proves the basis of the
induction. Assume the equality holds for k = n; we prove it for k = n + 1:
Q0 Q1 Q2 · · · Qn An+1 = Q0 Q1 Q2 · · · Qn
1 (Qn An+1 )
= Q0 Q1 Q2 · · · Qn
1 (Qn (Rn Qn ))
= Q0 Q1 Q2 · · · Qn
1 (Qn Rn )Qn
= (Q0 Q1 Q2 · · · Qn
1 An )Qn
= (AQ0 Q1 Q2 · · · Qn
1 )Qn
= AQ0 Q1 Q2 · · · Qn .
For the second equality, note that Ak = Qk Rk , so that
Q0 Q1 Q2 · · · Qk
Multiply both sides by Rk
(Q0 Q1 Q2 · · · Qk
1
1 Ak
= Q0 Q1 Q2 · · · Qk
1 Q k Rk
= AQ0 Q1 Q2 · · · Qk
1.
· · · R2 R1 to get
· · · R2 R1 ) = A(Q0 Q1 Q2 · · · Qk
1 Qk )(Rk Rk 1
1 )(Rk 1
· · · R2 R1 ),
as desired.
For the QR-decomposition of Ak+1 , we again use induction. For k = 0, the equation says that
Q0 R0 = A0+1 = A, which is true. Now assume the statement holds for k = n. Then using the first
equation in this exercise together with the fact that Ak = Qk Rk and the inductive hypothesis, we get
An+1 = A · An
= A(Q0 Q1 · · · Qn
= Q0 Q1 · · · Qn
= Q0 Q1 · · · Qn
1 )(Rn 1
1 An R n
· · · R1 R0 )
1 · · · R1 R0
1 Q n Rn R n 1
= (Q0 Q1 · · · Qn )(Rn Rn
1
· · · R1 R0
· · · R1 R0 ),
as desired.
5.4
Orthogonal Diagonalization of Symmetric Matrices
1. The characteristic polynomial of A is
det(A
I) =
Thus the eigenvalues of A are
row-reduce
1
⇥
4
1
1
4
= 3 and
A

A
5I
A basis for the eigenspace is given by
8 + 15 = (
2
= 5, we row-reduce


⇤
1
1 0
1
0 =
!
1
1 0
0

5)(
1
.
1
To find the eigenspace corresponding to
⇥
2
3)
= 5. To find the eigenspace corresponding to


⇤
1 1 0
1 1 0
0 =
!
.
1 1 0
0 0 0
3I
A basis for the eigenspace is given by
=
2
1
.
1
1
0
0
.
0
1
= 3, we
406
CHAPTER 5. ORTHOGONALITY
These eigenvectors are orthogonal, as is guaranteed by Theorem
5.19, so to turn them into orthonormal
p
vectors, we normalize them. The norm of each vector is 2, so setting
"
#
p1
2
p1
2
Q=
p1
2
p1
2
we see that Q is an orthonormal matrix whose columns are the normalized eigenvectors for A, so that

3 0
1
T
Q AQ = Q AQ =
= D.
0 5
2. The characteristic polynomial of A is
det(A
1
I) =
Thus the eigenvalues of A are
we row-reduce
1
⇥
3
3
=
4 and
A + 4I
A basis for the eigenspace is given by

2
=
1
+2
8 = ( + 4)(
2).
= 2. To find the eigenspace corresponding to

⇤
3 3 0
1 1 0
0 =
!
.
3 3 0
0 0 0
2
1
=
4,

1
.
1
To find the eigenspace corresponding to
= 2, we row-reduce


⇥
⇤
3
3 0
1
A 2I 0 =
!
3
1 0
0

1
A basis for the eigenspace is given by
.
1
2
1
0
0
.
0
These eigenvectors are orthogonal, as is guaranteed by Theorem
5.19, so to turn them into orthonormal
p
vectors, we normalize them. The norm of each vector is 2, so setting
"
#
p1
2
p1
2
Q=
p1
2
p1
2
we see that Q is an orthonormal matrix whose columns are the normalized eigenvectors for A, so that

4 0
Q 1 AQ = QT AQ =
= D.
0 2
3. The characteristic polynomial of A is
det(A
1
I) = p
p
2
2
=
2
2=(
2)( + 1)
Thus the eigenvalues of A are 1 = 1 and 2 = 2. To find the eigenspace corresponding to
we row-reduce
"
#
"
#
p
p
h
i
2
2 0
1 22 0
!
.
A+I 0 = p
2 1 0
0 0 0

1
A basis for the eigenspace is given by p .
2
To find the eigenspace corresponding to
⇥
A
2I
= 2, we row-reduce
p

⇤
1
2 0
1
0 = p
!
0
2
1 0
2

p
2
0
0
.
0
1
=
1,
5.4. ORTHOGONAL DIAGONALIZATION OF SYMMETRIC MATRICES
407
p
2
.
1
A basis for the eigenspace is given by
These eigenvectors are orthogonal, as is guaranteed by Theorem
5.19, so to turn them into orthonormal
p
vectors, we normalize them. The norm of each vector is 3, so setting
p #
"
6
3
p1
3
p1
p3
6
3
Q=
we see that Q is an orthonormal matrix whose columns are the normalized eigenvectors for A, so that

1 0
1
T
Q AQ = Q AQ =
= D.
0 2
4. The characteristic polynomial of A is
det(A
9
I) =
Thus the eigenvalues of A are
we row-reduce
h
A
2
2
1
= 5 and
5I
i
0 =
A basis for the eigenspace is given by

2
"
2
=
6
15 + 50 = (
5)(
10).
= 10. To find the eigenspace corresponding to
#
"
#
1
4
2 0
1
0
2
!
.
2
1 0
0
0 0
1
= 5,
1
.
2
To find the eigenspace corresponding to
= 10, we row-reduce


⇥
⇤
1
2 0
1
A 10I 0 =
!
2
4 0
0

2
A basis for the eigenspace is given by
.
1
2
2
0
0
.
0
These eigenvectors are orthogonal, as is guaranteed by Theorem
5.19, so to turn them into orthonormal
p
vectors, we normalize them. The norm of each vector is 5, so setting
"
#
Q=
p1
5
p2
5
p2
5
p1
5
we see that Q is an orthonormal matrix whose columns are the normalized eigenvectors for A, so that

5 0
1
T
Q AQ = Q AQ =
= D.
0 10
5. The characteristic polynomial of A is
5
det(A
0
0
0
I) =
3
Thus the eigenvalues of A are
1 = 4, we row-reduce
h
A
0
3
1
4I
1
=
3
+7
2
2
40 =
(
4)(
5)( + 2).
1
= 4,
2
2
1
i
6
0 = 40
0
= 5, and
0
3
3
3
0
3
3
=
2. To find the eigenspace corresponding to
3
2
0
1
7
6
05 ! 40
0
0
0
1
0
0
1
0
3
0
7
05 .
0
408
CHAPTER 5. ORTHOGONALITY
2 3
0
A basis for the eigenspace is given by 415.
1
To find the eigenspace corresponding to
= 5, we row-reduce
3
2
0
0
0 0
0
⇥
⇤
A 5I 0 = 40
4
3 05 ! 40
0
3
4 0
0
2 3
1
A basis for the eigenspace is given by 405.
0
2
2
To find the eigenspace corresponding to
3
=
2, we row-reduce
3
2
0 0 0
1 0
3 3 05 ! 4 0 1
3 3 0
0 0
2
7
⇤
A + 2I 0 = 40
0
2 3
0
A basis for the eigenspace is given by 4 15.
1
⇥
1
0
0
0
1
0
0
1
0
3
0
05 .
0
3
0
05 .
0
These eigenvectors are orthogonal, as is guaranteed by Theorem 5.19, sopto turn them into orthonormal
vectors, we normalize them. The norm of the first and third vectors is 2, while the second is already
a unit vector, so setting
2
3
0 1
0
6 1
7
p
p1 7
Q=6
4 2 0
25
p1
p1
0
2
2
we see that Q is an orthonormal matrix whose columns
2
4
Q 1 AQ = QT AQ = 40
0
6. The characteristic polynomial of A is
2
det(A
I) =
3
3
0
0
4
2
4
Thus the eigenvalues of A are
3, we row-reduce
1 =
1
=
3
+6
are the normalized eigenvectors for A, so that
3
0
0
5
05 = D.
0
2
2
+ 13
42 =
( + 3)(
=
3,
2
2
= 2, and
3
3
2
0
1
7
6
7
6
05 ! 40
0
0
0
0
3
3
A basis for the eigenspace is given (after clearing fractions) by 4 55.
4
0
A + 3I
A
2I
5
i 6
0 =6
43
0
To find the eigenspace corresponding to
h
7).
= 7. To find the eigenspace corresponding to
3
4
5
4
h
2)(
2
3
0
5
4
4
5
0
1
0
2
= 2, we row-reduce
3
2
0
3
0
0
1 0
i 6
7
6
6
7
6
=
!
0
4 3 0 4 05
40 1
0 4 0 0
0 0
3
7
07
5.
2
2
4
3
0
0
3
0
7
07
5.
0
5.4. ORTHOGONAL DIAGONALIZATION OF SYMMETRIC MATRICES
409
2
3
4
A basis for the eigenspace is given (after clearing fractions) by 4 05.
3
To find the eigenspace corresponding to
2
h
A
7I
= 7, we row-reduce
3
2
5
3
0
0
1
i 6
7
6
6
7
6
0 =4 3
5
4 05 ! 40
0
4
5 0
0
2
3
0
1
3
4
5
4
0
0
0
0
3
3
A basis for the eigenspace is given (after clearing fractions) by 455.
4
3
7
07
5.
These eigenvectors are orthogonal, as is guaranteed by Theorem 5.19, so to p
turn them into
p orthonormal
vectors, we normalize them. The norms of these vectors are respectively 5 2, 5, and 5 2, so setting
2
3
6
Q=6
4
3
p
5 2
p1
2
4
p
5 2
4
5
3
p
5 27
p1 7
2 5
4
p
5 2
0
3
5
we see that Q is an orthonormal matrix whose columns are the normalized eigenvectors for A, so that
2
3
3 0 0
Q 1 AQ = QT AQ = 4 0 2 05 = D.
0 0 7
7. The characteristic polynomial of A is
1
det(A
I) =
Thus the eigenvalues of A are
1 = 0, we row-reduce
0
0
1
1
1
0 =
1
0
= 0,
2
+3
2
2 =
(
1)(
2).
1
= 1, and
2
1
i
6
A 0I 0 = 4 0
1
2 3
1
A basis for the eigenspace is given by 405.
1
h
3
0
1
0
3
1
0
1
= 2. To find the eigenspace corresponding to
3
2
0
1
7
6
05 ! 40
0
0
To find the eigenspace corresponding to 2 = 1, we row-reduce
2
3
2
0 0
1 0
1
⇥
⇤
A I 0 =4 0 0
0 05 ! 4 0
1 0
0 0
0
2 3
0
A basis for the eigenspace is given by 415.
0
To find the eigenspace corresponding to 3 = 2, we
2
1
0
⇥
⇤
A 2I 0 = 4 0
1
1
0
row-reduce
3
2
1 0
1
0 05 ! 4 0
1 0
0
0
1
0
1
0
0
0
0
0
3
0
05 .
0
0
1
0
0
1
0
3
0
7
05 .
0
1
0
0
3
0
05 .
0
410
CHAPTER 5. ORTHOGONALITY
2
3
1
A basis for the eigenspace is given by 4 05.
1
These eigenvectors are orthogonal, as is guaranteed by Theorem 5.19, sopto turn them into orthonormal
vectors, we normalize them. The norm of the first and third vectors is 2, while the second is already
a unit vector, so setting
2
3
p1
p1
0
27
6 2
Q=6
0
1
07
4
5
p1
p1
0
2
2
we see that Q is an orthonormal matrix whose columns are
2
0 0
Q 1 AQ = QT AQ = 40 1
0 0
the normalized eigenvectors for A, so that
3
0
05 = D.
2
8. The characteristic polynomial of A is
1
det(A
2
2
2
I) =
Thus the eigenvalues of A are
1, we row-reduce
1 =
⇥
Thus the eigenspace E
1
2
1
A+I
is
2
2
2
1
=
=
+3
2
+9 +5=
( + 1)2 (
5)
1
1 and
2
2
0 = 42
2
⇤
3
2
2
2
2
= 5. To find the eigenspace E
2
2
2
3
2
0
1
05 ! 40
0
0
1
0
0
1
0
0
1
corresponding to
3
0
05 .
0
3
0
2 3
2 31
s t
1
1
4 s 5 = span @x1 = 4 15 , x2 = 4 05A .
t
0
1
To find the eigenspace E5 corresponding to
2
4
⇥
⇤
A 5I 0 = 4 2
2
2 3
1
Thus a basis for E5 is v3 = 415.
1
2
= 5, we row-reduce
3
2
1 0
2
2 0
4
2 05 ! 40 1
2
4 0
0 0
1
1
0
3
0
05 .
0
E5 is orthogonal to E 1 as guaranteed by Theorem 10.19, so to turn this basis into an orthogonal basis
we must orthogonalize the basis for E 1 above. Use Gram-Schmidt:
2 3
2 3
2 3
2 3
1
✓
◆
1
1
1
6 27
v
·
x
1
1
2
4 15 = 6 1 7.
v 1 = 4 15 ,
v2 = x2
v1 = 4 05
4 25
v1 · v1
2
0
1
0
1
We double v2 to remove fractions, so we get the orthogonal basis
2 3
2 3
2 3
1
1
1
v1 = 4 15 , v2 = 4 15 , v3 = 415
0
2
1
5.4. ORTHOGONAL DIAGONALIZATION OF SYMMETRIC MATRICES
411
p p
p
The norms of these vectors are, respectively, 2, 6, and 3, so normalizing each vector and setting
Q equal to the matrix whose columns are the normalized eigenvectors gives
2
3
6
Q=6
4
so that
p1
2
p1
2
p1
6
p1
6
p2
6
0
p1
37
p1 7 ,
35
p1
3
2
3
0 0
1 05 = D.
0 5
1
Q 1 AQ = QT AQ = 4 0
0
9. The characteristic polynomial of A is
1
det(A
1
1
0
0
I) =
Thus the eigenvalues of A are
we row-reduce
⇥
A + 0I
Thus the eigenspace E0 is
0
0
1
0
0
1
0
0
1
1
1
= 0 and
2
1
⇤ 61
0 =6
40
0
2
1
1
0
0
=
4
4
3
+4
2
=
2
(
2)2 .
1
= 2. To find the eigenspace E0 corresponding to
3
2
3
0 0 0
1 1 0 0 0
6
7
0 0 07
7 ! 60 0 1 1 07 .
5
4
1 1 0
0 0 0 0 05
1 1 0
0 0 0 0 0
1
= 0,
2
3
0
2 3
2 31
s
1
0
6 s7
B
6 17
6 07 C
6 7 = span Bx1 = 6 7 , x2 = 6 7C .
4 t5
@
4 05
4 15 A
t
0
1
To find the eigenspace E2 corresponding
2
1
⇥
⇤ 6 1
A 2I 0 = 6
4 0
0
Thus the eigenspace E2 is
to
1
1
0
0
2
= 2, we row-reduce
3
2
0
0 0
1
60
0
0 07
7!6
40
1
1 05
1
1 0
0
1 0
0 1
0 0
0 0
2 3
0
2 3
2 31
s
1
0
6s7
B
617
607C
6 7 = span Bx3 = 6 7 , x4 = 6 7C .
4 t5
@
405
415A
t
0
1
0
1
0
0
3
0
07
7.
05
0
E0 is orthogonal to E2 as guaranteed by Theorem 10.19, and x1 and x2 are also orthogonal,
as are x3
p
and x4 . So these four vectors form an orthogonal basis. The norm of each vector is 2, so normalizing
each vector and setting Q equal to the matrix whose columns are the normalized eigenvectors gives
2
3
p1
p1
0
0
2
2
6
7
6 p1
p1
0
07
6
7
2
2
Q=6
7
6
p1
p1 7
0
0
4
2
25
p1
0
0 p12
2
412
CHAPTER 5. ORTHOGONALITY
so that
2
0
6
0
Q 1 AQ = QT AQ = 6
40
0
0
0
0
0
3
0
07
7 = D.
05
2
0
0
2
0
10. The characteristic polynomial of A is
2
det(A
I) =
0
0
0
1
0
0
1
0
0
Thus the eigenvalues of A are
we row-reduce
⇥
AI
Thus the eigenspace E1 is
1
0
0
1
0
=
6
3
+ 12
2
10 + 3 = (
1)3 (
3)
2
= 1 and
2
1 0
⇤ 60 0
0 =6
40 0
1 0
1
2
2
= 3. To find the
3
2
0 1 0
1
60
0 0 07
7!6
40
0 0 05
0 1 0
0
3
02
s
6 t 7
B6
6 7 = span B6
4u5
@4
s
Note that these vectors are mutually orthogonal.
To find the eigenspace E3 corresponding
2
1
⇥
⇤ 6 0
A 3I 0 = 6
4 0
1
2 3
1
607
6
Thus a basis for E3 is 4 7
.
05
1
4
to
2
0
2
0
0
3
1
07
7,
05
1
2 3
0
617
6 7,
405
0
eigenspace E1 corresponding to
3
0 0 1 0
0 0 0 07
7.
0 0 0 05
0 0 0 0
1
= 1,
2 31
0
607C
6 7C .
415A
0
= 3, we row-reduce
3
2
0
1 0
1
60
0
0 07
7!6
40
2
0 05
0
1 0
0
0
1
0
0
0
0
1
0
1
0
0
0
3
0
07
7.
05
0
Since the eigenvectors of E1 are orthogonal to those of E3 by Theorem 10.19, and the eigenvectors of
E3 are already
all we need to do is to normalize the vectors. The norms are,
p mutually orthogonal,
p
respectively, 2, 1, 1, and 2, so setting Q equal to the matrix whose columns are the normalized
eigenvectors gives
2 1
3
p
0 0 p12
2
6
7
6 0
1 0 07
6
7,
Q=6
0 1 07
4 0
5
1
p1
p
0 0
2
2
so that
2
1
60
1
T
Q AQ = Q AQ = 6
40
0
0
1
0
0
0
0
1
0
3
0
07
7 = D.
05
3
= (a
)2
11. The characteristic polynomial is
det(A
I) =
a
b
b
a
b2 .
5.4. ORTHOGONAL DIAGONALIZATION OF SYMMETRIC MATRICES
413
This is zero when a
= ±b, so the eigenvalues are a + b and a b. To find the corresponding
eigenspaces, row-reduce:


⇥
⇤
b
b 0
1
1 0
A (a + b)I 0 =
!
b
b 0
0
0 0


⇥
⇤
b b 0
1 1 0
A (a b)I 0 =
!
.
b b 0
0 0 0


1
1
So bases for the two eigenspaces (which are orthogonal by Theorem 5.19) are
and
. Each of
1
1
p
these vectors has norm 2, so normalizing each gives
"
#
Q=
p1
2
p1
2
p1
2
p1
2
so that
Q
1
AQ = QT AQ =

,
a+b
0
0
a
b
.
12. The characteristic polynomial is (expanding along the second row)
a
det(A
I) =
0
0
b
b
0
a
0
= (a
) (a
)2
b2 .
a
This is zero when a = or when a
= ±b, so the eigenvalues are a, a + b and a b. To find
corresponding eigenspaces, row-reduce:
3
2
3
2
0 0 b 0
1 0 0 0
⇥
⇤
A aI 0 = 40 0 0 05 ! 40 0 1 05
b 0 0 0
0 0 0 0
2
3
2
3
b
0
b 0
1 0
1 0
⇥
⇤
A (a + b)I 0 = 4 0
b
0 05 ! 40 1
0 05
b
0
b 0
0 0
0 0
2
3
2
3
1 0 1 0
b 0 b 0
⇥
⇤
A (a b)I 0 = 40 b 0 05 ! 40 1 0 05 .
b 0 b 0
0 0 0 0
2 3 2 3
2
0
1
So bases for the three eigenspaces (which are orthogonal by Theorem 5.19) are 415, 405 and 4
0
1
Normalizing the second and third vectors gives
2
3
p1
0 p12
2
6
7
Q=6
0
07
41
5,
1
1
p
0 p2
2
so that
2
a
Q 1 AQ = QT AQ = 40
0
0
a+b
0
3
0
0 5.
a b
the
3
1
05.
1
13. (a) Since A and B are each orthogonally diagonalizable, they are each symmetric by the Spectral
Theorem. Therefore A + B is also symmetric, so it too is orthogonally diagonalizable by the
Spectral Theorem.
414
CHAPTER 5. ORTHOGONALITY
(b) Since A is orthogonally diagonalizable, it is symmetric by the Spectral Theorem. Therefore cA is
also symmetric, so it too is orthogonally diagonalizable by the Spectral Theorem.
(c) Since A is orthogonally diagonalizable, it is symmetric by the Spectral Theorem. Therefore
A2 = AA is also symmetric, so it too is orthogonally diagonalizable by the Spectral Theorem.
14. If A is invertible and orthogonally diagonalizable, then QT AQ = D for some invertible diagonal matrix
D and orthogonal matrix Q, so that
D
Thus A
1
1
1
= QT AQ
1
=Q
A
1
1
QT
= QT A
1
T
QT
= QT A
1
Q.
is also orthogonally diagonalizable (and by the same orthogonal matrix).
15. Since A and B are orthogonally diagonalizable, they are both symmetric. Since AB = BA, Exercise
36 in Section 3.2 implies that AB is also symmetric, so it is orthogonally diagonalizable as well.
16. Let A be a symmetric matrix. Then it is orthogonally diagonalizable, say D = QT AQ. Then the
diagonal entries of D are the eigenvalues of A. First, if every eigenvalue of A is nonnegative, let
2
3
2p
3
··· 0
···
0
1
1
p
6
7
6
.. 7 .
..
D = 4 ... . . . ... 5 , and define D = 4 ...
5
.
p.
0 ···
0
·
·
·
n
n
p
Let B = Q DQT . Then B is diagonally orthogonalizable by construction, so it is symmetric. Further,
since Q is orthogonal,
p
p
p
p
p p
B 2 = Q DQT Q DQT = Q DQ 1 Q DQT = Q D DQT = QDQT = A.
Conversely, suppose A = B 2 for some symmetric matrix B. Then B is orthogonally diagonalizable,
say QT BQ = D, so that
QT AQ = QT B 2 Q = (QT BQ)(QT BQ) = D2 .
But D2 is a diagonal matrix whose diagonal entries are the squares of the diagonal entries of D, so
they are all nonnegative. Thus the eigenvalues of A are all nonnegative.
17. From Exercise 1 we have
1
= 3,
q1 =
"
p1
2
p1
2
#
,
Therefore
A=
T
1 q1 q1
+
= 5,
2
T
2 q2 q2
=3
"
#
"
1
2
1
2
q2 =
1
2
1
2
#
+5
"
"
p1
2
p1
2
1
2
1
2
#
.
1
2
1
2
#
.
p1
2
p1
2
#
.
18. From Exercise 2 we have
1
=
4,
q1 =
p1
2
p1
2
,
2
Therefore
A=
T
1 q1 q1
+
T
2 q2 q2
=
4
"
1
2
1
2
= 2,
1
2
1
2
q2 =
#
+2
"
"
1
2
1
2
1
2
1
2
#
.
5.4. ORTHOGONAL DIAGONALIZATION OF SYMMETRIC MATRICES
19. From Exercise 5 we have
1
2
0
3
6 1 7
p 7
q1 = 6
4 25 ,
= 4,
2
2 3
1
6 7
7
q2 = 6
405 ,
0
= 5,
p1
2
Therefore
A=
T
1 q1 q1
T
2 q2 q2
+
20. From Exercise 8 we have
1
=
1,
Therefore
A=
T
1 q1 q1
+
T
3 q3 q3
+
2
3
p1
27
p1 7 ,
25
6
q1 = 6
4
T
1 q2 q2
+
0
T
2 q3 q3
=
2
0
0
0
1
2
1
2
17
25
1
2
6
= 46
40
2
6
q2 = 6
4
2
6
6
4
3
0
7
3
2
1
1
2
1
2
0
0
3
0
7
07
5
0
0
0
3
1
2
1
2
0
6
+ 56
40
p1
67
p1 7 ,
65
p2
6
0
2
2
6
6
4
=
415
2
6
q2 = 6
4
2,
3
0
7
07
5
0
= 5,
1
6
1
6
1
3
2
p1
2
p1
2
,
q2 =
p1
2
p1
2
6
26
40
0
0
1
2
1
2
2
7
3
3
0
7
17 .
25
1
2
p1
6 13 7
6p 7 .
4 35
p1
3
3
1
6
1
6
1
3
3
p1 7 .
25
p1
2
0
q3 =
1
37
17
35
2
3
21. Since v1 and v2 are already orthogonal, normalize each of them to get
" #
"
#
q1 =
0
+
2
1
63
1
56
43
1
3
3
1
3
1
3
1
3
1
37
17 .
35
1
3
.
Now let Q be the matrix whose columns are q1 and q2 ; then the desired matrix is
"
#
"
#
"
#"
#"
# "
p1
p1
p1
p1
1 0
1 0
1
0
1
T
2
2
2
Q
Q =Q
Q = 12
=
p
p1
p1
p1
0 2
0 2
0
2
2
2
2
2
1
2
3
2
3
2
1
2
9
5
12
5
12
5
9
5
#
.
22. Since v1 and v2 are already orthogonal, normalize each of them to get
" #
"
#
q1 =
p1
5
p2
5
,
q2 =
p2
5
p1
5
.
Now let Q be the matrix whose columns are q1 and q2 ; then the desired matrix is
"
#
"
#
"
#"
#"
# "
p1
p2
p1
p2
3
0
3
0
3
0
1
T
5
5
5
5
Q
Q =Q
Q = 2
=
p
p1
p2
p1
0
3
0
3
0
3
5
5
5
5
#
.
p p
p
23. Note that v1 , v2 , and v3 are mutually orthogonal, with norms 2, 3, and 6 respectively. Let Q be
the matrix whose columns are q1 = kvv11 k , q2 = kvv22 k , and q3 = kvv33 k ; then the desired matrix is
2
3
2
3
1 0 0
1 0 0
6
7
6
7 T
1
7
6
7
Q6
40 2 05 Q = Q 40 2 05 Q
0
0
3
0
=
2
p1
6 12
6p
4 2
0
0
3
p1
3
p1
3
p1
3
32
p1
6 7 61
1
p 7 60
65 4
p2
0
6
0
2
0
32
0
76
6
07
54
3
p1
2
p1
3
p1
6
p1
2
p1
3
p1
6
3 2
0
7 6
p1 7 = 6
4
35
p2
6
5
3
2
3
1
3
2
3
5
3
1
3
3
1
37
17 .
35
8
3
416
CHAPTER 5. ORTHOGONALITY
p
p
p
24. Note that v1 , v2 , and v3 are mutually orthogonal, with norms 42, 3, and 14 respectively. Let Q
be the matrix whose columns are q1 = kvv11 k , q2 = kvv22 k , and q3 = kvv33 k ; then the desired matrix is
2
1
6
Q6
40
0
0
4
0
3
0
7
07
5Q
4
1
2
1
0
6
= Q6
40
4
0
2
6
=6
4
2
6
=6
4
0
p4
42
p5
42
p1
42
p1
3
p1
3
p1
3
44
21
50
21
10
21
3
0
7 T
07
5Q
4
32
p2
14 7 61
1
p 7 60
14 5 4
p3
0
14
32
0
76
6
07
54
4
0
4
0
3
50
21
43
42
25
42
10
21 7
25 7 .
42 5
163
42
p4
42
p1
3
p2
14
p5
42
p1
3
p1
14
3
p1
42 7
p1 7
35
p3
14
25. Since q is a unit vector, q · q = 1, so that
projW v =
✓
q·v
q·q
◆
q = (q · v)q = q(q · v) = q(qT v) = (qqT )v.
26. (a) By definition (see Section 5.2), the projection of v onto W is
projW v =
✓
q1 · v
q1 · q1
◆
q1 + · · · +
✓
qk · v
qk · qk
= (q1 · v)q1 + · · · + (qk · v)qk
◆
qk
= q1 (q1 · v) + · · · + qk (qk · v)
= q1 (qT1 v) + · · · + qk (qTk · v)
= (q1 qT1 + · · · + qk qTk )v.
Therefore the matrix of the projection is indeed q1 qT1 + · · · + qk qTk .
(b) First,
P T = (q1 qT1 + · · · + qk qTk )T = q1 qT1
T
+ · · · + qk qTk
T
= qT1
T
qT1 + · · · + qTk
T
qk = q1 qT1 + · · · + qk qTk = P.
Next, since qi · qj = 0 for i 6= j, and qi · qi = qTi qi = 1, we have
P 2 = (q1 qT1 )(q1 qT1 ) + · · · + (qk qTk )(qk qTk )
= q1 qT1 q1 qT1 + · · · + qk qTk qk qTk = q1 qT1 + · · · qk qTk = P.
(c) We have
⇥
QQT = q1
···
2
3
qT1
⇤6 7
qk 4 ... 5 = q1 qT1 + · · · qk qTk = P.
qTk
Thus rank(P ) = rank(QQT ) = rank(Q) by Theorem 3.28 in Section 3.5. But if q1 , . . . , qk is an
orthonormal basis, then rank(Q) = k, so that rank(P ) = k as well.
5.5. APPLICATIONS
417
27. Suppose the eigenvalues⇥ are 1 , . . . , ⇤ n and that v1 , . . . , vn are a corresponding orthonormal set of
eigenvectors. Let Q1 = v1 · · · vn . Then
2 T3
2 T3
v1
v1
⇥
⇤
⇤
6 7
6 7⇥
QT1 AQ1 = 4 ... 5 A v1 · · · vn = 4 ... 5 Av1 · · · Avn
vnT
vnT
1) ⇥ (n
where A1 is an (n
triangular.
2
3
v1T
6 7⇥
= 4 ... 5
vnT
1 v1
···
n vn
⇤
=

⇤
= T,
A1
1
0
1) matrix. The result now follows by induction since T is block upper
28. We know that if A is nilpotent, then its only eigenvalue is 0. So if A is nilpotent, then all of its
eigenvalues are real, so that by Exercise 27 there is an orthogonal matrix Q such that QT AQ is upper
triangular. But A and QT AQ have the same eigenvalues; since the eigenvalues of QT AQ are the
diagonal entries, all the diagonal entries must be zero.
5.5
Applications


⇤
2 3 ⇥
x y = 2x2 + 6xy + 4y 2 .
3 4


⇤
x
5
1 ⇥
x1 x2 = 5x21 + 2x1 x2 x22 .
2. f (x) = xT Ax = 1
x2 1
1
 
⇤
1
3
2 ⇥
T
1 6 = 3 · 12 4 · 1 · 6 + 4 · 62 = 123.
3. f (x) = x Ax =
6
2
4
2 32
3
x
1 0
3 ⇥
⇤
15 x y z = 1x2 + 2y 2 + 3z 2 + 0xy 6xz + 2yz = x2 + 2y 2 + 3z 2
4. f (x) = 4y 5 4 0 2
z
3 1
3
1. f (x) = xT Ax =
5. From Exercise 4,
x
y
02
f @4
2 32
1 2
6. f (x) = xT Ax = 425 42
3 0
7. The diagonal entries are
8. The diagonal entries are
9. The diagonal entries are
6xz + 2yz.
31
2
15A = 22 + 2 · ( 1)2 + 3 · 12 6 · 2 · 1 + 2 · ( 1) · 1 = 5.
1
3
2 0 ⇥
⇤
0 15 1 2 3 = 2 · 12 + 0 · 22 + 1 · 32 + 4 · 1 · 2 + 0 · 1 · 3 + 2 · 2 · 3 = 31.
1 1

1 3
1 and 2, and the corners are each 12 · 6 = 3, so A =
.
3 2

0 1
both zero, and the corners are each 12 · 1 = 12 , so A = 1 2 .
0
2

3
3
1
3
2 .
3 and 1, and the corners are each 2 · ( 3) = 2 , so A =
3
1
2
10. The diagonal entries are 1, 0, and
Thus
1, while the o↵-diagonal entries are
2
3
1
4
0
0
35 .
A = 44
0
3
1
1
2
· 8 = 4, 0, and
1
2
· ( 6) =
3.
418
CHAPTER 5. ORTHOGONALITY
11. The diagonal entries are 5,
1
2 · 4 = 2. Thus
1, and 2, while the o↵-diagonal entries are
12. The diagonal entries are 2,
1
2 · 0 = 0. Thus
3, and 1, while the o↵-diagonal entries are
2
5
A=4 1
2
2
13. The matrix of this form is A =

2
A=4 0
2
2
2
· 2 = 1,
1
2
· ( 4) =
2, and
1
2
· 0 = 0,
1
2
· ( 4) =
2, and
3
2
05 .
1
0
3
0
2
, so that its characteristic polynomial is
5
2
2
2
3
2
25 .
2
1
1
2
1
2
2
=
5
7 +6=(
1)(
6).
So the eigenvalues are 1 and 6; finding the eigenvectors by the usual method gives corresponding
2
1
eigenvectors v1 =
and v2 =
. Normalizing these and constructing Q gives
1
2
Q=
"
p2
5
p1
5
p1
5
p2
5
#
.
The new quadratic form is

⇤ 1
y2
0
⇥
T
f (y) = y Dy = y1
14. The matrix of this form is A =

1
4
y1
= y12 + 6y22 .
y2
4
, so that its characteristic polynomial is
1
1
4
4

0
6
=
1
2
2
15 = (
5)( + 3).
So the eigenvaluesare 5 and 3;
 finding the eigenvectors by the usual method gives corresponding
1
1
eigenvectors v1 =
and v2 =
. Normalizing these and constructing Q gives
1
1
Q=
"
p1
2
p1
2
p1
2
p1
2
#
.
The new quadratic form is
⇥
f (y) = yT Dy = y1
2
7
15. The matrix of this form is A = 44
4
4
1
8
4
4
1
7
4
8
y2

⇤ 5
0
0
3

y1
= 5y12
y2
3y22 .
3
4
85, so that its characteristic polynomial is
1
4
8
1
=(
9)2 ( + 9).
5.5. APPLICATIONS
419
So the eigenvalues are 9 (of algebraic multiplicity
the eigenvectors
2 32) and 9; 2finding
3
2 3by the usual
2
2
1
method gives corresponding eigenvectors v1 = 405 and v2 = 415 for = 9 and 4 25 for = 9.
1
0
2
Orthogonalizing v1 and v2 gives
2 3
2 3 2 23
✓
◆
2
2
v2 · v1
4 4 5 4 55
0 =
1 .
v20 = v2
v1 = 415
v1 · v1
5
4
0
1
5
Normalizing these and constructing Q gives
2
p2
6 5
Q=6
4 0
p1
5
The new quadratic form is
⇥
f (y) = yT Dy = y1
2
⇤ 9
y3 4 0
0
y2
2
0
9
0
1
3
7
27
.
35
2
3
32 3
0 y1
05 4y2 5 = 9y12 + 9y22
9 y3
9y32 .
3
2 0
1 05, so that its characteristic polynomial is
0 3
1
16. The matrix of this form is A = 4 2
0
1
2
2
0
3
2
p
3 5
5
p
3 5
4
p
3 5
0
0
1
0
3)2 ( + 1).
=(
3
So the eigenvalues are 3 (of algebraic multiplicity
2 2)
3 and 1; finding
2 3 the eigenvectors
2 3by the usual
1
0
1
method gives corresponding eigenvectors v1 = 4 15 and v2 = 405 for = 3 and 415 for = 1.
0
1
0
Since v1 and v2 are already orthogonal, we normalize all three vectors and construct Q to get
2
3
p1
p1
0
2
27
6
1
1 7
Q=6
4 p2 0 p2 5 .
0 1
0
The new quadratic form is
⇥
f (y) = yT Dy = y1
y2
2
1
1
1
0
3
0
32 3
0 y1
05 4y2 5 = 3y12 + 3y22
1 y3
y32 .
3
1 0
0 15, so that its characteristic polynomial is
1 1
1
17. The matrix of this form is A = 4 1
1
1
2
⇤ 3
y3 4 0
0
1
0
1
1
=(
2)(
1)( + 1).
420
CHAPTER 5. ORTHOGONALITY
So the eigenvalues 2are 3
2, 1, and2 3
1; finding
2 the
3 eigenvectors by the usual method gives corresponding
1
1
1
eigenvectors v1 = 4 15, v2 = 405, and 4 25. Normalize all three vectors and construct Q to get
1
1
1
2
3
p1
3
p1
3
p1
3
6
Q=6
4
The new quadratic form is
⇥
f (x0 ) = (x0 )T Dx0 = x0
2
0
18. The matrix of this form is A = 41
1
p1
2
0
p1
2
2
2
⇤
z 0 40
0
y0
p1
67
p2 7 .
65
p1
6
32 3
0 x0
05 4y 0 5 = 2(x0 )2 + (y 0 )2
1
z0
0
1
0
(z 0 )2 .
3
1
15, so that its characteristic polynomial is
0
1
0
1
1
1
1
1
1
=(
2)( + 1)2 .
1
So the eigenvalues 2are3 2 and 1; finding the2 eigenvectors
by 2the 3
usual method gives corresponding
3
1
2
0
eigenvectors v1 = 415 for = 2 and v2 = 4 15 and v3 = 4 15 for = 1. The latter two are
1
1
1
already orthogonal, so simply normalize all three vectors and construct Q to get
2
3
p1
p2
0
6
6 13
7
p
p1
p1 7 .
Q=6
4 3
6
25
p1
3
p1
6
p1
2
The new quadratic form is
⇥
f (x0 ) = (x0 )T Dx0 = x0
19. The matrix of this form is A =

1
0
y0
2
2
⇤
z 0 40
0
0
1
0
32 3
0 x0
05 4y 0 5 = 2(x0 )2
1
z0
(y 0 )2
(z 0 )2 .
0
, so that its characteristic polynomial is
2
1
0
0
2
= (1
)(2
).
Since the eigenvalues, 1 and 2, are both positive, this is a positive definite form.

1
1
20. The matrix of this form is A =
, so that its characteristic polynomial is
1
1
1
1
1
1
= (
2).
Since the eigenvalues, 0 and 2, are both nonnegative, but one is zero, this is a positive semidefinite
form.
5.5. APPLICATIONS
421
21. The matrix of this form is A =

2
1
1
, so that its characteristic polynomial is
2
2
1
1
Since the eigenvalues,
= ( + 3)( + 1).
2
1 and
3, are both negative, this is a negative definite form.

1 2
22. The matrix of this form is A =
, so that its characteristic polynomial is
2 1
1
2
2
=(
1
3)( + 1).
Since the eigenvalues,
1 and 3, have opposite signs, this is an indefinite form.
2
3
2 1 1
23. The matrix of this form is A = 41 2 15, so that its characteristic polynomial is
1 1 2
2
1
1
1
1
1
2
1
Since both eigenvalues, 1 and 4, are
2
1
24. The matrix of this form is A = 40
1
0
1
0
0
=
(
1)(
2).
1
are all nonnegative, but one of them is zero, this is a positive
3
0
05, so that its characteristic polynomial is
1
2
1
0
1
2
2
0
4).
2
1
Since the eigenvalues, 0, 1, and 2,
semidefinite form.
2
1
25. The matrix of this form is A = 42
0
1)2 (
(
positive, this is a positive definite form.
3
0 1
1 05, so that its characteristic polynomial is
0 1
1
0
1
=
0
0
1
0
=
( + 1)2 (
3).
1
Since the eigenvalues,
1 and 3, have opposite signs, this is an indefinite form.
2
3
1
1
1
1
15, so that its characteristic polynomial is
26. The matrix of this form is A = 4 1
1
1
1
1
1
1
1
Since the eigenvalues, 0 and
semidefinite form.
1
1
1
1
=
2
( + 3)
1
3, are both nonpositive but one of them is zero, this is a negative
422
CHAPTER 5. ORTHOGONALITY
27. By the Principal Axes Theorem, if A is the matrix of f (x), then there is an orthogonal matrix Q such
that QT AQ = D is a diagonal matrix, and then if y = QT x, we have
f (x) = xT Ax = yT Dy =
2
1 y1
+ ··· +
2
n yn .
• For 5.22(a), f (x) is positive definite if and only if 1 y12 + · · · + n yn2 > 0 for all y1 , . . . , yn . But
this happens if and only if i > 0 for all i, since if some i  0, take yi = 1 and all other yj = 0
to get a value for f (x) that is not positive.
• For 5.22(b), f (x) is positive semidefinite if and only if 1 y12 + · · · + n yn2
0 for all y1 , . . . , yn .
But this happens if and only if i
0 for all i, since if some i < 0, take yi = 1 and all other
yj = 0 to get a value for f (x) that is negative.
• For 5.22(c), f (x) is negative semidefinite if and only if 1 y12 + · · · + n yn2 < 0 for all y1 , . . . , yn .
But this happens if and only if i < 0 for all i, since if some i
0, take yi = 1 and all other
yj = 0 to get a value for f (x) that is nonnegative.
• For 5.22(d), f (x) is negative semidefinite if and only if 1 y12 + · · · + n yn2  0 for all y1 , . . . , yn .
But this happens if and only if i  0 for all i, since if some i > 0, take yi = 1 and all other
yj = 0 to get a value for f (x) that is positive.
• For 5.22(e), f (x) is indefinite if and only if 1 y12 + · · · + n yn2 can be either negative or positive.
But from the arguments above, we see that this can happen if and only if some i is positive and
some j is negative.
28. The given matrix represents the form ax2 + 2bxy + dy 2 , which can be written in the form given in the
hint:
✓
◆2 ✓
◆
b
b2
ax2 + 2bxy + dy 2 = a x + y + d
y2 .
a
a
This is a positive definite form if and only if the right-hand side is positive for all values of x and y not
2
both zero. First suppose that a > 0 and det A = ad b2 > 0. Then ad > b2 so that d > ba (using the
2
fact that a > 0), and thus d ba > 0. Therefore the first term on the right-hand side is always positive
if either x or y is nonzero since a > 0 and nonzero squares are always positive, while the second term
is always nonnegative since the constant factor is positive. Thus the form is positive definite. Next
assume the form is positive definite; we must show a > 0 and det A > 0. By way of contradiction,
assume a  0; then setting x = 1 and y = 0 we get a  0 for the value of the form, which contradicts
the assumption that it is positive definite. Thus a > 0. Next assume that det A = ad b2  0; then
2
2
(since a > 0) d  ba . Now setting x = ab and y = 1 gives d ba  0 for the value of the form, which
is again a contradiction. Thus a > 0 and det A > 0.
2
29. Since A = B T B, we have xT Ax = xT B T Bx = (Bx)T (Bx) = kBxk
0. Now, if xT Ax = 0, then
2
kBxk = 0. But since B is invertible, this implies that x = 0. Therefore xT Ax > 0 for all x 6= 0, so
that A is positive definite.
30. Since A is symmetric and positive definite, we can find an orthogonal matrix Q and a diagonal matrix
D such that A = QDQT , where
2
3
··· 0
1
6
7
D = 4 ... . . . ... 5 , where i > 0 for all i.
0
Let
···
n
2p
then D = C 2 = C T C. This gives
6
C = C T = 4 ...
0
1
···
..
.
···
3
0
.. 7 ;
5
p.
n
A = QDQT = QC T CQT = (CQT )T CQT .
5.5. APPLICATIONS
423
Since all the i are positive, C is invertible; since Q is orthogonal, it is invertible. Thus B = CQT is
invertible and we are done.
31. (a) xT (cA)x = c xT Ax > 0 because c > 0 and A is positive definite.
(b) xT A2 x = xT Ax
T
xT Ax = xT Ax
T
2
> 0.
T
(c) x (A + B)x = x Ax + x Bx > 0 since both A and B are positive definite.
(d) First note that A is invertible by Exercise 30, since it factors as A = B T B where B is invertible.
n o
Since A is positive definite, its eigenvalues { i } are positive; the eigenvalues of A 1 are 11 ,
which are also positive. Thus A
1
is positive definite.
32. From Exercise 30, we have A = QC T CQT , where Q is orthogonal and C is a diagonal matrix. But
then C T = C, so that A = QCCQT = QCQT QCQT = (QCQT )(QCQT ) = B 2 . B is symmetric
since B = QCQT is orthogonally diagonalizable, and it is positive definite since its eigenvalues are the
diagonal entries of C, which are all positive.
33. The eigenvalues of this form, from Exercise 20, are 1 = 2 and 2 = 0, so that by Theorem 5.23, the
maximum value of f subject to the constraint kxk = 1 is 2 and the minimum value is 0. These occur
at the unit eigenvectors of A, which are found by row-reducing:


⇥
⇤
1
1 0
1 1 0
A 2I 0 =
!
1
1 0
0 0 0


⇥
⇤
1
1 0
1
1 0
A 0I 0 =
!
.
1
1 0
0
0 0
So normalized eigenvectors are
1
v1 = p
2
"
#
1
1
=
"
p1
2
p1
2
#
,
" # " #
p1
1 1
v2 = p
= 12 .
p
2 1
2
Therefore the maximum value of 2 occurs at x = ±v1 , and the minimum of 0 occurs at x = ±v2 .
34. The eigenvalues of this form, from Exercise 22, are 1 = 3 and 2 = 1, so that by Theorem 5.23, the
maximum value of f subject to the constraint kxk = 1 is 3 and the minimum value is 1. These occur
at the unit eigenvectors of A, which are found by row-reducing:


⇥
⇤
2
2 0
1
1 0
A 3I 0 =
!
2
2 0
0
0 0


⇥
⇤
2 2 0
1 1 0
A+I 0 =
!
.
2 2 0
0 0 0
So normalized eigenvectors are
" # " #
p1
1 1
v1 = p
= 12 ,
p
2 1
2
1
v2 = p
2
"
1
1
#
=
"
p1
2
p1
2
#
Therefore the maximum value of 3 occurs at x = ±v1 , and the minimum of
.
1 occurs at x = ±v2 .
35. The eigenvalues of this form, from Exercise 23, are 1 = 4 and 2 = 1, so that by Theorem 5.23, the
maximum value of f subject to the constraint kxk = 1 is 4 and the minimum value is 1. These occur
at the unit eigenvectors of A, which are found by row-reducing:
2
3
2
3
2
1
1 0
1 0
1 0
⇥
⇤
A 4I 0 = 4 1
2
1 05 ! 40 1
1 05
1
1
2 0
0 0
0 0
2
3
2
3
1 1 1 0
1 1 1 0
⇥
⇤
A I 0 = 41 1 1 05 ! 40 0 0 05 .
1 1 1 0
0 0 0 0
424
CHAPTER 5. ORTHOGONALITY
So normalized eigenvectors are
2 3 2 3
p1
1
7 6 13 7
1 6
v1 = p 6
17 = 6 p 7 ,
3 4 5 4 13 5
p
1
3
2
3
2
2
3
1
6
1 6 7
v2 = p 6
=6
17
4
5
4
2
0
p1
27
p1 7 ,
25
0
1
3
2
6
1 6 7
v3 = p 6
=6
07
4
5
4
2
1
3
p1
27
07
5.
p1
2
Therefore the maximum value of 4 occurs at x = ±v1 , and the minimum of 1 occurs at x = ±v2 as
well as at x = ±v3 .
36. The eigenvalues of this form, from Exercise 24, are 1 = 2, 2 = 1, and 3 = 0, so that by Theorem
5.23, the maximum value of f subject to the constraint kxk = 1 is 2 and the minimum value is 0.
These occur at the corresponding unit eigenvectors of A, which are found by row-reducing:
2
3
2
3
1
0
1 0
1 0
1 0
⇥
⇤
A 2I 0 = 4 0
1
0 05 ! 40 1
0 05
1
0
1 0
0 0
0 0
2
3
2
3
1 0 1 0
1 0 1 0
⇥
⇤
A 0I 0 = 40 1 0 05 ! 40 1 0 05 .
1 0 1 0
0 0 0 0
So normalized eigenvectors are
2 3 2 3
p1
1
7 6 27
1 6
v1 = p 6
07 = 6 0 7 ,
24 5 4 1 5
p
1
2
2
1
3
2
p1
2
3
6
7
1 6 7
v2 = p 6
=6
07
0 7
5
4
4
5.
2
1
p
1
2
Therefore the maximum value of 2 occurs at x = ±v1 , and the minimum of 0 occurs at x = ±v2 .
37. Since
1
2
···
f (x) =
n,
2
1 y1
we have
+
2
2 y2
+ ··· +
2
n yn
2
n (y1
+ y22 + · · · + yn2 ) =
n
2
kyk =
n
since we are assuming the constraint kyk = 1.
38. If qn is a unit eigenvector corresponding to
n,
then Aqn =
f (qn ) = qTn Aqn = qTn
n qn
=
n
n qn ,
so that
qTn qn =
n.
Thus the quadratic form actually takes the value n , so by property (a) of Theorem 5.23, it must be
the minimum value of f (x), and it occurs when x = qn .
39. Divide this equation through by 25 to get
x2
y2
+
= 1,
25
5
which is an ellipse, from Figure 5.15.
40. Divide through by 4 and rearrange terms to get
x2
4
which is a hyperbola, from Figure 5.15.
y2
= 1,
4
5.5. APPLICATIONS
425
41. Rearrange terms to get
y = x2
1,
which is a parabola, from Figure 5.15.
42. Divide through by 8 and rearrange terms to get
x2
y2
+
= 1,
4
8
which is an ellipse, from Figure 5.15.
43. Rearrange terms to get
y2
3x2 = 1,
which is a hyperbola, from Figure 5.15.
44. This is a parabola, from Figure 5.15.
45. Group the x and y terms and then complete the square, giving
(x2
4x + 4) + (y 2
4x + 4) =
4+4+4
) (x
2)2 + (y
2)2 = 4.
2
2
This is a circle of radius 2 centered at (2, 2). Letting x0 = x 2 and y 0 = y 2, we have (x0 ) +(y 0 ) = 4.
y
4
3
y!
x!
2
1
1
2
3
4
x
46. Group the x and y terms and simplify:
(4x2
8x) + (2y 2 + 12y) =
)
6
4(x2
2x) + 2(y 2 + 6y) =
6.
Now complete the square and simplify again:
4(x2
2x + 1) + 2(y 2 + 6y + 9) =
2
4(x
6 + 4 + 18
2
1) + 2(y + 3) = 16
1)2
(x
4
+
(y + 3)2
= 1.
8
This is an ellipse centered at (1, 3). Letting x0 = x
)
)
1 and y 0 = y + 3, we have
(x0 )2
4
+
(y 0 )2
8
= 1.
426
CHAPTER 5. ORTHOGONALITY
y
1
"1
2
3
x
"1
y!
"2
x!
"3
"4
"5
"6
47. Group the x and y terms to get 9x2
9x
2
✓
1
4 y +y+
4
2
◆
= 37
This is a hyperbola centered at 0,
4(y 2 + y) = 37. Now complete the square and simplify:
)
1
1
2
9x
✓
1
4 y+
2
2
◆2
= 36
y + 12
9
x2
4
)
. Letting x0 = x and y 0 = y + 12 , we have
( x0 )
2
4
2
= 1.
(y 0 )
9
2
= 1.
y
y!
3
2
1
"3
"2
1
"1
2
3
x
x!
"1
"2
"3
48. Rearrange terms to get 3y
3y
13 = x2 + 10x; then complete the square:
13 + 25 = x2 + 10x + 25 = (x + 5)2
)
3y + 12 = (x + 5)2
)
y+4=
1
(x + 5)2 .
3
5.5. APPLICATIONS
427
This is a parabola. Letting x0 = x + 5 and y 0 = y + 4, we have y 0 =
1
3
2
(x0 ) .
y
x
"9
"8
"7
"6
"5
"4
"3
"2
"1
"1
y!
"2
"3
x!
"4
2y 2
49. Rearrange terms to get 4x =
square:
2x
4=
y2
8y; then divide by 2, giving 2x =
(y 2 + 4y + 4) =
This is a parabola. Letting x0 = x
(y + 2)2
)
x
4y. Finally, complete the
1
(y + 2)2 .
2
1
0 2
2 (y ) .
2=
2 and y 0 = y + 2, we have x0 =
y
1
"3
"2
y!
1
"1
2
3
x
"1
x!
"2
"3
"4
50. Complete the square and simplify:
2y 2
2(y
2
3x2
18x
10y + 25)
3(x + 6x + 9) + 11
2(y
5)
2
(y
5)
3(x + 3) = 12
6
2
(x + 3)
= 1.
4
This is a hyperbola. Letting x0 = x + 3 and y 0 = y
5, we have
)
(y 0 )
2
( x0 )
6
4
y
y!
12
10
8
6
x!
4
2
"12
"10
"8
"6
"4
)
50 + 27 = 0
2
2
)
20y + 11 = 0
2
2
"2
"2
x
2
= 1.
428
CHAPTER 5. ORTHOGONALITY
51. The matrix of this form is
A=
"
1
1
2
1
2
1
#
,
which has characteristic polynomial
(1
)
2
1
=
4
3
2 + =
4
2
✓
1
2
◆✓
3
2
◆
.
Therefore the associated diagonal matrix is
D=
"
3
2
0
0
1
2
#
,
so that the equation becomes
T
xT Ax = (x0 ) Dx0 =
3 0 2 1 0 2
(x ) + (y ) = 6.
2
2
Dividing through by 6 to simplify gives
2
2
(x0 )
(y 0 )
+
= 1,
4
12
which is an ellipse. To determine the rotation matrix Q, find unit eigenvectors corresponding to
and 2 = 12 by row-reduction to get
"
#
p1
2
p1
2
Q=
Then
e1 ! Qe1 =
"
p1
2
p1
2
#
p1
2
p1
2
.
e2 ! Qe2 =
,
"
p1
2
p1
2
#
,
so that the axis rotation is through a 45 angle.
y
3
y!
x!
2
1
"3
"2
1
"1
"1
"2
"3
52. The matrix of this form is
A=

4
5
5
,
4
2
3
x
1
=
3
2
5.5. APPLICATIONS
429
which has characteristic polynomial
(4
)2
25 =
2
8
9=(
9)( + 1).
Therefore the associated diagonal matrix is
D=

9
0
0
,
1
so that the equation becomes
T
xT Ax = (x0 ) Dx0 = 9 (x0 )
2
2
(y 0 ) = 9.
Dividing through by 9 to simplify gives
(x0 )
2
(y 0 )
= 1,
9
2
which is a hyperbola. To determine the rotation matrix Q, find unit eigenvectors corresponding to
1 = 9 and 2 = 1 by row-reduction to get
Q=
Then
e1 ! Qe1 =
"
p1
2
p1
2
"
#
p1
2
p1
2
p1
2
p1
2
#
.
e2 ! Qe2 =
,
"
p1
2
p1
2
#
,
so that the axis rotation is through a 45 angle.
y
3
y!
x!
2
1
"3
"2
1
"1
2
3
x
"1
"2
"3
53. The matrix of this form is
A=

4
3
3
,
4
which has characteristic polynomial
(4
)( 4
)
9=
2
25 = (
5)( + 5)
430
CHAPTER 5. ORTHOGONALITY
Therefore the associated diagonal matrix is
D=

5
0
0
,
5
so that the equation becomes
T
xT Ax = (x0 ) Dx0 = 5 (x0 )
2
2
5 (y 0 ) = 5.
Dividing through by 5 to simplify gives
(x0 )
2
2
(y 0 ) = 1,
which is a hyperbola. To determine the rotation matrix Q, find unit eigenvectors corresponding to
1 = 9 and 2 = 1 by row-reduction to get
"
#
Q=
Then
e1 ! Qe1 =
"
p3
10
p1
10
#
p3
10
p1
10
p1
10
p3
10
.
e2 ! Qe2 =
,
so that the axis rotation is through an angle whose tangent is
h
p1
10
p
1/ 10
p
3/ 10
p3
10
i
= 13 , or about 18.435 .
y
3
y!
2
x!
1
"3
"2
1
"1
2
3
x
"1
"2
"3
54. The matrix of this form is
A=

1=
2
3
1
1
,
3
which has characteristic polynomial
(3
)(3
)
6 +8=(
Therefore the associated diagonal matrix is
D=

4
0
0
,
2
,
4)(
2).
5.5. APPLICATIONS
431
so that the equation becomes
T
2
2
xT Ax = (x0 ) Dx0 = 4 (x0 ) + 2 (y 0 ) = 8.
Dividing through by 8 to simplify gives
2
2
(x0 )
(y 0 )
+
= 1,
2
4
which is an ellipse. To determine the rotation matrix Q, find unit eigenvectors corresponding to
and 2 = 2 by row-reduction to get
"
#
p1
2
p1
2
Q=
Then
e1 ! Qe1 =
so that the axis rotation is through a
"
p1
2
p1
2
#
p1
2
p1
2
=4
.
e2 ! Qe2 =
,
1
"
p1
2
p1
2
#
,
45 angle.
y
2
y!
1
"2
1
"1
2
x
"1
x!
"2
55. Writing the equation as xT Ax + Bx + C = 0, we have

⇥
p
3
2
A=
,
B = 28 2
2
3
A has characteristic polynomial
(3
)(3
)
4=
2
p ⇤
22 2 ,
6 +5=(
5)(
C = 84.
1).
Therefore the associated diagonal matrix is
D=

5
0
0
.
1
To determine the rotation matrix Q, find unit eigenvectors corresponding to
row-reduction to get
"
#
Q=
p1
2
p1
2
p1
2
p1
2
.
1
= 5 and
2
= 1 by
432
CHAPTER 5. ORTHOGONALITY
Thus
BQ =
h
p i
22 2
p
28 2
"
p1
2
p1
2
p1
2
p1
2
#
=
h
i
6 ,
50
so the equation becomes
⇣
2
5 (x0 )
5 (x0 )
2
2
50x0 + (y 0 )
6y 0 = 84 )
⌘ ⇣
⌘
2
10x0 + 25 + (y 0 )
6y 0 + 9 = 84 + 125 + 9
5(x
0
2
5) + (y
(x
0
5)
2
5 and y 00 = y 0
(y
+
10
Letting x00 = x0
0
2
3) = 50
0
3)
2
50
)
)
= 1.
3 gives
2
2
(x00 )
(y 00 )
+
= 1,
10
50
which is an ellipse.
56. Writing the equation as xT Ax + Bx + C = 0, we have

⇥
6
2
A=
,
B = 20
2
9
⇤
10 ,
A has characteristic polynomial
(6
)(9
)
2
4=
15 + 50 = (
C=
10)(
5.
15).
Therefore the associated diagonal matrix is
D=

10
0
0
.
5
To determine the rotation matrix Q, find unit eigenvectors corresponding to
row-reduction to get
"
#
Q=
Thus
BQ =
h
20
10
i
"
p1
5
p2
5
p2
5
p1
5
p1
5
p2
5
p2
5
p1
5
#
.
h
= 0
p i
10 5 ,
so the equation becomes
p
2
2
10 (x0 ) + 5 (y 0 )
10 5y 0 = 5 )
⇣
⌘
p
2
2
10 (x0 ) + 5 (y 0 )
2 5 y 0 + 5 = 5 + 25
⇣
p ⌘2
2
10 (x0 ) + 5 y 0
5 = 30 )
p 2
2
y0
5
(x0 )
+
= 1.
3
6
p
Letting x00 = x0 and y 00 = y 0
5 gives
2
2
(x00 )
(y 00 )
+
= 1,
3
6
which is an ellipse.
)
1
= 10 and
2
= 5 by
5.5. APPLICATIONS
433
57. Writing the equation as xT Ax + Bx + C = 0, we have

⇥ p
0 1
A=
,
B= 2 2
1 0
⇤
0 ,
A has characteristic polynomial
(
)(
)
1=
2
D=

1=(
C=
1.
1)( + 1)
Therefore the associated diagonal matrix is
1
0
0
.
1
To determine the rotation matrix Q, find unit eigenvectors corresponding to
row-reduction to get
"
#
p1
2
p1
2
Q=
Thus
h p
BQ = 2 2
i
0
"
p1
2
p1
2
p1
2
p1
2
p1
2
p1
2
1
= 1 and
2
=
1 by
.
#
h
= 2
i
2 ,
so the equation becomes
⇣
(x0 )
2
2
(y 0 ) + 2x0 + 2y 0 = 1 )
⇣
⌘
2
2
(x0 ) + 2x0 + 1
(y 0 )
2y 0 + 1 = 1 + 1
⌘
(x0 + 1)
Letting x00 = x0 + 1 and y 00 = y 0
2
(y 0
2
(y 00 ) = 1,
1
)
2
1) = 1.
1 gives
(x00 )
2
which is a hyperbola.
58. Writing the equation as xT Ax + Bx + C = 0, we have

⇥ p
1
1
A=
,
B= 4 2
1
1
A has characteristic polynomial
(1
)(1
)
2
1=
⇤
0 ,
2 = (
C=
4.
2).
Therefore the associated diagonal matrix is
D=

2
0
0
.
0
To determine the rotation matrix Q, find unit eigenvectors corresponding to
row-reduction to get
"
#
p1
2
p1
2
Q=
Thus
h p
BQ = 4 2
i
0
"
p1
2
p1
2
p1
2
p1
2
p1
2
p1
2
.
#
h
= 4
i
4 ,
1
= 2 and
2
= 0 by
434
CHAPTER 5. ORTHOGONALITY
so the equation becomes
2
2 (x0 ) + 4x0 + 4y 0 = 4
2y 0 =
0 2
)
(x )
2x0 + 2 )
⇣
⌘
2
2y 0 1 =
(x0 ) + 2x0 + 1
)
✓
◆
1
2
2 y0
= (x0 + 1)
)
2
1
1 0
2
y0
=
(x + 1) .
2
2
Letting x00 = x0 + 1 and y 00 = y 0
1
2
gives
1 00 2
(x ) ,
2
y 00 =
which is a parabola.
59. x2 y 2 = (x + y)(x y) = 0 means that x = y or x =
pair of lines y = x and y = x.
y, so the graph of this degenerate conic is the
2
1
!2
1
!1
2
!1
!2
60. Rearranging terms gives x2 + 2y 2 = 2, which has no solutions since the left-hand side is a sum of two
squares, so is always nonnegative. Therefore this is an imaginary conic.
61. The left-hand side is a sum of two squares, so it equals zero if and only if x = y = 0. The graph of this
degenerate conic is the single point (0, 0).
1.
0.5
!1.
0.5
!0.5
!0.5
!1.
1.
5.5. APPLICATIONS
435
62. x2 + 2xy + y 2 = (x + y)2 . Then (x + y)2 = 0 is satisfied only when x + y = 0, so the graph of this
degenerate conic is the line y = x.
2
1
!2
1
!1
2
!1
!2
63. Rearranging terms gives
x2
p
2xy + y 2 = 2 2(y
p
y)2 = 2 2(y
x), or (x
This is satisfied when y = x. If y 6= x, then we can divide through by y
p
p
y x = 2 2, or y = x + 2 2.
x).
x to get
p
So the graph of this degenerate conic is the union of the lines y = x and y = x + 2 2.
4
3
2
1
!2
1
!1
2
!1
!2
64. Writing the equation as xT Ax + Bx + C = 0, we have

⇥ p
2 1
A=
,
B= 2 2
1 2
p ⇤
2 2 ,
C = 6.
436
CHAPTER 5. ORTHOGONALITY
A has characteristic polynomial
(2
)(2
)
1=
2
Therefore the associated diagonal matrix is
D=
4 +3=(

3
0
3)(
1).
0
.
1
To determine the rotation matrix Q, find unit eigenvectors corresponding to
row-reduction to get
"
#
Q=
Thus
h p
BQ = 2 2
so the equation becomes
p1
2
p1
2
p i
2 2
2
p1
2
p1
2
"
1
= 3 and
2
= 1 by
.
p1
2
p1
2
p1
2
p1
2
#
h
= 0
i
4 ,
2
3 (x0 ) + (y 0 ) + 4y 0 = 6 )
⇣
⌘
2
2
3 (x0 ) + (y 0 ) + 4y 0 + 4 = 6 + 4
2
2
3 (x0 ) + (y 0 + 2) =
)
2.
The left-hand side is always nonnegative, so this equation has no solutions and this is an imaginary
conic.
65. Let the eigenvalues of A be
1
and
2.
Then there is an orthogonal matrix Q such that

0
QT AQ = D = 1
.
0
2
Then by the Principal Axes Theorem, xT Ax = x0T Dx0 =
divide through by k to get the standard equation
1
k
(x0 )2 +
2
k
1 (x
0 2
) +
2 (y
0 2
) = k. If k 6= 0, then we may
(y 0 )2 = 1.
(a) Since det A = det D = 1 2 < 0, the coefficients of (x0 )2 and (y 0 )2 in the standard equation have
opposite signs, so this curve is a hyperbola.
(b) Since det A = det D = 1 2 > 0, the coefficients of (x0 )2 and (y 0 )2 in the standard equation have
the same sign. If they are both positive, the curve is an ellipse; if they are positive and equal,
then the ellipse is in fact a circle. If they are both negative, then the equation has no solutions
so is an imaginary conic.
(c) Since det A = det D = 1 2 = 0, then at least one of the coefficients of (x0 )2 and (y 0 )2 in the
standard equation
is zero. If the remaining coefficient is positive, then we get the two parallel
q
lines x0 = ± ki , while if it is zero or negative, the equation has no solutions and we get an
imaginary conic.
(d) If k = 0 but det A = det D = 1 2 6= 0, then the equation is
1 (x
and both
1
and
2
0 2
) +
2 (y
0 2
) = 0,
are nonzero. Then solving for y 0 gives
r
y0 = ±
1
x0 .
2
If 1 and 2 have the same sign, then this equation has only the solution x0 = y 0 = 0, so the
graph is just the origin. If they have opposite signs, then the graph is two straight lines whose
equations are above.
5.5. APPLICATIONS
437
(e) If k = 0 and det A = det D =
= 0, then the equation is
1 2
1 (x
0 2
) +
2 (y
0 2
) = 0,
and at least one of 1 and 2 is zero. If 1 = 0, we get x0 = 0, so the graph is the y 0 -axis, a
straight line. Similarly, if 2 = 0, we get y 0 = 0, so the graph is the x0 -axis, also a straight line
(if both 1 and 2 are zero, then A was the zero matrix to start, so it does not correspond to a
quadratic form).
66. The matrix of this form is
2
4
A = 42
2
which has characteristic polynomial
3
2
+ 12
3
2
25 ,
4
2
4
2
36 + 32 =
2)2 (
(
8)
Therefore the associated diagonal matrix is
2
8
D = 40
0
so that the equation becomes
T
2
3
0
05 ,
2
0
2
0
2
2
2
2
xT Ax = (x0 ) Dx0 = 8 (x0 ) + 2 (y 0 ) + 2 (z 0 ) = 8, or (x0 ) +
This is an ellipsoid.
67. The matrix of this form is
+3
0
1
2
3=
(x + 1)(x
1
A = 40
0
which has characteristic polynomial
3
2
2
+
3
0
25 ,
1
1)(x
3)
Therefore the associated diagonal matrix is
2
1
D=4 0
0
so that the equation becomes
T
xT Ax = (x0 ) Dx0 =
0
1
0
3
0
05 ,
3
2
2
2
1
2
3
2
25 ,
1
2
(x0 ) + (y 0 ) + 3 (z 0 ) = 1,
which is a hyperboloid of one sheet.
68. The matrix of this form is
2
1
A=4 2
2
which has characteristic polynomial
3
3
2
+ 9 + 27 =
(x + 3)2 (x
3)
2
(y 0 )
(z 0 )
+
= 1.
4
4
438
CHAPTER 5. ORTHOGONALITY
Therefore the associated diagonal matrix is
2
so that the equation becomes
T
xT Ax = (x0 ) Dx0 =
3
0 0
3 05 ,
0 3
3
D=4 0
0
3 (x0 )
2
2
2
(x0 )
(y 0 )
+
4
4
2
3 (y 0 ) + 3 (z 0 ) = 12, or
2
2
(z 0 )
=
4
1.
This is a hyperboloid of two sheets.
69. Writing the form as xT Ax + Bx + C
2
0
A = 41
0
The characteristic polynomial of A is
eigenspace gives
2 3
0
6 7
6
v0 = 407
5,
= 0, we have
3
1 0
⇥
0 05 , B = 0
0 0
3
=
(
v1 =
1
Then setting
2
0
6
Q=6
40
1
we get
2
⇤
1 ,
0
1)( + 1). Computing unit eigenvectors for each
3
p1
6 27
6 p1 7 ,
4 25
v
1
0
3
p1
2
p1
2
p1
27
p1 7 ,
2 5
0
Then the form becomes
2
2
6
=6
4
0
6
D=6
40
0
0
⇥
BQ = 1
C = 0.
0
1
0
⇤
0 .
0
xT Ax + Bx + C = 0 ) x0T Dx0 + BQx0 = 0 ) (y 0 )
2
3
p1
27
p1 7 .
2 5
0
3
0
7
07
5
1
2
(z 0 ) + x0 = 0 ) x0 = (z 0 )
This is a hyperbolic paraboloid.
70. Writing the form as xT Ax + Bx + C = 0, we have
2
3
16
0
12
⇥
05 , B = 60
A = 4 0 100
12
0
9
0
⇤
80 ,
C = 0.
The characteristic polynomial of A is
3
+ 125
2
2500 =
(
25)(
Computing unit eigenvectors for each eigenspace gives
2 3
2 3
3
5
6 7
v25 = 4 05 ,
4
5
Then setting
2
3
5
6
Q = 40
4
5
v100
2 3
0
6 7
= 415 .
0
0
25
0
3
0
7
0 5
100
4
5
6 7
v0 = 4 0 5 ,
3
5
4
5
0
3
5
3
0
7
15 ,
0
2
0
6
D = 40
0
100).
2
2
(y 0 ) .
5.5. APPLICATIONS
439
we get
BQ =
Then the form becomes
⇥
100
⇤
0 .
0
2
xT Ax + Bx + C = 0 ) x0T Dx0 + BQx0 = 0 ) 25 (y 0 ) + 100 (z 0 )
2
2
100x0 = 0 ) x0 =
(y 0 )
2
+ (z 0 ) .
4
This is an elliptic paraboloid.
71. Writing the form as xT Ax + Bx + C = 0, we have
2
3
1 2
1
⇥
15 , B = 1
A=4 2 1
1 1
2
⇤
1 ,
1
C = 0.
The characteristic polynomial of A is
3
+9 =
(
3)( + 3).
Computing unit eigenvectors for each eigenspace gives
2
3
2 3
Then setting
6
v0 = 6
4
p1
37
p1 7 ,
35
p1
3
2
6
Q=6
4
we get
p1
3
p1
3
p1
3
v3 =
v
3
0
3
p1
2
p1
2
2
p1
6 7
p1 7 ,
65
p2
6
0
BQ =
Then the form becomes
p1
6 27
6 p1 7 ,
4 25
⇥p
3
0
6
=6
4
3
0
⇤
0 .
0
2
2
3 (z 0 ) +
3
p1
67
p1 7 .
65
p2
6
3
0
7
07
5
3
0
6
D=6
40
0
xT Ax + Bx + C = 0 ) x0T Dx0 + BQx0 = 0 ) 3 (y 0 )
2
p
2
3 x0 = 0 ) x0 =
This is a hyperbolic paraboloid.
72. Writing the form as xT Ax + Bx = 15, we have
2
3
10 0
20
05 ,
A = 4 0 25
20 0
10
⇥ p
B = 20 2
p ⇤
20 2 .
50
The characteristic polynomial of A is
3
+ 45
2
200
7500 =
( + 10)(
Computing unit eigenvectors for each eigenspace gives
2 3
2 3
p1
0
6 27
6 7
7
6
7
v 10 = 6
,
v
=
25
405
415 ,
1
p
0
2
25)(
2
6
v30 = 6
4
30).
3
p1
27
07
5.
p1
2
(y 0 )
p
1/ 3
2
(z 0 )
p
1/ 3
440
CHAPTER 5. ORTHOGONALITY
Then setting
2
p1
6 2
Q=6
40
2
6
D=6
4
0 7
5,
1
p1
2
we get
3
p1
27
0
p1
2
0
⇥
BQ = 40
Then the form becomes
10
0
0
25
0
0
0
3
7
07
5
30
⇤
0 .
50
xT Ax + Bx = 15 )
x0T Dx0 + BQx0 = 15 )
2
2
2
10 (x0 ) + 25 (y 0 ) + 30 (z 0 ) + 40x0 + 50y 0 = 15 )
⇣
⌘
⇣
⌘
2
2
2
10 (x0 ) + 4x0 + 25 (y 0 ) + 2y 0 + 30 (z 0 ) = 15 )
2
2
2
10 (x0 + 2) + 25 (y 0 + 1) + 30 (z 0 ) = 0 )
2
2
(x0 + 2) =
2
(y 0 + 1)
(z 0 )
+
.
2/5
1/3
Letting x00 = x0 + 2, y 00 = y 0 + 1, z 00 = z 0 gives
2
2
(x00 ) =
2
(y 00 )
(z 00 )
+
.
2/5
1/3
This is an elliptic cone.
73. Writing the form as xT Ax + Bx = 6, we have
2
11
A=4 1
4
1
11
4
3
4
45 ,
14
B=
⇥
12
12
⇤
12 .
The characteristic polynomial of A is
3
+ 36
2
396 + 1296 =
(
6)(
12)(
18).
Computing unit eigenvectors for each eigenspace gives
2
6
v6 = 6
4
Then setting
we get
3
p1
37
p1 7 ,
35
p1
3
2
6
Q=6
4
p1
3
p1
3
p1
3
v12 =
2
3
2
p1
6 27
6 p1 7 ,
4 25
6
v18 = 6
4
0
p1
2
p1
2
0
3
p1
6 7
p1 7 ,
65
p2
6
⇥ p
BQ = 12 3
2
6
6
D=6
40
0
0
⇤
0 .
0
12
0
3
p1
67
p1 7 .
65
p2
6
0
3
7
07
5
18
5.5. APPLICATIONS
441
Then the form becomes
xT Ax + Bx = 6 )
x0T Dx0 + BQx0 = 6 )
p
2
2
2
6 (x0 ) + 12 (y 0 ) + 18 (z 0 ) + 12 3x0 6 )
⇣
⌘
p
2
2
2
6 (x0 ) + 2 3 x0 + 12 (y 0 ) + 18 (z 0 ) = 6 )
⇣
⌘
p
2
2
2
6 (x0 ) + 2 3 x0 + 3 + 12 (y 0 ) + 18 (z 0 ) = 24 )
p 2
2
2
x0 + 3
(y 0 )
(z 0 )
+
+
= 1.
4
2
4/3
Letting x00 = x0 +
p
3, y 00 = y 0 , z 00 = z 0 gives
2
2
2
(x00 )
(y 00 )
(z 00 )
+
+
= 1.
4
2
4/3
This is an ellipsoid.
74. The strategy is to reduce the problem to one in which we can apply part (b) of Exercise 65. Using the
⇥
⇤
1
hint, let v be an eigenvector corresponding to = a bi, and let P = Re v Im v . Set B = P P T
.
Now, B is a real matrix; it is also symmetric since
⇣
⌘
⇣
⌘ 1 ⇣
⌘ 1
1 T
T
T
1
BT = P P T
= PPT
= PT PT
= PPT
= B.
Also,
det B = det
⇣
PPT
1
⌘
=
1
1
1
=
=
2 > 0.
det (P P T )
det P det P T
(det P )
Thus Exercise 65(b) applies to B, so that xT Bx = k is an ellipse if both eigenvalues of B are positive
(since k > 0). But by Exercise 29 in this section, B 1 is positive definite, so that B is as well (Exercise
31(d)), and therefore has positive eigenvalues. Hence the trajectories of xT B are ellipses.
Next, let
C=

a
b
b
,
a
so that by Theorem 4.43, A = P CP 1 . Note that


 2
 2
a b a
b
a + b2 ab ab
| |
CT C =
=
=
b a b
a
ab ab a2 + b2
0
0
| |
=

1
0
0
= I.
1
Now,
1
(Ax)T B(Ax) = xT AT P P T
1 T
= xT P CP
1 T
= xT P
T
1
= xT P T
1
=x
T
P
= xT P P T
(Ax)
1
PT
CT P T P T
T
C CP
1
P
1
1
1
P
1
1
P CP
CP
1
x
x
x
x
x
T
= x Bx.
Thus the quadratics xT Bx = k and (Ax)T B(Ax) = k are the same, so that the trajectories of Ax are
ellipses as well.
442
CHAPTER 5. ORTHOGONALITY
Chapter Review
1. (a) True. See Theorem 5.1 and the definition preceding it.
(b) True. See Theorem 5.15; the Gram-Schmidt process allows us to turn any basis into an orthonormal basis, showing that any subspace has such a basis.
(c) True. See the definition preceding Theorem 5.5 in Section 5.1.
(d) True. See Theorem 5.5.
(e) False. If Q is an orthogonal matrix, then |det Q| = 1. However, there are many matrices with
determinant 1 that are not orthogonal. For example,


1 1
2 5
,
.
0 1
1 3
?
?
(f ) True. By Theorem 5.10, (row(A)) = null(A). But if (row(A)) = Rn , then null(A) = Rn so
that Ax = 0 for every x, and therefore A = O (since Aei = 0 for every i, showing that every row
of A is zero).
(g) False. For example, let W be the subset of R2 spanned by e1 , and let v = e2 . Then projW (v) = 0,
but v 6= 0.
(h) True. If A is orthogonal, then A 1 = AT ; but if A is also symmetric, then AT = A. Thus
A 1 = A, so that A2 = AA 1 = I.
(i) False. For example, any diagonal matrix is orthogonally diagonalizable, since it is already diagonal.
But if it has a zero entry on the diagonal, it is not invertible.
(j) True. For example, the diagonal matrix with 1 , 2 , . . . , n on the diagonal has this property.
2. The first pair of vectors is already orthogonal, since 1 · 4 + 2 · 1 + 3 · ( 2) = 0. The other two pairs are
orthogonal if and only if
1·a+2·b+3·3=0
4·a+1·b
)
2·3=0
a + 2b =
4a + b =
Computing and row-reducing the augmented matrix gives


1 2
9
1 0
!
4 1
6
0 1
So the unique solution is a = 3 and b =
form an orthogonal set.
3. Let
2 3
1
u 1 = 4 05 ,
1
9
6.
3
6
6; these are the only values of a and b for which these vectors
2
3
1
u2 = 4 15 ,
1
2
3
1
u3 = 4 25 .
1
Note that these vectors form an orthogonal set. Therefore using Theorem 5.2, we write
v = c1 u1 + c2 u2 + c3 u3
✓
◆
✓
◆
✓
◆
v · u1
v · u2
v · u3
=
u1 +
u2 +
u3
u1 · u1
u2 · u2
u3 · u3
7·1 3·0+2·1
7 · 1 3 · 1 + 2 · ( 1)
7 · ( 1) 3 · 2 + 2 · 1
=
u1 +
u2 +
u3
1·1+0·0+1·1
1 · 1 + 1 · 1 1 · ( 1)
1 · ( 1) + 2 · 2 + 1 · 1
9
2
11
= u1 + u2
u3 .
2
3
6
So the coordinate vector is
2
3
6
[v]B = 4
9
2
2 7
.
3 5
11
6
Chapter Review
443
4. We first find v2 . Since v1 and v2 =

x
are orthogonal, we have
y
3
4
x + y = 0,
5
5
But also v2 is a unit vector, so that x2 + y 2 =
h
i
4
3
• If v2 =
, then
5
5
• If v2 =
h
4
5
3
5
i
or
16 2
9 y
v=
1
3v1 + v2 =
2
v=
1
3v1 + v2 =
2
4
y.
3
x=
25 2
9 y
+ y2 =
= 1. Therefore y = ± 35 and x = ⌥ 45 .
3
" #
3
5
4
5
1
+
2
"
4
5
3
5
#
3
" #
1
+
2
"
4
5
3
5
#
=
"
11
5
21
10
#
.
=
"
7
5
27
10
#
.
, then
3
5
4
5
5. Using Theorem 5.5, we show that the matrix is orthogonal by showing that
2
32
3 2
6
4
6
2
3
p1
p
1 0
7
7
7
7
5
7
5
6 1
7 62
7 6
T
2
15
p
p 76
p 7 = 60 1
0
QQ = 6
0
4
4
5
5 5 47
7 55
4
15
2
3
2
2
p
p
p
p
p
0 0
7
7 5
7 5
7 5
5
7 5
that QQT = I:
3
0
7
07
5 = I.
1
6. Let Q be the given matrix. If Q is to be orthogonal, then by Theorem 5.5, we must have
"
#"
# "
# "
#
1
1
a 12 b
+ a2 12 b + ac
1 0
T
2
4
QQ =
= 1
=
.
b c a c
b2 + c 2
0 1
2 b + ac
Therefore we must have
a2 +
The first equation gives a = ±
±
p
1
= 1,
4
3
2 .
1
ac + b = 0,
2
b2 + c2 = 1.
Substituting this into the second equation gives
p
3
1
c + b = 0,
2
2
p
b = ⌥ 3c
)
b2 = 3c2 .
)
But b2 + c2 = 1 and b2 = 3c2 means that 4c2 = 1, so that c = ± 12 . So there are two possibilities:
• If a =
• If a =
p
3
2 ,
then b =
p
3 c, so we get the two matrices
"
p
3
2 ,
then b =
p
p
3
2
1
2
1
2p
3
2
#
"
,
1
p2
3
2
p
3
2
1
2
#
.
3 c, so we get the two matrices
"
1
p2
3
2
p
3
2
1
2
#
,
"
1
2p
3
2
p
3
2
1
2
#
.
444
CHAPTER 5. ORTHOGONALITY
7. We must show that Qvi · Qvj = 0 for i 6= j and that Qvi · Qvi = 1. Since Q is orthogonal, we know
that QQT = QT Q = I. Then for any i and j (possibly equal) we have
T
Qvi · Qvj = (Qvi ) Qvj = viT QT Q vj = viT vj = vi · vj .
Thus if i 6= j, the dot product is zero since vi · vj = 0 for i 6= j, and if i = j, then the dot product is 1
since vi · vi = 1.
8. The given condition means that for all vectors x and y,
x·y
Qx · Qy
=
.
kxk kyk
kQxk kQyk
Let qi be the ith column of Q. Then qi = Qei . Since the ei are orthonormal and the qi are unit
vectors, we have
(
0 i 6= j
qi · qj
ei · ej
qi · qj =
=
= ei · ej =
kqi k kqj k
kei k kej k
1 i = j.
It follows that Q is an orthogonal matrix.

⇥
5
9. W is the column space of A =
, so that W ? is the null space of AT = 5
2
augmented matrix:
⇥
⇤
⇥
⇤
5 2 0 ! 1 25 0 ,
⇤
2 . Row-reduce the
so that a basis for the null space, and thus for W ? , is given by
" #
" #
2
2
5 , or
.
1
5
2
3
1
⇥
10. W is the column space of A = 4 25, so that W ? is the null space of AT = 1
1
augmented matrix is already row-reduced:
⇥
⇤
1 2
1 0 ,
2
⇤
1 . The
so that the null space, which is W ? , is the set
2
3
2 3
2 3
02 3 2 31
2s + t
2
1
2
1
4 s 5 = s 4 15 + t 405 = span @4 15 , 405A .
t
0
1
0
1
These two vectors form a basis for W ? .
11. Let A be the matrix whose columns are the given basis of W . Then by Theorem 5.10, W ? =
?
(col(A)) = null(AT ). Row-reducing AT , we get


1
1
4 0
1 0
1 0
!
,
0
1
3 0
0 1
3 0
so that a basis for W ? is
2
3
1
4 35 .
1
Chapter Review
445
12. Let A be the matrix whose columns are the given basis of W . Then by Theorem 5.10, W ? =
?
(col(A)) = null(AT ). Row-reducing AT , we get
"
so that
W?
1
1
#
"
0
1
!
0
0
1 1 1
1 1 2
82
39 8 2
s 32 t >
>
>
> >
>
<6 1
7= < 6
t
2
6
7
6
= 4
5> = > s 4
s
>
>
>
>
:
; :
t
13. Row-reducing A gives
2
0
1
#
0
,
0
3
2
1
2
1
0
82
3
2 3 39
1
>
2 >
>
>
<6
6 1 7=
07
7 + t 6 2 7 = span 6
4 05 >
4
15
>
>
>
;
:
0
1
1
60
A!6
40
0
0
1
0
0
2
0
0
0
3
2
0
0
3
1
07
7,
15
0
39
3 >
>
6 17=
6 7 .
4 05>
>
;
2
2
3
4
17
7,
05
0
so that a basis for row(A) is given by the nonzero rows of the reduced matrix, or
⇥
1
0
2
3
A basis for col(A) is given by the columns of
one basis is
82
>
>
<6
6
4
>
>
:
⇤ ⇥
4 , 0
1
0
2
1
⇤
.
A corresponding to leading 1s in the reduced matrix, so
3
1
17
7,
25
3
39
1 >
>
6 27=
6 7 .
4 15>
>
;
5
2
A basis for null(A) is given by the solution set of the matrix, which, from the reduced form, is
82
82
39
2s 3t 4u >
>
>
>
>
>
>
>
>
7>
>
>
2t
u
<6
=
<6
6
7
6
6
7 = span 6
s
6
7
6
>
>
>
>
>
4
5>
4
t
>
>
>
>
>
>
:
;
:
u
3
2
07
7
17
7,
05
0
2
0
1
0
0
0
5
3
0
0
0
6
6
6
6
4
3
3
27
7
07
7,
15
0
2
6
6
6
6
4
Finally, to determine null(AT ), we row-reduce AT :
2
6
6
6
6
4
Then
1
1
2
1
3
1 2
2 1
2 4
1 8
2 9
3
2
3
1
60
57
7
6
6
67
7 ! 60
40
15
7
0
82
82
39
5s t >
>
>
>
>
>
<6
=
<6
7
3s
+
2t
T
6
7
6
null(A ) = 4
=
span
4
s5>
>
>
>
>
>
:
;
:
t
3
1
27
7
07
7.
05
0
3
5
37
7,
15
0
14. The orthogonal decomposition of v with respect to W is
v = projW (v) + (v
39
4 >
>
>
>
17
7=
07
7> .
05>
>
>
;
1
projW (v)),
39
1 >
>
6 27=
6 7 .
4 05>
>
;
1
2
446
CHAPTER 5. ORTHOGONALITY
so we must find projW (v). Let the three given vectors in W be w1 , w2 , and w3 . Then
✓
◆
✓
◆
✓
◆
w1 · v
w2 · v
w3 · v
projW (v) =
w1 +
w2 +
w3
w1 · w1
w2 · w2
w3 · w3
2 3
2
0
7
6
0 · 1 + 1 · 0 + 1 · ( 1) + 1 · 2 6
617 + 1 · 1 + 0 · 0 + 1 · ( 1) 1 · 2 6
=
4
5
1
0·0+1·1+1·1+1·1
1 · 1 + 0 · 0 + 1 · 1 1 · ( 1) 4
1
2 3
3
7
3 · 1 + 1 · 0 2 · ( 1) + 1 · 2 6
6 17
+
4
5
2
3 · 3 + 1 · 1 2 · ( 2) + 1 · 1
1
2 3
2 3
2 3
0
1
3
7 2 6 07
6 17
16
7
1
7
6 7+
6 7
= 6
3 415 3 4 15 15 4 25
1
1
1
2 11 3
Then
6
6
=6
4
15
47
5 7.
19 7
5
15
22
15
v
2
6
6
projW (v) = 6
4
15. (a) Using Gram-Schmidt, we get
2 3
1
617
7
v1 = x1 = 6
415
1
v2 = x2
2 3
1
6 17
7
=6
4 15
0
3
1
07
7
7
15
2
2 3
1
617
7
projv1 (x2 )v1 = 6
415
0
2 3 2 13
1
4
17
7 6
36
1
6
6 7 = 6 47
7
4 415 4 14 5
1
3
2
6
6
6
4
3
11
15
47
57
19 7
5
15
22
15
2
6
6
=6
4
3
4
15
47
57 .
4 7
5
15
8
15
2 3
1
7
1·1+1·1+1·1+1·06
617
4
1 · 1 + 1 · 1 + 1 · 1 + 1 · 1 15
1
4
v3 = x3 projv1 (x3 )v1 projv2 (x3 )v2
2 3
0
6 17 1 · 0 + 1 · 1 + 1 · 1 + 1 · 1
7
=6
4 15 1 · 1 + 1 · 1 + 1 · 1 + 1 · 1 v 1
1
2 13 2 23
2 3
2 3
0
1
4
3
6 17
17
617 3 617 1 6
6 47 6 37
6
7
6
7
=4 5
+ 6 7 = 6 7.
1
4 415 3 4 14 5 4 13 5
1
1
3
0
4
1
4
·0+
1
16
1
4
·1+
1
+ 16
+
1
3
4 ·1
4
1
9
16 + 16
·1
v2
3
1
07
7
15
1
Chapter Review
447
(b) To find a QR factorization, we normalize the vectors from (a) and use those as the column vectors
of Q. Then we set R = QT A. Normalizing v1 , v2 , and v3 yields
2 3 2 3
1
1
6 7 6 21 7
6 7
v1
1 6 17
7 = 627
q1 =
= 6
6
7
17
kv1 k
2 4 15 6
425
1
1
2
2 3 2
p 6
v2
2 36
6
q2 =
=
kv2 k
3 6
4
q3 =
2
p 6
v3
66
6
=
kv3 k
2 6
4
1
4
17
7
47
17
45
3
4
3
2
3
17
7
37
17
35
p
6
6
=6
6
4
2
6
6
=6
6
4
0
3
3
p6 7
37
p6 7
37
5
p6
3
2
p 3
6
p3 7
67
p6 7 .
67
6 5
0
Then
Q=
2
1
6 21
62
6
61
42
1
2
p
3
p6
3
p6
3
6p
3
2
p
3
6
p3 7
6 7
p6 7 ,
6 7
6 5
0
and
2
6
R = QT A = 6
4
1
p2
3
6p
6
3
1
p2
3
p6
6
6
1
p2
3
p6
6
6
2
3 1
1
2p 7 6
61
37 6
2 56
41
0
1
1
1
1
0
3
2
0
2
7
6
17
7 = 60
4
17
5
0
1
3
p2
3
2
0
3
2p
3
7
37 .
p6 5
6
3
Then A = QR is a QR factorization of A.
16. Note that the two given vectors,
2 3
2 3
1
0
6 07
6 17
7
6 7
v1 = 6
425 and v2 = 4 15 ,
2
1
are indeed orthogonal. We extend these to a basis by adding e3 and e4 . We can verify that this is a
basis by computing the determinant of the matrix whose column vectors are the vi ; this determinant
is 1, so the matrix is invertible and therefore the vectors are linearly independent, so they do in fact
448
CHAPTER 5. ORTHOGONALITY
form a basis for R4 . We apply Gram-Schmidt to v3 and v4 :
v3 = x3 projv1 (x3 )v1 projv2 (x3 )v2
2 3
0
6 07 0 · 1 + 0 · 0 + 1 · 2 + 0 · 2
6 7
=6 7
v1
12 + 02 + 22 + 22
4 15
0
2 3
0
6 07
6 7
=6 7
4 15
0
2 3
1
6
2 607
7
6 7
9 425
2
2
16
6
6
34
3 2
0
6
17
7 6
7=6
15 4
1
0 · 0 + 0 · 1 + 1 · 1 + 0 · ( 1)
v2
02 + 12 + 12 + ( 1)2
3
2
9
17
37
27
5
9
1
9
v4 = x4 projv1 (x4 )v1 projv2 (x4 )v2
2 3
0
6 07 0 · 1 + 0 · 0 + 0 · 2 + 1 · 2
6 7
=6 7
v1
12 + 02 + 22 + 22
4 05
projv3 (x4 )v3
0 · 0 + 0 · 1 + 0 · 1 + 0 · ( 1)
v2
02 + 12 + 12 + ( 1)2
1
2
9
1
2
1
+0·
3 +0· 9 +1·
9
v3
4
1
4
1
+
+
+
81
9
81
81
2 3
2 3
2 23 2 13
1
0
9
3
7
6 7
6 17 6 17
26
607 1 6 17 1 6 3 7 6 6 7
6 7 + 6 7 + 6 7 = 6 7.
9 425 3 4 15 2 4 29 5 4 05
1
1
2
1
9
6
0·
2 3
0
6 07
6 7
=6 7
4 05
1
17. If x1 + x2 + x3 + x4 = 0, then x4 =
x1
8
>
>
<
x2
x3 , so that a basis for W is given by
3
2 3
2 39
1
0
0 >
>
6 07
6 17
6 07 =
7 , x2 = 6 7 , x3 = 6 7 .
B = x1 = 6
4 05
4 05
4 15 >
>
>
>
:
;
1
1
1
2
Then using Gram-Schmidt,
2
6
6
v1 = x1 = 6
4
v2 = x2
2
3
1
07
7
7
05
1
2
6
6
projv1 (x2 )v1 = 6
4
3
0
6 17
6 7
=6 7
4 05
1
2
3
2
3
0
17
7
7
05
1
3
1
1
2
7 6 7
16
6 07 6 17
6 7=6 7
2 4 05 4 05
1
1
2
2
1 · 0 + 0 · 1 + 0 · 0 1 · ( 1) 6
6
6
12 + 02 + 02 + ( 1)2
4
3
1
07
7
7
05
1
Chapter Review
449
v3 = x3 projv1 (x3 )v1 projv2 (x3 )v2
2 3
2 3
0
1
6 07 1 · 0 + 0 · 0 + 0 · 1 1 · ( 1) 6 07
6 7
6 7
=6 7
6 7
12 + 02 + 02 + ( 1)2
4 15
4 05
1
1
2 3
2 3
2 13 2 13
0
1
2
3
6 07 1 6 07 1 6 17 6 1 7
6 7
6 7
6 7 6 37
=6 7
6 7
6 7 = 6 7.
4 15 2 4 05 3 4 05 4 15
1
1
2
1
1
2
·0+1·0+0·1
1
4 +1+0+
1
2
1
4
2
1
2
3
7
· ( 1) 6
6 17
6 7
4 05
1
2
1
3
18. (a) The characteristic polynomial of A is
2
det(A
I) =
1
1
1
1
1
2
1
3
=
+6
2
9 =
(
2
⇥
I
so the eigenvalues are 0 and 3. To find the eigenspaces, we row-reduce A
2
3
2
3
2 1
1 0
1 0
1 0
⇥
⇤
A 0I 0 = 4 1 2
1 05 ! 40 1
1 05
1 1
2 0
0 0
0 0
3
2
3
2
1
1
1 0
1
1 1 0
⇥
⇤
A 3I 0 = 4 1
1
1 05 ! 40
0 0 05 .
1
1
1 0
0
0 0 0
Thus the eigenspaces are
8
2 39
1 =
<
E0 = span v1 = 4 15 ,
:
;
1
3)2 ,
⇤
0 :
8
<
2 3
2 39
1
1 =
E3 = span v2 = 415 , x3 = 4 05 .
:
;
0
1
The vectors in E0 are orthogonal to those in E3 since they correspond to di↵erent eigenvalues,
but the basis vectors above in E3 are not orthogonal, so we use Gram-Schmidt to convert the
second basis vector into one orthogonal to the first:
2 3
2 3 2 3
1
1
1
2
6 7 1 · ( 1) + 1 · 0 + 0 · 1 6 7 6 1 7
v 3 = 4 05
415 = 4 2 5 .
1·1+1·1+0·0
1
0
1
Finally, normalize v1 , v2 , and v3 and let Q be the matrix whose columns are those vectors; then
Q is the orthogonal matrix
2
3
6
Q=6
4
p1
3
p1
3
p1
3
p1
2
p1
2
0
p1
67
p1 7 .
6 5
p2
6
Finally, an orthogonal diagonalization of A is given by
2
0 0
QT AQ = D = 40 3
0 0
3
0
05 .
3
450
CHAPTER 5. ORTHOGONALITY
(b) From part (a), we have
1
= 0 and
2
3
6
q1 = 6
4
2
=
p1
37
p1 7 ,
35
p1
3
3
= 3, with
2 3
q2 =
T
1 q1 q1
T
2 q2 q2
+
3
p1
67
p1 7 .
65
p2
6
6
q3 = 6
4
0
The spectral decomposition of A is
A=
2
p1
6 27
6 p1 7 ,
4 25
T
3 q3 q3
+
= 3q2 qT2 + 3q3 qT3
2 3
2
"
#
p1
1
1
2
p
6 7 p
6
1 7
2
2 + 36
= 36
4 p2 5 0
4
0
2
3
2
1
1
1
0
2
2
6
7
6 6
1
1
7
6 1
= 36
4 2 2 05 + 3 4 6
1
0 0 0
3
3
p1
67 h
p1 7
65
p2
6
1
6
1
6
1
3
p1
6
p1
6
p2
6
i
3
1
37
17 .
35
2
3
19. See Example
2 3 5.20, or Exercises 23 and 24 in Section 5.4. The basis vector for E 2 , which we name
1
v3 = 4 15, is orthogonal to the other two since they correspond to di↵erent eigenvalues. However,
0
we must orthogonalize the basis for E0 using Gram-Schmidt:
2 3
2 3
2 3 2 3
1
1
1
0
6 7
6 7 1·1+1·1+1·06 7 6 7
v1 = 415 ,
v2 = 415
=
415 405 .
12 + 1 2 + 0 2
0
1
0
1
p p
Now v1 , v2 , and v3 are mutually orthogonal, with norms 2, 2, and 1 respectively. Let Q be the
matrix whose columns are q1 = kvv11 k , q2 = kvv22 k , and q3 = kvv33 k ; then the desired matrix is
2
1
6
Q6
40
0
0
1
0
3
0
7
07
5Q
2
1
2
1
0
6
= Q6
40
2
6
=6
4
1
0
0
p1
2
p1
2
0
0
0
1
3
0
7 T
07
5Q
2
32
p1
1
27 6
1
p 7 60
25 4
20. vi viT is symmetric since
vi viT
0
0
T
= viT
32
p1
0
76 2
6
07
54 0
p1
2
2
0
1
0
T
p1
2
0
p1
2
3 2
1
0
7 6 2
6 3
17
5=4 2
0
0
3
2
1
2
0
3
0
7
07
5.
1
viT = viT vi ,
so it is equal to its own transpose. Since a scalar times a symmetric matrix is a symmetric matrix, and
the sum of symmetric matrices is symmetric, it follows that A is symmetric. Next,
!
n
n
n
X
X
X
T
Avj =
ci v i vi v j =
ci vi viT vj =
ci vi (vi · vj ) = cj vj
i=1
i=1
i=1
since {vi } is a set of orthonormal vectors. This shows that each cj is an eigenvalue for A with
corresponding eigenvector vj . Since there are n of them, these are all the eigenvalues and eigenvectors.
Chapter 6
Vector Spaces
6.1
Vector Spaces and Subspaces
⇢
x
. V is the set of all vectors in R2 whose first and second components are the same.
x
We verify all ten axioms of a vector space:





x
y
x
y
x+y
1. If
,
2 V then
+
=
2 V since its first and second components are the same.
x
y
x
y
x+y






x
y
x+y
y+x
y
x
2.
+
=
=
=
+
.
x
y
x+y
y+x
y
x
✓
 ◆ 






✓
 ◆
x
y
z
x+y
z
x+y+z
x
y+z
x
y
z
3.
+
+
=
+
=
=
+
=
+
+
.
x
y
z
x+y
z
x+y+z
x
y+z
x
y
z





0
x
0
x+0
x
4.
2 V , and
+
=
=
2V.
0
x
0
x+0
x






x
x
x
x
x x
0
5. If
2 V , then
2 V , and
+
=
=
.
x
x
x
x
x x
0



x
x
cx
6. If
2 V , then c
=
2 V since its first and second components are the same.
x
x
cx
✓
 ◆




✓
 ◆
x
y
x+y
c(x + y)
c(y + x)
y+x
y
x
7. c
+
=c
=
=
=c
=c
+
.
x
y
x+y
c(x + y)
c(y + x)
y+x
y
x







x
(c + d)x
cx + dx
cx
dx
x
x
=
=
=
+
=c
+d
.
8. (c + d)
x
(c + d)x
cx + dx
cx
dx
x
x
✓  ◆




x
dx
c(dx)
(cd)x
x
=c
=
=
= (cd)
.
9. c d
x
dx
c(dx)
(cd)x
x



x
1x
x
=
=
.
10. 1
x
1x
x
⇢
x
2. V =
: x, y 0 fails to satisfy Axioms 5 and 6, so it is not a vector space:
y




x
0
x
cx
5. If
6=
2 V , choose c < 0; then c
=
2
/ V since at least one of cx and cy is negative.
y
0
y
cy



x
0
x
6. If
6=
2 V , then
2
/ V since x < 0.
x
0
x
1. Let V =
451
452
CHAPTER 6. VECTOR SPACES
3. V =
⇢
x
y
: xy
0
fails to satisfy Axiom 1, so it is not a vector space:

x
y
and
are both in V , but
y
x




x
y
x y
x y
+
=
=
2
/V
y
x
y x
(x y)
1. Choose x 6= y with xy
0. Then

since (x y)( (x y)) = (x y)2 < 0 because x 6= y.
⇢
x
4. V =
: x y fails to satisfy Axioms 5 and 6, so it is not a vector space:
y


x
x
5. If
2 V and x 6= y, then x > y, so that x < y. Therefore
2
/ V.
y
y



x
cx
x
6. Again let
2 V and x 6= y, and choose c = 1. Then
=
2
/ V as above.
y
cy
y
.
5. R2 with the given scalar multiplication fails to satisfy Axiom 8, so it is not a vector space:

x
8. Choose
2 R2 with y 6= 0, and let c and d be nonzero scalars. Then
y








x
(c + d)x
cx + dx
x
x
cx
dx
cx + dx
(c + d)
=
=
, while c
+d
=
+
=
.
y
y
y
y
y
y
y
2y
These two are unequal since we chose y 6= 0.
6. R2 with the given addition fails to satisfy Axiom 7, so it is not a vector space:


x1
x
8. Choose
, 2 2 R2 and let c be a scalar not equal 1. Then
y1
y2
✓

◆


x1
x
x + x2 + 1
cx1 + cx2 + c
c
+ 2
=c 1
=
y1
y2
y 1 + y2 + 1
cy1 + cy2 + c





x
x
cx1
cx2
cx1 + cx2 + 1
c 1 +c 2 =
+
=
.
y1
y2
cy1
cy2
cy1 + cy2 + 1
These two are unequal since we chose c 6= 1.
7. All axioms apply, so that this set is a vector space. Let x, y, and z by positive real numbers, so that
x, y, z 2 V , and c and d be any real numbers. Then
1. x
y = xy > 0, so that x
2. x
y = xy = yx = y
3. (x
y)
z = (xy)
4. 0 = 1, since x
5.
x.
y 2V.
z = (xy)z = x(yz) = x
0 = x1 = 1x = 0
x is the positive real number
6. c
7. c
c
x = x > 0, so that c
(x
8. (c + d)
9. c
10. 1
(d
c
x=x
x) = c
x
x = x = x.
=x x =x
d
x = x
d c
c
=x
1
x
( x) = x ·
y c = (c
x)
(c
y).
x)
(c
y).
x2V.
c d
(y
since x
y) = (xy) = x y = xc
c+d
c c
1
x,
(yz) = x
z).
x = x.
d
x = (c
cd
c d
= (x ) = d
c
x =d
= 1 = 0.
(c
x).
6.1. VECTOR SPACES AND SUBSPACES
453
8. This is not a vector space, since Axiom 6 fails. Choose any rational number q, and let c = ⇡. Then q⇡
is not rational, so that V is not closed under scalar multiplication.
9. All
so that this is a vector space. Let x, y, z, u, v, w, c, and d be real numbers. Then
 axioms apply,

x y
u v
and
2 V , and
0 z
0 w
1. The sum of upper triangular matrices is again upper triangular.
2. Matrix addition is commutative.
3. Matrix addition is associative.


x y
x
4. The zero matrix O is upper triangular, and
+O =
0 z
0


x y
x
y
.
5.
is the matrix
0 z
0
z


x y
cx cy
6. c
=
2 V since it is upper triangular.
0 z
0 cz
y
.
z
7.
c
✓
x
0

y
u
+
z
0
v
w
◆
✓
◆
x+u y+v
=c
0
z+w

c(x + u) c(y + v)
=
0
c(z + w)

cx + cu cy + cv
=
0
cz + cw


cx cy
cu cv
+
=
0 cz
0 cw


x y
u v
=c
+c
.
0 z
0 w
8.
(c + d)
✓ 
◆

x y
dx
9. c d
=c
0 z
0


x y
x y
10. 1
=
.
0 z
0 z

dy
c(dx)
=
dz
0

x
0

y
(c + d)x (c + d)y
=
z
0
(c + d)z

cx + dx cy + dy
=
0
cz + dz


cx cy
dx dy
=
+
0 cz
0 dz


x y
x y
=c
+d
.
0 z
0 z

c(dy)
(cd)x
=
c(dz)
0

(cd)y
x
= (cd)
(cd)z
0
y
.
z
10. This set V is not a vector space, since it is not closed under addition. For example,





1 0
0 0
1 0
0 0
1 0
,
2 V, but
+
=
2
/V
0 0
0 1
0 0
0 1
0 1
since the product of the diagonal entries is not zero.
454
CHAPTER 6. VECTOR SPACES
11. This set V is a vector space. Let A be a skew-symmetric matrix and c a constant. Then
1. The sum of skew-symmetric matrices is skew-symmetric.
2. Matrix addition is commutative.
3. Matrix addition is associative.
4. The zero matrix O is skew-symmetric, and A + O = A.
5. If A is skew-symmetric, then so is
A, since ( A)ij =
aij = aji = (A)ji =
6. If A is skew-symmetric, then so is cA, since (cA)ij = c(A)ij =
c(A)ji =
( A)ji .
(cA)ji .
7. Scalar multiplication distributes over matrix addition.
8, 9. These properties hold for general matrices, and skew-symmetric matrices are closed under these
operations, so then in V .
10. 1A = A is skew-symmetric.
12. Axioms 1, 2, and 6 were verified in Example 6.3. For the remaining examples, using the notation from
Example 3, and letting r(x) = c0 + c1 x + c2 x2 ,
3.
(p(x) + q(x)) + r(x) = ((a0 + b0 ) + (a1 + b1 )x + (a2 + b2 )x2 ) + r(x)
= (a0 + b0 + c0 ) + (a1 + b1 + c1 )x + (a2 + b2 + c2 )x2
= (a0 + a1 x + a2 x2 ) + ((b0 + c0 ) + (b1 + c1 )x + (b2 + c2 )x2 )
= p(x) + (q(x) + r(x)).
4. The zero polynomial 0 + 0x + 0x2 is the zero vector, since
0 + p(x) + (0 + a0 ) + (0 + a1 )x + (0 + a2 )x2 = a0 + a1 x + a2 x2 = p(x).
5.
p(x) is the polynomial
a0
a1 x
a2 x2 .
7.
c(p(x) + q(x)) = c (a0 + a1 x + a2 x2 ) + (b0 + b1 x + b2 x2 )
= c (a0 + b0 ) + (a1 + b1 )x + (a2 + b2 )x2
= c(a0 + b0 ) + c(a1 + b1 )x + c(a2 + b2 )x2
= (ca0 + cb0 ) + (ca1 + cb1 )x + (ca2 + cb2 )x2
= ca0 + ca1 x + ca2 x2 + cb0 + cb1 x + cb2 x2
= c a 0 + a 1 x + a 2 x 2 + c b0 + b 1 x + b2 x 2
= cp(x) + cq(x).
8.
(c + d)p(x) = (c + d) a0 + a1 x + a2 x2
= (c + d)a0 + (c + d)a1 x + (c + d)a2 x2
= (ca0 + da0 ) + (ca1 + da1 )x + (ca2 + da2 )x2
= (ca0 + ca1 x + ca2 x2 ) + (da0 + da1 x + da2 x2 )
= c a0 + a1 x + a2 x2 + d a0 + a1 x + a2 x2
= cp(x) + dp(x).
6.1. VECTOR SPACES AND SUBSPACES
455
9.
c(dp(x)) = c da0 + da1 x + da2 x2
= c(da0 ) + c(da1 )x + c(da2 )x2
= (cd)a0 + (cd)a1 x + (cd)a2 x2
= (cd) a0 + a1 x + a2 x2
= (cd)p(x).
10. 1p(x) = 1a0 + 1a1 x + 1a2 x2 = a0 + a1 x + a2 x2 = p(x).
13. Axioms 1 and 6 were verified in Example 6.4. For the rest, using the notation from Example 4,
2, 3. Addition of functions is commutative and associative.
4. The zero function f (x) = 0 for all x is the zero vector, since (0 + f )(x) = 0(x) + f (x) = f (x).
5. The function f such that ( f )(x) = f (x) is such that ( f + f )(x) = ( f )(x) + f (x) =
f (x) + f (x) = 0, so their sum is the zero function.
7. The function c(f + g) is defined by (c(f + g))(x) = c((f + g)(x)) by the definition of scalar multiplication. But (f + g)(x) = f (x) + g(x), so we get (c(f + g))(x) = c(f (x) + g(x)) = cf (x) + cg(x) =
(cf + cg)(x).
8. Again using the definition of scalar multiplication, the function (c+d)f is defined by ((c+d)f )(x) =
(c + d)(f (x)) = cf (x) + df (x) = (cf + df )(x).
9. The function df is defined by (df )(x) = df (x), so c(df ) is defined by (c(df ))(x) = c((df )(x)) =
cdf (x) = ((cd)f )(x).
10. (1f )(x) = 1f (x) = f (x).
14. This is not a complex vector space, since Axiom 6 fails. Let z 2 C be any nonzero complex number,
and let c 2 C be any complex number with nonzero real and imaginary parts. Then



z
cz
cz
c
=
6=
.
z̄
cz̄
cz
15. This is a vector space over C; the proof is identical to the proof that Mmn (R) is a real vector space.


z
w
16. This is a complex vector space. Let 1 , 1 2 C2 , and let c, d 2 C. Then
z2
w2
1, 2, 3, 4, 5. These follow since addition is the same as in the usual vector space C2 .


z
c̄z1
6. c 1 =
2 C2 .
z2
c̄z2
7.
✓

◆



z1
w
z +w+1
c̄(z1 + w1 )
c̄z1 + c̄w1
c
+ 1
=c 1
=
=
z2
w2
z2 + w 2
c̄(z2 + w2 )
c̄z2 + c̄w2




c̄z1
c̄w1
z
w
=
+
=c 1 +c 1 .
c̄z2
c̄w2
z2
w2






z1
(c + d)z1
(c + d)z1
cz1
dz1
z
z
=
=
=
+
=c 1 +d 1 .
z2
cz2
z2
z2
(c + d)z2
dz2
(c + d)z2




✓  ◆
¯
¯
z
(cd)z1
c̄dz
dz
z
9. (cd) 1 =
= ¯1 =c ¯1 =c d 1 .
z2
c̄
dz
dz
z2
(cd)z2
2
2




z
1̄z1
1z1
z
10. 1 1 =
=
= 1 .
z2
1̄z2
1z2
z2
8. (c + d)

456
CHAPTER 6. VECTOR SPACES
17. This is not a complex vector space, since Axiom 6 fails. Let x 2 Rn be any nonzero vector, and let
c 2 C be any complex number with nonzero imaginary part. Then cx 2
/ Rn , so that Rn is not closed
under scalar multiplication.
18. This set V is a vector space over Z2 . Let x, y 2 Zn2 each have an even number of 1s, say x has 2r ones
and y has 2s ones.
1. x + y has a 1 where either x or y has a 1, except that if both x and y have a 1, x + y has a zero. Say
this happens in k di↵erent places. Then the total number of 1s in x+y is 2r+2s 2k = 2(r+s k),
which is even, so V is closed under addition.
2. Since addition in Z2 is commutative, we see that x + y = y + x since we are simply adding
corresponding components.
3. Since addition in Z2 is associative, we see that (x + y) + z = x + (y + z) since we are simply adding
corresponding components.
4. 0 2 Zn2 has an even number (zero) of ones, so 0 2 V , and x + 0 = x.
5. Since 1 =
1 in Z2 , we can take
x = x; then
x + x = x + x = 0.
6. If c 2 Z2 , then cx is equal to either 0 or x depending on whether c = 0 or c = 1. In either case,
x 2 V implies that cx 2 V .
7, 8. Since multiplication distributes over addition in Z2 , these axioms follow since operations are
component by component.
9. Since multiplication in Z2 is distributive, these axioms follow since operations are component by
component.
10. 1x = x since multiplication is component by component.
19. This set V is not a vector space:
1. For example, let x and y be vectors with exactly one 1. Then x + y has either 0 or 2 ones (it has
zero if the locations of the 1s in x and y are the same), so x + y 2
/ V and V is not closed under
addition.
4. 0 2
/ V since 0 does not have an odd number of 1s.
6. If c = 0, then cx = 0 2
/ V if x 2 V , so that V is not closed under scalar multiplication.
20. This is a vector space over Zp ; the proof is identical to the proof that Mmn (R) is a real vector space.
21. This is not a vector space. Axioms 1 through 7 are all satisfied, as is Axiom 10, but Axioms 8 and 9
fail since addition and multiplication are not the same in Z6 as they are in Z3 . For example:
8. Let c = 2, d = 1, and u = 1. Then (c + d)u = (2 + 1)u = 0u = 0 since addition of scalars is
performed in Z3 , while cu + du = 2 · 1 + 1 · 1 = 2 + 1 = 3 since here the addition is performed in
Z6 .
9. Let c = d = 2 and u = 1. Then (cd)u = 1 · 1 = 1 since the scalar multiplication is done in Z3 , while
c(du) = 2(2 · 1) = 2 · 2 = 4 since here the multiplication is done in Z6 .
22. Since 0 = 0 + 0, we have
0u = (0 + 0)u = 0u + 0u
by Axiom 8. By Axiom 5, we add
0u to both sides of this equation, giving
0u + 0u = ( 0u + 0u) + 0u, so that by Axiom 5 0 = 0 + 0u.
Finally, 0 + 0u = 0u by Axiom 4, so that 0 = 0u.
23. To show that ( 1)u = u, it suffices to show that ( 1)u + u = 0, by Axiom 5. But ( 1)u + u =
( 1)u + 1u by Axiom 10. Then use Axiom 8 to rewrite this as ( 1 + 1)u = 0u. By part (a) of Theorem
6.1, 0u = 0 and we are done.
6.1. VECTOR SPACES AND SUBSPACES
457
24. This is a subspace. We check each of the conditions in Theorem 6.2:
2 3 2 3 2
3
a
b
a+b
1. 405 + 405 = 4 0 5 2 W since its first and third components are equal and its second component
a
b
a+b
is zero.
2 3 2 3
a
ca
2. c 405 = 4 0 5 2 W .
a
cz
25. This is a subspace. We check each of the conditions in Theorem 6.2:
2 3 2 3 2
3 2
3
a
b
a+b
(a + b)
1. 4 a5 + 4 b5 = 4 a b 5 = 4 (a + b)5 2 W .
2a
2b
2a + 2b
2(a + b)
2 3 2
3 2
3
a
ac
(ac)
2. c 4 a5 = 4 ac5 = 4 (ac)5 2 W .
2a
2ac
2(ac)
26. This is not a subspace. It fails both conditions in Theorem 6.2:
2
3 2
3 2
3
a
c
a+c
5+4
5=4
5, which is not in W .
b
d
b+d
1. 4
a+b+1
c+d+1
(a + c) + (b + d) + 2
2
3 2
3 2
3
a
ca
ca
5=4
5=4
5, which is not in W if c 6= 1.
b
cb
cb
2. c 4
a+b+1
c(a + b + 1)
ca + cb + c
27. This is not a subspace. It fails both conditions in Theorem 6.2:
2 3 2 3 2
3
a
c
a+c
1. 4 b 5 + 4 d 5 = 4 b + d 5. If a and c do not have the same sign, then |a| + |c| =
6 |a + c|, so that
|a|
|c|
|a| + |c|
the sum is not in W .
2 3 2
3
a
ca
2. c 4 b 5 = 4 cb 5. If c < 0, then c |a| =
6 |ca|, so that the result is not in W .
|a|
c |a|
28. This is a subspace. We check each of the conditions in Theorem 6.2:




a b
c d
a+c
b+d
a+c
b+d
1.
+
=
=
2 W.
b 2a
d 2c
b + d 2a + 2c
b + d 2(a + c)



a b
ca cb
ac
bc
2. c
=
=
2 W.
b 2a
cb 2ac
bc 2(ac)
29. This is not a subspace; condition 1 in Theorem 6.2 fails. For example, if a, d


a 0
a
a
,
2 W,
0 d
d
d
but
since 0 < ad.

a
0
0
+
d

a
d

a
0
=
d
d
a
2
/W
0
0, then
458
CHAPTER 6. VECTOR SPACES
30. This is not a subspace; both conditions of Theorem 6.2 fail. For example:
Let A = B = I; then det A = det B = 1, so that A, B 2 W , but det(A + B) = det 2I = 2, so
A+B 2
/ W.
Let A = I and c be any constant other than ±1. Then det(cA) = cn det A 6= det A = 1, so that
cA 2
/ W.
31. Since the sum of two diagonal matrices is again diagonal, and a scalar times a diagonal matrix is
diagonal, this subset satisfies both conditions of Theorem 6.2, so it is a subspace.
32. This is not a subspace. For example, it fails the second condition: let A be idempotent and c be any
constant other than 0 or 1. Then
(cA)2 = c2 A2 = c2 A 6= cA
so that cA is not idempotent, so is not in W .
33. This is a subspace. We check each of the conditions in Theorem 6.2:
1. Suppose that A, C 2 W . Then (A + C)B = AB + CB = BA + BC = B(A + C), so that A + C 2 W .
2. Suppose that A 2 W and c is any scalar. Then (cA)B = c(AB) = c(BA) = B(cA), so that cA 2 W .
34. This is a subspace. We check each of the conditions in Theorem 6.2:
1. Let v = bx + cx2 , v0 = b0 x + c0 x2 . Then
v + v0 = bx + cx2 + b0 x + c0 x2 = (b + b0 )x + (c + c0 )x2 2 W.
2. Let v = bx + cx2 and d be any scalar. Then dv = d(bx + cx2 ) = (db)x + (dc)x2 2 W .
35. This is a subspace. We check each of the conditions in Theorem 6.2:
1. Let v = a + bx + cx2 , v0 = a0 + b0 x + c0 x2 2 W . Then
v + v0 = a + bx + cx2 + a0 + b0 x + c0 x2 = (a + a0 ) + (b + b0 )x + (c + c0 )x2 .
Then (a + a0 ) + (b + b0 ) + (c + c0 ) = (a + b + c) + (a0 + b0 + c0 ) = 0, so that v + v0 2 W .
2. Let v = a + bx + cx2 and d be any scalar. Then dv = d(a + bx + cx2 ) = (da) + (db)x + (dc)x2 , and
da + db + dc = d(a + b + c) = d · 0 = 0, so that dv 2 W .
36. This is not a subspace; it fails condition 1 in Theorem 6.2. For example, let v = 1 + x and v0 = x2 .
Then both v and v0 are in V , but v + v0 = 1 + x + x2 2
/ V since the product of its coefficients is not
zero.
37. This is not a subspace; it fails both conditions of Theorem 6.2.
1. For example let v = x3 and v0 =
x3 . Then v + v0 = 0, which is not a degree 3 polynomial.
2. Let c = 0; then cv = 0 for any v, but 0 2
/ W , so that this condition fails.
38. This is a subspace. We check each of the conditions in Theorem 6.2:
1. Let f, g 2 W . Then (f + g)( x) = f ( x) + g( x) = f (x) + g(x) = (f + g)(x), so that f + g 2 W .
2. Let f 2 W and c be any scalar. Then (cf )( x) = cf ( x) = cf (x) = (cf )(x), so that cf 2 W .
39. This is a subspace. We check each of the conditions in Theorem 6.2:
1. Let f, g 2 W . Then (f + g)( x) = f ( x) + g( x) =
so that f + g 2 W .
f (x)
2. Let f 2 W and c be any scalar. Then (cf )( x) = cf ( x) =
g(x) =
(f (x) + g(x)) =
cf (x) =
(f + g)(x),
(cf )(x), so that cf 2 W .
6.1. VECTOR SPACES AND SUBSPACES
459
40. This is not a subspace; it fails both conditions in Theorem 6.2. For example, if f, g 2 W , then
(f + g)(0) = f (0) + g(0) = 1 + 1 6= 1, so that f + g 2
/ W . Similarly, if c is a scalar not equal to 1, then
(cf )(0) = cf (0) = c 6= 1, so that cf 2
/ W.
41. This is a subspace. We check each of the conditions in Theorem 6.2:
1. Let f, g 2 W . Then (f + g)(0) = f (0) + g(0) = 0 + 0 = 0, so that f + g 2 W .
2. Let f 2 W and c be any scalar. Then (cf )(0) = cf (0) = c · 0 = 0, so that cf 2 W .
42. If f and g 2 W are integrable and c is a constant, then
Z
Z
Z
(f + g) = f + g,
Z
(cf ) = c
Z
f,
so that both f + g and cf 2 W . Thus W is a subspace.
43. This is not a subspace; it fails condition 2 in Theorem 6.2. For example, let f 2 W be such that
f 0 (x) > 0 everywhere, and choose c = 1; then (cf )0 (x) = cf 0 (x) < 0, so that cf 2
/ W.
44. If f, g 2 W and c is a constant, then
(f + g)00 = f 00 + g 00 ,
(cf )00 = cf 00 ,
so that both f + g and cf have continuous second derivatives, so they are both in W .
45. This is not a subspace; it fails both conditions in Theorem 6.2.
1. f (x) =
1
x2
2 W , but
f (x) =
1
x2
2
/ W , since limx!0 ( f (x)) =
1.
2. Take c = 0 and f 2 W ; then limx!0 (cf )(x) = limx!0 0 = 0, so that cf 2
/ W.
46. Suppose that v, v0 2 U \ W , and let c be a scalar. Then v, v0 2 U , so that v + v0 and cv 2 U .
Similarly, v, v0 2 W , so that v + v0 and cv 2 W . Therefore v + v0 2 U \ W and cv 2 U \ W , so that
U \ W is a subspace.
⇢
⇢
x
0
47. For example, let U =
, the x-axis, and let W =
, the y-axis. Then U [ W consists of all
0
y
2
vectors in R in which at least one component is zero. Choose a, b 6= 0; then





a
0
a
0
a
2 U,
2 W, but
+
=
2
/ U [W
0
b
0
b
b
since neither a nor b is zero.
48. (a) U + W is the xy-plane, since
2 3
2 3
x
0
U = 4 0 5 , W = 4y 5 ,
0
0
so that any point in the xy-plane is
2 3 2 3 2 3
a
a
0
4 b 5 = 405 + 4 b 5 2 U + W.
0
0
0
However, a point not in the xy-plane has nonzero z-coordinate, so it cannot be a sum of an
element of U and an element of W . Thus U + W is exactly the xy-plane.
460
CHAPTER 6. VECTOR SPACES
(b) Suppose x, x0 2 U + W ; say x = u + w and x0 = u0 + w0 , where u, u0 2 U and w, w0 2 W . Then
x + x0 = u + w + u0 + w0 = (u + u0 ) + (w + w0 ) 2 U + W
since both U and W are subspaces. If c is a scalar, then
cx = c(u + w) = cu + cw 2 U + W,
again since both U and W are subspaces.
49. We verify all ten axioms of a vector space. Note that both U and V are vector spaces, so they satisfy
all the vector space axioms. These facts will be used without comment below. Let (u1 , v1 ), (u2 , v2 ),
and (u3 , v3 ) be elements of U ⇥ V , and let c and d be scalars. Then
1. (u1 , v1 ) + (u2 , v2 ) = (u1 + u2 , v1 + v2 ) 2 U ⇥ V .
2. (u1 , v1 ) + (u2 , v2 ) = (u1 + u2 , v1 + v2 ) = (u2 + u1 , v2 + v1 ) = (u2 , v2 ) + (u1 , v1 ).
3.
((u1 , v1 ) + (u2 , v2 )) + (u3 , v3 ) = (u1 + u2 , v1 + v2 ) + (u3 , v3 )
= ((u1 + u2 ) + u3 , (v1 + v2 ) + v3 )
= (u1 + (u2 + u3 ), v1 + (v2 + v3 ))
= (u1 , v1 ) + (u2 + u3 , v2 + v3 )
= (u1 , v1 ) + ((u2 , v2 ) + (u3 , v3 )).
4. (0, 0) 2 U ⇥ V , and (u1 , v1 ) + (0, 0) = (u1 + 0, v1 + 0) = (u1 , v1 ).
5. (u1 , v1 ) + ( u1 , v1 ) = (u1
u1 , v1
v1 ) = (0, 0).
6. c(u1 , v1 ) = (cu1 , cv1 ) 2 U ⇥ V .
7.
c ((u1 , v1 ) + (u2 , v2 )) = c(u1 + u2 , v1 + v2 )
= (c(u1 + u2 ), c(v1 + v2 ))
= (cu1 + cu2 , cv1 + cv2 )
= (cu1 , cv1 ) + (cu2 , cv2 )
= c(u1 , v1 ) + c(u2 , v2 ).
8.
(c + d)(u1 , v1 ) = ((c + d)u1 , (c + d)v1 )
= (cu1 + du1 , cv1 + dv1 )
= (cu1 , cv1 ) + (du1 , dv1 )
= c(u1 , v1 ) + d(u1 , v1 ).
9. c (d(u1 , v1 )) = c(du1 , dv1 ) = (c(du1 ), c(du2 )) = ((cd)u1 , (cd)v1 ) = (cd)(u1 , v1 ).
10. 1(u1 , v1 ) = (1u1 , 1v1 ) = (u1 , v1 ).
50. Suppose that (w, w), (w0 , w0 ) 2
. Then
(w, w) + (w0 , w0 ) = (w + w0 , w + w0 ),
which is in
since its two components are the same. Let c be a scalar. Then
c(w, w) = (cw, cw),
which is also in since its components are the same. Therefore
6.2, so is a subspace.
satisfies both conditions of Theorem
6.1. VECTOR SPACES AND SUBSPACES
461
51. If aA + bB = C, then
a
Then a
b = 2 and b

1
1

1
1
+b
1
1

1
a+b
=
0
b a

b
1
=
a
3
a
2
.
4
a = 3, which is impossible. Therefore C is not in the span of A and B.
52. If aA + bB = C, then
a

1
1

1
1
+b
1
1

1
a+b
=
0
b a

b
3
=
a
5
a
Then a = 1 from the lower right. The upper left then gives
and b a = 5, so that A + 4B = C.
5
.
1
1+b = 3, so that b = 4. Then a b =
5
53. Suppose that
a(1
2x) + b(x
x2 ) + c( 2 + 3x + x2 ) = (a
2c) + ( 2a + b + 3c)x + ( b + c)x2 = 3
5x
x2 .
Then
a
2c =
3
2a + b + 3c =
5
b+ c=
1.
The third equation gives b = c + 1; substituting this in the second equation gives 2a + 4c = 6,
which is 2 times the first equation. So there is a family of solutions, which can be parametrized as
a = 2t + 3, b = t + 1, c = t. For example, with t = 0 we get
s(x) = 3p(x) + q(x),
and s(x) 2 span(p(x), q(x), r(x)).
54. Suppose that
a(1
2x) + b(x
x2 ) + c( 2 + 3x + x2 ) = (a
2c) + ( 2a + b + 3c)x + ( b + c)x2 = 1 + x + x2 .
Then
a
2c = 1
2a + b + 3c = 1
b + c = 1.
The third equation gives b = c 1; substituting into the second equation gives 2a + 4c 1 = 1,
or a 2c = 1. This conflicts with the first equation, so the system has no solution. Thus s(x) 2
/
span(p(x), q(x), r(x)).
55. Since sin2 x + cos2 x = 1, we have 1 = f (x) + g(x) 2 span(sin2 x, cos2 x).
56. Since cos 2x = cos2 x
sin2 x = g(x)
f (x), we have cos 2x =
f (x) + g(x) 2 span(sin2 x, cos2 x).
57. Suppose that sin 2x = af (x) + bg(x) = a sin2 x + b cos2 x for all x. Then
Let x = 0; then 0 = a sin2 0 + b cos2 0 = b, so that b = 0.
⇣ ⇡⌘
⇡
⇡
⇡
Let x = ; then sin 2 ·
= sin ⇡ = 0 = a sin2 + b cos2 = a, so that a = 0.
2
2
2
2
Thus the only solution is a = b = 0, but this is not a valid solution since sin 2x 6= 0. So sin 2x 2
/
span(sin2 x, cos2 x).
462
CHAPTER 6. VECTOR SPACES
58. Suppose that sin x = af (x) + bg(x) = a sin2 x + b cos2 x for all x. Then
Let x = 0; then 0 = a sin2 0 + b cos2 0 = b, so that b = 0.
⇡
⇡
⇡
⇡
Let x = ; then sin = 1 = a sin2 + b cos2 = a, so that a = 1.
2
2
2
2
Thus the only possibility is sin x = sin2 x. But this is not a valid equation; for example, if x =
then the left-hand side is 1 while the right-hand side is 1. So sin x 2
/ span(sin2 x, cos2 x).
59. Let

1
V1 =
0
1
,
1

0
V2 =
1
1
,
0

1
V3 =
1
0
,
1
V4 =

0
1
3⇡
2 ,
1
.
0
Now, V4 = V3 V1 , so that span(V1 , V2 , V3 , V4 ) = span(V1 , V2 , V3 ). If span(V1 , V2 , V3 ) = M22 , then
every matrix is in the span, so in particular E11 is. But


a+c a+b
1 0
aV1 + bV2 + cV3 = E11 )
=
.
b+c a+c
0 0
Now the diagonal entries give a + c = 0 and a + c = 1, which has no solution. So E11 is not in the span
and thus span(V1 , V2 , V3 ) 6= M22 .
60. Let
V1 =

1
1
0
,
0
V2 =
Solving the four independent systems

1
1
1
,
0
V3 =

1
1
1
,
1
V4 =

0
1
1
.
0
E11 = aV1 + bV2 + cV3 + dV4
E12 = aV1 + bV2 + cV3 + dV4
E21 = aV1 + bV2 + cV3 + dV4
E22 = aV1 + bV2 + cV3 + dV4
gives
E11 = 2V1
E21 =
V4
E12 =
V 1 + V2
V 1 + V 2 + V4
V2
E22 =
V 2 + V3 .
Therefore span(V1 , V2 , V3 , V4 ) ◆ span(E11 , E12 , E21 , E22 ) = M22 .
61. Let p(x) = 1 + x, q(x) = x + x2 , and r(x) = 1 + x2 . Then
a
+ c = 1
1
= 0 ) a= , b=
1 = ap(x) + bq(x) + cs(x) ) a + b
2
b + c = 0
1
1
, c=
2
2
a
+ c = 0
1
1
= 1 ) a= , b= , c=
x = ap(x) + bq(x) + cs(x) ) a + b
2
2
b + c = 0
a
+ c = 0
= 0 ) a=
x2 = ap(x) + bq(x) + cs(x) ) a + b
b + c = 1
1
2
1
1
1
, b= , c= .
2
2
2
Thus span(p(x), q(x), r(x)) ◆ span(1, x, x2 ) = P2 .
62. Let p(x) = 1 + x + 2x2 , q(x) = 2 + x + 2x2 , and r(x) =
x = ap(x) + bq(x) + cs(x) )
1 + x + 2x2 . Then
a + 2b
c = 0
a + b + c = 1
2a + 2b + 2c = 0 .
But the second pair of equations are inconsistent, so that x 2
/ span(p(x), q(x), r(x)).
span(p(x), q(x), r(x)) 6= P2 .
Therefore
6.2. LINEAR INDEPENDENCE, BASIS, AND DIMENSION
463
63. Suppose 0 and 00 satisfy Axiom 4. Then 0 + 00 = 0 since 00 satisfies Axiom 4, and also 0 + 00 = 00
since 0 satisfies Axiom 4. Putting these two equalities together gives 0 = 00 , so that 0 is unique.
64. Suppose v0 and v00 satisfy Axiom 5 for v. Then using Axiom 5,
v00 = v00 + 0 = v00 + (v + v0 ) = (v00 + v) + v0 = 0 + v0 = v0 .
6.2
Linear Independence, Basis, and Dimension
1. We wish to determine if there are constants c1 , c2 , and c3 , not all zero, such that



1
1
1
1
1 0
c1
+ c2
+ c3
= 0.
0
1
1
0
3 2
This is equivalent to asking for solutions to the linear system
c1 + c2 + c3 = 0
c1
c2
=0
c2 + 3c3 = 0
c1
+ 2c3 = 0.
Row-reducing the corresponding augmented matrix, we get
2
3
2
1
1 1 0
1 0
6 1
7
60 1
1
0
0
6
7!6
4 0
40 0
1 3 05
1
0 2 0
0 0
3
0
07
7.
05
0
0
0
1
0
This system has only the trivial solution c1 = c2 = c3 = 0, so that these matrices are linearly
independent.
2. We wish to determine if there are constants c1 , c2 , and c3 , not all zero, such that



2
3
1
1
1 3
c1
+ c2
+ c3
= 0.
4
2
3
3
1 5
This is equivalent to asking for solutions to the linear system
2c1 + c2
3c1
c3 = 0
c2 + 3c3 = 0
4c1 + 3c2 + c3 = 0
2c1 + 3c2 + 5c3 = 0.
Row-reducing the corresponding augmented matrix, we get
2
3
2
2
1
1 0
1 0
6 3
6
1
3 07
6
7 ! 60 1
4 4
40 0
3
1 05
2
3
5 0
0 0
2
3
0
0
3
0
07
7.
05
0
This system has a nontrivial solution c1 = 2t, c2 = 3t, c3 = t; setting t = 1 for example gives



2
3
1
1
1 3
2
3
+
= 0.
4
2
3
3
1 5
so that the matrices are linearly dependent.
464
CHAPTER 6. VECTOR SPACES
3. We wish to determine if there are constants c1 , c2 , c3 , and c4 , not all zero, such that




1 1
3 0
0 2
1 0
c1
+ c2
+ c3
+ c4
= 0.
2 2
1 1
3 1
1 7
This is equivalent to asking for solutions to the linear system
c1 + 3c2
c1
2c1 + c2
c4 = 0
+ 2c3
=0
3c3
c4 = 0
2c1 + c2 + c3 + 7c4 = 0.
Row-reducing the corresponding
2
1
6 1
6
4 2
2
augmented matrix, we get
3
2
3
0
1 0
1 0
60 1
0
2
0 07
7!6
40 0
1
3
1 05
1
1
7 0
0 0
0
0
1
0
4
1
2
0
3
0
07
7.
05
0
This system has a nontrivial solution c1 = 4t, c2 = t, c3 = 2t, c4 = t; setting t = 1 for example
gives




1 1
3 0
0 2
1 0
= 0.
4
+2
+
2 2
1 1
3 1
1 7
so that the matrices are linearly dependent.
4. We wish to determine if there are constants c1 , c2 , c3 , and c4 , not all zero, such that




1 1
1 0
0 1
1 1
c1
+ c2
+ c3
+ c4
= 0.
0 1
1 1
1 1
1 0
This is equivalent to asking for solutions to the linear system
c1 + c2
c1
+ c4 = 0
+ c3 + c4 = 0
c2 + c3 + c4 = 0
c1 + c2 + c3
Row-reducing the corresponding augmented
2
1 1 0 1
61 0 1 1
6
40 1 1 1
1 1 1 0
matrix, we
3
2
1
0
60
07
7!6
40
05
0
0
= 0.
get
0
1
0
0
0
0
1
0
0
0
0
1
3
0
07
7.
05
0
Since the only solution is the trivial solution, it follows that these four matrices are linearly independent
over M22 .
5. Following, for instance, Example 6.26, we wish to determine if there are constants c1 and c2 , not both
zero, such that
c1 (x) + c2 (1 + x) = 0.
This is equivalent to asking for solutions to the linear system
c2 = 0
c1 + c2 = 0.
By forward substitution, we see that the only solution is the trivial solution, c1 = c2 = 0. It follows
that these polynomials are linearly independent over P2 .
6.2. LINEAR INDEPENDENCE, BASIS, AND DIMENSION
465
6. Following, for instance, Example 6.26, we wish to determine if there are constants c1 , c2 , and c3 , not
all zero, such that
c1 (1 + x) + c2 (1 + x2 ) + c3 (1
x + x2 ) = (c1 + c2 + c3 ) + (c1
c3 )x + (c2 + c3 )x2 = 0.
This is equivalent to asking for solutions to the linear system
c1 + c2 + c3 = 0
c1
c3 = 0
c2 + c3 = 0.
Row-reducing the corresponding augmented matrix, we get
2
3
2
1 1
1 0
1 0 0
41 0
1 05 ! 40 1 0
0 1
1 0
0 0 1
3
0
05 .
0
Since the only solution is the trivial solution, it follows that these three polynomials are linearly
independent over P2 .
7. Following, for instance, Example 6.26, we wish to determine if there are constants c1 , c2 , and c3 , not
all zero, such that
c1 (x) + c2 (2x
x2 ) + c3 (3x + 2x2 ) = (c1 + 2c2 + 3c3 )x + ( c2 + 2c3 )x2 = 0.
This is a homogeneous system of two equations in three unknowns, so we know it has a nontrivial
solution and these polynomials are linearly dependent. To find an expression showing the linear
dependence, we must find solutions to the linear system
c1 + 2c2 + 3c3 = 0
c2 + 2c3 = 0.
Row-reducing the corresponding augmented matrix, we get


1
2 3 0
1 0
!
0
1 2 0
0 1
The general solution is c1 =
7
2
0
.
0
7t, c2 = 2t, c3 = t. Setting t = 1, for example, we get
x2 ) + (3x + 2x2 ) = 0.
7(x) + 2(2x
8. Following, for instance, Example 6.26, we wish to determine if there are constants c1 , c2 , c3 , and c4 ,
not all zero, such that
c1 (2x)+c2 (x x2 )+c3 (1+x3 )+c4 (2 x2 +x3 ) = (c3 +2c4 )+(2c1 +c2 )x+( c2 c4 )x2 +(c3 +c4 )x3 = 0.
This is equivalent to asking for solutions to the linear system
c3 + 2c4 = 0
2c1 + c2
=0
c2
c4 = 0
c3 + c4 = 0.
Row-reducing the corresponding augmented matrix, we get
2
3
2
0
0 1
2 0
1 0
62
7
60 1
1
0
0
0
6
7!6
40
40 0
1 0
1 05
0
0 1
1 0
0 0
0
0
1
0
0
0
0
1
3
0
07
7.
05
0
Since the only solution is the trivial solution, it follows that these four polynomials are linearly independent over P3 .
466
CHAPTER 6. VECTOR SPACES
9. Following, for instance, Example 6.26, we wish to determine if there are constants c1 , c2 , c3 , and c4 ,
not all zero, such that
c1 (1
2x) + c2 (3x + x2
x3 ) + c3 (1 + x2 + 2x3 ) + c4 (3 + 2x + 3x3 )
= (c1 + c3 + 3c4 ) + ( 2c1 + 3c2 + 2c4 )x + (c2 + c3 )x2 + ( c2 + 2c3 + 3c4 )x3 = 0.
This is equivalent to asking for solutions to the linear system
c1
+ c3 + 3c4 = 0
2c1 + 3c2
+ 2c4 = 0
c2 + c3
=0
c2 + 2c3 + 3c4 = 0.
Row-reducing the corresponding augmented matrix, we get
2
3
2
1
0 1 3 0
1 0
7
60 1
6 2
3
0
2
0
6
7!6
4 0
40 0
1 1 0 05
0
1 2 3 0
0 0
0
0
1
0
0
0
0
1
3
0
07
7.
05
0
Since the only solution is the trivial solution, it follows that these four polynomials are linearly independent over P3 .
10. We wish to find solutions to a + b sin x + c cos x = 0 that hold for all x. Setting x = 0 gives a + c = 0,
so that a = c; setting x = ⇡ gives a c = 0, so that a = c. This forces a = c = 0; then setting x = ⇡2
gives b = 0. So the only solution is the trivial solution, and these functions are linearly independent in
F.
11. Since sin2 x + cos2 x = 1, the functions are linearly dependent.
12. Suppose that aex + be x = 0 for all x. Setting x = 0 gives ae0 + be0 = a + b = 0, while setting x = ln 2
gives 2a + 12 b = 0. Since the coefficient matrix,
"
#
1 1
,
2 12
has nonzero determinant, the only solution to the system is the trivial solution, so that these functions
are linearly independent in F .
13. Using properties of the logarithm, we have ln(2x) = ln 2 + ln x and ln(x2 ) = 2 ln x, so we want to see
if {1, ln 2 + ln x, 2 ln x} are linearly independent over F . But clearly they are not, since by inspection
( 2 ln 2) · 1 + 2 · (ln 2 + ln x)
2 ln x = 0.
So these functions are linearly dependent in F .
14. We wish to find solutions to a sin x + b sin 2x + c sin 3x = 0 that hold for all x. Setting
⇡
2
⇡
x=
6
⇡
x=
3
x=
)
a
c=0
p
1
3
)
a+
b=0
2p
2p
3
3
)
a
b = 0.
2
2
⇣
p ⌘
Adding the second and third equations gives 12 + 23 a = 0, so that a = 0; then the first and third
equations force b = c = 0. So the only solution is the trivial solution, and thus these functions are
linearly independent in F .
6.2. LINEAR INDEPENDENCE, BASIS, AND DIMENSION
467
15. We want to prove that if the Wronskian is not identically zero, then f and g are linearly independent.
We do this by proving the contrapositive: assume f and g are linearly dependent. If either f or g is
zero, then clearly the Wronskian is zero as well. So assume that neither f nor g is zero; since they are
linearly dependent, each is a nonzero multiple of the other. Say f (x) = ag(x), so that f 0 (x) = ag 0 (x).
Now, the Wronskian is
f (x)
f 0 (x)
g(x)
= f (x)g 0 (x)
g 0 (x)
g(x)f 0 (x) = ag(x)g 0 (x)
g(x)(ag 0 (x)) = ag(x)g 0 (x)
ag(x)g 0 (x) = 0.
Thus the Wronskian is identically zero, proving the result.
1 sin x
cos x
cos x
sin x
sin x =
W (x) = 0 cos x
= cos2 x sin2 x =
sin x
cos x
0
sin x
cos x
the Wronskian is not zero, the functions are linearly independent.
Exercise 11:
1
sin2 x
cos2 x
2 sin x cos x
2 sin x cos x
W (x) = 0
0 2(cos2 x sin2 x) 2(sin2 x cos2 x)
16. Exercise 10:
1
= 0
0
=
sin2 x
sin 2x
2 cos 2x
sin 2x
2 cos 2x
1 6= 0. Since
cos2 x
sin 2x
2 cos 2x
sin 2x
2 cos 2x
= 0.
Since the Wronskian is zero, the functions are linearly dependent.
ex
e x
Exercise 12: W (x) = x
= 1 1 = 2 6= 0. Since the Wronskian is not zero, the
e
e x
functions are linearly independent.
Exercise 13:
1 ln(2x) ln(x2 )
1
2
1
2
x
x
W (x) = 0
=
= 0.
1
2
x
x
1
2
x2
x2
0
x2
x2
Since the Wronskian is zero, the functions are linearly dependent.
sin x
sin 2x
sin 3x
2 cos 2x
3 cos 3x . Rather than computing the determinant and
Exercise 14: W (x) = cos x
sin x
4 sin 2x
9 sin 3x
showing it is not identically zero, it suffices to find some value of x such that W (x) 6= 0. Let
x = ⇡3 ; then
W
⇣⇡⌘
3
=
=
sin ⇡3
sin 2⇡
3
sin ⇡
cos ⇡3
sin ⇡3
p
3
2
1
2p
3
2
p
3
2p
3
2
2 cos 2⇡
3
4 sin 2⇡
3
p
3
0
2
3 cos ⇡
1
p
2 3
9 sin ⇡
3
0
p
3
2
p
2 3
✓
◆
3
27
=3
3+
=
6 0.
=
4
4
=3
468
CHAPTER 6. VECTOR SPACES
So the Wronskian is not identically zero and therefore the functions are linearly independent.
17. (a) Yes, it is. Suppose
a(u + v) + b(v + w) + c(u + w) = (a + c)u + (a + b)v + (b + c)w = 0.
Since {u, v, w} are linearly independent, it follows that a + c = 0, a + b = 0, and b + c = 0.
Subtracting the latter two from each other gives a c = 0; combining this with the first equation
shows that a = c = 0, and then b = 0. So the only solution to the system is the trivial one, so
these three vectors are linearly independent.
(b) No, they are not, since
(u
v) + (v
w)
(u
w) = 0.
18. Since dim M22 = 4 but B has only three elements, it cannot be a basis.
19. If
a

1
0

0
0
+b
1
1

1
1
+c
0
1

1
1
+d
1
1

1
0
=
1
0
0
,
0
then
a
+c+d= 0
b+c+d= 0
b+c+d= 0
a
+c
d = 0.
Subtracting the second and third equations shows that b = 0; subtracting the first and fourth shows
that d = 0. Then from the second equation, c = 0 as well, which forces a = 0. So the only solution
is the trivial solution, and these matrices are linearly independent. Since there are four of them, and
dim M22 = 4, it follows that they form a basis for M22 .
20. Since dim M22 = 4, if these vectors are linearly independent then they span and thus form a basis. To
see if they are linearly independent, we want to solve




1 0
0 1
1 1
1 0
c1
+ c2
+ c3
+ c4
= 0.
0 1
1 0
0 1
1 1
Solving this linear system gives c1 = 2c4 , c2 = c4 , and c3 = c4 . For example, letting c4 = 1 gives




1 0
0 1
1 1
1 0
2
+
+
= 0.
0 1
1 0
0 1
1 1
Thus the matrices are linearly dependent so do not form a basis (see Theorem 6.10).
21. Since dim M22 = 4 but B has five elements, it cannot be a basis.
22. Since dim P2 = 3, if these vectors are linearly independent then they span and thus form a basis. To
see if they are linearly independent, we want to solve
c1 x + c2 (1 + x) + c3 (x
x2 ) = c2 + (c1 + c2 + c3 )x
c3 x2 = 0.
Rather than setting up the augmented matrix and row-reducing it, simply note that the first and third
terms force c2 = c3 = 0; then the second term forces c1 = 0. So the only solution is the trivial solution,
and these three polynomials are linearly independent, so they form a basis for P2 .
6.2. LINEAR INDEPENDENCE, BASIS, AND DIMENSION
469
23. Since dim P2 = 3, if these vectors are linearly independent then they span and thus form a basis. To
see if they are linearly independent, we want to solve
c1 (1
x) + c2 (1
x2 ) + c3 (x
x2 ) = (c1 + c2 ) + ( c1 + c3 )x + ( c2
To solve the resulting linear system, we row-reduce its augmented
2
3
2
1
1
0 0
1 0
1
4 1
0
1 05 ! 4 0 1
1
0
1
1 0
0 0
0
This has general solution c1 = t, c2 =
(1
c3 )x2 = 0.
matrix:
3
0
05 .
0
t, c3 = t, so setting t = 1 we get
x)
x2 ) + (x
(1
x2 ) = 0.
Therefore the vectors are not linearly independent, so they cannot form a basis for P2 , which has
dimension 3.
24. Since dim P2 = 3 but there are only two vectors in B, they cannot form a basis.
25. Since dim P2 = 3 but there are four vectors in B, they must be linearly dependent, so cannot form a
basis.
26. We want to find constants c1 , c2 , c3 , and c4 such that
c1 E22 + c2 E21 + c3 E12 + c4 E11 = A =

1
3
2
.
4
Setting up the equivalent linear system gives
c1
=4
c2
=3
c3
=2
c4 = 1.
This system obviously has the solution c1 = 4, c2 = 3, c3 = 2, c4 = 1, so the coordinate vector is
2 3
4
637
7
[A]B = 6
425 .
1
27. We want to find constants c1 , c2 , c3 , and c4 such that




1 0
1 1
1 1
1
c1
+ c2
+ c3
+ c4
0 0
0 0
1 0
1

1
1
=
1
3
2
.
4
1, c1 =
1, so the coordinate vector
Setting up the equivalent linear system gives
c1 + c2 + c3 + c4 = 1
c2 + c3 + c4 = 2
c3 + c4 = 3
c4 = 4.
By backwards substitution, the solution is c4 = 4, c3 = 1, c2 =
is
2 3
1
6 17
7
[A]B = 6
4 15 .
4
470
CHAPTER 6. VECTOR SPACES
28. We want to find constants c1 , c2 , and c3 such that
p(x) = 1 + 2x + 3x2 = c1 (1 + x) + c2 (1
x) + c3 (x2 ) = (c1 + c2 ) + (c1
c2 )x + c3 x2 .
The corresponding linear system is
c1 + c2
c1
=1
c2
=2
c3 = 3.
Row-reducing the corresponding augmented
2
1
1 0
6
61
1 0
4
0
0 1
so that c1 = 32 , c2 =
1
2,
matrix gives
3
2
1
1 0
7
6
7
6
25 ! 4 0 1
3
0
0
0
0
1
3
3
27
17 ,
25
3
and c3 = 3. Thus the coordinate vector is
2 3
6
[p]B = 4
3
2
17
.
25
3
29. We want to find constants c1 , c2 , and c3 such that
p(x) = 2
x + 3x2 = c1 (1) + c2 (1 + x) + c3 ( 1 + x2 ) = (c1 + c2
Equating coefficients shows that c3 = 3 and c2 =
gives c1 = 6, so that the coordinate vector is
c 3 ) + c 2 x + c 3 x2 .
1; substituting into the equation c1 + c2
c3 = 2
2
3
6
[p]B = 4 15 .
3
30. To show B is a basis, we must show that it spans V and that the elements of B are linearly independent.
B spans V , since if v 2 V , then it can be written as a linear
P combination of vectors in B by hypothesis.
To show they are linearly independent, note that 0 = P 0ui ; since the representation of 0 is unique
by assumption, it follows that the only solution to 0 =
ci ui is the trivial solution, so that B = {ui }
is a linearly independent set. Thus B is a basis for V .
31. [c1 u1 + · · · + ck uk ]B = [c1 u1 ]B + · · · + [ck uk ]B = c1 [u1 ]B + · · · + ck [uk ]B .
32. Suppose c1 u1 + · · · + cn un = 0. Then clearly
0 = [c1 u1 + · · · + cn un ]B .
But by Exercise 31, the latter expression is equal to
c1 [u1 ]B + · · · + cn [un ]B ;
since [ui ]B are linearly independent in Rn , we get ci = 0 for all i, so that the only solution to the
original equation is the trivial solution. Thus the ui are linearly independent in V .
33. First suppose that span(u1 , . . . , um ) = V . Choose
2 3
c1
6 c2 7
6 7
x = 6 . 7 2 Rn .
4 .. 5
cn
6.2. LINEAR INDEPENDENCE, BASIS, AND DIMENSION
Then
P
471
ci ui 2 V since V = span(ui ). Therefore by Exercise 31
x = [c1 u1 + · · · + cm um ]B = c1 [u1 ]B + · · · + cm [um ]B ,
which shows that x 2 span(S).
For the converse, suppose that span(S) = Rn . Choose v 2 V ; then [v]B 2 Rn , and using Exercise 31
again,
hX
i
X
[v]B =
ci [ui ]B =
ci ui .
B
P
But since B is a basis, and the two elements
v
and
c
u
of
V
have
the same coordinate vector, they
i i
P
must be the same element. Thus v =
ci ui , so that {ui } spans V .
34. Using the standard basis {x2 , x, 1} for P2 , if p(x) = ax2 + bx + c 2 V , then p(0) = c = 0, so that c = 0
and therefore p(x) = ax2 + bx. Since x2 and x are linearly independent and span V , {x, x2 } is a basis,
so that dim V = 2.
35. Two polynomials in V are p(x) = 1 x and q(x) = 1 x2 , since p(1) = q(1) = 0. If we show these
polynomials are linearly independent and span V , we are done. They are linearly independent, since if
a(1
x) + b(1
x2 ) = (a + b)
ax
bx2 = 0,
then equating coefficients shows that a = b = 0. To see that they span, suppose r(x) = c+bx+ax2 2 V .
Then r(1) = a + b + c = 0, so that c = b a and therefore
r(x) =
(b + a) + bx + ax2 =
b(1
x)
a(1
x2 ),
so that r is a linear combination of p and q. Thus {p(x), q(x)} forms a basis for V , which has dimension
2.
36. Suppose p(x) = c + bx + ax2 ; then xp0 (x) = bx + 2ax2 . So p(x) = xp0 (x) means that c = 0 and a = 2a,
so that a = 0. Thus p(x) = bx. Then clearly V = span(x), so that {x} forms a basis, and dim V = 1.
37. Upper triangular matrices in M22 have the form

a b
= aE11 + bE12 + dE22 .
0 d
Since E11 , E12 , and E22 span V , and we know they are linearly independent, it follows that they form
a basis for V , which therefore has dimension 3.
38. Skew-symmetric matrices in M22 have the form


0 b
0
=b
b 0
1
1
= bA.
0
Since A spans V , and {A} is a linearly independent set, we see that {A} is a basis for V , which therefore
has dimension 1.
39. Suppose that
A=
Then

a
AB =
c
a+b
,
c+d

a
c
b
.
d
BA =

a+c
c
b+d
.
d
Then AB = BA means that a = a + c, so that c = 0, and also a + b = b + d, so that a = d. So the
matrix A has the form


a b
1 0
A=
=a
+ bE12 = aI + bE12 .
0 a
0 1
I and E12 are linearly independent since I = E11 + E22 , and {E11 , E22 , E12 } is a linearly independent
set. Since they span, we see that {I, E12 } forms a basis for V , which therefore has dimension 2.
472
CHAPTER 6. VECTOR SPACES
40. dim Mnn = n2 . For symmetric matrices, we have aij = aji . Therefore specifying the entries on or
above the diagonal determines the matrix, and these entries may take any value. Since there are
n + (n
1) + · · · + 2 + 1 =
n(n + 1)
2
entries on or above the diagonal, this is the dimension of the space of symmetric matrices.
41. dim Mnn = n2 . For skew-symmetric matrices, we have aij = aji . Therefore all the diagonal entries
are zero; further, specifying the entries above the diagonal determines the matrix, and these entries
may take any value. Since there are
(n
1) + · · · + 2 + 1 =
n(n
1)
2
entries above the diagonal, this is the dimension of the space of symmetric matrices.
42. Following the hint, Let B = {v1 , . . . , vk } be a basis for U \ W . By Theorem 6.10(e), this can be
extended to a basis C for U and to a basis D for W . Consider the set C [ D. Clearly it spans U + W ,
since an element of U + W is u + w for u 2 U , w 2 W ; since C and D are bases for U and W , the result
follows. Next we must show that this set is linearly independent. Suppose dim U = m and dim V = n;
then
C = {uk+1 , . . . , um } [ B,
D = {wk+1 , . . . , wn } [ B.
Suppose that a linear combination of C [ D is the zero vector. This means that
a1 v1 + · · · + ak vk + ck+1 uk+1 + · · · + cm um + dk+1 wk+1 + dn wn = A + C + D = 0.
This means that C = A D. Since A 2 U \ W and D 2 W , we see that
C 2 W . But also C 2 U , since it is a combination of basis vectors for U . Thus
impossible unless all of the ci are zero, since the uk+1 are elements of U not in
We are left with A + D = 0; this is a linear combination of basis elements of W ,
all the dj must be zero as well. Therefore C [ D is linearly independent.
A D 2 W so that
C 2 U \ W , which is
the subspace U \ W .
so that all the ai and
It follows that C [ D forms a basis for U + W , so it remains to count the elements in this set. Clearly
C \ D = B, for any basis element in common between C and D must be in U \ W , so it must be in B.
Write #X for the number of elements in the set X. Then
dim(U + W ) = #(C [ D) = #C + #D
#B = n + m
k = dim U + dim W
dim(U \ W ).
43. Let dim U = m, with basis B = {u1 , . . . , um }, and dim V = n, with basis C = {v1 , . . . , vn }.
(a) Claim that a basis for U ⇥ V is D = {(ui , 0), (0, vj )}. This will also show that dim(U ⇥ V ) = mn,
by counting the basis vectors. First we want to show that D spans. Let x = (u, v) 2 U ⇥ V .
Then
0
1
m
n
m
n
X
X
X
X
x = (u, v) = @
ci ui ,
dj v j A =
ci (ui , 0) +
dj (0, vj ),
i=1
j=1
i=1
j=1
proving that D spans U ⇥ V . To see that D is linearly independent, suppose that 0 is a linear
combination of the elements of D. This means that
0
1
m
n
m
n
X
X
X
X
0 = (0, 0) =
ci (ui , 0) +
dj (0, vj ) = @
ci ui ,
dj v j A .
i=1
j=1
i=1
j=1
Pm
But the ui are linearly independent, so that P
i=1 ci ui = 0 means that all of the ci are zero;
n
similarly the vj are linearly independent, so that j=1 dj vj = 0 means that all of the dj are zero.
Thus D is linearly independent.
So D is a basis for U ⇥ W , and therefore dim(U ⇥ W ) = mn.
6.2. LINEAR INDEPENDENCE, BASIS, AND DIMENSION
473
(b) Let dim W = k, with basis {w1 , . . . , wk }. Claim that D = {(wi , wi )} is a basis for ; proving
this claim also shows, by countingPthe basis vectors, that dim = dim W . To see that D spans
, choose (w, w) 2 ; then w =
ci wi , so that
⇣X
⌘ X
X
X
(w, w) =
c i wi ,
c i wi =
(ci wi , ci wi ) =
ci (wi , wi ).
Thus D spans
P
. To see that D is linearly independent, suppose that
ci (wi , wi ) = 0. Then
⇣X
⌘
X
X
X
0=
ci (wi , wi ) =
(ci wi , ci wi ) =
ci wi ,
c i wi .
P
This means that
ci wi = 0; since the wi are linearly independent, we see that all the ci are
zero. So D is linearly independent, showing that it is a basis for and thus that dim = dim W .
44. Suppose that P is finite dimensional, and that {p1 (x), . . . , pn (x)} is a basis. Let m be the largest power
of x that appears in any of the basis elements, and consider q(x) = xm+1 . Since any linear combination
of the pi has powers of x at most equal to m, it follows that q 2
/ span{pi }, so that P cannot have a
finite basis.
45. Start by adding all three standard basis vectors for P2 to the set, yielding {1 + x, 1 + x + x2 , 1, x, x2 }.
Since (1 + x) + x2 = 1 + x + x2 , we see that {1 + x, 1 + x + x2 , x2 } is linearly dependent, so discard x2 ,
leaving {1 + x, 1 + x + x2 , 1, x}. Next, since (1 + x) = 1 + x, the set {1 + x, 1, x} is linearly dependent,
so discard x, leaving B = {1 + x, 1 + x + x2 , 1}. This is a set of three vectors, so to prove that it is a
basis, it suffices to show that it spans. But by construction, the standard basis vectors not in B are
linear combinations of elements of B. Hence B is a basis for P2 .
46. Start by adding the standard basis vectors Eij to the set, giving
⇢


0 1
1 1
B= C=
,D=
, E11 , E12 , E21 , E22 .
0 1
0 1
Since E11 = D C is the di↵erence of the two given matrices, we discard it from B, leaving us with
five elements. Next, C = E12 + E22 , so discard E22 , leaving us with B = {C, D, E12 , E21 }. This has
four elements, and dim M22 = 4, so to prove that this is a basis, it suffices to prove that it spans. But
by construction, the standard basis vectors not in B are linear combinations of elements of B. So B
spans and therefore forms a basis for M22 .
47. Start by adding the standard basis vectors Eij to the set, giving
⇢



1 0
0 1
0
1
B= I=
,C=
,D=
, E11 , E12 , E21 , E22 .
0 1
1 0
1
0
Since E21 = 12 (C + D), we discard it from B. Next, I = E11 + E22 , so discard E22 from B. Finally,
E12 = 12 (C D), so we discard it from B. This leaves us with B = {I, C, D, E11 }. This has four
elements, and dim M22 = 4, so to prove that this is a basis, it suffices to prove that it spans. But by
construction, the standard basis vectors not in B are linear combinations of elements of B. So B spans
and therefore forms a basis for M22 .
48. Start by adding the standard basis vectors Eij to the set, giving
⇢


1 0
0 1
B= I=
,C=
, E11 , E12 , E21 , E22 .
0 1
1 0
Since I + C = E11 + E12 + E21 + E22 , these are a linearly dependent set; discard E22 from B. Next,
C = E12 + E21 , so discard E21 from B. This leaves us with B = {I, C, E11 , E12 }. This has four
elements, and dim M22 = 4, so to prove that this is a basis, it suffices to prove that it spans. But by
construction, the standard basis vectors not in B are linear combinations of elements of B. So B spans
and therefore forms a basis for M22 .
474
CHAPTER 6. VECTOR SPACES
49. The standard basis for P1 is {1, x}. If we show that each of these vectors is in the span of the given
vectors, then those vectors span P1 . But 1 2 span(1, 1 + x, 2x), and x = (1 + x) 1 2 span(1, 1 + x, 2x).
Thus
span(1, 1 + x, 2x) ◆ span(1, x) = P1 ,
so that a basis for the span is {1, x}.
50. Since (1 2x) + (2x x2 ) = 1 x2 , these three elements are linearly independent, so we can discard
1 x2 from the set without changing the span, leaving {1 2x, 2x x2 , 1 + x2 }. Claim these are
linearly independent. Suppose that
a(1
2x) + b(2x
x2 ) + c(1 + x2 ) = (a + c) + ( 2a + 2b)x + ( b + c)x2 = 0.
Then a = c and a = b from the constant term and the coefficient of x, so that b = c. But then
b = c together with b + c = 0 shows that b = c = 0, so that a = 0. Hence these three polynomials
are linearly independent, so they form a basis for the span of the original set.
51. Since (1 x) + (x x2 ) = 1 x2 and (1 x) (x x2 ) = 1 2x + x2 , we can remove both 1 x2 and
1 2x + x2 from the set without changing the span. This leaves us with {1 x, x x2 }. Claim these
are linearly independent. Suppose that
a(1
x) + b(x
x2 ) = a + ( a + b)x
bx2 = 0.
Then clearly a = b = 0. Hence these two polynomials are linearly independent, so they form a basis
for the span of the original set.
52. Let

1
V1 =
0
0
,
1

0
V2 =
1
1
,
0
V3 =

1
1
1
,
1
V4 =

1
1
1
.
1
Then V3 + V4 = 0, so we can discard V4 from the set without changing the span. Next, V3 = V2 V1 ,
so we can discard V3 from the set without changing the span. So we are left with {V1 , V2 }; these are
linearly independent since they are not scalar multiples of one another. So this set forms a basis for
the span of the given set.
53. Since cos 2x = cos2 x sin2 x, we can discard cos 2x from the set without changing the span, leaving us
with {sin2 x, cos2 x}. Claim that this set is linearly independent. For suppose that a sin2 x+b cos2 x = 0
holds for all x. Setting x = 0 gives b = 0, while setting x = ⇡2 gives a = 0. Thus a = b = 0 and these
functions are linearly independent. So {sin2 x, cos2 x} forms a basis for the span of the original set.
54. Suppose that av + c1 v1 + · · · + cn vn = 0, so that c1 v1 + · · · + cn vn = av. If a 6= 0, then dividing
through by a shows that v 2 span(S), a contradiction. Therefore we must have a = 0, and we are
left with
c1 v1 + · · · + cn vn = 0,
which implies (since the vi are linearly independent) that all of the ci are zero. Therefore all coefficients
must be zero, so that S 0 is linearly independent.
55. If vn 2 span(v1 , . . . , vn
1 ),
then there are constants such that
v n = c1 v1 + · · · + cn
1 vn 1 .
Now choose x 2 V = span(S). Then
x = a1 v 1 + · · · + an
= a1 v 1 + · · · + an
1 vn 1
+ an v n
1 vn
+ an (c1 v1 + · · · + cn
1
= (a1 + an c1 )v1 + · · · + (an
1
+ an cn
1 vn 1 )
1 )vn 1 ,
showing that x 2 span(S 0 ). Then V ✓ span(S 0 ) ✓ span(S) = V , so that span(S 0 ) = V .
6.2. LINEAR INDEPENDENCE, BASIS, AND DIMENSION
475
56. Use induction on n. If S = {v1 }, then S is linearly independent so is a basis. Next, suppose the
statement is true for all sets that span V and have no more than n 1 elements. Let S = {v1 , . . . , vn }.
If S is linearly independent, then it is a basis and we are done. Otherwise it is a linearly dependent
set, so we can write some element of S (say, vn after a possible reordering) as a linear combination of
the other, so that vn 2 span{v1 , · · · , vn 1 }. Then by Exercise 55, we can remove vn from S, giving
S 0 = {v1 , . . . , vn 1 } that also spans V . By the inductive assumption, S 0 can be made into a basis for
V by removing some (possibly no) elements, so S can.
57. Suppose x 2 V . Since S = {v1 , . . . , vn } is a basis, there are scalars a1 , . . . , an such that
x = a1 v 1 + a2 v 2 + · · · + an v n .
Since the ci are nonzero, we also have
a1
a2
an
x = (c1 v1 ) + v2 + · · · +
(cn vn ),
c1
c2
cn
so that T = {c1 v1 , . . . , cn vn } spans V . To show T is linearly independent, the argument is much the
same. Suppose that
b1 (c1 v1 ) + b2 (c2 v2 ) + · · · + bn (cn vn ) = (b1 c1 )v1 + (b2 c2 )v2 + · · · + (bn cn )vn = v0 .
Since S is linearly independent, it follows that bi ci = 0 for all i. Since all of the ci are nonzero, all of
the bi must be zero, so that T is linearly independent. Therefore T is also a basis for V .
58. Let
S = {v1 , v2 , . . . , vn },
T = {v1 , v1 + v2 , v1 + v2 + v3 , . . . , v1 + v2 + · · · + vn }.
By Exercise 21 in Section 2.3, to show that span(T ) = span(S) = V it suffices to show that every
element of S is in span(T ) and vice versa. But
vi = (v1 + v2 + · · · + vi
1
+ vi )
(v1 + v2 + · · · + vi
v1 + · · · + vi = (v1 ) + (v2 ) + · · · + (vi ) 2 span(S).
1)
2 span(T ) and
Thus the two spans are equal. Since S is a basis, and T has the same number of elements as S, it must
also be a basis, by Theorem 6.10(d).
59. (a)
(x
(a0
(x
p1 (x) =
(a1
(x
p2 (x) =
(a2
p0 (x) =
a1 )(x a2 )
(x
=
a1 )(a0 a2 )
(1
a0 )(x a2 )
(x
=
a0 )(a1 a2 )
(2
a0 )(x a1 )
(x
=
a0 )(a2 a1 )
(3
2)(x
2)(1
1)(x
1)(2
1)(x
1)(3
3)
x2
=
3)
3)
x2
=
3)
2)
x2
=
2)
5x + 6
1
5
= x2
x+3
2
2
2
4x + 3
= x2 + 4x 3
1
3x + 2
1
3
= x2
x + 1.
2
2
2
(b) If i = j, then substituting aj for x in the definition of pi gives a numerator equal to the denominator, so that pi (aj ) = 1 for i = j. If i 6= j, however, then substituting aj for x in the definition
of pi gives zero, since the term x aj in the numerator of pi becomes zero. Thus
(
0 if i 6= j
pi (aj ) =
1 if i = j.
60. (a) Note first that the degree of each pi is n, since there are n factors in the numerator, so that
pi 2 Pn for all i. Suppose that there are constants ci such that
c0 p0 (x) + c1 p1 (x) + · · · + cn pn (x) = 0.
Then this equation must hold for all x. Substitute ai for x. Then by Exercise 59(b), all pj (ai )
for j 6= i vanish, leaving only ci pi (ai ) = ci . Therefore ci = 0. Since this holds for every i, we see
that all of the ci are zero, so that the pi are linearly independent.
476
CHAPTER 6. VECTOR SPACES
(b) Since dim Pn = n + 1 and there are n + 1 polynomials p0 , p1 , . . . , pn that span, they must form
a basis, by Theorem 6.10(d).
61. (a) This is similar to part (a) of Exercise 60. Suppose that q(x) 2 P, and that
q(x) = c0 p0 (x) + c1 p1 (x) + · · · + cn pn (x).
Then substituting ai for x gives q(ai ) = ci pi (ai ) = ci , since pj (ai ) = 0 for j 6= i by Exercise
59(b). Since the pi form a basis for P, any element of P has a unique representation as a linear
combination of the pi , so that unique representation must be the one we just found:
q(x) = q(a0 )p0 (x) + q(a1 )p1 (x) + · · · + q(an )pn (x).
(b) q(x) passes through the given n + 1 points if and only if q(ai ) = c1 . But this is exactly what we
proved in part (a). Since the representation is unique in part (a), this is the only such degree n
polynomial.
(c) (i) Here a0 = 1, a1 = 2, a2 = 3, so we use the Lagrange polynomials determined in Exercise
59(a) and therefore
q(x) = c0 p0 (x) + c1 p1 (x) + c2 p2 (x)
✓
◆
1 2 5
=6
x
x+3
x2 + 4x
2
2
= 3x2
(ii) Here a0 =
16x + 19.
3
2
✓
1 2
x
2
3
x+1
2
◆
1, a1 = 0, a2 = 3. We first compute the Lagrange polynomials:
(x
(a0
(x
p1 (x) =
(a1
(x
p2 (x) =
(a2
a1 )(x a2 )
(x
=
a1 )(a0 a2 )
( 1
a0 )(x a2 )
(x
=
a0 )(a1 a2 )
(0
a0 )(x a1 )
(x
=
a0 )(a2 a1 )
(3
p0 (x) =
(
(
(
(
0)(x 3)
x2 3x
1
3
=
= x2
x
0)( 1 3)
4
4
4
1))(x 3)
x2 2x 3
1 2 2
=
=
x + x+1
1))(0 3)
3
3
3
2
1))(x 0)
x +x
1 2
1
=
=
x + x.
1))(3 0)
12
12
12
Using these polynomials, we get
q(x) = c0 p0 (x) + c1 p1 (x) + c2 p2 (x)
✓
◆
✓
◆
✓
◆
1 2 3
1 2 2
1 2
1
= 10
x
x +5
x + x+1 +2
x + x
4
4
3
3
12
12
✓
◆
✓
◆
5 5 1
15
10
1
=
+
x2 +
+
+
x+5
2 3 6
2
3
6
= x2
4x + 5.
62. Suppose that (a0 , 0), (a1 , 0), . . . , (an , 0) be the n + 1 distinct zeros. Then by Exercise 61(a), the only
polynomial passing through these n + 1 points is
q(x) = 0p0 (x) + 0p1 (x) + · · · + 0pn (x) = 0,
so it is the zero polynomial.
63. An n ⇥ n matrix is invertible if and only if it row-reduces to the identity matrix, so that its columns
are linearly independent. Since n linearly independent columns in Znp form a basis for Znp , finding the
number of invertible matrices in Mnn (Zp ) is the same as finding the number of ways to construct a
basis for Znp .
To construct a basis, we can start by choosing any nonzero vector v1 2 Znp . There are pn 1 ways
to choose this vector (since Znp has pn elements, and we are excluding only one). For the second basis
Exploration: Magic Squares
477
element v2 , we can choose any vector not in span(v1 ) = {a1 v1 : a1 2 Zp }. There are therefore pn p
choices for v2 . Likewise, suppose we have chosen v1 , . . . , vk 1 . Then we can choose for vk any vector
not in the span of the previous vectors, so any vector not in
span(v1 , . . . , vk
There are pn
pk
1
1)
= {a1 v1 + a2 v2 + · · · + ak
1 vk 1
: a 1 , a2 , . . . , a k
1
2 Zp }.
ways of choosing vk .
Thus the total number of bases is (pn
1)(pn
p)(pn
p2 )(· · · )(pn
pn
1
).
Exploration: Magic Squares
1. A classical magic square M contains each of the numbers 1, 2, . . . , n2 exactly once. Since wt(M ) is the
common sum of any row (or column), and there are n rows, we have, using the comments preceding
Exercise 51 in Section 2.4)
wt(M ) =
1
1 n2 (n2 + 1)
n(n2 + 1)
(1 + 2 + · · · + n2 ) = ·
=
.
n
n
2
2
2
2. A classical 3 ⇥ 3 magic square will have common sum 3(3 2+1) = 15. Therefore that no two of 7, 8,
and 9 can be in the same row, column, or diagonal, since the sum of any two of these numbers is at
least 15. The central square cannot be 1, since then 1 + 2 + k = 15 is impossible. Similarly, it cannot
be any of 2 through 4, since then some row, column, or diagonal contains the numbers 1 and 2, which
is impossible. The central square cannot be 7, 8, or 9, since then some line will contain two of those
numbers. Finally, it cannot be 6, since then some line will contain both 6 and 9. So the central square
must be 5. Also, 7 and 1 cannot be in the same line, since the third number would have to be 7.
Finally, 9 cannot be in a corner, since it would then participate in three lines adding to 15. But there
are only two possible sets of numbers, one of which is 9, adding to 15: 9, 5, 1 and 9, 4, and 2. Thus 9
must be on a side. Keeping these constraints in mind and trying a few configurations gives for example
2
3
2
3
8 1 6
2 7 6
4 3 5 75
49 5 15
4 9 2
4 3 8
These examples are essentially the same — the second one is a rotated and reflected version of the
first.
3. Since the magic squares in Exercise 2 have a weight of 15, subtracting 14
3 from each entry will subtract
14 from each row and column while maintaining common sums, so the result will have weight 1. For
example, using the second magic square above, we get
2
3
6
4
8
3
13
3
2
3
7
3
1
3
5
3
4
3
11 7
.
3 5
10
3
In general, starting from any 3 ⇥ 3 classical magic square, to construct a magic square of weight w,
add w 315 to each entry in the magic square. This clearly preserves the common sum property, and
the new sum of each row is 15 + 3 w 315 = w.
4. (a) We must show that if A and B are magic squares, then so is A + B, and if c is any scalar, then
cA is also a magic square. Suppose wt(A) = a and wt(B) = b. Then the sum of the elements in
any row of A + B is the sum of the elements in that row of A plus the sum of the elements in that
row of B, which is therefore a + b. The same argument shows that the sum of the elements in any
column or either diagonal of A + B is also a + b, so that A + B 2 Mag 3 with weight a + b. In the
same way, the sum of the elements in any row or column, or either diagonal, of cA is c times the
sum of the elements in that line of A, so it is equal to ca. Thus cA 2 Mag 3 with weight ca.
478
CHAPTER 6. VECTOR SPACES
(b) If A, B 2 Mag 03 , then wt(A) = wt(B) = 0. From part (a), wt(A + B) = 0 as well, and also
wt(cA) = c wt(A) = 0. Thus A + B, cA 2 Mag 03 , so by Theorem 6.2, Mag 03 is a subspace of Mag 3 .
5. Let M be any 3 ⇥ 3 magic square, with wt(M ) = w. Using Exercise 3, we define M0 to be the magic
square of weight 0 derived from M by adding 3w to every element of M . This is the same as adding
the matrix
2
3
6
4
to M . Thus
w
3
w
3
w
3
w
3
w7
35
w
3
w
J
3
=
w
w
J, so that M = M0 + J.
3
3
M0 = M
Thus k =
w
3
w
3
w
3
w
3.
6. The linear equations express the fact that every row and column sum each diagonal sum is zero:
a+b+c
=0
d+e+f
=0
g+h+i=0
a
+d
b
+g
+e
c
a
matrix gives
0
1
0
0
0
1
0
0
0
0
1
1
0
0
0
1
+h
+f
0
0
1
0
1
0
0
0
0
0
1
0
0
1
1
0
+e
3
2
0
1
60
07
7
6
60
07
7
6
7
60
07
!6
7
60
07
6
7
60
07
6
40
05
0
0
=0
+i=0
+e
c
Row-reducing the associated
2
1 1 1 0 0
60 0 0 1 1
6
60 0 0 0 0
6
61 0 0 1 0
6
60 1 0 0 1
6
60 0 1 0 0
6
41 0 0 0 1
0 0 1 0 1
=0
+i=0
+g
0
1
0
0
0
0
0
0
=0
0
0
1
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
1
0
0
1
1
1
0
1
1
0
1
0
1
2
0
2
1
0
So there are two free variables in the general solution, which gives the magic square
2
3
s
t
s+t
4 t + 2s 0
t 2s5 .
s t t
s
7. From Exercise 6, a basis for Mag 03 is given by
2
3
2
1 0
1
0
25 , B = 4 1
A=4 2 0
1 0
1
1
1
0
1
3
1
15 ,
0
3
0
07
7
07
7
07
7.
07
7
07
7
05
0
so that Mag 03 has dimension 2. Note that these matrices are linearly independent since they are not
scalar multiples of each other. (The matrix in Exercise 6 is not the same as that given in the hint. To
show that the two are just di↵erent forms of one another, make the substitutions u = s, v = s + t;
then s = v and t = v s = u + v, so the matrix becomes
2
3
u
u v
v
4 u+v
0
u v5 ,
v
u+v
u
which is the matrix given in the hint.)
Exploration: Magic Squares
479
8. From Exercise 5, an arbitrary element M 2 Mag 3 can be written
M0 2 Mag 3 ,
M = M0 + kJ,
2
1
J = 41
1
3
1
15 .
1
1
1
1
Since M0 has basis {A, B} (see Exercise 7), it follows that M3 is spanned by S = {A, B, J}. To see that
this is a basis, we must prove that these matrices are linearly independent. Now, we have seen that
A and B are linearly independent, so if S is linearly dependent, then span(S) = span(A, B) = Mag 03 .
But this is impossible, since S spans Mag 3 , which contains matrices not in Mag 03 . Thus S is linearly
independent, so it forms a basis for Mag 3 , which therefore has dimension 3.
9. Using the matrix M from exercise 5, adding both diagonals and the middle column gives
(a + e + i) + (c + e + g) + (b + e + h) = (a + b + c) + (g + h + i) + 3e,
which is the sum of the first and third rows, plus 3 times the central entry. Since each row, column,
and diagonal sum is w, we get w + w + w = w + w + 3e, so that e = w3 .
10. The sum of the squares in the entries of M are
s2 + ( s
t)2 + t2 + ( s + t)2 + 02 + (s
t)2 + ( t)2 + (s + t)2 + ( s)2
= s2 + (s2 + 2st + t2 ) + t2 + (s2
2st + t2 ) + (s2
2st + t2 ) + t2 + (s2 + 2st + t2 ) + s2
= 6s2 + 6t2 .
We have another way of computing this sum, however. Since M was obtained from a classical 3 ⇥ 3
magic square by subtracting 5 from each entry (see Exercise 5), the sum of the squares of the entries
of M is also
(1
2
5) + (2
2
5) + · · · + (9
2
5) = 60.
p
So the equation is 6s2 + 6t2 = 60, or s2 + t2 = 10, which is a circle of radius 10 centered at the origin.
The only way to write 10 as a sum of two integral squares is (±1)2 + (±3)2 , so this equation has the
eight integral solutions
(s, t) = ( 3, 1), ( 3, 1), ( 1, 3), ( 1, 3), (3, 1), (3, 1), (1, 3), (1, 3).
Substituting those values of s and t into
2
s+5
M + 5J = 4 s + t + 5
t+5
s
t+5
5
s+t+5
3
t+5
s t + 55
s+5
gives the eight classical magic squares
2
2
47
6
2
8
41
6
9
5
1
3
5
7
3
4
35 ,
8
3
4
95 ,
2
2
2
49
4
2
8
43
4
7
5
3
1
5
9
3
6
15 ,
8
3
6
75 ,
2
2
4
43
8
2
6
41
8
9
5
1
7
5
3
3
2
75 ,
6
3
2
95 ,
4
2
4
49
2
2
6
47
2
3
5
7
1
5
9
3
8
15 ,
6
3
8
35 .
4
These matrices can all be obtained from one another by some combination of row and column interchanges and matrix transposition. A plot of the circle, with these eight points marked, is
480
CHAPTER 6. VECTOR SPACES
t
3
2
1
!3
!2
1
!1
2
s
3
!1
!2
!3
6.3
Change of Basis
1. (a)



2
1
0
= b1
+ b2
3
0
1



2
1
1
= c1
+ c2
3
1
1
)
(b) By Theorem 6.13, [ C | B ] ! [ I | PC B ]:
"
#
"
1
1 1 0
1 0
!
1
1 0 1
0 1
(c) [x]C = PC
B [x]B
=
"
1
2
1
2
1
2
1
2
#" # "
2
=
3
b1 = 2
b2 = 3
)
5
2
1
2
#
1
2
1
2
c1 =
c2 =
1
2
1
2
#
)
5
2
1
2
)

[x]B =
)
PC
2
3
[x]C =
B
=
"
"
5
2
1
2
1
2
1
2
#
1
2
1
2
#
, which is the same as the answer from part (a).
(d) By Theorem 6.13, [ B | C ] ! [ I | PB C ]:


1 0 1
1
1
1
) PB C =
0 1 1
1
1
1
"
#" # " #
5
1
1
2 = 2 , which is the same as the answer from part (a).
(e) [x]B = PB C [x]C =
1
1
1
3
2
2. (a)



4
1
1
= b1
+ b2
1
0
1



4
0
2
= c1
+ c2
1
1
3
(b) By Theorem 6.13, [ C | B ] ! [ I | PC B ]:
"
#
"
0 2 1 1
1 0
!
1 3 0 1
0 1
)
b1 = 5
b2 = 1
)
[x]B =
)
c1 = 7
c2 = 2
)
[x]C =
3
2
1
2
1
2
1
2
#
)
PC
B
=
"


3
2
1
2
5
1
7
2
1
2
1
2
#
6.3. CHANGE OF BASIS
(c) [x]C = PC
B [x]B
481
=
"
3
2
1
2
1
2
1
2
#"
5
1
#
=
"
#
7
2
, which is the same as the answer from part (a).
(d) By Theorem 6.13, [ B | C ] ! [ I | PB C ]:



1 1 0 2
1 0
1
1
1
1
!
) PB C =
0 1 1 3
0 1
1
3
1
3



1
1
7
5
(e) [x]B = PB C [x]C =
=
, which is the same as the answer from part (a).
1
3
2
1
3. (a)
2
6
4
2
6
4
3
2 3
2 3
2 3
1
1
0
0
7
6 7
6 7
6 7
=
b
+
b
+
b
05
1 405
2 4 15
3 405
1
0
0
1
3
2 3
2 3
2 3
1
1
0
0
7
6 7
6 7
6 7
05 = c1 415 + c2 415 + c3 405
1
1
1
1
(b) By Theorem 6.13,
2
1 0 0
6
41 1 0
1 1 1
(c)
[ C | B ] ! [ I | PC
3
2
1 0 0
1 0
7
6
0 1 05 ! 40 1
0 0 1
0 0
[x]C = PC
1
0
1
)
)
c1 =
c2 =
c3 =
1
1
1
)
0
0
1
3
0 0
7
1 05
1 1
1
1
0
2
1
6
=6
4 1
0
which is the same as the answer from
(d) By Theorem 6.13, [ B | C ] ! [ I | PB
2
1 0 0 1
40 1 0 1
0 0 1 1
(e)
[x]C = PB
2
6
[x]B = 4
2
6
[x]C = 4
]:
B
B [x]B
)
b1 =
b2 =
b3 =
part (a).
C ]:
3
0 0
1 05 )
1 1
C [x]C
2
1
6
=6
41
1
0
1
1
which is the same as the answer from part (a).
)
PC
B
2
1
6
=4 1
0
3
1
7
05
1
3
1
7
15 .
1
3
0 0
7
1 05 .
1 1
32 3 2 3
0 0
1
1
76 7 6 7
6 07 = 6 17
1 07
54 5 4 5
1 1
1
1
PB
0
C
32
2
1
= 41
1
1
3
0
05 .
1
0
1
1
3
2
1
3
76 7 6 7
6 7 6 7
07
5 4 15 = 4 05 ,
1
1
1
4. (a)
2 3
2 3
2 3
2 3
3
0
0
1
6 7
6 7
6 7
6 7
617 = b1 617 + b2 607 + b3 607
4 5
4 5
4 5
4 5
5
0
1
0
2 3
2 3
2 3
2 3
3
1
0
1
6 7
6 7
6 7
6 7
617 = c1 617 + c2 617 + c3 607
4 5
4 5
4 5
4 5
5
0
1
1
b1 = 1
)
b3 = 3
c1 =
)
)
b2 = 5
c2 =
c3 =
1
2
3
2
7
2
)
2 3
1
6 7
7
[x]B = 6
4 55
3
2
6
[x]C = 6
4
3
1
27
37 .
25
7
2
482
CHAPTER 6. VECTOR SPACES
[C |B ]![I
3
2
0 1
1
7
6
6
0 07
5 ! 40
1 0
0
(b) By Theorem 6.13,
2
1 0 1 0
6
61 1 0 1
4
0 1 1 0
| PC
]:
B
0
0
1
0
0
1
1
2
1
2
1
2
(c)
[x]C = PC
2
1
2
1
2
1
2
6
=6
4
B [x]B
3
1
2
1
2
1
2
1
27
17
25
1
2
32 3
1
2
1
2
1
2
1
2 7 617
17 6 7
2 5 455
1
3
2
which is the same as the answer from part (a).
[ B | C ] ! [ I | PB C
3
2
1 1 0 1
1
0 1 1 05 ! 4 0
0 0 1 1
0
(d) By Theorem 6.13,
2
0 0
41 0
0 1
PC
B
2
6
=6
4
6
=6
4
1
2
1
2
1
2
1
2
1
2
1
2
3
1
27
17 .
25
1
2
3
1
27
3 7,
2 5
7
2
]:
0 0
1 0
0 1
(e)
[x]C = PB
)
2
1
0
1
2
1
1
1
0
which is the same as the answer from part (a).
5. Note that in terms of the standard basis,
⇢

1
0
B=
,
0
1
C=
,
)
32
0
76
6
17
54
1
1
6
=6
40
C [x]C
3
0
15
1
1
1
0
⇢
0
,
1
3
1
27
3 7
2 5
7
2

PB
C
2
1
= 40
1
1
1
0
3
0
15
1
2 3
1
6 7
6
= 4 57
5
3
1
1
.
[2
x]B =
(a)
x = b1 (1) + b2 (x)
2
x = c1 (x) + c2 (1 + x)
(b) By Theorem 6.13, [ C | B ] ! [ I | PC


0 1 1 0
1
!
1 1 0 1
0
(c) [2
x]C = PC
B [2
x]B =

1
1
1
0

x]B = PB
C [2
x]C =

0
1
1
1

B
)
0
1
C
2
1
)
c1 =
c2 =
3
2
)
1
0
)
PC

2
1

3
x]C =
.
2
[2
]:
1
1
2
=
1
(d) By Theorem 6.13, [ B | C ] ! [ I | PB

1 0 0
0 1 1
(e) [2
b1 =
b2 =
)
2
]:
1
1
3
=
2


1
1
1
.
0
3
, which agrees with the answer in part (a).
2
)

B
=
PC
B
=

0
1
1
1
2
, which agrees with the answer in part (a).
1
6.3. CHANGE OF BASIS
483
6. Note that in terms of the standard basis,
⇢

1
1
B=
,
1
1
C=
,
⇢
0
,
2

4
0
.
(a)
1 + 3x = b1 (1 + x) + b2 (1
(c) [1 + 3x]C = PC
B [1
+ 3x]B =
1
2
1
4
1
2
1
4
(e) [1 + 3x]B = PB
C [1
+ 3x]C =
1
)
1
2
1
4
#
" #
1
#"
2
1
=
3
2
1
4
1
1
1
4
=
7. Note that in terms of the standard basis,
8 2 3 2 3 2 39
0
0 =
< 1
B = 415 , 415 , 405 ,
:
;
1
1
1
(a)
(b) By Theorem 6.13, [ C | B ] ! [
2
1 0
6
60 1
4
0 0
"
2
1
(c) [1 + x2 ]C = PC
2
B
0
1
0
0
1
1
1
1
1
1
2
4
B [1 + x ]B = 1
1
0
1
1
[1 + 3x]B =
" #
2
1
#
3
2
1
4
)
#
PC
=
B
"
1
2
1
4
1
2
1
4
#
.
, which agrees with the answer in part (a).
)
PC
B
=

1
1
2
2
, which agrees with the answer in part (a).
8 2 3 2 3 2 39
0
0 =
< 1
C = 405 , 415 , 405 .
:
;
0
0
1
c1 = 1
c2 = 0
c3 = 1
)
I | PC
)
[1 + 3x]C =
2
2
1 + x2 = b1 (1 + x + x2 ) + b2 (x + x2 ) + b3 (x2 )
1 + x2 = c1 (1) + c2 (x) + c3 (x2 )
#
1
2
1
4
0
2
1
"
]:
B
#" #
2 32
1
3
2
1
4
c2 =
(d) By Theorem 6.13, [ B | C ] ! [ I | PB C ]:


1
1 0 4
1 0
!
1
1 2 0
0 1
"
b2 =
c1 =
)
(b) By Theorem 6.13, [ C | B ] ! [ I | PC
"
#
"
0 4 1
1
1
!
2 0 1
1
0
"
)
x)
1 + 3x = c1 (2x) + c2 (4)
b1 = 2
)
b1 =
b2 =
b3 =
1
1
1
)
2
3
1
[1 + x2 ]B = 4 15
1
2 3
1
[1 + x2 ]C = 405 .
1
)
]:
3
0
7
07
5
1
)
PC
B
2
1
6
=6
41
1
0
1
1
3
0
7
07
5.
1
32 3 2 3
0
1
1
05 4 15 = 405, which agrees with the answer in part (a).
1
1
1
484
CHAPTER 6. VECTOR SPACES
[ B | C ] ! [ I | PB C ]:
3
2
1 0 0
1 0 0
1
0
0 1 05 ! 40 1 0
1
1
0 0 1
0 0 1
0
1
2
32 3 2
1
0 0 1
2
4 1
1 05 405 = 4
C [1 + x ]C =
0
1 1 1
(d) By Theorem 6.13,
2
1 0 0
41 1 0
1 1 1
(e) [1 + x2 ]B = PB
(a).
8. Note that in terms of the standard basis,
8 2 3 2 3 2 39
1
0 =
< 0
B = 415 , 405 , 415 ,
:
;
0
1
1
8 2 3 2 3 2 39
1
0 =
< 1
C = 405 , 415 , 405 .
:
;
0
0
1
(a)
2x
x2 = b1 (x) + b2 (1 + x2 ) + b3 (x + x2 )
)
4
2x
x2 = c1 (1) + c2 (1 + x) + c3 (x2 )
c1 = 6
c2 = 2
c3 = 1
2x
[ C | B ] ! [ I | PC
3
2
0 1 0
1 0
7
6
7
6
!
1 0 15
40 1
0
x 2 ] C = PC
answer in part (a).
(d) By Theorem 6.13, [
2
0 1 0 1
41 0 1 0
0 1 1 0
(e) [4
2x
2
6
6
4
2
6
6
4
1
B [4
1
0
C [4
2x
B
)
]:
0
1
1
0
1
0
1
0
1
1
= 4 1
0
1
0
1
0
x2 ] B
2x
B |C ]![I |
3
2
1 0
1
1 05 ! 40
0 1
0
x 2 ] B = PB
answer in part (a).
9. (a)
b1 = 3
b2 = 4
b3 = 5
4
(b) By Theorem 6.13,
2
1 1 0
6
60 1 0
4
0 0 1
(c) [4
3
2
0
1
05 ) PC B = 4 1
1
0
3
1
15, which agrees with the
1
2
PB
C
0
1
0
0
0
1
]:
1
1
1
2
1
1
1
x2 ] C = 4 1
1
2
1
1
2
)
3
1
7
17
5
1
32
1
15 4
1
)
[4
[4
2x
PC
B
answer in part
2
3
3
2x x2 ]B = 4 45
5
2 3
6
x 2 ] C = 4 25
1
2
1
6
=6
4 1
0
3
1
7
17
5.
1
1
0
1
3
2 3
3
6
45 = 4 25, which agrees with the
5
1
3
1
05
1
32
1
05 4
1
3
2 3
2 3
2 3
2 3
4
1
0
0
0
607
617
6 07
607
27
7 = b1 6 7 + b 2 6 7 + b 3 6 7 + b 4 6 7 )
405
405
4 15
405
05
1
0
0
0
1
3
2 3
2 3
2 3
2 3
4
1
2
1
1
6 27
617
6 17
607
27
7 = b1 6 7 + b 2 6 7 + b 3 6 7 + b 4 6 7 )
4 05
415
4 05
405
05
1
1
0
1
1
)
3
0 0
1 05
1 1
2
1
2
1
) PC B = 4 1
1
1
3 2 3
6
3
25 = 4 45, which agrees
1
5
b1
b2
b3
b4
=
=
=
=
b1
b2
b3
b4
4
2
0
1
=
=
=
=
)
5
2
0
3
9
2
3
1
05
1
with the
2
3
4
6 27
7
[A]B = 6
4 05
1
2 53
2
)
6 07
6 7
[A]C = 6 7.
4 35
9
2
6.3. CHANGE OF BASIS
485
(b) By Theorem 6.13, [ C | B ] ! [ I | PC
2
1
6
6 2
6
6 0
4
1
2
1
1
1
0
0
1
1
0
0
1
0
1
0
0
0
0
1
0
1
1
0
0
0
3
2
1
7
6
6
07
7 ! 60
7
60
05
4
1
0
0
]:
B
0
0
0
1
2
0
1
1
0
0
0
0
1
0
1
0
1
1
1
1
3
2
1
2
0
0
)
1
2
3
7
07
7
17
5
1
2
PC
B
2
1
2
6
6 0
=6
6 1
4
3
2
(c)
[A]C = PC
B [A]B
2
1
2
0
0
1
1
6 0
6
=6
4 1
3
2
1
2
1
1
1
2
32
0
1
0
1
1
1
1
2
3 2 53
4
2
6 07
27
7 6 7
7 = 6 7,
05 4 35
9
1
2
6
07
76
76
15 4
1
2
which agrees with the answer in part (a).
(d) By Theorem 6.13, [ B | C ] ! [ I | PB
2
1
6
60
6
60
4
0
C
0
0
0
1
2
1
1
0
0
2
1
1
0
1
0
0
1
0
0
0
1
1
0
1
(e)
[A]B = PB
]:
1
3
7
07
7
07
5
1
2
1
6
6 2
6
C [A]C = 6
4 0
1
)
PB
2
1
1
1
1
0
0
1
1
C
32
2
1
6
6 2
=6
6 0
4
1
5
2
3
2
1
1
1
1
0
0
1
2
76 7 6
6 7 6
07
7 6 07 = 6
7
7 6
05 6
4 35 4
9
1
2
1
3
7
07
7.
07
5
1
3
4
7
27
7,
07
5
1
which agrees with the result in part (a).
10. (a)
2 3
2 3
2 3
2 3
2 3
1
1
0
1
1
6 7
6 7
6 7
6 7
6 7
617
6 7
6 7
6 7
6 7
6 7 = b1 607 + b2 617 + b3 617 + b4 607
617
607
6 17
607
617
4 5
4 5
4 5
4 5
4 5
1
1
0
0
0
2 3
2 3
2 3
2 3
2 3
1
1
1
1
0
6 7
6 7
6 7
6 7
6 7
617
617
6 17
6 07
617
6 7 = b 1 6 7 + b 2 6 7 + b 3 6 7 + b4 6 7
617
607
6 17
6 17
617
4 5
4 5
4 5
4 5
4 5
1
1
0
1
1
b1 = 1
)
b2 = 1
b3 = 0
)
b4 = 0
b1 =
)
b2 =
b3 =
b4 =
1
3
1
3
1
3
1
3
2 3
1
6 7
617
7
[A]B = 6
607
4 5
0
2 3
1
)
6 31 7
637
7
[A]C = 6
617 .
435
1
3
1
2
3
7
07
7
17
5
1
2
486
CHAPTER 6. VECTOR SPACES
(b) By Theorem 6.13, [ C | B ] ! [ I | PC
2
1
6
61
6
60
4
1
1
1
0
1
0
1
1
0
1
0
1
1
1
1
1
0
1
0
0
1
1
1
0
0
]:
B
3
2
1
1
7
6
6
07
7 ! 60
7
60
15
4
0
0
0
0
0
1
0
0
0
1
0
0
0
1
2
3
1
3
2
3
1
3
1
3
2
3
1
3
2
3
3
2
3
2
3
1
3
1
3
)
PC
1
3
27
7
37
27
35
1
3
B
2
2
3
1
3
2
3
1
3
6
6
=6
6
4
3
1
3
2
3
1
3
2
3
2
3
2
3
1
3
1
3
1
3
27
7
37
.
27
35
1
3
1
0
1
1
1
2
1
2
1
2
1
2
1
2
1
2
1
2
1
2
1
2
(c)
[A]C = PC
2
2
3
1
3
2
3
1
3
6
6
6
B [A]B = 6
4
1
3
2
3
1
3
2
3
32 3
2
3
2
3
1
3
1
3
1
1
3
6 7
27
7
6
7
3 7 6 17
7
6
7
2
3 5 4 05
1
0
3
2 3
1
6 31 7
637
7
=6
617
435
1
3
which agrees with the answer in part (a).
(d) By Theorem 6.13, [ B | C ] ! [ I | PB
2
1
6
60
6
60
4
1
0
1
1
1
1
1
1
1
0
1
1
0
1
0
1
0
1
1
0
0
0
1
0
1
3
2
0
1
7
6
6
17
7 ! 60
7
60
15
4
1
0
C
]:
0
0
0
1
0
1
1
0
0
0
1
0
0
0
1
1
2
1
2
1
2
1
2
1
2
1
2
1
2
1
2
1
2
)
1
3
37
7
27
17
25
1
2
PB
C
2
6
6
=6
6
4
(e)
[A]B = PB
2
6
6
6
C [A]C = 6
4
1
0
1
1
2
1
2
1
2
1
2
1
2
1
2
1
2
1
2
1
2
32 3
1
3
7
6
37 617
7
27 637
17 617
25 435
1
1
2
3
1
which agrees with the result in part (a).
11. Note that in terms of the standard basis {sin x, cos x},
B=
⇢
1
,
1

0
1
,
C=
⇢
1
,
1

1
1
.
2 3
1
6 7
6 17
7
=6
6 07
4 5
0
3
37
7
27
.
17
25
1
2
6.3. CHANGE OF BASIS
487
(a)
2 sin x
)
2 sin x
[2 sin x
3 cos x = c1 (sin x + cos x) + c2 (sin x
)
[2 sin x
(b) By Theorem 6.13, [ C | B ] ! [ I | PC B ]:
"
#
"
1
1 1 0
1 0 1
!
1
1 1 1
0 1 0
(c) [2 sin x
b1 =
b2 =
" #
2
3 cos x]B =
5
3 cos x = b1 (sin x + cos x) + b2 (cos x)
3 cos x]C = PC
1
2
1
2
#
"
1
2
1
2
1
3 cos x]B =
0
B [2 sin x
3 cos x]C =
PC
#"
"
B
1
2
5
2
#
.
"
1
2
5
2
#
#
1
2
1
2
1
=
0
#
"
2
=
5
1
2
5
2
c1 =
c2 =
)
cos x)
)
2
5
)
.
, which agrees with the
answer in part (a).
(d) By Theorem 6.13, [ B | C ] ! [ I | PB C ]:


1 0 1
1
1 0
!
1 1 1
1
0 1
(e) [2 sin x
3 cos x]B = PB
1
0
1
2
)
"
#"
1
2
1
3 cos x]C =
0
C [2 sin x
PC
1
2
5
2
#
=

"
#
2
, which agrees with the
5
B
=
1
0
1
2
answer in part (a).
12. Note that in terms of the standard basis
⇢
1
B=
,
1
{sin x, cos x},

⇢

0
1
1
,
C=
,
1
1
1
.
(a)
)
sin x = b1 (sin x + cos x) + b2 (cos x)
sin x = c1 (cos x
(c) [sin x]C = PC
1
1
1
0
1
1
1
1
B [sin x]B
b2 =
)
sin x) + c2 (sin x + cos x)
(b) By Theorem 6.13, [ C | B ] ! [ I | PC
"
b1 = 1
=
"
0
1
1
2
1
2
#
#"
!
B
"
#
1
1
1
)
c1 =
c2 =
[sin x]B =
1
2
1
2
)
"
1
1
#
[sin x]C =
"
1
2
1
2
#
]:
1
0
0
0
1
1
=
"
1
2
1
2
#
1
2
1
2
#
)
PC
B
=
"
0
1
1
2
1
2
#
.
, which agrees with the answer in part (a).
488
CHAPTER 6. VECTOR SPACES
(d) By Theorem 6.13, [ B | C ] ! [ I | PB C ]:


1 0
1 1
1 0
!
1 1
1 1
0 1
(e) [sin x]B = PB
C [sin x]C
=
"
1
2
#"
1
0
1
2
1
2
#
1
2
=
"
1
1
1
0
#
)
PC
B
=

1
2
1
0
, which agrees with the answer in part (a).
13. From Section 3.6, Example 3.58, if B is the usual basis and C is the basis whose basis vectors are the
rotations of the usual basis vectors through 60 , the matrix of the rotation is
"
# "
p #
3
1
cos 60
sin 60
2
2
p
PC B =
=
.
3
1
sin 60 cos 60
2
2
Then
(a)
[x]C = PC
(b) Since PC
B
B [x]B
"
=
1
p2
3
2
3
2
1
2
#" #
3
"
1
p2
3
2
p
2
"
=
3
2
p #
3
p .
3
2 3
+
1
is an orthogonal matrix, we have
[x]B = (PC
B)
1
[x]C = (PC
B)
T
[x]C =
p
3
2
1
2
#"
" p
#
2 3+2
= p
.
4
2 3 2
4
#
14. From Section 3.6, Example 3.58, if B is the usual basis and C is the basis whose basis vectors are the
rotations of the usual basis vectors through 135 , the matrix of the rotation is
"
# " p
p #
2
2
cos 135
sin 135
2
p
p2
PC B =
=
.
2
2
sin 135 cos 135
2
2
Then
(a)
[x]C = PC
(b) Since PC
B
B [x]B
=
"
p
p
2
p2
2
2
2
p2
2
2
#" #
3
2
=
"
p
2
2
p
5 2
2
#
.
is an orthogonal matrix, we have
[x]B = (PC
B)
1
[x]C = (PC
B)
T
[x]C =
"
p
p
2
p2
2
2
2
p2
2
2
#"
#
4
"
#
0
= p .
4
4 2
are the coordinates of B in the basis C, so we have




1
1
2
1
[u1 ]C =
, so that u1 = 1
1
=
1
2
3
1




1
1
2
3
[u2 ]C =
, so that u2 = 1
+2
=
.
2
2
3
4
15. By definition, the columns of PC
B
Therefore
B = {u1 , u2 } =
⇢
1
,
1

3
4
.
6.3. CHANGE OF BASIS
489
16. Let C = {p1 (x), p2 (x), p3 (x)}. By definition, the columns of PC B are the coordinates of B in the basis
C, so that
2 3
1
[x]C = 4 05 ) x = p1 (x) p3 (x),
1
2 3
0
[1 + x]C = 425 ) 1 + x = 2p2 (x) + p3 (x),
1
2 3
0
[1 x + x2 ]C = 415 ) 1 x + x2 = p2 (x) + p3 (x).
1
Subtract the third equation from the second, giving p2 (x) = 2x x2 ; substituting back into the third
equation gives p3 (x) = 1 3x + 2x2 ; then from the first equation, p1 (x) = 1 2x + 2x2 . So
C = {p1 (x), p2 (x), p3 (x)} = {1
2x + 2x2 , 2x
x2 , 1
3x + 2x2 }.
17. Since a = 1 in this case and we are considering a quadratic, we have for the Taylor basis of P2
B = {1, x
1, (x
1)2 } = {1, x
1, x2
2x + 1}.
Let C = {1, x, x2 } be the standard basis for P2 . Then
2
To find the change-of-basis matrix
2
1
⇥
⇤
B C = 40
0
Therefore
3
1
[p(x)]C = 4 25 .
5
⇥
⇤
⇥
⇤
PB C , we reduce B C ! I PB C :
3
2
1
1 1 0 0
1 0 0 1 1
1
2 0 1 05 ! 4 0 1 0 0 1
0
1 0 0 1
0 0 1 0 0
[p(x)]C = PB
2
1
4
C [p(x)]B = 0
0
1
1
0
2
1)
5(x
so that the Taylor polynomial centered at 1 is
18. Since a =
8(x
32 3 2 3
1
1
2
25 4 25 = 4 85 ,
1
5
5
3
1
25 .
1
1)2 .
2 in this case and we are considering a quadratic, we have for the Taylor basis of P2
B = {1, x + 2, (x + 2)2 } = {1, x + 2, x2 + 4x + 4}.
Let C = {1, x, x2 } be the standard basis for P2 . Then
2
To find the change-of-basis matrix
2
1
⇥
⇤
B C = 40
0
PB
2
1
0
C,
4
4
1
3
1
[p(x)]C = 4 25 .
5
⇥
⇤
⇥
⇤
we reduce B C ! I PB C :
3
2
1 0 0
1 0 0 1
2
0 1 05 ! 40 1 0 0
1
0 0 1
0 0 1 0
0
3
4
45 .
1
490
CHAPTER 6. VECTOR SPACES
Therefore
2
1
4
C [p(x)]B = 0
0
[p(x)]C = PB
so that the Taylor polynomial centered at
2 is
5(x + 2)2 .
23 + 22(x + 2)
19. Since a =
32 3 2
3
4
1
23
45 4 25 = 4 225 ,
1
5
5
2
1
0
1 in this case and we are considering a cubic, we have for the Taylor basis of P3
B = {1, x + 1, (x + 1)2 , (x + 1)3 } = {1, x + 1, x2 + 2x + 1, x3 + 3x2 + 3x + 1}.
Let C = {1, x, x2 , x3 } be the standard basis for P2 . Then
2 3
0
6 07
7
[p(x)]C = 6
4 05 .
1
To find the change-of-basis matrix PB
⇥
2
1
⇤ 60
C =6
40
0
B
Therefore
1
1
0
0
1
2
1
0
1
3
3
1
C,
1
0
0
0
⇥
we reduce B
0
1
0
0
3
2
0
1
60
07
7!6
40
05
1
0
0
0
1
0
2
1
60
6
C [p(x)]B = 4
0
0
[p(x)]C = PB
so that the Taylor polynomial centered at
1
2
B=
0
1
0
0
0
0
1
0
PB
0
0
0
1
C
1
0
0
0
⇤
:
1
1
0
0
32 3 2
1
0
607 6
37
76 7 = 6
35 405 4
1
1
1
2
1
0
1 is
3
1
37
7.
35
1
1
2
1
0
3
1
37
7,
35
1
3(x + 1)2 + (x + 1)3 .
1 + 3(x + 1)
20. Since a =
1
1
0
0
⇤
⇥
C ! I
in this case and we are considering a cubic, we have for the Taylor basis of P3
(
1, x
✓
1
, x
2
1
2
◆2 ✓
, x
1
2
◆3 )
=
⇢
1, x
1 2
,x
2
1
x + , x3
4
Let C = {1, x, x2 , x3 } be the standard basis for P2 . Then
2 3
0
6 07
7
[p(x)]C = 6
4 05 .
1
To find the change-of-basis matrix PB
h
B
2
1
i 60
6
C =6
40
0
1
2
1
4
1
0
0
1
1
0
C,
1
8
3
4
3
2
1
⇥
we reduce B
1
0
0
0
0
1
0
0
0
0
1
0
⇤
⇥
C ! I
3
2
0
1
60
07
7
6
7!6
40
05
1
0
0
1
0
0
PB
0
0
1
0
C
0
0
0
1
⇤
3 2 3
x + x
2
4
1
8
:
1
0
0
0
1
2
1
4
1
0
0
1
1
0
3
1
8
37
47 .
37
5
2
1
.
6.4. LINEAR TRANSFORMATIONS
491
Therefore
2
1
60
6
C [p(x)]B = 6
40
0
[p(x)]C = PB
1
2
1
4
1
0
0
1
1
0
1
2
◆2
1
2
so that the Taylor polynomial centered at is
✓
◆
✓
1 3
1
3
+
x
+
x
8 4
2
2
21. Since [x]D = PD
C [x]C
and [x]C = PC
[x]D = PD
C [x]C
B [x]B ,
we get
= PD
C (PC
B [x]B )
32 3
213
1
0
8
37 6 7
4 7 607
6 7
37
5 405
2
1
8
637
6 7
= 6 43 7 ,
425
1
1
✓
+ x
1
2
◆3
= (PD
C PC
.
B )[x]B .
Since the change-of-basis matrix is unique (its columns are determined by the two bases), it follows
that
PD C PC B = PD B .
22. By Theorem 6.10 (c), to show that C is a basis it suffices to show that it is linearly independent, since
C has n elements and dim V = n. Since P is invertible, the columns of P are linearly independent. But
the given equation says that [ui ]B = pi , since the elements of that column are the coordinates of ui with
respect to the basis B. Since the columns of P are linearly independent, so is {[u1 ]B , [u2 ]B , . . . , [un ]B }.
But then by Theorem 6.7 in Section 6.2, it follows that {u1 , u2 , . . . , un } is linearly independent, so
that C is a basis for V .
To show that P = PB C , recall that the columns of PB C are the coordinate vectors of C with respect
to the basis B. But looking at the given equation, we see that this is exactly what the columns of P
are. Thus P = PB C .
6.4
Linear Transformations
1. T is a linear transformation:
T
✓
and
T
a
c

b
a0
+ 0
d
c
✓ 
a
↵
c
b
d
◆
b0
d0
=T
◆
=T
✓
↵a
↵c

a + a0
c + c0
↵b
↵d
◆
b + b0
d + d0
=

T
✓
w
y
x
z
◆
+T
✓
w0
y0
3
a + a 0 + b + b0
6
7
0
7
=6
4
5
0
0
0
c+c +d+d
2
3 2 0
3
a+b
a + b0

6 0 7 6 0 7
7+6
7=T a
=6
4 0 5 4 0 5
c
c+d
c 0 + d0

↵a + ↵b
0
a+b
=↵
0
↵c + ↵d
0
2. T is not a linear transformation:
✓
 0
◆
✓
w x
w x0
w + w0
T
+ 0
=T
0
y z
y z
y + y0
while
2
x0
z0
◆
◆
x + x0
z + z0
=
=


1
x

=
w
y
x+x
z
+
1
2
x+x
0
y
Since the two are not equal, T is not a linear transformation.

0
a
= ↵T
c+d
c
w + w0
1
y
0
0
y

y
1
0
y
w+w
0
x
 0
b
a
+T 0
d
c
0
0
1
w0
z0
1
z
2
z
z0
b
.
d
z0
b0
d0
492
CHAPTER 6. VECTOR SPACES
3. T is a linear transformation. If A, C 2 Mnn and ↵ is a scalar, then
T (A + C) = (A + C)B = AB + CB = T (A) + T (C)
T (↵A) = (↵A)B = ↵(AB) = ↵T (A).
4. T is a linear transformation. If A, C 2 Mnn and ↵ is a scalar, then
T (A + C) = (A + C)B
B(A + C) = AB + CB
BA
BC = (AB
BA) + (CB
BC)
= T (A) + T (C)
T (↵A) = (↵A)B
B(↵A) = ↵(AB)
↵(BA) = ↵(AB
BA) = ↵T (A).
5. T is a linear transformation, since by Exercise 44 in Section 3.2, tr(A + B) = tr(A) + tr(B) and
tr(↵A) = ↵ tr(A).
6. T is not a linear transformation. For example, in M22 , let A = I; then T (2A) = T (2I) = 2 · 2 = 4,
while 2T (A) = 2T (I) = 2(1 · 1) = 2.
7. T is not a linear transformation. For example, let A be any invertible matrix in Mnn ; then
invertible as well, so that rank(A) = rank( A) = n. Then
A is
T (A) + T ( A) = rank(A) + rank( A) = 2n while T (A + ( A)) = T (O) = rank(O) = 0.
8. Since T (0) = 1 + x + x2 6= 0, T is not a linear transformation by Theorem 6.14(a).
9. T is a linear transformation. Let p(x) = a + bx + cx2 and p0 (x) = a0 + b0 x + c0 x2 be elements of P2 ,
and let ↵ be a scalar. Then
(T (p + p0 ))(d) = (T ((a + a0 ) + (b + b0 )x + (c + c0 )x2 ))(d)
= (a + a0 ) + (b + b0 )(d + 1) + (b + b0 )(d + 1)2
= a + b(d + 1) + b(d + 1)2 + a0 + b0 (d + 1) + b0 (d + 1)2
= (T (p))(d) + (T (p0 ))(d),
and
T (↵p)(d) = (T (↵a + ↵bx + ↵cx2 ))(d) = ↵a + ↵b(d + 1) + ↵b(d + 1)2
= ↵ a + b(d + 1) + b(d + 1)2 = ↵(T (p))(d).
10. T is a linear transformation. Let f, g 2 F and ↵ be a scalar. Then
(T (f + g))(x) = (f + g)(x2 ) = f (x2 ) + g(x2 ) = (T (f ))(x) + (T (g))(x)
(T (↵f ))(x) = (↵f )(x2 ) = ↵f (x2 ) = ↵(T (f ))(x).
11. T is not a linear transformation. For example, let f be any function in F that takes on a value other
than 0 or 1. Then
2
2
(T (2f ))(x) = ((2f )(x)) = (2f (x)) = 4f (x)2 6= 2f (x)2 = 2(T (f ))(x),
where the inequality holds because of the conditions on f .
12. This is similar to Exercise 10; T is a linear transformation. Let f, g 2 F and ↵ be a scalar. Then
(T (f + g))(c) = (f + g)(c) = f (c) + g(c) = (T (f ))(c) + (T (g))(c)
(T (↵f ))(c) = (↵f )(c) = ↵f (c) = ↵(T (f ))(c)
6.4. LINEAR TRANSFORMATIONS
493
13. For T , we have (with ↵ a scalar)
✓
 ◆

a
c
a+c
T
+
=T
= (a + c) + (a + c + b + d)x = (a + (a + b)x) + (c + (c + d)x)
b
d
b+d


a
c
=T
+T
b
d
✓  ◆


a
↵a
a
T ↵
=T
= ↵a + (↵a + ↵b)x = ↵(a + (a + b)x) = ↵T
.
b
↵b
b
For S, suppose that p(x), q(x) 2 P1 and ↵ is a scalar; then
T (p(x) + q(x)) = x(p(x) + q(x)) = xp(x) + xq(x) = T (p(x)) + T (q(x))
T (↵p(x)) = x(↵p(x)) = ↵(xp(x)) = ↵T (p(x)).
14. Using the comment following the definition of a linear transformation, we have
2 3
2 3 2 3 2 3 2 3

✓ 
 ◆


1
3
5
6
11
5
1
0
1
0
T
=T 5
+2
= 5T
+ 2T
= 5 4 25 + 2 405 = 4 105 + 405 = 4105
2
0
1
0
1
1
4
5
8
3
In a similar way,
2 3
2 3 2 3 2 3 2
3
✓ 
 ◆


1
3
a
3b
a + 3b
a
1
0
1
0
T
=T a
+b
= aT
+ bT
= a 4 25 + b 405 = 4 2a 5 + 4 05 = 4 2a 5
b
0
1
0
1
1
4
a
4b
a + 4b

15. We must first find the coordinates of
that


7
in the basis
9


7
1
3
=c
+d
9
1
1
)
⇢
1
,
1
c + 3d =
c
d =

7
9
3
1
)
. So we want to find c, d such
c =
d =
5
4.
Then using the comment following the definition of a linear transformation, we have

✓ 

◆


7
1
3
1
3
T
=T 5
4
= 5T
4T
= 5(1 2x) 4(x + 2x2 ) = 8x2
9
1
1
1
1
14x + 5.
The second part is much the same, except that we need to find c, d such that



c = a+3b
a
1
3
c + 3d = a
4
=c
+d
)
)
b
1
1
c
d = b
d = a4 b .
Then using the comment following the definition of a linear transformation, we have

✓


◆
a + 3b 1
a b
a
3
T
=T
+
b
1
1
4
4


a + 3b
a b
1
3
=
T
+
T
1
1
4
4
a + 3b
a b
=
(1 2x) +
(x + 2x2 )
4
4
a b 2 a + 7b
a + 3b
=
x
x+
.
2
4
4
16. Since T is linear,
T (6 + x
4x2 ) = T (6) + T (x) + T ( 4x2 ) = 6T (1) + T (x)
= 6(3
2x) + (4x
4T (x2 )
x2 )
4(2 + 2x2 ) = 10
8x
9x2 .
494
CHAPTER 6. VECTOR SPACES
Similarly,
T (a + bx + cx2 ) = T (a) + T (bx) + T (cx2 ) = aT (1) + bT (x) + cT (x2 )
= a(3
2x) + b(4x
x2 ) + c(2 + 2x2 ) = (3a + 2c) + (4b
2a)x + (2c
b)x2 .
17. We do the second part first, then substitute for a, b, and c to get the first part. We must find the
coordinates of a + bx + cx2 in the basis 1 + x, x + x2 , 1 + x2 . So we want to find r, s, t such that
a + bx + cx2 = r(1 + x) + s(x + x2 ) + t(1 + x2 ) = (r + t) + (r + s)x + (s + t)x2
r
+ t = a
r + s
= b
s + t = c
)
)
r =
s =
t =
a+b c
2
a+b+c
2
a b+c
.
2
Then using the comment following the definition of a linear transformation, we have
✓
◆
a+b c
a+b+c
a b+c
2
2
2
T (a + bx + cx ) = T
(1 + x) +
(x + x ) +
(1 + x )
2
2
2
a+b c
a+b+c
a b+c
=
T (1 + x) +
T (x + x2 ) +
T (1 + x2 )
2
2
2
a+b c
a+b+c
a b+c
=
(1 + x2 ) +
(x x2 ) +
(1 + x + x2 )
2
2
2
3a b c 2
= a + cx +
x
2
For the second part, substitute a = 4, b =
T (4
1, and c = 3 to get
x + 3x2 ) = 4 + 3x +
12 + 1
2
3
x2 = 4 + 3x + 5x2 .
18. We do the second part first, then substitute for a, b, c, and d to get the first part. We must find the
coordinates of





a b
1 0
1 1
1 1
1 1
in the basis
,
,
,
.
c d
0 0
0 0
1 0
1 1
So we want to find r, s, t, u such that

a
c

b
1
=r
d
0

0
1
+s
0
0

1
1
+t
0
1

1
1
+u
0
1
1
1
Then using the comment following the definition of a

✓


a b
1 0
1
T
= T (a b)
+ (b c)
c d
0 0
0


1 0
1
= (a b)T
+ (b c)T
0 0
0
= (a
b) · 1 + (b
= a + b + c + d.
c) · 2 + (c
r + s + t + u =
s + t + u =
)
t + u =
u =
a
r = a b
b
s = b c
)
c
t = c d
d
u =
d.
linear transformation, we have


◆
1
1 1
1 1
+ (c d)
+ +d
0
1 0
1 1


1
1 1
1 1
+ (c d)T
+ +dT
0
1 0
1 1
d) · 3 + d · 4
For the second part, substitute a = 1, b = 3, c = 4, d = 2 to get

1 3
T
= 1 + 3 + 4 + 2 = 10.
4 2
6.4. LINEAR TRANSFORMATIONS
495
19. Let
T (E11 ) = a,
T (E12 ) = b,
T (E21 ) = c,
T (E22 ) = d.
Then

w x
T
= t (wE11 + xE12 + yE21 + zE22 ) = wT E11 + xT E12 + yT E21 + zT E22 = aw + bx + cy + dz.
y z
20. Writing
2
3
2 3
2 3
0
2
3
4 65 = a 415 + b 405
8
0
2
means that we must have b = 4, and then a = 6, so that if T were a linear transformation, then
2 3
2 3
2 3
0
2
3
T 4 65 = 6T 415 4 405 = 6(1 + x) 4(2 x + x2 ) = 2 + 10x 4x2 6= 2 + 2x2 .
8
0
2
This is a contradiction, so T cannot be linear.
21. Choose v 2 V . Then v = 0
of a linear transformation,
v by definition, so using part (a) of Theorem 6.14 and the definition
T ( v) = T (0
v) = T (0)
T (v) = 0
T (v) =
T (v).
Pn
22. Choose v 2 V . Then there are scalars c1 , . . . , cn such that v = i=1 ci vi since the vi are a basis.
Since T is linear, using the comment following the definition of a linear transformation we get
!
n
n
n
X
X
X
T (v) = T
ci vi =
ci T (vi ) =
ci vi = v.
i=1
i=1
i=1
Since T (v) = v for every vector v 2 V , T is the identity transformation.
Pn
23. We must show that if p(x) 2 Pn , then T p(x) = p0 (x). Suppose that p(x) = k=0 ak xk . Then since T
is linear, using the comment following the definition of a linear transformation we get
!
n
n
n
n
X
X
X
X
k
T p(x) = T
ak x
=
ak T (xk ) = a0 T (1) +
ak T (xk ) =
kak xk 1 = p0 (x).
k=0
k=0
k=1
k=1
24. (a) We prove the contrapositive: suppose that {v1 , . . . , vn } is linearly dependent in V ; then we have
c1 v1 + · · · + cn vn = 0,
where not all of the ci are zero. Then
T (c1 v1 + · · · + cn vn ) = c1 T (v1 ) + · · · + cn T (vn ) = T (0) = 0,
showing that {T (v1 ), . . . , T (vn )} is linearly dependent in W . Therefore if {T (v1 ), . . . , T (vn )} is
linearly independent in W , we must also have that {v1 , . . . , vn } is linearly independent in V .
(b) For example, the zero transformation from V to W , which takes every element of V to 0, shows
that the converse is false. There are many other examples as well, however. For instance, let


x
x
2
2
T :R !R :
7!
.
y
0
Then e1 and e2 are linearly independent in V , but


1
0
T (e1 ) =
and T (e2 ) =
0
0
are linearly dependent.
496
CHAPTER 6. VECTOR SPACES
25. We have
(S
(S
✓  ◆


2
2
5
4
1
=S T
=S
=
1
1
1
0
6

✓  ◆


x
x
2x + y
2x
y
T)
=S T
=S
=
.
y
y
y
0 2x + 2y
T)

Since the domain of T is R2 but the codomain of S is M22 , T
compute it.
S does not make sense and we cannot
26. We have
(S
x2 ) = S(T (3 + 2x
T )(3 + 2x
2
2)x + 2 · ( 2)x = 2
= 2 + (2
(S
x2 )) = S(2 + 2 · ( 1)x) = S(2
2
4x
2x)
2
2
T )(a + bx + cx ) = S(T (a + bx + cx )) = S(b + 2cx) = b + (b + 2c)x + 2(2c)x2
= b + (b + 2c)x + 4cx2 .
Since the domain of S is P1 , which equals the codomain of T , we can compute T
(T
S:
S)(a + bx) = T (S(a + bx)) = T (a + (a + b)x + 2bx2 ) = (a + b) + 2 · 2bx = (a + b) + 4bx.
27. The Chain Rule says that (f
g)0 (x) = g 0 (x)f 0 (g(x)). Let g(x) = x + 1 and f (x) = p(x). Then
(S
T )(p(x)) = S(T (p(x))) = S(p0 (x)) = p0 (x + 1)
(T
S)(p(x)) = T (S(p(x))) = T (p(x + 1)) = T ((p g)(x)) = (p g)0 (x) = g 0 (x)p0 (g(x)) = p0 (x + 1)
since the derivative of g(x) is 1.
28. Use the Chain Rule from the previous exercise, and let g(x) = x + 1:
(S
T )(p(x)) = S(T (p(x))) = S(xp0 (x)) = (x + 1)p0 (x + 1)
(T
S)(p(x)) = T (S(p(x))) = T (p(x + 1)) = T ((p g)(x)) = x(p g)0 (x) = xg 0 (x)p0 (g(x)) = xp0 (x + 1)
since the derivative of g(x) is 1.
29. We must verify that S T = IR2 and T S = IR2 .

✓  ◆



x
x
x y
4(x y) + ( 3x + 4y)
x
(S T )
=S T
=S
=
=
y
y
3x + 4y
3(x y) + ( 3x + 4y)
y

✓  ◆



x
x
4x + y
(4x + y) (3x + y)
x
(T S)
=T S
=T
=
=
.
y
y
3x + y
3(4x + y) + 4(3x + y)
y

x
Thus (S T )
= (T
y
we are done.
30. We must verify that S



x
x
x
S)
=
for any vector
, so both compositions are the identity and
y
y
y
T = IP1 and T
S = IP1 .
✓
b
+ (a + 2b)x
2
◆
(S
T )(a + bx) = S(T (a + bx)) = S
(T
S)(a + bx) = T (S(a + bx)) = T (( 4a + b) + 2ax) =
Thus (S T )(a + bx) = (T
identity and we are done.
=
✓
4·
◆
b
b
+ (a + 2b) + 2 · x = a + bx
2
2
2a
+ (( 4a + b) + 2 · 2a)x = a + bx.
2
S)(a + bx) = a + bx for any a + bx 2 P1 , so both compositions are the
6.4. LINEAR TRANSFORMATIONS
497
31. Suppose T 0 and T 00 are such that T 0 T = T 00 T = IV and T T 0 = T T 00 = IW . To show that
T 0 = T 00 , it suffices to show that T 0 (w) = T 00 (w) for any vector w 2 W . But
T 0 (w) = IV
T 0 (w) = T 00 T
T 0 (w) = T 00 ((T
T 0 )w) = T 00 IW (w) = T 00 (w).
32. (a) First, if T (v) = ±v, then clearly {v, T v} is linearly dependent. For the converse, if {v, T (v)} is
linearly dependent, then for some c, d not both zero cv + dT v = 0. Then
T (cv + dT v) = cT v + d(T
since T
T )v = cT v + dv = T 0 = 0
T = IV . Thus
cv + dT v = 0
cT v + dv = 0.
Subtracting these two equations gives c(T v v) d(T v v) = (c d)(T v v) = 0. But this
means that either c = d 6= 0 or that T v = v. If c = d =
6 0, then divide through by c in the
equation cv + dT v = 0 to get v + T v = 0, so that T v = v. Hence T v = ±v.


x
x
(b) Let T
=
. Then
y
y
(T
so that T
T = IV , and
T)


x
=T
y
x
and
y

✓  ◆



x
x
( x)
x
T
=T
=
=
,
y
y
( y)
y
x
are linearly dependent.
y
33. (a) First, if T (v) = v, then clearly {v, T v} = {v, v} is linearly dependent, and if T (v) = 0, then
{v, T v} = {v, 0} is linearly dependent. For the converse, suppose that {v, T v} is linearly dependent. Then there exist c, d, not both zero, such that
cv + dT v = 0.
As in Exercise 32, we get
T (cv + dT v) = cT v + d(T
since T
T )v = cT v + dT v = T 0 = 0
T = T . Thus
(c + d)T v = 0
cv +
dT v = 0.
Subtracting the two equations gives c(v T v) = 0, so that either T v = v or c = 0. But if c = 0,
then cv + dT v = dT v = 0, so that since c and d were not both zero, d cannot be zero, so that
T v = 0. Thus either T v = v or T v = 0.


x
x
(b) For example, let T
=
. This is a projection operator, so we already know that T T = T ;
y
0
to show this via computation,

✓  ◆



x
x
x
x
x
(T T )
=T T
=T
=
=T
.
y
y
0
0
y
Then for example T



1
1
0
=
while T
= 0.
0
0
1
498
CHAPTER 6. VECTOR SPACES
34. To show that S+T is a linear transformation, we must show it obeys the rules of a linear transformation.
Using the definition of S + T and of cT , as well as the fact that both S and T are linear, we have
(S + T )(u + v) = S(u + v) + T (u + v) = Su + Sv + T u + T v = (Su + T u) + (Sv + T v)
= (S + T )u + (S + T )v
(c(S + T ))v = (cS + cT )v = (cS)v + (cT )v = c(Sv) + c(T v) = c(Sv + T v) = c(S + T )v.
35. We show that L = L (V, W ) satisfies the ten axioms of a vector space. We use freely below the fact
that V and W are both vector spaces. Also, note that S = T in L (V, W ) simply means that Sv = T v
for all v 2 V .
1. From Exercise 34, S, T 2 L implies that S + T 2 L .
2. (S + T )v = Sv + T v = T v + Sv = (T + S)v.
3. ((R + S) + T )v = (R + S)v + T v = Rv + Sv + T v = Rv + (S + T )v = (R + (S + T ))v.
4. Let Z : V ! W be the map Z(v) = 0 for all v 2 V. Then Z is the zero map, so it is linear and
thus in L , and
(S + Z)v = Sv + Zv = Sv + 0 = Sv.
5. If S 2 L , let
S : V ! W be the map ( S)v =
Sv. By Exercise 34,
(S + ( S))v = Sv + ( S)v = Sv
S 2 L , and
Sv = 0.
6. By Exercise 34, if S 2 L , then cS 2 L .
7. (c(S + T ))v = c(S + T )v = c(Sv + T v) = cSv + cT v = (cS)v + (cT )v = ((cS) + (cT ))v.
8. ((c + d)S)v = (c + d)Sv = cSv + dSv = (cS)v + (cT )v = ((cS) + (cT ))v.
9. (c(dS))v = c(dS)v = c(dSv) = (cd)Sv = ((cd)S)v.
10. By Exercise 34, (1S)v = 1Sv = Sv.
36. To show that these linear transformations are equal, it suffices to show that they have the same value
on every element of V .
(a) (R (S + T ))v = R((S + T )v) = R(Sv + T v) = R(Sv) + R(T v) = (R S)v + (R T )v.
(b)
(c(R S))v = c(R S)v = cR(Sv) = (cR)(Sv) = ((cR) S)v
(c(R S))v = c(R S)v = cR(Sv) = (cR)(Sv) = R((cS)v) = (R (cS))v.
6.5
The Kernel and Range of a Linear Transformation
1. (a) Only (ii) is in ker(T ), since


1 2
1 0
(i) T
=
6= O,
1 3
0 3


0 4
0 0
(ii) T
=
= O,
2 0
0 0


3
0
3
0
(iii) T
=
6= O.
0
3
0
3
(b) From the definition of T , a matrix is in range(T ) if and only if its two o↵-diagonal elements are
zero. Thus only (iii) is in range(T ).
2. (a) T is explicitly defined by
T
Thus

a
c
b
=a+d
d
6.5. THE KERNEL AND RANGE OF A LINEAR TRANSFORMATION
(i) T


0
2

1
(iii) T
0
(ii) T
1
1
499
2
=1+3=4
3
4
=0+0=0
0
3
= 1 + ( 1) = 0,
1
and therefore (ii) and (iii) are in ker(T ) but (i) is not.
(b) All three of these are in range(T ). For example,

0 0
(i) T
= 0 + 0 = 0,
0 0

2 1
(ii) T
= 2 + 0 = 2,
4 0
p
p
p
2
⇡
(iii) T 2
= 22 + 0 = 22 .
12 0
(c) ker(T ) is the set of all matrices in M22 whose diagonal elements are negatives of each other.
range(T ) is all real numbers.
3. (a) We have

1 1
0
(i) T (1 + x) = T (1 + 1x + 0x ) =
=
6= 0,
1+0
1


0 1
1
2
2
(ii) T (x x ) = T (0 + 1x 1x ) =
=
6= 0,
1 + ( 1)
0


1 1
0
(iii) T (1 + x x2 ) = T (1 + 1x 1x2 ) =
=
= 0,
1 + ( 1)
0
2

so only (iii) is in range(T ).
(b) All of them are. For example,

0
2
(i) T (1 + x x ) =
, from part (a),
0


1 0
1
(ii) T (1) =
=
,
0+0
0


0 0
0
(iii) T (x2 ) =
=
.
0+1
1
(c) ker(T ) consists of all polynomials a + bx + cx2 such that a = b and b = c; parametrizing via
a = t gives ker(T ) = {a + ax ax2 }, so that ker(T ) consists of all multiples of 1 + x x2 . range(T )
is all of R2 , since it is a subspace of R2 containing both e1 and e2 (see items (ii) and (iii) of part
(b)).
4. (a) Since xp0 (x) = 0 if and only if p0 (x) = 0, we see that only (i) is in ker(T ). For the other two, we
have T (x) = xx0 = x and T (x2 ) = x(x2 )0 = 2x2 .
(b) A polynomial in range(T ) must be divisible by x; that is, it must have no constant term. Thus
(i) is not in range(T ). However, x = T (x) from part (a), and also from part (a), we see that
T 21 x2 = x2 . Thus (ii) and (iii) are in range(T ).
(c) ker(T ) consists of all polynomials whose derivative vanishes, so it consists of constant polynomials.
range(T ) consists of all polynomials with zero constant terms, sinceR given any such polynomial,
we may write it as xg(x) for some polynomial g(x); then clearly T
g(x) dx = xg(x).
⇢
0 b
5. Since ker(T ) =
, a basis for ker(T ) is
c 0
⇢

0 1
0 0
,
.
0 0
1 0
500
CHAPTER 6. VECTOR SPACES
Similarly, since range(T ) =
⇢
a
0
0
d
, a basis for range(T ) is
⇢
1
0

0
0
,
0
0
0
1
.
Thus dim ker(T ) = 2 and dim range(T ) = 2. The Rank Theorem says that
4 = dim M22 = nullity(T ) + rank(T ) = dim ker(T ) + dim range(T ) = 2 + 2,
which is true.
6. Since
ker(T ) =
a basis for ker(T ) is
⇢
⇢

0
0
,
1
0
1
0
a
c
b
a
,

1
0
,
0
1
0
0
.
From Exercise 2, range(T ) = R. Thus dim ker(T ) = 3 and dim range(T ) = 1. The Rank Theorem says
that
4 = dim M22 = nullity(T ) + rank(T ) = dim ker(T ) + dim range(T ) = 3 + 1,
which is true.
7. Since ker(T ) = a + ax ax2 , a basis for ker(T ) is {1 + x x2 }. From Exercise 3, range(T ) = R2 .
Thus dim ker(T ) = 1 and dim range(T ) = 2. The Rank Theorem says that
3 = dim P2 = nullity(T ) + rank(T ) = dim ker(T ) + dim range(T ) = 1 + 2,
which is true.
8. Since ker(T ) is the constant polynomials, it has a basis {1}. From Exercise 4, range(T ) is the set of
polynomials with nonzero constant terms, or {bx + cx2 }, so it has a basis {x, x2 }. Thus dim ker(T ) = 1
and dim range(T ) = 2. The Rank Theorem says that
3 = dim P2 = nullity(T ) + rank(T ) = dim ker(T ) + dim range(T ) = 1 + 2,
which is true.
9. Since

1
T
0

0
1
=
0
0

0
1
=
,
0
0

0
T
1

0
0
=
0
1

0
0
=
,
0
1
it follows that range(T ) = R2 , since it is a subspace of R2 containing both e1 and e2 . Thus rank(T ) =
dim range(T ) = 2, so that by the Rank Theorem,
nullity(T ) = dim M22
10. Since
T (1

1
x) =
1
rank(T ) = 4

0
1
=
,
1
0
2 = 2.
T (x) =

0
,
1
it follows that range(T ) = R2 , since it is a subspace of R2 containing both e1 and e2 . Thus rank(T ) =
dim range(T ) = 2, so that by the Rank Theorem,
nullity(T ) = dim P2
rank(T ) = 3
2 = 1.
6.5. THE KERNEL AND RANGE OF A LINEAR TRANSFORMATION
11. If A =

a
c
501
b
, then
d

a
T (A) = AB =
c
b
d


1
a
=
1
c
1
1

a+b
a
=
c+d
c
b
d
Thus T (A) = O if and only if a = b and c = d, so that
⇢
a a
ker(T ) =
c c
so that a basis for ker(T ) is
⇢

1
0
,
0
1
1
0
0
1
b
d
(a
(c
b)
.
d)
,
.
Hence nullity(T ) = dim ker(T ) = 2, so that by the Rank Theorem
12. If A =

a
c
rank(T ) = dim M22
nullity(T ) = 4


b
a
=
d
c
2 = 2.
b
, then
d
T (A) = AB
BA =

a
c
b
d
1
0
1
1

1
0
1
1

a
c

a+b
c+d
So A 2 ker(T ) if and only if c = 0 and a = d, so that



a b
1 0
0
A=
=a
+b
0 a
0 1
0
a+c
c
b+d
=
d

c
0
a
d
c
.
1
.
0
These two matrices are linearly independent since they are not multiples of one another, so that
nullity(T ) = 2. Since dim M22 = 4, we see that by the Rank Theorem
rank(T ) = dim M22
nullity(T ) = 4
2 = 2.
13. The derivative of p(x) = a + bx + cx2 is b + 2cx; then p0 (0) = b, so this is zero if b = 0. Thus
ker(T ) = {a + cx2 }, so that a basis for ker(T ) is {1, x2 }. Then nullity(T ) = dim ker(T ) = 2, so that by
the Rank Theorem,
rank(T ) = dim P2 nullity(T ) = 3 2 = 1.
14. We have
T (A) = A
2
AT = 4a12
a31
0
a12
a21
a13
a21
0
a32
a23
a13
a23
0
3
a31
a32 5 ,
so that A 2 ker(T ) if and only if a12 = a21 , a13 = a31 , and a23 = a32 ; that is, if and only if A is
symmetric. By Exercise 40 in Section 6.2, nullity(T ) = dim ker(T ) = 3·4
2 = 6. Then by the Rank
Theorem,
rank(T ) = dim M33 nullity(T ) = 9 6 = 3.
15. (a) Yes, T is one-to-one. If
T
then 2a



a
2a b
0
=
=
,
b
a + 2b
0
b = 0 and a + 2b = 0. 
The only solution to this pair of equations is a = b = 0, so that
0
the only element of the kernel is
and thus T is one-to-one.
0
(b) Since T is one-to-one and the dimension of the domain and codomain are both 2, T must be onto
by Theorem 6.21.
502
CHAPTER 6. VECTOR SPACES
16. (a) Yes, T is one-to-one. If
T

a
= (a
b
2b) + (3a + b)x + (a + b)x2 = 0,
then a + b = 0 from the third term. Since 3a + b = 0 as well, subtracting gives 2a = 0, so that
a
a = 0. The third term then forces b = 0. Thus
= 0 and T is one-to-one.
b
(b) T cannot be onto. R2 has dimension 2, so by the rank theorem 2 = dim R2 = rank(T ) +
nullity(T ) rank(T ). It follows that the dimension of the range of T is at most 2 (in fact, since
from part (a) nullity(T ) = 0, the dimension of the range is equal to 2). But dim P2 = 3, so T is
not onto.
17. (a) If T (a + bx + cx2 ) = 0, then
2a
b
=0
a+b
3c = 0
a
Row-reducing the augmented matrix gives
2
2
1
0
4 1
1
3
1
0
1
+ c = 0.
3
2
0
1
05 ! 4 0
0
0
0
1
0
1
2
0
This has nontrivial solutions a = t, b = 2t, c = t, so that
3
0
05 .
0
ker(T ) = {t + 2tx + tx2 } = {t(1 + 2x + x2 )}.
So T is not one-to-one.
(b) Since dim P2 = dim R3 = 3, the fact that T is not one-to-one implies, by Theorem 6.21, that T
is not onto.
18. (a) By Exercise 10, nullity(T ) = 1, so that ker(T ) 6= {0}, so that T is not one-to-one.
(b) By Exercise 10, range(T ) = R2 , so that T is onto by definition.
2 3
a
19. (a) If T 4 b 5 = O, then a b = a + b = 0 and b c = b + c = 0; these equations give a = b = c = 0 as
c
the only solution. Thus T is one-to-one.
(b) By the Rank Theorem,
dim range(T ) = dim R3
dim ker(T ) = 3
0 = 3.
However, dim M22 = 4 > 3 = dim range(T ), so that T cannot be onto.
2 3
a
20. (a) If T 4 b 5 = O, then a + b + c = 0, b 2c = 0, and a c = 0. The latter two equations give a = c
c
and b = 2c, so that from the first equation 4c = 0 and thus c = 0. It follows that a = b = 0. Thus
T is one-to-one.
3
(b) By Exercise 40 in Section 6.2, dim W = 2·3
2 = 3 = dim R . Since the dimensions of the domain
and codomain are equal and T is one-to-one, Theorem 6.21 implies that T is also onto.
6.5. THE KERNEL AND RANGE OF A LINEAR TRANSFORMATION
503
21. These vector spaces are isomorphic. Bases for the two vector spaces are
R3 : {e1 , e2 , e3 }.
D3 : {E11 , E22 , E33 },
Since both have dimension 3, we know that D3 ⇠
= R3 . Define
T (aE11 + bE22 + cE33 ) = a e1 + b e2 + c e3 .
Then T is a linear transformation from the way we have defined it. Since the ei are linearly independent,
it follows that a e1 + b e2 + c e3 = 0 if and only if a = b = c = 0. Thus ker(T ) = {0} so that T is
one-to-one. Since dim D3 = dim R3 , we know that T is onto as well, so it is an isomorphism.
22. By Exercise 40 in Section 6.2, dim S3 = 3·4
2 = 6. Since an upper triangular matrix has three entries
that are forced to be zero (those below the diagonal), and the other six may take any value, we see
that dim U3 = 6 as well. Thus S3 ⇠
= U3 . Define
2
3
2
3
a11 a12 a13
a11 a12 a13
T : S3 ! D3 : 4a21 a22 a23 5 7! 4 0 a22 a23 5 .
a31 a32 a33
0
0 a33
Then any matrix in ker(T ) must have a11 = a12 = a13 = a22 = a23 = a33 = 0, so that it must
itself be the zero matrix. Thus ker(T ) = {0} so that T is one-to-one. Since T is one-to-one and
dim S3 = dim U3 = 6, it follows that T is onto and is therefore an isomorphism.
23. By Exercises 40 and 41 in Section 6.2, dim S3 =
are unequal, S3 ⇠
6 S30 .
=
3·4
2
= 6 while dim S30 =
2·3
2
= 3. Since their dimensions
24. Since a basis for P2 is {1, x, x2 }, we have dim P2 = 3. If p(x) 2 W , say p(x) = a + bx + cx2 + dx3 ,
then p(0) = 0 implies that a = 0, so that p(x) = bx + cx2 + dx3 ; a basis for this subspace is {x, x2 , x3 }
so that dim W = 3 as well. Therefore P2 ⇠
= W . Define
T : P2 ! W : a + bx + cx2 7! ax + bx2 + cx3 .
Then T (a + bx + cx2 ) = ax + bx2 + cx3 = 0 implies that a = b = c = 0. Thus T is one-to-one; since
the dimensions of domain and codomain are equal, T is also onto, so it is an isomorphism.
25. A basis for C as a vector space over R is {1, i}, so that dim C = 2. A basis for R2 is {e1 , e2 }, so that
dim R2 = 2. Thus C ⇠
= R2 . Define

a
2
T : C ! R : a + bi 7!
.
b
Then T (a + bi) = 0 means that a = b = 0. Thus T is one-to-one; since the dimensions of domain and
codomain are equal, T is also onto, so it is an isomorphism.
26. V consists of all 2 ⇥ 2 matrices whose upper-left and lower-right entries are negatives of each other, so
that
⇢
a
b
V =
.
c
a
Thus a basis for V is
⇢
1
0
0
,
1

0
0
1
,
0

0
1
0
0
,
and therefore dim V = 3. Since dim W = dim R2 = 2, these vector spaces are not isomorphic.
27. Since the dimensions of the domain and codomain are the same, it suffices to show that T is one-to-one.
Let p(x) = a0 + a1 x + · · · + an xn . Then
p(x) + p0 (x) = (a0 + a1 x + · · · + an
1x
n 1
+ an xn ) + (a1 + 2a2 x + · · · + nan xn
= (a0 + a1 ) + (a1 + 2a2 )x + · · · + (an
1
+ nan )x
n 1
n
+ an x .
1
)
504
CHAPTER 6. VECTOR SPACES
If p(x) + p0 (x) = 0, then the coefficient of xn is zero, so that an = 0. But then the coefficient of xn 1
is 0 = an 1 + nan = an 1 , so that an 1 = 0. We continue this process, determining that ai = 0 for all
i. Thus p(x) = 0, so that T is one-to-one and is thus an isomorphism.
28. Since the dimensions of the domain and codomain are the same, it suffices to show that T is one-to-one.
Now, T (xk ) = (x 2)k . Suppose that
T (a0 + a1 x + · · · + an xn ) = a0 + a1 (x
2) + · · · + an (x
2)n = 0.
If we show that {1, x 2, (x 2)2 , . . . , (x 2)n } is a basis for Pn , then it follows that all the ai are
zero, so that ker(T ) = {0} and T is one-to-one. To prove they form a basis, it suffices to show they
are linearly independent, since there are n + 1 elements in the set and dim Pn = n + 1. We do this
by induction. For n = 0, this is clear. For n = 1, suppose that a + b(x 2) = (a 2b) + bx = 0.
Then b = 0 and thus a = 0, so that linear independence for n = 1 is established. Now suppose that
{1, x 2, . . . , (x 2)k } is a linearly independent set, and suppose that
a0 + a1 (x
2) + · · · + ak
1 (x
2)k
1
+ ak (x
2)k = 0.
If this polynomial were expanded, the only term involving xk comes from the last term of the sum,
and it is ak xk . Therefor ak = 0, and then the other ai = 0 by induction. Thus this set forms a basis
for Pn , so that T is one-to-one and thus is an isomorphism.
29. Since the dimensions of the domain and codomain are the same, it suffices to show that T is one-to-one.
That is, we want to show that xn p x1 = 0 implies that p(x) = 0. Suppose
p(x) = a0 + a1 x + · · · + an xn .
Then
✓ ◆
✓
✓ ◆
✓ ◆◆
1
1
1
n
x p
= x a0 + a1
+ · · · + an
= a0 xn + a1 xn
x
x
xn
n
Therefore if xn p
isomorphism.
1
x
1
+ · · · + an .
= 0, all of the ai are zero so that p(x) = 0. Thus T is one-to-one and is an
30. (a) As the hint suggests, define a map T : C [0, 1] ! C [2, 3] by (T (f ))(x) = f (x
Thus for example (T (f ))(2.5) = f (2.5 2) = f (0.5). T is linear, since
(T (f + g))(x) = (f + g)(x
(T (cf ))(x) = (cf )(x
2) = f (x
2) = cf (x
2) + g(x
2) for f 2 C [0, 1].
2) = (T (f ))(x) + (T (g))(x)
2) = c(T (f ))(x).
We must show that T is one-to-one and onto. First suppose that T (f ) = 0; that is, suppose that
f 2 C [0, 1] and that f (x 2) = 0 for all x 2 [2, 3]. Let y = x 2; then x 2 [2, 3] means y 2 [0, 1],
and we get f (y) = 0 for all y 2 [0, 1]. But this means that f = 0 in C [0, 1]. Thus T is one-to-one.
To see that it is onto, choose f 2 C [2, 3], and let g(x) = f (x + 2). Then
(T (g))(x) = g(x
2) = f (x + 2
2) = f (x),
so that T (g) = f and thus T is onto. Since T is one-to-one and onto, it is an isomorphism.
(b) This is similar to part (a). Define a map T : C [0, 1] ! C [a, a + 1] by (T (f ))(x) = f (x a) for
f 2 C [0, 1]. T is proved linear in a similar fashion to part (a). We must show that T is one-to-one
and onto. First suppose that T (f ) = 0; that is, suppose that f 2 C [0, 1] and that f (x a) = 0
for all x 2 [a, a + 1]. Let y = x a; then x 2 [a, a + 1] means y 2 [0, 1], and we get f (y) = 0 for
all y 2 [0, 1]. But this means that f = 0 in C [0, 1]. Thus T is one-to-one. To see that it is onto,
choose f 2 C [a, a + 1], and let g(x) = f (x + a). Then
(T (g))(x) = g(x
a) = f (x + a
a) = f (x),
so that T (g) = f and thus T is onto. Since T is one-to-one and onto, it is an isomorphism.
6.5. THE KERNEL AND RANGE OF A LINEAR TRANSFORMATION
505
31. This is similar to Exercise 30. Define a map T : C [0, 1] ! C [0, 2] by (T (f ))(x) = f 12 x for f 2 C [0, 1].
Thus for example (T (f ))(1) = f 12 · 1 = f 12 . T is proved linear in a similar fashion to part (a)
of Exercise 30. We must show that T is one-to-one and onto. First suppose that T (f ) = 0; that is,
suppose that f 2 C [0, 1] and that f 12 x = 0 for all x 2 [0, 2]. Let y = 12 x; then x 2 [0, 2] means
y 2 [0, 1], and we get f (y) = 0 for all y 2 [0, 1]. But this means that f = 0 in C [0, 1]. Thus T is
one-to-one. To see that it is onto, choose f 2 C [0, 2], and let g(x) = f (2x). Then
✓ ◆
✓
◆
1
1
(T (g))(x) = g
x = f 2 · x = f (x),
2
2
so that T (g) = f and thus T is onto. Since T is one-to-one and onto, it is an isomorphism.
32. This is again similar to Exercises
30 and 31,
⇣
⌘ and is the most general form. Define a map T : C [a, b] !
b a
C [c, d] by (T (f ))(x) = f d c (x c) + a for f 2 C [a, b]. Then (T (f ))(c) = f (a) and (T (f ))(d) =
b a + a = b. T is proved linear in a similar fashion to part (a) of Exercise 30. We must show that
T ⇣is one-to-one and
⌘ onto. First suppose that T (f ) = 0; that is, suppose that f 2 C [a, b] and that
f
b a
d c (x
c) + a = 0 for all x 2 [c, d]. Let y =
b a
d c (x
c) + a; then x 2 [c, d] means y 2 [a, b], and
we get f (y) = 0 for all y 2 [a, b]. But this means that
⇣ f = 0 in C [a,
⌘ b]. Thus T is one-to-one. To see
d c
that it is onto, choose f 2 C [c, d], and let g(x) = f b a (x a) + c . Then
(T (g))(x) = g
✓
b
d
a
(x
c
c) + a
◆
=f
✓
d
b
c
a
✓✓
b
d
a
(x
c
c) + a
◆
◆
◆
a + c = f (x),
so that T (g) = f and thus T is onto. Since T is one-to-one and onto, it is an isomorphism.
33. (a) Suppose that (S T )(u) = 0. Then (S T )(u) = S(T u) = 0. Since S is one-to-one, we must have
T u = 0. But then since T is one-to-one, it follows that u = 0. Thus S T is one-to-one.
(b) Choose w 2 W . Since S is onto, we can find some v 2 V such that Sv = w. But T is onto as
well, so there is some u 2 U such that T u = v. Then
(S
so that S
T )(u) = S(T u) = Sv = w,
T is onto.
34. (a) Suppose that T u = 0. Then since S0 = 0, we have
(S
T )u = S(T u) = S0 = 0.
But S T is one-to-one, so we must have u = 0, so that T is one-to-one as well. Intuitively, if T
takes two elements of U to the same element of V , then certainly S T takes those two elements
of U to the same element of W .
(b) Choose w 2 W . Since S
Then
T is onto, there is some u 2 U such that (S
Sv = S(T u) = (S
so that S is onto. Intuitively, if S
T )u = w. Let v = T u.
T )u = w,
T “hits” every element of W , then surely S must.
35. (a) Suppose that T : V ! W and that dim V < dim W . Then range(T ) is a subspace of W , and the
Rank Theorem asserts that
dim range(T ) = dim V
nullity(T )  dim V < dim W.
Theorem 6.11(b) in Section 6.2 states that if U is a subspace of W , then dim U = dim W if and
only if U = W . But here dim range(T ) 6= dim W , and range(T ) is a subspace of W , so that
range(T ) 6= W and T cannot be onto.
506
CHAPTER 6. VECTOR SPACES
(b) We want to show that ker(T ) 6= {0}. Since range(T ) is a subspace of W , then dim W
dim range(T ) = rank(T ). Then the Rank Theorem asserts that
dim W + nullity(T )
so that dim ker(T ) = nullity(T )
one-to-one.
dim V
rank(T ) + nullity(T ) = dim V,
dim W > 0. Thus ker(T ) 6= {0}, so that T is not
36. Since dim Pn = dim Rn+1 = n + 1, we need only show that T is a one-to-one linear transformation. It
is linear since
2
3 2
3 2
3 2
3
(p + q)(a0 )
p(a0 ) + q(a0 )
p(a0 )
q(a0 )
6 (p + q)(a1 ) 7 6 p(a1 ) + q(a1 ) 7 6 p(a1 ) 7 6 q(a1 ) 7
6
7 6
7 6
7 6
7
(T (p + q))(x) = 6
7=6
7 = 6 .. 7 + 6 .. 7 = (T (p))(x) + (T (q))(x)
..
..
4
5 4
5 4 . 5 4 . 5
.
.
(p + q)(an )
p(an ) + q(an )
p(an )
q(an )
2
3 2
3
2
3
(cp)(a0 )
cp(a0 )
p(a0 )
6 (cp)(a1 ) 7 6 cp(a1 ) 7
6 p(a1 ) 7
6
7 6
7
6
7
(T (cp))(x) = 6
7 = 6 .. 7 = c 6 .. 7 = c(T (p))(x).
..
4
5
4
5
4
.
.
. 5
(cp)(an )
cp(an )
p(an )
Now suppose T (p) = 0. That means that
p(a0 ) = p(a1 ) = · · · = p(an ) = 0,
so that p has n + 1 zeroes since the ai are distinct. But Exercise 62 in Section 6.2 implies that p(x)
must be the zero polynomial, so that ker(T ) = {0} and thus T is one-to-one.
37. By the Rank Theorem, since T : V ! V and T 2 : V ! V , we have
rank(T ) + nullity(T ) = dim V = rank(T 2 ) + nullity(T 2 ).
Since rank(T ) = rank(T 2 ), it follows that nullity(T ) = nullity(T 2 ), and thus that dim ker(T ) =
dim ker(T 2 ). So if we show ker(T ) ✓ ker(T 2 ), we can conclude that ker(T ) = ker(T 2 ). Choose
v 2 ker(T ); then T 2 v = T (T v) = T 0 = 0. Hence ker(T ) = ker(T 2 ).
Now choose v 2 range(T ) \ ker(T ); we want to show v = 0. Since v 2 range(T ), there is w 2 V such
that v = T w. Then since v 2 ker(T ),
0 = T v = T (T w) = T 2 w,
so that w 2 ker(T 2 ) = ker(T ). So w is also in ker(T ). As a result, v = T w = 0, as we were to show.
38. (a) We have
T ((u1 , w1 ) + (u2 , w2 )) = T (u1 + u2 , w1 + w2 ) = u1 + u2
= (u1
w1 ) + (u2
T (c(u, w)) = T (cu, cw) = cu
w1
w2
w2 ) = T (u1 , w1 ) + T (u2 , w2 )
cw = c(u
w) = cT (u, w).
(b) By definition, U + W = {u + w : u 2 U, w 2 W }. So if u + w 2 U + W , we have
T (u,
w) = u + w,
so that U + W ⇢ range(T ). But since T (u, w) = u w 2 U + W , we also have range(T ) ⇢ U + W .
Thus range(T ) = U + W .
6.6. THE MATRIX OF A LINEAR TRANSFORMATION
507
(c) Suppose v = (u, w) 2 ker(T ). Then u w = 0, so that u = w and thus u 2 U \ W , and
v = (u, u). So
ker(T ) = {(u, u) : u 2 U } ⇢ U ⇥ W
is a subspace of U ⇥ W . Then
S : U \ W ! ker(T ) : u 7! (u, u)
is clearly a linear transformation that is one-to-one and onto. Thus S is an isomorphism between
U \ W and ker T .
(d) From Exercise 43(b) in Section 6.2, dim(U ⇥ W ) = dim U + dim W . Then the Rank Theorem
gives
rank(T ) + nullity(T ) = dim(U ⇥ W ) = dim U + dim W.
But from parts (a) and (b), we have
nullity(T ) = dim ker(T ) = dim(U \ W ).
rank(T ) = dim range(T ) = dim(U + W ),
Thus
rank(T ) + nullity(T ) = dim(U + W ) + dim(U \ W ) = dim U + dim W.
dim(U \ W ).
Rearranging this equation gives dim(U + W ) = dim U + dim W
6.6
The Matrix of a Linear Transformation
1. Computing T (v) directly gives
T (v) = T (4 + 2x) = 2
To compute the matrix [T ]C
[T (1)]C = [0
x]C =

0
,
1
B,
4x.
we compute the value of T on each basis element of B:
[T (x)]C = [1
0x]C =

1
0
)
[T ]C
Then
[T ]C
B [4
+ 2x]B =

0
1
1
0

4
=
2

=
B
2
= [2
4
⇥
[T (1)]C
⇤
[T (x)]C =

0
1
1
.
0
4x]C = [T (4 + 2x)]C .
2. Computing T (v) directly gives (as in Exercise 1)
T (v) = T (4 + 2x) = 2
4x.
In terms of the basis B, we have 4 + 2x = 3(1 + x) + (1 x), so that

3
[4 + 2x]B =
.
1
Now,
[T (1 + x)]C = [1
x]C =

1
,
1
[T (1
x)]C = [ 1
[T ]C
Then
[T ]C
B [4 + 2x]B =

1
1
1
1

3
=
1
B

x]C =
=
⇥

1
,
1
[T (1 + x)]C
2
= [2
4
)
[T (1
⇤
x)]C =
4x]C = [T (4 + 2x)]C .

1
1
1
.
1
508
CHAPTER 6. VECTOR SPACES
3. Computing T (v) directly gives
T (v) = T (a + bx + cx2 ) = a + b(x + 2) + c(x + 2)2 .
In terms of the standard basis B, we have
2 3
a
[a + bx + cx2 ]B = 4 b 5 .
c
Now,
2 3
1
[T (1)]C = [1]C = 405 ,
0
2 3
0
[T (x)]C = [x + 2]C = 415 ,
0
[T ]C
B
=
⇥
2 3
0
[T (x2 )]C = [(x + 2)2 ]C = 405
1
[T (1)]C
[T (x)]C
[T (x2 )]C
⇤
Then
[T ]C
2
1
2
4
0
[a
+
bx
+
cx
]
=
B
B
0
0
1
0
)
2
1
= 40
0
0
1
0
3
0
05 .
1
32 3 2 3
0 a
a
05 4 b 5 = 4 b 5 = [a + b(x + 2) + c(x + 2)2 ]C = [T (a + bx + cx2 )]C .
1
c
c
4. Computing T (v) directly gives
T (v) = T (a + bx + cx2 ) = a + b(x + 2) + c(x + 2)2 = (a + 2b + 4c) + (b + 4c)x + cx2 .
To compute [a + bx + cx2 ]B , we want to solve
a + bx + cx2 = r + s(x + 2) + t(x + 2)2 = (r + 2s + 4t) + (s + 4t)x + tx2
for r, s, and t. The x2 terms give t = c; then from the linear term b = s + 4c, so that s = b
a = r + 2(b 4c) + 4c = r + 2b 4c, so that r = a 2b + 4c. Hence
2
3
a 2b + 4c
[a + bx + cx2 ]B = 4 b 4c 5 .
c
Now,
2 3
1
[T (1)]C = [1]C = 405 ,
0
2 3
4
[T (x + 2)]C = [x + 4]C = 415 ,
0
3
16
[T ((x + 2)2 )]C = [(x + 4)2 ]C = [x2 + 8x + 16]C = 4 8 5 ,
1
so that
[T ]C
B
=
⇥
[T (1)]C
[T (x + 2)]C
[T ((x + 2)2 )]C
2
⇤
2
1
= 40
0
4
1
0
3
16
8 5.
1
4c; finally,
6.6. THE MATRIX OF A LINEAR TRANSFORMATION
509
Then
[T ]C
2
32
3
1 4 16 a 2b + 4c
2
4
8 5 4 b 4c 5
B [a + bx + cx ]B = 0 1
0 0 1
c
2
3
(a 2b + 4c) + 4(b 4c) + 16c
5
(b 4c) + 8c
=4
c
2
3
a + 2b + 4c
= 4 b + 4c 5
c
= [(a + 2b + 4c) + (b + 4c)x + cx2 ]C = [T (a + bx + cx2 )]C .
5. Directly,
T (a + bx + cx2 ) =
Using the standard basis B, we have
Now,
[T (1)]C =

1
1
so that

=
C
[T ]C
Then
[T ]C
B [a
+ bx + cx2 ]B =

1
1
⇥
[T (x)]C =
[T (1)]C

0
1
=
C
[T (x)]C
T (a + bx + cx2 ) =
Using the basis B, we have
[T (x2 )]C =

0
1
so that
[T ]C
=
C
B
Then
[T ]C
B [a
+ bx + cx2 ]B =


=
⇥
1
1

0
,
1
[T (x2 )]C =

⇤
1
[T (x2 )]C =
1
0
1
2 3


a
0 0 4 5
a
a
b =
=
1 1
a+b+c
a+b+c
c
6. Directly, as in Exercise 5,
Now,
a
.
a+b+c
2 3
a
2
4
[a + bx + cx ]B = b 5 .
c
1
,
1
=
B



[T (x)]C =
[T (x2 )]C

[T (x)]C
0
1
C

0
,
1
C
= [T (a + bx + cx2 )]C .
a
.
a+b+c
=
C
=
0
.
1
2 3
c
[a + bx + cx2 ]B = 4 b 5 .
a
1
,
1
0
1

1
,
1
⇤
[T (1)]C =
[T (1)]C =

1
1

1
1
=
C

0
,
1
1 0
.
1 1
2 3


c
1 0 4 5
(b + c)
a
b =
=
1 1
a+b+c
a+b+c
a
C
= [T (a + bx + cx2 )]C .
510
CHAPTER 6. VECTOR SPACES
7. Computing directly gives
Tv = T
We first compute the coordinates of


2
3 2 3
7+2·7
7
7
= 4 ( 7) 5 = 475 .
7
7
7
7
in B. We want to solve
7



7
1
3
=r
+s
7
2
1
for r and s; that is, we want r and s such that 7 = r + 3s and 7 = 2r
equations gives r = 2 and s = 3. Thus


7
2
=
.
7 B
3
Now,
 ✓ ◆
1
T
2
C
so that
[T ]C
Then
[T ]C
B
B

2
3
2 3
5
6
= 4 15 = 4 35 ,
2 C
2
  ✓ ◆
1
T
=
2
7
7
B
2
6
=4 3
2
C
 ✓
◆
3
T
1
 ✓
◆
3
T
1
C
C
s. Solving this pair of
2
3
2 3
1
4
= 4 35 = 4 25 ,
1 C
1
2
6
=4 3
2
3
4
25 .
1
3
2 3 2 3
 
4 
0
7
2
7
5
4
5
4
5
2
= 0 = 7 = T
3
7
1
7
7 C
.
C
8. Computing directly gives (as in the statement of Exercise 7)
2
3

a + 2b
a
Tv = T
= 4 a 5.
b
b

a
We first compute the coordinates of
in B. We want to solve
b



a
1
3
=r
+s
b
2
1
for r and s; that is, we want r and s such that a = r + 3s and b = 2r
gives r = 17 (a + 3b) and s = 17 (2a b). Thus
" #
"
#
1
a
(a + 3b)
7
= 1
.
b
b)
7 (2a
s. Solving this pair of equations
B
Now,
 ✓ ◆
1
T
2
C
so that
[T ]C
B
2
3
2 3
5
6
= 4 15 = 4 35 ,
2 C
2
  ✓ ◆
1
T
=
2
C
 ✓
◆
3
T
1
 ✓
◆
3
T
1
C
C
2
3
2 3
1
4
= 4 35 = 4 25 ,
1 C
1
2
6
=4 3
2
3
4
25 .
1
6.6. THE MATRIX OF A LINEAR TRANSFORMATION
511
Then
[T ]C
2
3
" #
#
6
4 "1
a
6
7 7 (a + 3b)
=4 3
25 1
B
b
b)
7 (2a
B
2
1
2
3
6
4
b)
7 (a + 3b) + 7 (2a
6
7
= 4 37 (a + 3b) 27 (2a b)5
2
1
b)
7 (a + 3b)
7 (2a
2
3 2
3
" " ##
2a + 2b
a + 2b
a
6
7 6
7
= 4 a b5 = 4 a 5 = T
.
b
C
b
b
C
9. Computing directly gives
T v = T A = AT =

a
b
c
.
d
Since A = aE11 + bE12 + cE21 + dE22 , the coordinate vector of A in B is
2 3
a
6b7
7
[A]B = 6
4c5 .
d
Now,
T
[T E11 ]C = [E11
]C
T
[T E12 ]C = [E12
]C
T
[T E21 ]C = [E21
]C
T
[T E22 ]C = [E22
]C
Therefore
[T ]C
B
Then
[T ]C
=
⇥
[T E11 ]C
2
1
60
6
B [A]B = 4
0
0
2 3
2 3
1
1
607
607
7
6 7
= [E11 ]C = 6
405 = 405 ,
0 C
0
2 3
0

607
0 0
= [E21 ]C =
= 6 7,
1 0 C 415
0
2 3
0

617
0 1
= [E12 ]C =
= 6 7,
0 0 C 405
0
2 3
0

607
0 0
= [E22 ]C =
= 6 7.
0 1 C 405
1
[T E12 ]C
0
0
1
0
0
1
0
0
[T E21 ]C
2
1
⇤ 60
[T E22 ]C = 6
40
0
32 3 2 3
0
a
a

6b7 6c7
07
76 7 = 6 7 = a
05 4 c 5 4 b 5
b
1
d
d
c
d
C
0
0
1
0
0
1
0
0
3
0
07
7.
05
1
= [T A]C .
512
CHAPTER 6. VECTOR SPACES
10. Computing directly gives
T
Tv = TA = A =

a
b
c
.
d
Since A = aE11 + bE12 + cE21 + dE22 , the coordinate vector of A in B is
2 3
d
6c7
7
[A]B = 6
4b5 .
a
Now,
T
[T E22 ]C = [E22
]C = [E22 ]C
T
[T E21 ]C = [E21
]C = [E12 ]C
T
[T E12 ]C = [E12
]C = [E21 ]C
T
[T E11 ]C = [E11
]C = [E11 ]C
Therefore
[T ]C
B
=
⇥
[T E22 ]C
Then
[T ]C
2
0
60
6
B [A]B = 4
1
0
[T E21 ]C
1
0
0
0
2 3
0
607
0 0
=
= 6 7,
0 1 C 415
0
2 3
1

607
0 1
=
= 6 7,
0 0 C 405
0
2 3
0

617
0 0
=
= 6 7,
1 0 C 405
0
2 3
2 3
1
0
6 07
6 07
7
6 7
=6
4 05 = 4 05 .
0 C
1

2
0
⇤ 60
[T E11 ]C = 6
41
0
[T E12 ]C
32 3 2 3
0
d
c

6c7 6b7
07
76 7 = 6 7 = a
05 4 b 5 4 d 5
b
1
a
a
0
1
0
0
c
d
C
1
0
0
0
0
1
0
0
3
0
07
7.
05
1
= [T A]C .
11. Computing directly gives
Tv = TA =

a
c
b
d

1
1
1
1

1
1
1
1

a b
c d

a b
=
c d
b
d
a
c

a
c
c
a
Since A = aE11 + bE12 + cE21 + dE22 , the coordinate vector of A in B is
2 3
a
6b7
7
[A]B = 6
4c5 .
d
b
d

d
c
=
b
a
b
d
d
b
a
.
c
6.6. THE MATRIX OF A LINEAR TRANSFORMATION
513
Now,
[T E11 ]C = [E11 B
[T E12 ]C = [E12 B
[T E21 ]C = [E21 B
[T E22 ]C = [E22 B
BE11 ]C =

BE12 ]C =

BE21 ]C =

BE22 ]C =

0
1
1
0
1
0
0
1
1
0
0
1
0
1
1
0
2
C
B
=
⇥
Then
[T ]C
[T E11 ]C
2
0
6 1
6
B [A]B = 4
1
0
[T E12 ]C
1
0
0
1
1
0
0
1
[T E21 ]C
6
=6
4
2
C
6
=6
4
2
C
6
=6
4
2
C
Therefore
[T ]C
3
0
17
7,
15
0
3
1
07
7,
05
1
3
1
07
7,
05
1
3
0
17
7.
15
0
6
=6
4
2
0
⇤ 6 1
[T E22 ]C = 6
4 1
0
32 3 2
0 a
c
6 7 6
17
7 6 b 7 = 6d
15 4 c 5 4a
0
d
b
3
b

a7
7= c
d5
a
c
1
0
0
1
b
d
d
b
1
0
0
1
a
c
C
3
0
17
7.
15
0
= [T A]C .
12. Computing directly gives

a
Tv = TA =
c
b
d

a
b

c
0
=
d
c b
b
c
0
.
Using the given basis B for M22 , the coordinate vector of A is
2 3
a
6b7
7
[A]B = 6
4c5 .
d
Now,
[T E11 ]C = [E11
[T E21 ]C = [E21
2 3
0
607
T
7
E11 ]C = 6
405 ,
0
T
E21
]C = [E21
[T E12 ]C = [E12
2
3
0
6 17
7
E12 ]C = 6
4 15 ,
0
T
E12
]C = [E12
[T E22 ]C = [E22
2
3
0
6 17
7
E21 ]C = 6
4 15 ,
0
2 3
0
607
T
7
E22 ]C = 6
405 .
0
514
CHAPTER 6. VECTOR SPACES
Therefore
[T ]C
Then
[T ]C
=
B
⇥
[T E22 ]C
2
0
60
6
B [A]B = 4
0
0
[T E21 ]C
0
1
1
0
2
0
⇤ 60
[T E11 ]C = 6
40
0
[T E12 ]C
32 3 2
3
0 a
0

6 7 6
7
07
0
7 6 b 7 = 6b c7 =
05 4 c 5 4c b5
c b
0
d
0
0
1
1
0
13. (a) If a sin x + b cos x 2 W , then
0
1
1
0
b
0
1
1
0
c
0
C
3
0
07
7.
05
0
= [T A]C .
D(a sin x + b cos x) = D(a sin x) + D(b cos x) = aD sin x + bD cos x = a cos x
(b) Since
[sin x]B =

1
,
0
[cos x]B =

b sin x 2 W.
0
,
1
then
[D sin x]B = [cos x]B =
(c) [f (x)]B = [3 sin x

0
,
1
5 cos x]B =
[D cos x]B = [ sin x]B =


1
,
0
1
0

)
[D]B =

0
1
1
.
0
3
, so
5
[D(f (x))]B = [D]B [f (x)]B =

0
1

3
5
=
,
5
3
so that D(f (x)) = f 0 (x) = 5 sin x + 3 cos x. Directly, since (sin x)0 = cos x and (cos x)0 =
we also get f 0 (x) = 3 cos x 5( sin x) = 5 sin x + 3 cos x.
14. (a) If ae2x + be
2x
sin x,
2 W , then
D(ae2x + be
2x
) = D(ae2x ) + D(be
(b) Since
2x
) = aDe2x + bDe

1
[e ]B =
,
0
2x
2x
= (2a)e2x + ( 2b)e
]B =


0
,
2
)


1
2
=
,
3
6
[e
2x
2x
]B =
2x
2 W.
0
,
1
then
[De2x ]B = [2e2x ]B =
(c) [f (x)]B = [e2x
3e
2x

2
,
0
]B =

2x
[De
]B = [ 2e

2
0
0
.
2
1
, so
3

2
[D(f (x))]B = [D]B [f (x)]B =
0
so that D(f (x)) = f 0 (x) = 2e2x + 6e
15. (a) Since
[D]B =
2 3
1
[e2x ]B = 405 ,
0
3x
0
2
. Directly, f 0 (x) = e2x · 2
2 3
0
[e2x cos x]B = 415 ,
0
3e
2x
· ( 2) = 2e2x + 6e
2 3
0
[e2x sin x]B = 405 ,
1
3x
.
6.6. THE MATRIX OF A LINEAR TRANSFORMATION
515
then
2 3
2
[De2x ]B = [2e2x ]B = 405 ,
0
3
0
[D(e2x cos x)]B = [2e2x cos x e2x sin x]B = 4 25 ,
1
2 3
0
[D(e2x sin x)]B = [2e2x sin x + e2x cos x]B = 415 .
2
Thus
2
3
0 0
2 15 .
1 2
2
[D]B = 40
0
(b) [f (x)]B = [3e2x
2
2
3
3
e2x cos x + 2e2x sin x]B = 4 15, so
2
2
2
[D(f (x))]B = [D]B [f (x)]B = 40
0
32 3 2 3
0 0
3
6
2 15 4 15 = 405 ,
1 2
2
5
so that D(f (x)) = f 0 (x) = 6e2x + 5e2x sin x. Directly,
f 0 (x) = 3e2x · 2
(2e2x cos x
e2x sin x) + (4e2x sin x + 2e2x cos x) = 6e2x + 5e2x sin x.
16. (a) Since
2 3
1
607
7
[cos x]B = 6
405 ,
0
2 3
0
617
7
[sin x]B = 6
405 ,
0
2 3
0
607
7
[x cos x]B = 6
415 ,
0
2 3
0
6 07
7
[x sin x]B = 6
4 05 ,
1
then
2
6
[D cos x]B = [ sin x]B = 6
4
[D(x cos x)]B = [cos x
Thus
2
6
x sin x]B = 6
4
2
3
2 3
0
1
607
17
7 , [D sin x]B = [cos x]B = 6 7 ,
405
05
0
0
3
2 3
1
0
6 17
07
7 , [D(x sin x)]B = [sin x + x cos x]B = 6 7 .
4 15
05
1
0
0
6 1
[D]B = 6
4 0
0
1
0
0
0
3
1 0
0 17
7.
0 15
1 0
516
CHAPTER 6. VECTOR SPACES
2 3
1
607
6
(b) [f (x)]B = [cos x + 2x cos x]B = 4 7
, so
25
0
2
0
6 1
[D(f (x))]B = [D]B [f (x)]B = 6
4 0
0
so that D(f (x)) = f 0 (x) = 2 cos x
f 0 (x) =
sin x
32 3 2
1 0 1
6 7 6
0 17
7 6 07 = 6
0 15 425 4
1 0 0
1
0
0
0
3
2
17
7,
05
2
2x sin x. Directly,
sin x + 2 cos x
2x sin x = 2 cos x
sin x
2x sin x.
17. (a) Let p(x) = a + bx; then
(S
✓
T )p(x) = S(T p(x)) = S
Then
[(S
T )(1)]D =

1
1
p(0)
p(1)
◆

1
,
1
=
D
so that
[S
(b) We have
[T (1)]C =

1
1
so that
T ]D
=
C

 
1
S
0
so that
=
D

1
2
=
D

B
= [S]D
=
B
1
1

1
1
C
=

=

B
1
2

2
1
=

=
D

2b
.
b
2
,
1
2
.
1

0
1
C
0
,
1
0
.
1
 
0
S
1
1
,
2
C [T ]C
T )(x)]D =
[T (x)]C =
Then
T ]D


a
a 2(a + b)
a
=
=
a+b
2a (a + b)
a

1
,
1
[S]D
[S

[(S
=
B
[T ]C
Next,
=S
=
D

2
1
=
D

2
,
1
2
.
1
1
2
2
1

1
1
0
=
1

1
1
2
,
1
which matches the result from part (a).
18. (a) Let p(x) = a + bx; then
(S T )p(x) = S(T p(x)) = S(a + b(x + 1)) = S((a + b) + bx) = (a + b) + b(x + 1) = (a + 2b) + bx.
Then
[(S
so that
2 3
2 3
1
1
T )(1)]D = 405 = 405 ,
0 D
0
[S
T ]D
B
[(S
2
1
= 40
0
2 3
2 3
2
2
T )(x)]D = 415 = 415 ,
0 D
0
3
2
15 .
0
6.6. THE MATRIX OF A LINEAR TRANSFORMATION
(b) We have
2 3
2 3
1
1
[T (1)]C = [1]C = 405 = 405 ,
0 C
0
so that
517
2 3
2 3
1
1
[T (x)]C = [x + 1]C = 415 = 415 ,
0 C
0
[T ]C
B
Next,
2
3
1
15 .
0
1
= 40
0
2 3
1
[S(1)]D = [1]D = 405 ,
0
2 3
1
[S(x)]D = [x + 1]D = 415 ,
0
2 3
1
[S(x2 )]D = [(x + 1)2 ]D = [x2 + 2x + 1]D = 425 ,
1
so that
[S]D
2
1
= 40
0
C
Then
[S
T ]D
B
= [S]D
C [T ]C
B
which matches the result from part (a).
3
1
25 .
1
1
1
0
2
1
= 40
0
32
1
1
25 40
1
0
1
1
0
3 2
1
1
15 = 40
0
0
3
2
15 ,
0
19. In Exercise 1, the basis for P1 in both the domain and codomain was already the standard basis E,
so we can use the matrix computed there:

0 1
[T ]E E =
.
1 0
Since this matrix has determinant 1, it is invertible, so that T is invertible as well. Then
[T
So T
1
1
(a + bx)]E
(a + bx) =
E
= ([T ]E
E)
1

a
=
b

=

0
1
1
0
1


a
0
=
b
1
1
0

a
=
b

b
.
a
b + ax.
20. From Exercise 5,
[T ]C
B
1
1
0
1
0
.
1
Since this matrix is not square, it is not invertible, so that T is not invertible.
21. We first compute [T ]E
2 3
1
[T (1)]E = [1]E = 405 ,
0
so that
where E = {1, x, x2 } is the standard basis:
2 3
2 3
2
4
[T (x)]E = [x + 2]E = 415 , [T (x2 )]E = [(x + 2)2 ]E = [x2 + 4x + 4]E = 445 ,
0
1
E,
[T ]E
E
2
1
= 40
0
2
1
0
3
4
45 .
1
518
CHAPTER 6. VECTOR SPACES
Since this is an upper triangular matrix with nonzero diagonal entries, it is invertible, so that T is
invertible as well. To find T 1 , we have
[T
So T
1
(a + bx + cx2 )]E
1
(a + bx) = (a
22. We first compute [T ]E
= ([T ]E
E
2 3 2
a
1
14 5
b = 40
E)
c
0
2b + 4c) + (b
E,
3
4
45
1
2
1
0
2 3
a
4b5
c
2
1
= 40
0
1
4c)x + cx2 = a + b(x
2
1
0
2) + c(x
32 3 2
3
4 a
a 2b + 4c
45 4 b 5 = 4 b 4c 5 .
1
c
c
2)2 = p(x
2).
where E = {1, x, x2 } is the standard basis:
2 3
0
[T (1)]E = [0]E = 405 ,
0
2 3
1
[T (x)]E = [1]E = 405 ,
0
so that
[T ]E
E
2
0
= 40
0
1
0
0
2 3
0
[T (x2 )]E = [2x]E = 425 ,
0
3
0
25 .
0
Since this is an upper triangular matrix with a zero entry on the diagonal, it has determinant zero, so
is not invertible; therefore, T is not invertible either.
23. We first compute [T ]E
E,
where E = {1, x, x2 } is the standard basis:
2 3
1
[T (1)]E = [1 + 0]E = [1]E = 405 ,
0
2 3
1
[T (x)]E = [x + 1]E = 415 ,
0
so that
[T ]E
E
2
1
= 40
0
1
1
0
2 3
0
[T (x2 )]E = [x2 + 2x]E = 425 ,
1
3
0
25 .
1
Since this is an upper triangular matrix with nonzero diagonal entries, it is invertible, so that T is
invertible as well. To find T 1 , we have
[T
So T
1
2
(a + bx + cx )]E
1
E
(a + bx + cx2 ) = (a
= ([T ]E
E)
1
2 3 2
a
1
4 b 5 = 40
c
0
b + 2c) + (b
1
1
0
2c)x + cx2 .
3
0
25
1
2 3
a
4b5
c
2
1
= 40
0
1
1
1
0
32 3 2
3
2 a
a b + 2c
25 4 b 5 = 4 b 2c 5 .
1
c
c
6.6. THE MATRIX OF A LINEAR TRANSFORMATION
24. We first compute [T ]E
so that
E,
519
where E = {E11 , E12 , E21 , E22 } is the standard basis:
2 3
3
✓

◆

6 27
1 0 3 2
3 2
[T E11 ]E =
=
= 6 7,
0 0 2 1 E
0 0 E 4 05
0
2 3
2
✓

◆

6 17
0 1 3 2
2 1
[T E12 ]E =
=
= 6 7,
0 0 2 1 E
0 0 E 4 05
0
2 3
0
✓

◆

6 07
0 0 3 2
0 0
6
[T E21 ]E =
=
= 7,
1 0 2 1 E
3 2 E 4 35
2
2 3
0
✓

◆

6 07
0 0 3 2
0 0
6
[T E22 ]E =
=
= 7,
0 1 2 1 E
2 1 E 4 25
1
[T ]E
E
2
3
62
=6
40
0
2
1
0
0
3
0
07
7.
25
1
0
0
3
2
This matrix is block triangular, and each diagonal block is invertible, so it is invertible and therefore
T is invertible as well. To find T 1 , we have

T
So T
1
1


a
c
a
c
b
d
E
b
=
d
E

= ([T ]E
a + 2b
c + 2d
2 3 2
a
3
7 62
b
16
6 7=6
E)
4 c 5 40
d
0
2a
2c
3b
.
3d
2
1
0
0
0
0
3
2
3
0
07
7
25
1
1
2
2 3
a
6b7
6 7
4c5
d
1
6 2
=6
4 0
0
2
3
0
0
0
0
1
2
25. From Exercise 11, since the bases given there were both the standard basis,
2
3
0
1
1
0
6 1
0
0
17
7.
[T ]E E = 6
4 1
0
0
15
0
1
1
0
32 3 2
3
0
a
a + 2b
6 7 6
7
07
7 6 b 7 = 6 2a 3b 7 .
25 4 c 5 4 c + 2d5
3
d
2c 3d
This matrix has determinant zero (for example, the second and third rows are multiples of one another),
so that T is not invertible.
26. From Exercise 12, since the bases given there were both the standard basis,
2
3
0
0
0 0
60
1
1 07
7.
[T ]E E = 6
40
1
1 05
0
0
0 0
This matrix has determinant zero, so that T is not invertible.
520
CHAPTER 6. VECTOR SPACES
27. In Exercise 13, the basis is B = {sin x, cos x}. Using the method of Example 6.83,
Z

a
(a sin x + b cos x) dx = [D]B 1 B
.
b
B
Using [D]B
R
So (sin x
B
from Exercise 13, we get
Z

0
(sin x 3 cos x) dx =
1
B
3 cos x) dx =
3 sin x
1
0
1

1
=
3

0
1
1
0

1
=
3

3
.
1
cos x + C. This is the same answer as integrating directly.
28. In Exercise 14, the basis is B = {e2x , e 2x }. Using the method of Example 6.83,
Z

a
ae2x + be 2x dx = [D]B 1 B
.
b
B
Using [D]B
B
from Exercise 14, we get
"
Z
5e
So
R
2x
dx
B
5e
2x
dx =
5
2x
2e
#
0
2
2
=
0
1
" # "
1
0
= 2
0
5
#" # " #
0 0
0
=
.
1
5
5
2
2
+ C. This is the same answer as integrating directly.
29. In Exercise 15, the basis is B = {e2x , e2x cos x, e2x sin x}. Using the method of Example 6.83,
Z

a
ae2x + be 2x dx = [D]B 1 B
.
b
B
Using [D]B
Z
So
R
B
from Exercise 15, we get
e2x cos x
2e2x sin x dx
B
2
3
0 0
7
2 15
1 2
2
6
= 40
0
1
2
3 2
1
0
2
6 7 6
4 15 = 4 0
2
0
3 2 3
0
0
7 6 47
17 6
=
.
1
5
4
5
4
5
55
2
3
2
5
5
0
0
2
5
1
5
32
e2x cos x 2e2x sin x dx = 45 e2x cos x 35 e2x sin x + C. Checking via di↵erentiation gives
✓
◆0
4 2x
3 2x
4
3
e cos x
e sin x + C =
2e2x cos x e2x sin x
2e2x sin x + e2x cos x
5
5
5
5
= e2x cos x
2e2x sin x.
30. In Exercise 16, the basis is B = {cos x, sin x, x cos x, x sin x}. Using the method of Example 6.83,
2 3
a
Z
6b7
1
7
(a cos x + b sin x + cx cos x + dx sin x) dx = [D]B B 6
4c5 .
B
d
Using [D]B
So
R
from Exercise 16, we get
2
0
Z
6 1
(x cos x + x sin x) dx = 6
4 0
B
0
B
(x cos x + x sin x) dx = cos x + sin x
(cos x + sin x
3
1 0
0 17
7
0 15
1 0
1
0
0
0
1
2 3 2
0
0
607 61
6 7=6
415 40
1
0
1
0
0
0
1
0
0
1
32 3 2
0 0
6 7 6
17
7 607 = 6
15 415 4
0 1
3
1
17
7
15
1
x cos x + x sin x + C. Checking via di↵erentiation gives
0
x cos x + x sin x + C) =
sin x + cos x
= x cos x + x sin x.
cos x + x sin x + sin x + x cos x
6.6. THE MATRIX OF A LINEAR TRANSFORMATION
31. With E = {e1 , e2 }, we have

4·0
[T e1 ]E =
1+5·0
=
E

0
1
=
E

0
,
1
521
[T e2 ]E =
Therefore
[T ]E =

0
1

4·1
0+5·1
=
E

4
5

=
E
4
.
5
4
.
5
Using the method of Example 6.86(b), the eigenvectors of [T ]E will give us a diagonalizing basis C.
Using methods from Chapter 4, we find the eigenvalues and corresponding eigenvectors to be


1
4
,
.
1 = 4, v1 =
2 = 1, v2 =
1
1
Thus
C=
⇢
1
,
1

32. With E = {e1 , e2 }, we have

1 0
[T e1 ]E =
1+0
4
1
=
E
, and [T ]C =

1
1
=
E


1
1
1
,
1
Therefore
[T ]E =
1
4
1
[T ]E
[T e2 ]E =

1
1


0 1
0+1

4
4
=
1
0
1
1
=
E

1
1
0
1

=
E
1
.
1
1
.
1
Using the method of Example 6.86(b), the eigenvectors of [T ]E will give us a diagonalizing basis C.
Using methods from Chapter 4, we find the eigenvalues and corresponding eigenvectors to be


i
i
,
i, v2 =
.
1 = 1 + i, v1 =
2 =1
1
1
Since the eigenvectors are not real, T is not diagonalizable.
33. With E = {1, x}, we have

4
[T (1)]E = [4 + x]E =
1
E


4
=
,
1
2
[T (x)]E = [2 + 3x]E =
3
Therefore
[T ]E =

4
1
=
E

2
.
3
2
.
3
Using the method of Example 6.86(b), the eigenvectors of [T ]E will give us a diagonalizing basis C.
Using methods from Chapter 4, we find the eigenvalues and corresponding eigenvectors to be


2
1
,
.
1 = 5, v1 =
2 = 2, v2 =
1
1
Thus
C=
⇢
2
,
1

34. With E = {1, x, x2 }, we have
1
1
, and [T ]C =
2 3
2 3
1
1
[T (1)]E = [1]E = 405 = 405 ,
0 E
0

2
1
1
1
1
[T ]E

2
1

1
5
=
1
0
0
.
2
2 3
2 3
1
1
[T (x)]E = [x + 1]E = 415 = 415 ,
0 E
0
2 3
2 3
1
1
[T (x2 )]E = [(x + 1)2 ]E = [x2 + 2x + 1]E = 425 = 425 .
1 E
1
522
CHAPTER 6. VECTOR SPACES
Therefore
2
1
[T ]E = 40
0
1
1
0
3
1
25 .
1
Using the method of Example 6.86(b), the eigenvectors of [T ]E will give us a diagonalizing basis C. Since
this is an upper triangular matrix with 1s on the diagonal, the only eigenvalue
is 1; using methods from
2 3
1
Chapter 4 we find the eigenspace to be one-dimensional with basis 405. Thus T is not diagonalizable.
0
35. With E = {1, x}, we have

1
[T (1)]E = [1 + x · 0]E = [1]E =
0
E


1
=
,
0
0
[T (x)]E = [x + x · 1]E = [2x]E =
2
Therefore
[T ]E =

1
0
=
E

0
.
2
0
.
2
Thus T is already diagonal with respect to the standard basis E.
36. With E = {1, x, x2 }, we have
2 3
2 3
1
1
[T (1)]E = [1]E = 405 = 405 ,
0 E
0
2 3
2 3
2
2
[T (x)]E = [3x + 2]E = 435 = 435 ,
0 E
0
Therefore
3
2 3
4
9
[T (x2 )]E = [(3x + 2)2 ]E = [9x2 + 12x + 4]E = 4125 = 4125 .
9 E
4
2
1
[T ]E = 40
0
2
3
0
2
3
4
125 .
9
Using the method of Example 6.86(b), the eigenvectors of [T ]E will give us a diagonalizing basis C.
Since this is an upper triangular matrix with 1s on the diagonal, the eigenvalues are 1, 3, and 4; using
methods from Chapter 4 we find the corresponding eigenvectors to be
2 3
2 3
2 3
1
1
1
4
5
4
5
4
0
1
25 .
=
1,
v
=
,
=
3,
v
=
,
=
9,
v
=
1
1
2
2
3
3
0
0
1
Thus
8 2 3 2 3 2 39
2
1
1 =
1
< 1
C = 405 , 415 , 425 , and [T ]C = 40
:
;
0
0
1
0
3
1
25
1
1
2
1
[T ]E 40
0
1
1
0
3 2
1
1
25 = 40
1
0
0
3
0
3
0
05 .
9

d1
d2
, and let T be the reflection in `. Let d0 =
. As in Example
d2
d1
6.85, we may as well assume for simplicity that d is a unit vector; then d0 is as well. Continuing
37. Let ` have direction vector d =

1
1
0
6.6. THE MATRIX OF A LINEAR TRANSFORMATION
523
to follow Example 6.85, D = {d, d0 } is a basis for R2 . Since d0 is orthogonal to d, it follows that
T d0 = d0 . Thus


1
0
[T (d)]D = [d]D =
,
[T (d0 )]D = [ d0 ]D =
.
0
1
Thus
[T ]D =

1
0
0
.
1
Now, [T ]E = PE D [T ]D PD E , so we must find PE D . The columns of this matrix are the expression
of the basis vectors of D in the standard basis E, so we have
"
#
d1
d2
PE D =
.
d2
d1
Since we assumed d was a unit vector, this is an orthogonal matrix, so that
[T ]E = PE
D [T ]D PD
E
= PE
D [T ]D PE
1
D
=

= PE
d1
d2
T
D [T ]D PE D
d2
d1

1
0
0
1

 2
d2
d
d22
= 1
d1
2d1 d2
d1
d2
2d1 d2
.
d22 d21
Note that this is the same as the answer from Exercise 26 in Section 3.6, with the assumption that d
is a unit vector.
38. Let T be the projection onto W . From Example 5.11, we have an orthogonal basis for R2 ,
8
2 3
2 3
2 39
1
1
1 =
<
u1 = 415 , u2 = 4 15 , u3 = 4 15 ,
:
;
0
1
2
where {u1 , u2 } is a basis for W and {u3 } is a basis for W ? . Then
2 3
2 3
2 3
1
0
0
[T (u1 )]D = 405 , [T (u2 )]D = 415 , [T (u3 )]D = 405 ,
so that
0
0
0
Further, since the columns of PE
E,
2
PE
Thus
2
[T ]E = PE
2
1
6
= 41
0
3
3
If v = 4 15, then
2
D
D
1
[T ]D = 40
0
0
1
0
3
0
05 .
0
are the expression of the basis vectors of D in the standard basis
1
6
= 41
0
1
1
1
3
1
7
15
2
PE
1
E = PE D [T ]D PE
32
32
1
1 1 0 0
2
76
76 1
15 40 1 05 4 3
1
2 0 0 0
6
1
)
D [T ]D PD
1
1
1
2
2
6
projW v = [T ]E [v]E = 4
5
6
1
6
1
3
1
6
5
6
1
3
2
1
2
1
3
1
6
1
2
1
3
1
6
3
0
17
.
35
3 2
0
6
17
=4
35
5
6
1
6
1
3
1
6
5
6
1
3
6
=4
D
D
1
2
1
3
1
6
1
3
32
1
3
17 6
35 4
1
3
which is the same as was found in Example 11 in Section 5.2.
3 2
3
7 6
15 = 4
2
1
3
3
5
3
17
,
35
2
3
3
1
3
17
.
35
1
3
524
CHAPTER 6. VECTOR SPACES
39. Let B = {v1 , . . . , vn }, and suppose that A is such that A[v]B = [T v]C for all v 2 V . Then A[vi ]B =
[T vi ]C . But [vi ]B = ei since the vi are the basis vectors of B, so this equation says that Aei = [T vi ]C .
The left-hand side of this equation is the ith column of A, while the right-hand side is the ith column
of [T ]C B . Thus those ith columns are equal; since this holds for all i, we get A = [T ]C B .
40. Suppose that [v]B 2 null([T ]C
B ).
That means that
0 = [T ]C
B [v]B
= [T v]C ,
so that [T v]C = 0 and thus v 2 ker(T ). Similarly, if v 2 ker(T ), then
0 = [T v]C = [T ]C
B [v]B ,
so that [v]B 2 null([T ]C B ). Then the map V ! Rn : v 7! [v]B is an isomorphism mapping ker(T ) to
null([T ]C B ), so the two subspaces are isomorphic and thus have the same dimension. So nullity(T ) =
nullity([T ]C B ).
41. Suppose B = {b1 , . . . , bn }. Then
[T ]C
B [bi ]B
= [T (bi )]C
by definition of the matrix of T with respect to these bases, and denote by ai the columns of [T ]C B .
Then rank([T ]C B ) = dim col[T ]C B , which is the number of linearly independent ai , and rank(T ) =
dim range(T ), which is the number of linearly independent T (vi ). Now, since
ai = [T ]C
B [bi ]B
= [T (bi )]C
we see that {ai1 , ai2 , . . . , aim } are linearly independent if and only if {T (bi1 ), . . . , T (b)im } are linearly
independent, so col(A) = span(ai1 , ai2 , . . . , aim ) if and only if range(T ) = span(T (bi1 ), . . . , T (b)im )
and therefore rank(A) = rank(T ).
42. T is diagonalizable if and only if there is some basis C and matrix P such that [T ]C = P 1 AP is
diagonal. Thus if T is diagonalizable, so is A. For the reverse, if A is diagonalizable, then the basis C
whose elements are the columns of P is a basis in which [T ]C is diagonal.
43. Let dim V = n and dim W = m. Then [T ]C B is an m ⇥ n matrix, and by Theorem 3.26 in Section
3.5, rank([T ]C B ) + nullity([T ]C B ) = n. But by Exercises 40 and 41, rank([T ]C B ) = rank(T ) and
nullity([T ]C B ) = nullity(T ); since dim V = n, we get
dim V = rank(T ) + nullity(T ).
44. Claim that [T ]C 0
B0
= PC 0
C [T ]C
B PB
B0 .
To prove this, it is enough to show that for all x 2 V ,
[x]C 0 = PC 0
C [T ]C
B PB
B0 [x]B0
since the matrix of T with respect to B 0 and C 0 is unique by Exercise 39. And indeed
PC 0
C [T ]C
B PB
B0 [x]B0
= PC 0
C [T ]C
45. For T 2 L(V, W ), define '(T ) = [T ]C
it is linear:
B
B
(PB
B0 [x]B0 )
= PC 0
C
([T ]C
B [x]B )
= PC 0
C [x]C
= [x]C 0 .
2 Mmn . Claim that ' is an isomorphism. We must first show
'(S + T ) = [(S + T )]C B
⇥
⇤
= [(S + T )b1 ]C · · · [(S + T )bn ]C
⇥
⇤
= [Sb1 ]C + [T b1 ]C · · · [Sbn ]C + [T bn ]C
⇥
⇤ ⇥
⇤
= [Sb1 ]C · · · [Sbn ]C + [T b1 ]C · · · [T bn ]C
= '(S) + '(T ).
Exploration: Tiles, Lattices, and the Crystallographic Restriction
525
Similarly
'(cS) = [(cS)]C B
⇥
⇤
= [(cS)b1 ]C · · · [(cS)bn ]C
⇥
⇤
= c[Sb1 ]C · · · c[Sbn ]C
⇥
⇤
= c [Sb1 ]C · · · [Sbn ]C
= c'(S).
Next we show that ker ' = 0, so that ' is one-to-one. Suppose that '(T ) = 0. That means that
[T ]C B = [0]mn , so that nullity([T ]C B ) = n. By Exercise 40, this is equal to nullity(T ). Thus
nullity(T ) = n = dim V , so that T is the zero linear transformation, showing that ' is one-to-one.
Finally, since dim Mmn = mn, we will have shown that ' is an isomorphism if we prove that
dim L(V, W ) = mn. But if {v1 , . . . , vn } and {w1 , . . . , wm } are bases for V and W , then a linear
map from V to W sends each vi to some linear combination of the wj , so it is a sum of maps that send
vi to wk and the other vi to zero. There are mn of these maps, which are clearly linearly independent,
so that dim L(V, W ) = mn.
46. From Exercise 45 with W = R, we have dim W = dim R = 1, so that M1n ⇠
= V . Thus
V ⇤ = L(V, W ) = L(V, R) ⇠
= M1n ⇠
= V.
Exploration: Tiles, Lattices, and the Crystallographic Restriction
1. Let P be the pattern in Figure 6.15. We are given, from the figure, that P +u = P and that P +v = P .
Therefore if a is a positive integer, then P + au = (P + u) + (a 1)u = P + (a 1)u, and by induction
this is just P . Similarly if b is a positive integer then P + bv = (P + v) + (b 1)v = P + (b 1)v,
which again by induction is equal to P . If a is a negative integer, then since P + ( a)u = P , it follows
that P = P ( a)u = P + au, and similarly for b < 0 and v.
2. The lattice points correspond to centers
of the trefoils in Figure 6.15.
1.5
1.
0.5
!3
!2
v
u
!1
1
2
3
!0.5
!1.
!1.5
✓
3. Let ✓ be the smallest positive angle through which the lattice exhibits rotational symmetry. Let RO
denote the operation of rotation through an angle ✓ > 0 around a center of rotation O, and let P be a
pattern or lattice. Now we have
n✓
✓
RO
(P ) = RO
n
(P ) = P, n 2 Z, n > 0,
since rotation through n✓ is the same as applying a rotation through ✓ n times. If n < 0, then rotating
through n✓ gives a figure which, if rotated back through ( n)✓, is P , so it must be P as well. This
shows that P is rotationally symmetric under any integer multiple of ✓.
Next, let m be the smallest positive integer such that m✓ 360 , and suppose that m✓ = 360 + '.
360 +'
'
360
is an integer, so
Note that RO
(P ) = P , so that RO
(P ) = RO
(P ). We want to show that 360
✓
it suffices to show that ' = 0. Now, since m is minimal, we know that
so that m✓
' > m✓
✓ and thus
m✓
' = 360 > (m
'>
✓, so that ' < ✓. But
360
m✓
P = RO
(P ) = RO
+'
1)✓,
'
(P ) = RO
(P ),
and we see that the lattice is rotationally symmetric under a rotation of
the smallest positive such angle, it follows that ' = 0. Thus ✓ = 360
m .
< ✓0 . Since '
0 and ✓ is
526
CHAPTER 6. VECTOR SPACES
4. The smallest positive angle of rotational symmetry in the lattice of Exercise 2 is 60 ; this rotational
symmetry is also present in the original figure, in Figure 6.14 or 6.15.
5. Some examples are below:
n!1
n!2
1
!3
!2
1
v
u 1
!1
2
3
!3
!2
u 1
!1
!1
n!4
n!6
3
!3
!2
u 1
!1
2
3
1
u 1
!1
2
!1
!1
v
u 1
v
!2
1
v
v
1
!3
n!3
2
3
!3
!2
!1
!1
2
3
!1
It is impossible to draw a lattice with 8-fold rotational symmetry, as the rest of this section will show.
6. (a) With ✓ = 60 , we have
u=
and
[R✓ ]E =
"
" #
1
0
,
v=
cos ✓
sin ✓
sin ✓
cos ✓
#
"
=
1
p2
3
2
"
#
1
p2
3
2
,
p
3
2
1
2
#
.
(b) From the discussion in part (a),
R✓ (u) =
R✓ (v) =
With B = {u, v}, we have

0
[R✓ (u)]B = [v]B =
,
1
"
"
1
p2
3
2
1
p2
3
2
p
#" #
"
#
3
1
1
2
p2
=
= v,
3
1
0
2
2
# " #
p #"
3
1
1
2
p2
p2 =
=
3
3
1
2
2
2
[R✓ (v)]B = [ u + v]B =

1
1
u + v.
)
[R✓ ]B =

0
1
1
.
1
7. Since the lattice is invariant under R✓ , u and v are mapped to lattice points under this rotation. But
each lattice point is by definition au + bv where a and b are integers. Thus



a
b
a b
[R✓ (u)]B = [au + cv]B =
, [R✓ (v)]B = [bu + dv]B =
)
[R✓ ]B =
,
c
d
c d
and the matrix entries are integers.
8. Let E = {e1 , e2 } and B = {u, v}. Then by Theorem 6.29 in Section 6.6, [R✓ ]E = P 1 [R✓ ]B P . But this
means that [R✓ ]E ⇠ [R✓ ]B (that is, the two matrices are similar). By Exercise 35 in Section 4.4, this
implies that
2 cos ✓ = tr[R✓ ]E = tr[R✓ ]B = a + d,
which is an integer.
6.7. APPLICATIONS
527
9. If 2 cos ✓ = m is an integer, then m must be 0, ±1, or ±2, so the angles are angles whose cosines are
0, ± 21 , or ±1; these angles are
✓ = 60 , 90 , 120 , 180 , 240 , 270 , 300 , 360 ,
corresponding to
n = 1, 2, 3, 4, or 6.
These are the only possible n-fold rotational symmetries.
10. Left to the reader.
6.7
Applications
1. Here a = 3, so by Theorem 6.32, {e3t } is a basis for solutions, which therefore have the form
y(t) = ce3t . Since y(1) = ce3 = 2, we have c = 2e 3 , so that y(t) = 2e 3 e3t = 2e3t 3 .
2. Here a = 1, so by Theorem 6.32, {e t } is a basis for solutions, which therefore have the form x(t) = ce t .
Since x(1) = ce 1 = 1, we have c = e, so that x(t) = ee t = e1 t .
3. y 00
7y 0 + 12y = 0 corresponds to the characteristic equation 2 7 + 12 = (
3)(
4) = 0. Then
=
3
and
=
4,
so
Theorem
6.33
tells
us
that
the
general
solution
has
the
form
y(t)
=
c1 e3t + c2 e4t .
1
2
Then
1 e4
c
=
1
y(0) = 1
c1 + c2 = 1
e3 e4
)
)
3
4
e
c
+
e
c
=
1
y(1) = 1
1
2
e3 1
c2 = 3
e
e4
so that
y(t) =
1
e3
e4
e4 e3t + e3
1
1 e4t .
4. x00 +x0 12x = 0 corresponds to the characteristic equation 2 + 12 = ( 3)( +4) = 0. Then 1 = 3
and 2 = 4, so Theorem 6.33 tells us that the general solution has the form x(t) = c1 e3t + c2 e 4t .
Then
1
c1 =
x(0) = 0
c1 + c2 = 0
7
)
)
3c1
4c2 = 1
1
x0 (0) = 1
c2 =
7
so that
y(t) =
5. f 00
2
f0
=
1
1 3t
e
7
e
4t
.
p
f = 0 corresponds to the characteristic equation 2
1 = 0, so that 1 = 1+2 5 and
p
p
p
5
(1+ 5)t/2
+ c2 e(1 5)t/2 . So
2 . Therefore by Theorem 6.33, the general solution is f (t) = c1 e
f (0) = 0
f (1) = 1
)
p
e(1+
5)/2
so that
f (t) =
e(
p
c1 +
c1 + e(1
5 1)/2
p
e 5 1
⇣
p
5)/2
e(1+
p
c2 = 0
c2 = 1
5)t/2
e(1
c1 =
)
p
5)t/2
c2 =
⌘
.
e(
p
5 1)/2
p
e 5 1
p
e( 5 1)/2
p
e 5 1
528
CHAPTER 6. VECTOR SPACES
6. g 00 2g = 0 corresponds to the characteristic equation 2 2p= 0, so that
1 =
p
Therefore by Theorem 6.33, the general solution is g(t) = c1 et 2 + c2 e t 2 . So
g(0) = 1
)
g(1) = 0
e
p
2
c1 +
c1 + e
so that
g(t) =
7. y 00
1
e2
p
1
2
1
p
⇣
2
c1 =
c2 = 1
c2 = 0
et
p
2
)
+ e2
p
c2 =
p ⌘
2 t 2
p
2 and
2
=
p
2.
1
p
e2 2
p
e2 2
p
e2 2
1
1
.
2y 0 + y = 0 corresponds to the characteristic equation 2 2 + 1 = (
1)2 = 0, so that
t
t
= 2 = 1. Therefore by Theorem 6.33, the general solution is y(t) = c1 e + c2 te . So
y(0) = 1
y(1) = 1
c1
= 1
ec1 + ec2 = 1
)
so that
y(t) = et +
1
e
e
c1 = 1
1
c2 =
)
tet = et + (1
e)tet
1
e
e
.
8. x00 + 4x0 + 4x = 0 corresponds to the characteristic equation 2 + 4 + 4 = ( + 2)2 = 0, so that
2. Therefore by Theorem 6.33, the general solution is x(t) = c1 e 2t + c2 te 2t . So
1 = 2 =
x(0) = 1
c1
= 1
2c1 + c2 = 1
)
0
x (0) = 1
c1 = 1
)
c2 = 3
so that
y(t) = e
2t
2t
+ 3te
.
9. y 00 k 2 y = 0 corresponds to the characteristic equation 2 k 2 = 0, so that
Therefore by Theorem 6.33, the general solution is y(t) = c1 ekt + c2 e kt . So
y(0) = 1
y 0 (0) = 1
so that
y(t) =
10. y 00
1
= k and
2
=
k.
k+1
2k
k 1
c2 =
2k
c1 =
c1 + c2 = 1
kc1
kc2 = 1
)
1
)
1
(k + 1)ekt + (k
2k
1)e
kt
.
2ky 0 + k 2 y = 0 corresponds to the characteristic equation 2 2k + k 2 = (
k)2 = 0, so that
kt
kt
= 2 = k. Therefore by Theorem 6.33, the general solution is y(t) = c1 e + c2 te . So
y(0) = 1
y(1) = 0
c1
= 1
c1 + c2 = 0
)
c1 = 1
)
c2 =
1
so that
y(t) = ekt
11. f 00
2
te
kt
.
2f 0 + 5f = 0 corresponds to the characteristic equation 2 2 + 5 = 0, so that 1 = 1 + 2i and
= 1 2i. Therefore by Theorem 6.33, the general solution is f (t) = c1 et cos 2t + c2 et sin 2t. Then
so that
f (0) = 1
⇣⇡⌘
f
=0
4
)
c1
= 1
e⇡/4 c2 = 0
y(t) = et cos 2t.
)
c1 = 1
c2 = 0
6.7. APPLICATIONS
529
12. h00
4h0 + 5h = 0 corresponds to the characteristic equation 2 4 + 5 = 0, so that 1 = 2 + i and
i. Therefore by Theorem 6.33, the general solution is h(t) = c1 e2t cos t + c2 e2t sin t. Then
2 = 2
0
h (t) = c1 (2e2t cos t e2t sin t) + c2 (2e2t sin t + e2t cos t), and
h(0) = 0
0
h (0) =
1
)
c1
=
2c1 + c2 =
0
1
c1 = 0
)
c2 =
1
so that
y(t) =
e2t sin t.
13. (a) We start with p(t) = cekt ; then p(0) = ce0k = c = 100, so that c = 100. Next, p(3) = 100e3k =
1600, so that e3k = 16 and k = ln316 . Hence
t/3
p(t) = 100e(ln 16/3)t = 100 eln 16
= 100 · 16t/3 .
(b) The population doubles when p(t) = 100e(ln 16/3)t = 200. Simplifying gives
ln 16
3 ln 2
3 ln 2
3
t = ln 2, so that t =
=
= .
3
ln 16
4 ln 2
4
The population doubles in
3
4
of an hour, or 45 minutes.
(c) We want to find t such that p(t) = 100e(ln 16/3)t = 1000000. Simplifying gives
ln 16
3 ln 10000
12 ln 10
3 ln 10
t = ln 10000, so that t =
=
=
⇡ 9.966 hours.
3
ln 16
4 ln 2
ln 2
14. (a) Start with p(t) = cekt ; then p(0) = ce0k = c = 76. Next, p(1) = 76e1k = 92, so that k = ln 92
76 ,
and therefore p(t) = 76et ln(92/76) . For the year 2000, we get
p(10) = 76e10 ln(92/76) ⇡ 513.5.
This is almost twice the actual population in 2000.
(b) Start with p(t) = cekt ; then p(0) = ce0k = c = 203. Next, p(1) = 203e1k = 227, so that k = ln 227
203 ,
t ln(227/203)
and therefore p(t) = 203e
. For the year 2000, we get
p(3) = 203e3 ln(227/203) ⇡ 283.8.
This is a pretty good approximation.
(c) Since the first estimate is very high but the second estimate is close, we can conclude that population growth was lower than exponential in the early 20th century, but has been close to exponential
since 1970. This could be attributed for example to three major wars during the period from 1900
to 1970.
15. (a) Let m(t) = ae ct . Then m(0) = ae c·0 = a = 50, so that m(t) = 50e
1590 years, we know that 25 mg will remain after that time, so that
1590c
= 25.
Simplifying gives 1590c = ln 12 = ln 2, so that c =
After 1000 years, the amount remaining is
ln 2
1590 .
m(1590) = 50e
m(1000) = 50e
1000 ln 2/1590
(b) Only 10 mg will remain when m(t) = 10. Solving 50e
t
ln 2
1
= ln =
1590
5
ln 5, so that t =
ct
. Since the half-life is
So finally, m(t) = 50e
⇡ 32.33 mg.
t ln 2/1590
= 10 gives
1590 ln 5
⇡ 3692 years.
ln 2
t ln 2/1590
.
530
CHAPTER 6. VECTOR SPACES
16. With m(t) = ae ct , a half-life of 5730 years means that m(5730) = ae 5730c = a2 , so that e 5730c =
ln 2
and thus c = ln(1/2)
5730 = 5730 . To find out how old the posts are, we want to find t such that
1
2
5730 ln 0.45
⇡ 6600 years.
ln 2
p
p
17. Following Example 6.92, the general equation of the spring motion is x(t) = c1 cos K t + c2 sin K t.
Then x(0) = 10 = c1 cos 0 + c2 sin 0 = c1 , so that c1 = 10. Then
p
p
p
5 10 cos 10 K
p
x(10) = 5 = 10 cos 10 K + c2 sin 10 K , so that c2 =
.
sin 10 K
m(t) = ae
t ln 2/5730
= 0.45a, which gives t =
Therefore the equation of the spring, giving its length at time t, is
p
p
p
5 10 cos 10 K
p
c2 = 10 cos K t +
sin K t.
sin 10 K
k
and m = 50, we have K = 50
, so that
r
r
p
p
k
k
x(t) = c1 cos K t + c2 sin K t = c1 cos
t + c2 sin
t.
50
50
18. From the analysis in Example 6.92, since K =
Since the period of sin bt and cos bt are each
q
k
so that 50
= ⇡5 and thus k = 2⇡ 2 .
k
m
2⇡
b ,
the fact that the period is 10 means that p2⇡
k/50
= 10,
19. The di↵erential equation for the pendulum motion is the same as the equation of spring motion, so we
apply the analysis from Example 6.92. Recall that the period of sin bt and cos bt are each 2⇡
b .
(a) Since ✓00 +
g
L✓
g
L
= 0, we have K =
✓(t) = c1 cos
p
and
K t + c2 sin
p
K t = c1 cos
Recall that the period of sin bt and cos bt are each
q
2⇡
this case, since L = 1, we get P = 2⇡ g1 = p
g.
2⇡
b .
r
g
t + c2 sin
L
r
g
t.
L
Then the period is P = p2⇡ = 2⇡
gL
q
L
g.
In
(b) Since ✓1 does not appear in the formula, the period does not depend on ✓1 . It depends only on the
gravitational constant g and the length of the pendulum L. (Note that it also does not depend
on the weight of the pendulum).
20. We want to show that if y1 and y2 are two solutions, and c is a constant, then y1 + y2 and cy1 are also
solutions. But this follows directly from properties of the derivative:
(y1 + y2 )00 + a(y1 + y2 )0 + b(y1 + y2 ) = (y100 + ay10 + by1 ) + (y200 + ay20 + by2 ) = 0 + 0 = 0
(cy1 )00 + a(cy1 )0 + b(cy1 ) = cy100 + acy10 + bcy1 = c(y100 + ay10 + by1 ) = c · 0 = 0.
Thus the solution set is a subspace of F .
21. Since we know that dim S = 2, it suffices to show that e t and te t are solutions (are in S) and that
they are linearly independent. Note that is a solution of 2 + a + b = 0. Then
e
t 00
+a e
t 0
+ be
t
=
2
e
t
+a e
t
+ be
t
= e t(
2
+ a + b) = 0.
For te t , note that
te
te
t 0
t 00
=e
= e
t
+ te t ,
t
+ te
t 0
= e
t
+ e
t
+
2
te
t
=2 e
t
+
2
te t .
Chapter Review
531
Then
te
t 00
+ a te
t 0
+ bte
t
=2 e
t
2
+
= (2 + a)e
te
t
t
+a e
+ te t (
2
t
+ te
t
+ bte
t
+ a + b)
= (2 + a)e t .
p
a± a2 4b
,
2
But the solutions of 2 + a + b are =
a2 4b = 0, so that = a2 , so that 2 =
and te t is a solution.
so if the two roots are equal, we must have
a and then 2 + a = 0. So the remaining term vanishes,
To see that the two are linearly independent, suppose c1 e t + c2 te t is zero for all values of t. Setting
t = 0 gives c1 = 0, so that c2 te t is zero. Then set t = 1 to get c2 = 0. Thus the two are linearly
independent, so they form a basis for the solution set S.
22. Suppose that c1 ept cos qt + c2 ept sin qt is identically zero. Setting t = 0 gives c1 = 0, so that c2 ept sin qt
is identically zero, which clearly forces c2 = 0, proving linear independence.
Chapter Review
1. (a) False. All we can say is that dim V  n, so that any spanning set for V contains at most n vectors.
For example, let V = R, v1 = e1 and v2 = 2e1 . Then V = span(v1 , v2 ), but V has spanning sets
consisting of only one vector.
(b) True. See Exercise 17(a) in Section 6.2.
(c) True. For example,

1
I=
0
0
,
1

0
J=
1
1
,
0

1
L=
0
Then R I = E11 , L I = E12 , J + I L = E21 , J + I
invertible matrices spans; since dim M22 = 4, it is a basis.
1
,
1
R=

1
1
1
.
0
R = E22 . Therefore this set of four
(d) False. By Exercise 2 in Section 6.5, the kernel of the map tr : M22 ! R has dimension 3, so that
matrices with trace 0 span a subspace of M22 of dimension 3.
(e) False. In general, kx + yk =
6 kxk + kyk. For example, take x = e1 and y = e2 in R2 .
(f ) True. If T is one-to-one and {ui } is a basis for V , then {T ui } is a linearly independent set by
Theorem 6.22. It spans range(T ), so it is a basis for range(T ). Thus if T is onto, range(T ) = W ,
so that {T ui } is a basis for W and therefore dim V = dim W .
(g) False. This is only true if T is onto. For example, let T : R ! R be T (x) = 0. Then ker T = R
but the codomain is not {0} (although the range is).
(h) True. If nullity(T ) = 4, then rank T = dim M33 nullity(T ) = 9
five-dimensional subspace of P4 , so it is the entire space.
4 = 5. So range(T ) is a
(i) True. Since dim P3 = 4, it suffices to show that dim V = 4. But polynomials in V are of the
form a + bx + cx2 + dx3 + ex4 where a + b + c + d + e = 0, so they are of the form ( b c d
e) + bx + cx2 + dx3 + ex4 , where b, c, d, and e are arbitrary. Thus a basis for V is {x, x2 , x3 , x4 },
so dim V = 4.
(j) False. See Example 6.80 in Section 6.6.
2. This is the subspace {0}, since the only vector


x
0
.
satisfying x2 + 3y 2 = 0 is
y
0
3. The conditions a + b = a + c and a + c = c + d imply that b = c and a = d, so that
⇢
a b
W =
.
b a
532
CHAPTER 6. VECTOR SPACES
(Note that matrices of this form satisfy all of the original equalities.) To show that this is a subspace,
we must show that it is closed under sums and scalar multiplication. Thus suppose

 0
a b
a b0
,
2 W,
c a scalar.
b a
b0 a 0
Then

a
b
 0

b
a b0
a + a 0 b + b0
+ 0
2W
0 =
a
b a
b + b0 a + a 0


a b
ca cb
c
=
2 W.
b a
cb ca
4. To show that this is a subspace, we must show that it is closed under sums and scalar multiplication.
Suppose p(x), q(x) 2 W . Then
✓ ◆
✓ ✓ ◆
✓ ◆◆
✓ ◆
✓ ◆
1
1
1
1
1
3
3
3
3
x (p + q)
=x p
+q
=x p
+x q
= p(x) + q(x) = (p + q)(x),
x
x
x
x
x
so that p + q 2 W . Next, if c is a scalar, then
✓ ◆
✓ ◆
1
1
x3 (cp)
= cx3 p
= cp(x) = (cp)(x),
x
x
so that cp 2 W . Thus W is a subspace.
5. To show that this is a subspace, we must show that it is closed under sums and scalar multiplication.
Suppose f (x), g(x) 2 W . Then
(f + g)(x + ⇡) = f (x + ⇡) + g(x + ⇡) = f (x) + g(x) = (f + g)(x),
so that f + g 2 W . Next, if c is a scalar, then
(cf )(x + ⇡) = cf (x + ⇡) = cf (x) = (cf )(x),
so that cf 2 W . Thus W is a subspace.
6. They are linearly dependent. By the double angle formula, cos 2x = 1
1
cos 2x
2 sin2 x = 1
cos 2x
2 sin2 x, so that
2
3 sin2 x = 0.
3
7. Suppose that cA + dB = O. Taking the transpose of both sides gives cAT + dB T = O; since A is
symmetric and B is skew-symmetric, this is just cA dB = O. Adding these two equations gives
2cA = O. Since A is nonzero, we must have c = 0, so that dB = O. Then since B is nonzero, we must
have d = 0. Thus A and B are linearly independent.
8. The condition a + d = b + c means that a = b + c d, so that
⇢
⇢
a b
b+c
W =
: a+d=b+c =
c d
c
d
b
d
.
Then define B by finding three matrices in W in each of which exactly one of b, c, and d is 1 while the
other are zero:
⇢



1 1
1 0
1 0
B= B=
, C=
, D=
.
0 0
1 0
0 1
Consider the expression
bB + cC + dD =

b+c
c
d
b
.
d
If this is O, then clearly b = c = d = 0, so that B, C, and D are linearly independent. But also,
bB + cC + dD is a general matrix in W , so that B, C, and D span W . Thus {B, C, D} is a basis for
W.
Chapter Review
533
9. Suppose p(x) = a + bx + cx2 + dx3 + ex4 + f x5 2 P5 with p( x) = p(x). Then
p(1) = p( 1) ) a + b + c + d + e + f = a
b+c
d+e
p(2) = p( 2) ) a + 2b + 4c + 8d + 16e + 32f = a
2b + 4c
p(3) = p( 3) ) a + 3b + 9c + 27d + 81e + 243f = a
f ) b+d+f =0
8d + 16e
3b + 9c
32f ) b + 4d + 16f = 0
27d + 81e
243f
) b + 9d + 81f = 0.
These three equations in the unknowns b, d, f have only the trivial solution, so that b = d = f = 0,
and p(x) = a + cx2 + dx4 . Any polynomial of this form clearly has p(x) = p( x), since ( x)2 = x2
and ( x)4 = x4 . So {1, x2 , x4 } is a basis for W .
10. Let B = {1, 1 + x, 1 + x + x2 } = {b1 , b2 , b3 } and C = {1 + x, x + x2 , 1 + x2 } = {c1 , c2 , c3 }. Then
2 3
2 3
2 3
0
1
1
[c1 ]B = 415 ,
[c2 ]B = 4 05 ,
[c3 ]B = 4 15 ,
0
1
1
so that by Theorem 6.13
PB
Then
PC
C
2
0
= 41
0
2
1
2
1
2
1
2
6
= PB 1 C = 4
B
3
1
15 .
1
1
0
1
1
0
0
3
1
2
17
.
25
1
2
11. To show that T is linear, we must show that T (x + z) = T x + T z and that T (cx) = cT x, where c is a
scalar.
T (x + z) = y(x + z)T y = y xT + zT y = yxT y + yzT y = T x + T z
T (cx) = y(cx)T y = cyxT y = cT x.
Thus T is linear.
12. T is not a linear transformation. Let A 2 Mnn and c be a scalar other than 0 or 1. Then
T (cA) = (cA)T (cA) = c2 AT A = c2 T (A).
Since c 6= 0, 1, it follows that c2 T (A) 6= cT (A), so that T is not linear.
13. To show that T is linear, we must show that (T (p + q))(x) = T p(x) + T q(x) and that (T (cp))(x) =
cT p(x), where c is a scalar.
(T (p + q))(x) = (p + q)(2x
(T (cp))(x) = (cp)(2x
1) = p(2x
1) = cp(2x
1) + q(2x
1) = T p(x) + T q(x)
1) = T p(x).
Thus T is linear.
14. We first write 5
5
3x + 2x2 in terms of 1, 1 + x, and 1 + x + x2 :
3x + 2x2 = a(1) + b(1 + x) + c(1 + x + x2 ) = (a + b + c) + (b + c)x + cx2 .
Then c = 2, so that (from the x term) b =
T (5
3x + 2x2 ) = 8T (1)
5, and then a = 8. Thus
5T (1 + x) + 2T (1 + x + x2 )

1
=8
0
0
1
5

1
0

1
0
+2
1
1

1
3
=
0
2
7
.
3
534
CHAPTER 6. VECTOR SPACES
15. Since T : Mnn ! R is onto, we know from the rank theorem that
n2 = dim Mnn = rank(T ) + nullity(T ) = 1 + nullity(T ),
so that nullity(T ) = n2
1.
16. (a) We want a map T : M22 ! M22 such that T (A) = O if and only if A is upper triangular. If

a b
A=
,
c d
an obvious map to try is
T (A) =

0
c
0
.
0
Clearly the matrices sent to O by T are exactly the upper triangular matrices (together with O),
which is the subspace W . To see that T is linear, note that we can rewrite T as T (A) = a21 E21 .
Then
T (A + B) = (A + B)21 E21 = (a21 + b21 )E21 = a21 E21 + b21 E21 = T (A) + T (B)
T (cA) = (cA)21 E21 = ca21 E21 = cT (A).
(b) For this map, we want the image of any matrix A to be upper triangular. With A as above, an
obvious map to try is the one that sets the entry below the main diagonal to zero:

a b
T (A) =
.
0 d
It is obvious that range T = W , since any upper triangular matrix is the image of itself under T ,
and every matrix of the form T (A) is upper triangular by construction. To see that T is linear,
write T (A) = A a21 E21 ; then
T (A + B) = (A + B)
= (A
T (cA) = (cA)
(A + B)21 E21 = (A + B)
a21 E21 ) + (B
(a21 + b21 )E21
b21 E21 ) = T (A) + T (B)
(cA)21 E21 = cA
ca21 E21 = c(A
a21 E21 ) = cT (A).
17. We have
2 3
1
607
1 0
7
T (1) =
= 1 · E11 + 0 · E12 + 0 · E21 + 1 · E22 ) [T (1)]C = 6
405
0 1
1



1 1
1 0
0 1
T (x) = T ((x + 1) 1) = T (x + 1) T (1) =
=
0 1
0 1
0 0
2 3
0
6 17
7
= 0 · E11 + 1 · E12 + 0 · E21 + 0 · E22 ) [T (x)]C = 6
4 05
0


0
1
1
T (x2 ) = T ((1 + x + x2 ) (1 + x)) = T (1 + x + x2 ) T (1 + x) =
1
0
0
2 3
1
6 27
7
= 1 · E11 2 · E12 + 1 · E21 1 · E22 ) [T (1)]C = 6
4 15 .
1

1
=
1

1
1
2
1
Chapter Review
535
Since these vectors are the columns of [T ]C
[T ]C
B,
B
we have
2
1 0
60 1
=6
40 0
1 0
3
1
27
7.
15
1
18. We must show that S spans V and that S is linearly independent. First, since every vector can be
written as a linear combination of elements of S in one way, S spans V . Next, suppose
c1 v1 + c2 v2 + · · · + cn vn = 0.
Since we know that we can write 0 as the linear combination of the elements of S with zero coefficients,
and since the representation of 0 is unique by hypothesis, it follows that ci = 0 for all i, so that S is
linearly independent. Thus S is a basis for V .
19. If u 2 U , then T u 2 ker(S), so that (S
T )u = S(T u) = 0. Thus S
T is the zero map.
20. Since {T (v1 ), T (v2 ), . . . , T (vn )} is a basis for V , it follows that range(T ) = V since any element of V
can be expressed as a linear combination of elements in range(T ). By Theorem 6.30 (p) and (t), this
means that T is invertible.
Chapter 7
Distance and Approximation
7.1
Inner Product Spaces
1. With hu, vi = 2u1 v1 + 3u2 v2 , we get
(a) hu, vi = 2 · 1 · 4 + 3 · ( 2) · 3 = 10.
p
p
p
(b) kuk = hu, ui = 2 · 1 · 1 + 3 · ( 2) · ( 2) = 14.

p
p
3
(c) d(u, v) = ku vk =
= 2 · ( 3) · ( 3) + 3 · ( 5) · ( 5) = 93.
5

6 2
2. With A =
, we get
2 3


⇥
⇤ 6 2 4
T
2
(a) hu, vi = u Av = 1
= 4.
2 3 3
s


p
p
⇥
⇤ 6 2
1
1
2
(b) kuk = hu, ui =
= 10.
2 3
2
s



p
⇥
⇤ 6 2
3
3
3
5
(c) d(u, v) = ku vk =
=
= 189.
5
2 3
5

a
3. We want a vector w =
such that
b
hu, wi = 2 · 1 · a + 3 · ( 2) · b = 2a 6b = 0.

3t
3
Any vector of the form
is such a vector; for example,
.
t
1

a
4. We want a vector w =
such that
b


⇥
⇤ 6 2 a
T
2
hu, wi = u aw = 1
= 2a 4b = 0.
2 3 b

2
There are many such vectors, for example w =
.
1

5.
(a)
(b)
(c)
⌦
↵
hp(x), q(x)i = 3 2x, 1 + x + x2 = 3 · 1 + ( 2) · 1 + 0 · 1 = 1.
p
p
p
kp(x)k = hp(x), p(x)i = 3 · 3 + ( 2) · ( 2) + 0 · 0 = 13.
p
p
d(p(x), q(x)) = kp(x) q(x)k = 2 3x x2 = 22 + ( 3)2 + ( 1)2 = 14.
537
538
CHAPTER 7. DISTANCE AND APPROXIMATION
6.
hp(x), q(x)i =
(a)
✓
Z
1
2x)(1 + x + x2 ) dx =
(3
0
Z
1
( 2x3 + x2 + x + 3) dx
0
◆ 1
1 4 1 3 1 2
10
=
x + x + x + 3x
=
.
2
3
2
3
0
s
s✓
Z 1
p
kp(x)k = hp(x), p(x)i =
(3 2x)2 dx =
9x
(b)
0
d(p(x), q(x)) = kp(x)
(c)
s✓
=
q(x)k = 2
x2 =
3x
1 5 3 4 5 3
x + x + x
5
2
3
6x2
s
+ 4x
◆
Z
1
=
0
6x2
4
+ x3
3
◆
1
=
0
r
13
.
3
1
(2
3x
x2 )2 dx
0
r
41
.
30
7. We want a polynomial r(x) = a + bx + cx2 such that
hp(x), r(x)i = 3 · a
2 · b + 0 · c = 3a
There are many such polynomials; two examples are 2
2b = 0.
6x + 7x2 .
3x and 4
8. We want a polynomial r(x) = a + bx + cx2 such that
Z 1
hp(x), r(x)i =
(3 2x)(a + bx + cx2 ) dx
=
=
Z
0
1
(3a + (3b
2b)x2
2a)x + (3c
2cx3 ) dx
0
✓
3ax +
3b
2a
2
x +
3c
2
5
1
= 2a + b + c = 0.
6
2
There are many such polynomials. Three examples are
9. Recall the double-angle formulas sin2 x =
1
2
1
2
2b
3
x
c 4
x
2
2
1 + 4x2 , 5
cos 2x and cos2 x =
◆
1
0
12x, and 12x
1
2
+
1
2
20x2 .
cos 2x. Then
(a)
hf, gi =
Z
2⇡
f (x)g(x) dx =
0
=
(b) kf k =
(c)
R 2⇡
0
Z
f (x)2 dx =
d(f, g) = kf
gk =
2⇡
0
R 2⇡
0
Z
✓
1
2
Z
2⇡
sin2 x + sin x cos x dx
0
1
cos 2x + sin x cos x
2
sin2 x dx =
R 2⇡
0
2⇡
( cos x)2 dx =
0
10. We want to find a function g such that
⇥1
⇤2⇡
2
2 sin x 0 = 0, so choose g(x) = cos x.
R 2⇡
0
1
2
Z
1
2
◆
dx =
cos 2x dx =
2⇡
cos2 x dx =
0

Z
1
x
2
⇥1
2x
2⇡
0
✓
1
1
sin 2x + sin2 x
4
2
1
4
sin 2x
⇤2⇡
0
2⇡
= ⇡.
0
= ⇡.
◆
1 1
+ cos 2x dx
2 2

1
1
=
x + sin 2x
2
4
2⇡
= ⇡.
0
sin x · g(x) dx = 0. Note from the previous exercise that
7.1. INNER PRODUCT SPACES
539
11. We must show it satisfies all four properties of an inner product. First,
hp(x), q(x)i = p(a)q(a) + p(b)q(b) + p(c)q(c) = q(a)p(a) + q(b)p(b) + q(c)p(c) = hq(x), p(x)i .
Next,
hp(x), q(x) + r(x)i = p(a)(q(a) + r(a)) + p(b)(q(b) + r(b)) + p(c)(q(c) + r(c))
= p(a)q(a) + p(a)r(a) + p(b)q(b) + p(b)r(b) + p(c)q(c) + p(c)r(c)
= p(a)q(a) + p(b)q(b) + p(c)q(c) + p(a)r(a) + p(b)r(b) + p(c)r(c)
= hp(x), q(x)i + hp(x), r(x)i .
Third,
hdp(x), q(x)i = dp(a)q(a) + dp(b)q(b) + dp(c)q(c) = d(p(a)q(a) + p(b)q(b) + p(c)q(c)) = d hp(x), q(x)i .
Finally,
hp(x), p(x)i = p(a)p(a) + p(b)q(b) + p(c)q(c) = p(a)2 + p(b)2 + p(c)2
0,
and if equality holds, then p(a) = p(b) = p(c) = 0 since squares are always nonnegative. But p 2 P2 ,
so if it is nonzero, it has at most two roots. Since a, b, and c are distinct, it follows that p(x) = 0.
12. We must show it satisfies all four properties of an inner product. First,
hp(x), q(x)i = p(0)q(0) + p(1)q(1) + p(2)q(2) = q(0)p(0) + q(1)p(1) + q(2)p(2) = hq(x), p(x)i .
Next,
hp(x), q(x) + r(x)i = p(0)(q(0) + r(0)) + p(1)(q(1) + r(1)) + p(2)(q(2) + r(2))
= p(0)q(0) + p(0)r(0) + p(1)q(1) + p(1)r(1) + p(2)q(2) + p(2)r(2)
= p(0)q(0) + p(1)q(1) + p(2)q(2) + p(0)r(0) + p(1)r(1) + p(2)r(2)
= hp(x), q(x)i + hp(x), r(x)i .
Third,
hcp(x), q(x)i = cp(0)q(0) + cp(1)q(1) + cp(2)q(2) = c(p(0)q(0) + p(1)q(1) + p(2)q(2)) = c hp(x), q(x)i .
Finally,
hp(x), p(x)i = p(0)p(0) + p(1)q(1) + p(2)q(2) = p(0)2 + p(1)2 + p(2)2
0,
and if equality holds, then 0, 1, and 2 are all roots of p since squares are always nonnegative. But
p 2 P2 , so if it is nonzero, it has at most two roots. It follows that p(x) = 0.

0
13. The fourth axiom does not hold. For example, let u =
. Then hu, ui = 0 but u 6= 0.
1

1
14. The fourth axiom does not hold. For example, let u =
. Then hu, ui = 1 · 1 1 · 1 = 0, but u 6= 0.
1

1
Alternatively, if u =
, then u · u = 3 < 0, so the fourth axiom again fails.
2

1
15. The fourth axiom does not hold. For example, let u =
. Then hu, ui = 0 · 1 + 1 · 0 = 0, but u 6= 0.
0

1
Alternatively, if u =
, then u · u = 4 < 0, so the fourth axiom again fails.
2
16. The fourth axiom does not hold. For example, if p is any nonzero polynomial with zero constant term,
then hp(x), p(x)i = 0 but p(x) is not the zero polynomial.
540
CHAPTER 7. DISTANCE AND APPROXIMATION
17. The fourth axiom does not hold. For example, let p(x) = 1
but p(x) is not the zero polynomial.
x; then p(1) = 0 so that hp(x), p(x)i = 0,
18. The fourth axiom does not hold.
 For example E12 E22 = O, so that hE12 , E22 i = det O = 0, but
1 1
E11 6= O. Alternatively, let A =
; then A 6= O but hA, Ai = 0.
1 1
19. Examining Example 7.3, the matrix should be
⇥
u1

⇤ 4
u2
1
1
4


4
1
1
; and indeed
4

⇤ v1
u1 + 4u2
= 4u1 v1 + u2 v1 + u1 v2 + 4u2 v2 ,
v2
⇥
v1
= 4u1 + u2
v2
as desired.
20. Examining Example 7.3, the matrix should be
⇥
u1
u2

⇤ 1
2
2
5

⇥
v1
= u1 + 2u2
v2

1
2
2
; and indeed
5
2u1 + 5u2

⇤ v1
= u1 v1 + 2u2 v1 + 2u1 v2 + 5u2 v2 ,
v2
as desired.
21. The unit circle is
⇢
u=

u1
u2
1
: u21 + u22 = 1 .
4
u2
2
1
!1.0
0.5
!0.5
!1
!2
1.0
u1
7.1. INNER PRODUCT SPACES
22. The unit circle is
541
⇢
u=

u1
u2
: 4u21 + 2u1 u2 + 4u22 = 1 .
u2
0.6
0.4
0.2
!0.6
!0.4
0.2
!0.2
0.4
0.6
u1
!0.2
!0.4
!0.6
23. We have, using properties of the inner product,
hu, cvi = hcv, ui
(Property 1)
= c hu, vi
(Property 1).
= c hv, ui
(Property 3)
24. Let w be any vector. Then hu, 0i = hu, 0wi which, by Theorem 7.1(b), is equal to 0 hu, wi = 0.
Similarly, h0, vi = h0w, vi which, by property 3 of inner products, is equal to 0 hw, vi = 0.
25. Using properties of the inner product and Theorem 7.1 as needed,
hu + w, v
wi = hu, v
wi + hw, v
wi
= hu, vi + hu, wi + hw, vi + hw, wi
= hu, vi
=1
=
hu, wi + hv, wi
5+0
4
4=
2
kwk
hw, wi
8.
26. Using properties of the inner product and Theorem 7.1 as needed,
h2v
w, 3u + 2wi = h2v, 3u + 2wi + h w, 3u + 2wi
= h2v, 3ui + h2v, 2wi + h w, 3ui + h w, 2wi
= 6 hv, ui + 4 hv, wi
3 hw, ui
=6·1+4·0
2 · 22 =
= 6 hu, vi + 4 hv, wi
3·5
3 hu, wi
2 hw, wi
2
2 kwk
17.
27. Using properties of the inner product and Theorem 7.1 as needed,
p
ku + vk = hu + v, u + vi
p
= hu, ui + hu, vi + hv, ui + hv, vi
q
2
2
= kuk + 2 hu, vi + kvk
p
p
= 1 + 2 · 1 + 3 = 6.
542
CHAPTER 7. DISTANCE AND APPROXIMATION
28. Using properties of the inner product and Theorem 7.1 as needed,
p
k2u 3v + wk = h2u 3v + w, 2u 3v + wi
p
= 4 hu, ui 12 hu, vi 6 hv, wi + 4 hu, wi + 9 hv, vi + hw, wi
q
2
2
2
= 4 kuk
12 hu, vi 6 hv, wi + 4 hu, wi + 9 kvk + kwk
r
⇣p ⌘2
= 4 · 12 12 · 1 6 · 0 + 4 · 5 + 9 ·
3 + 22
p
= 43.
29. u+v = w means that u+v w = 0; this is true if and only if hu + v
this inner product gives
hu + v
wi = hu, ui + 2 hu, vi
w, u + v
2
= kuk + 2 · 1
=1+2
2·5
2 hu, wi
w, u + v
wi = 0. Computing
2 hv, wi + hv, vi + h w, wi
2
2
2 · 0 + kvk + kwk
10 + 3 + 2 = 0.
2
30. Since kuk = 1, we also have 1 = kuk = hu, ui. Similarly, 1 = hv, vi. But then
0  hu + v, u + vi = hu, ui + 2 hu, vi + hv, vi = 2 + 2 hu, vi .
Solving this inequality gives 2 hu, vi
2, so that hu, vi
1.
31. We have
hu + v, u
vi = hu, ui + hu, vi
hv, ui
hv, vi .
Since inner products are commutative, the middle two terms cancel, leaving us with
hu, ui
32. We have
2
hv, vi = kuk
2
kvk .
2
ku + vk = hu + v, u + vi = hu, ui + hu, vi + hv, ui + hv, vi .
Since inner products are commutative, the middle two terms are the same, and we get
2
2
hu, ui + 2 hu, vi + hv, vi = kuk + 2 hu, vi + kvk .
33. From Exercise 32, we have
2
2
2
ku + vk = kuk + 2 hu, vi + kvk .
Substituting
(1)
v for v gives another identity,
ku
2
2
2
vk = kuk + 2 hu, vi + k vk = kuk
Adding these two gives
2
ku + vk + ku
2
2
2
2
2 hu, vi + kvk .
2
vk = 2 kuk + 2 kvk ,
and dividing through by 2 gives the required identity.
34. Solve (1) and (2) in the previous exercise for hu, vi, giving
⌘
1⇣
2
2
2
ku + vk
kuk
kvk
2
⌘
1⇣
2
2
2
hu, vi =
ku vk + kuk + kvk .
2
hu, vi =
(2)
7.1. INNER PRODUCT SPACES
543
Add these two equations, giving
2 hu, vi =
dividing through by 2 gives
hu, vi =
1⇣
2
ku + vk
2
1⇣
2
ku + vk
4
2
ku
vk
⌘
=
2
ku
vk
⌘
1
2
ku + vk
4
;
1
ku
4
2
vk .
35. Use Exercise 34. First, if ku + vk = ku vk, then Exercise 34 implies that hu, vi = 0. Second, suppose
that u and v are orthogonal. Then again from Exercise 34,
0=
1
2
ku + vk
4
1
ku
4
1
(ku + vk + ku
4
2
vk =
vk) (ku + vk
ku
vk) ,
so that one of the two factors must be zero. If the first factor is zero, then both norms are zero, so
they are equal. If the second factor is zero, then we have ku + vk ku vk = 0 and again the norms
are equal.
36. From (2) in Exercise 33, we have
This is equal to
q
d(u, v) = ku
2
vk =
q
kuk
2
2
2 hu, vi + kvk .
2
kuk + kvk if and only if hu, vi = 0, which is to say when u and v are orthogonal.
37. The inner product we are working with is hu, vi = 2u1 v1 + 3u2 v2 , so that

1
v1 =
0





2·1·1+3·1·0 1
1
1
1
0
v2 =
=
=
.
1
1
0
1
2·1·1+3·0·0 0
⇢

1
0
The orthogonal basis is
,
.
0
1
38. The inner product we are working with is

4
hu, vi = uT Av = uT
2
Then

1
0
" #
1
v2 =
1
v1 =
4·1·1
4·1·1
2·1·1
2·1·0
2
= 4u1 v1
7
2u1 v2
2u2 v1 + 7u2 v2 .
" # " #
1
2·1·0+7·0·1 1
=
2·1·0+7·0·0 0
1
" # " #
1
1 1
= 2 .
2 0
1
⌦
↵
39. The inner product we are working with is a0 + a1 x + a2 x2 , b0 + b1 x + b2 x2 = a0 b0 + a1 b1 + a2 b2 .
Then
v1 = 1
1 · (1 + x)
1 = (1 + x)
1·1
1 · (1 + x + x2 )
v3 = (1 + x + x2 )
1
1·1
v2 = (1 + x)
The orthogonal basis is {1, x, x2 }.
1=x
x · (1 + x + x2 )
x = (1 + x + x2 )
x·x
1
x = x2 .
544
CHAPTER 7. DISTANCE AND APPROXIMATION
40. The inner product we are working with is hp(x), q(x)i =
v1 = 1
v2 = (1 + x)
2
v3 = (1 + x + x )
= (1 + x + x2 )
= (1 + x + x2 )
= (1 + x + x2 )
=
1
6
R1
1 · (1 + x)
1 = (1 + x)
1·1
0
R1
0
p(x)q(x) dx. Then
(1 + x) dx
1 = (1 + x)
R1
1 dx
0
3
1=
2
1
+x
2
◆
1
2 ✓
1 · (1 + x + x2 )
1
2 + x · (1 + x + x )
1
+x
1
1
1·1
2
2 +x ·
2 +x
R1
R1
✓
◆
1
2
(1 + x + x2 ) dx
1
2 + x (1 + x + x ) dx
0
0
1
+x
R1
R1
1
1
2
1 dx
2 +x ·
2 +x
0
0
✓
◆
11
1/6
1
1
+x
6
1/12
2
✓
◆
11
1
1 2
+x
6
2
x + x2 .
41. The inner product here is hf, gi =
R1
1
f (x)g(x) dx.
(a) To normalize the first three Legendre polynomials, we find the norm of each:
k1k =
x2
sZ
sZ
1
1
1 · 1 dx =
sZ
sZ
1
1 dx =
1
p
2
s
r
1
1 3
2
kxk =
x · x dx =
dx =
x
=
3
3
1
1
1
sZ ✓
sZ ✓
◆
✓
◆
◆
1
1
1
1
1
2 2 1
2
2
4
=
x
x
dx =
x
x +
dx
3
3
3
3
9
1
1
s
p
1
1 5 2 3 1
2 2
=
x
x + x
= p .
5
9
9
3 5
1
1
1
x2
So the normalized Legendre polynomials are
1
1
=p ,
k1k
2
p
x
6
=
x,
kxk
2
x2
x2
1
3
1
3
p ✓
3 5
= p
x2
2 2
1
3
◆
p
5
= p (3x2
2 2
1).
7.1. INNER PRODUCT SPACES
545
(b) For the computation, we will use the unnormalized Legendre polynomials given in Example 3.8,
1, x, and x2 13 , and normalize the fourth one at the end. With x4 = x3 , we have
x4 · v1
x4 · v2
x4 · v3
v4 = x4
v1
v2
v3
v1 · v1
v2 · v2
v3 · v3
R1 3
R1 3
R1 3 2 1
✓
◆
x · 1 dx
x · x dx
x x
1
3 dx
1
1
1
3
2
=x
1
x R1
x
R1
R1
3
1 · 1 dx
x · x dx
x2 13 x2 13 dx
1
1
1
R1 3
R1 4
R1
✓
◆
x dx
x dx
x5 13 x3 dx
1
1
= x3
1 R 11
x R1 1
x2
R1
2 2
1
2
4
3
1 dx
x dx
x
3 x + 9 dx
1
1
1
⇥ 1 4 ⇤1
⇥ 1 5 ⇤1
⇥1 6
⇤
1
✓
◆
1 4
1
4x
5x
6x
12 x
1
1
1
= x3
1 ⇥
x
x2
⇤
⇥
⇤1
1
1
1
2
1
2
3
x3
x5
x3 + x
3
= x3
0
2/5
x
2/3
5
1
9
9
1
0
3
x.
5
The norm of this polynomial is
sZ ✓
◆✓
1
3
kv4 k =
x3
x
x3
5
1
= x3
3
x
5
◆
dx =
sZ
1
1
✓
◆
6 4
9
x + x2 dx
5
25
s
1 7
6 5
3
=
x
x + x3
7
25
25
x6
Thus the normalized Legendre polynomial is
✓
◆ p
35
3
14
3
p
x
x =
(5x3
5
4
2 14
42. (a) Starting with `0 (x) = 1, `1 (x) = x, `2 (x) = x2
`0 (1) = 1
`1 (1) = 1
`2 (1) = 1
`3 (1) = 1
1
=
3
3
5
2
3
2
5
1
3,
L0 (x) = 1
)
L1 (x) = x
)
L2 (x) =
✓
3
x2
2
✓
5
L3 (x) =
x3
2
)
1
p
2 14
=
.
35
3x).
3
5 x,
`3 (x) = x3
)
1
we have
◆
1
3
1
= x2
3
2
2
◆
3
5 3 3
x = x
x.
5
2
2
(b) For L2 and L3 , the given recurrence says
2·2 1
xL1 (x)
2
2·3 1
L3 (x) =
xL2 (x)
3
L2 (x) =
2
1
3 2 1
x
2
2 ✓ 2
3 1
5
3 2
L1 (x) = x
x
3
3
2
L0 (x) =
1
2
◆
2
5
x = x3
3
2
so that the recurrence indeed works for n = 2 and n = 3. Then
✓
◆
✓
◆
2·4 1
4 1
7
5 3 3
3 3 2 1
L4 (x) =
xL3 (x)
L2 (x) = x
x
x
x
4
4
4
2
2
4 2
2
35 4 15 2 3
=
x
x +
8
4
8
✓
◆
✓
2·5 1
5 1
9
35 4 15 2 3
4 5 3
L5 (x) =
xL4 (x)
L3 (x) = x
x
x +
x
5
5
5
8
4
8
5 2
63 5 35 3 15
=
x
x + x.
8
4
8
3
x,
2
3
x
2
◆
546
CHAPTER 7. DISTANCE AND APPROXIMATION
43. Let B = {ui } be an orthogonal basis for W . Then
hprojW (v), uj i =
⌧X
hui , vi
ui , uj
hui , ui i
=
⌧
huj , vi
uj , uj
huj , uj i
huj , vi
huj , uj i = huj , vi .
huj , uj i
=
Then
hperpW (v), uj i = hv
projW (v), uj i = hv, uj i
hprojW (v), uj i = hv, uj i
huj , vi = 0.
Thus perpW (v) is orthogonal to every vector in B, so that perpW (v) is orthogonal to every w 2 W .
44. (a) We have
2
2
htu + v, tu + vi = t2 hu, ui + 2t hu, vi + hv, vi = kuk t2 + 2 hu, vi t + kvk
2
0.
2
This is a quadratic at2 + bt + c in t, where a = kuk , b = 2 hu, vi, and c = kvk .
2
(b) Since a = kuk > 0, this is an upward-opening parabola, so its vertex is its lowest point. That
b
vertex is at 2a
, so the lowest point, which must be nonnegative, is
a
✓
b
2a
◆2
+b
✓
b
2a
◆
+c=
ab2
4a2
b2 , so that
Since 4a > 0, we must have 4ac
b2
b2
+c=
2a
p
ac
1
2
2b2 + 4ac
4ac b2
=
4a
4a
0.
|b|.
(c) Substituting back in the values of a, b, and c from part (a) gives
p
Thus kuk kvk
ac =
q
2
1
1
|b| = |2 hu, vi| = |hu, vi| .
2
2
2
kuk kvk = kuk kvk
|hu, vi|, which is the Cauchy-Schwarz inequality.
Exploration: Vectors and Matrices with Complex Entries
2
1. Recall that if z 2 C, then zz = |z| . Then for v as given, we have
kvk =
p
v·v =
p
v 1 v1 + v 2 v2 + · · · + v n vn =
2
q
2
2
|v1 | + · · · + |vn | .
Finally, we must show that |z| = z 2 . But
2
2
|z| = (zz) = zzz 2 = z 2 z 2 = z 2 ,
so the above is in fact equal to
2. (a) u · v = i · (2
p
|v12 | + · · · + |vn2 |.
3i) + 1 · (1 + 5i) = i(2 3i) + (1 + 5i) =
p
p
p
(b) By Exercise 1, kuk = |i2 | + |12 | = | 1| + |1| = 2.
2 + 3i.
(c) By Exercise 1,
kvk =
p
|(2
(d) d(u, v) = ku
p
3 5.
3i)2 | + |(1 + 5i)2 | =
vk =

2 + 4i
5i
=
p
q
(2 + 3i)(2
2
3i) + (1
2
| 2 + 4i| + | 5i| =
5i)(1 + 5i) =
p
( 2
p
13 + 26 =
p
39.
4i)( 2 + 4i) + (5i)( 5i) =
Exploration: Vectors and Matrices with Complex Entries
(e) If w =

547
a
is orthogonal to u, then
b
u·w =i·a+1·b=
ai + b = 0,
so that b = ai. Thus for example setting a = 1, we get

1
.
i
(f ) If w =

a
is orthogonal to v, then
b
u·w =2
So b =
2+3i
1 5i a;
3i · a + 1 + 5i · b = (2 + 3i)a + (1
setting for example a = 1
5i, we get b =

1
2
2
5i)b = 0.
3i, and one possible vector is
5i
3i
3. (a) v · u = v 1 u1 + · · · + v n un = v1 u1 + · · · + vn un = v1 u1 + · · · + vn un = u1 v1 + · · · + un vn = u · v.
(b)
u · (v + w) = u1 (v1 + w1 ) + u2 (v2 + w2 ) + · · · un (vn + wn )
= u1 v 1 + u1 w 1 + u2 v 2 + u2 w 2 + · · · + un v n + un w n
= (u1 v1 + u2 v2 + · · · + un vn ) + (u1 w1 + u2 w2 + · · · + un wn )
= u · v + u · w.
(c) First,
(cu) · v = cu1 v1 + · · · + cun vn = cu1 v1 + · · · + cun vn = c(u1 v1 + · · · + nn vn ) = cu · v.
For the second equality, we have, using part (a),
u · (cv) = (cu) · v = cu · v = cu · v = cv · u.
2
(d) By Exercise 1, u · u = kuk = u21 + · + u2n
0. Further, equality holds if and only if each
2
u2i = |ui | = 0, so all the ui must be zero and therefore u = 0.
4. Just apply the definition: conjugate A and take its transpose:

T
i i
⇤
(a) A = A =
.
2i 3

T
2
5 2i
(b) A⇤ = A =
.
5 + 2i
1
2
3
2+i
4
T
0 5.
(c) A⇤ = A = 41 3i
2
3 + 4i
2
3
3i 1 + i 1 i
T
4
0 5.
(d) A⇤ = A = 4 0
1 i
i
i
548
CHAPTER 7. DISTANCE AND APPROXIMATION
5. These facts follow directly from standard properties of complex numbers together with the fact that
matrix conjugation is element by element:
⇥ ⇤
(a) A = aij = [aij ] = A.
⇥
⇤ ⇥
⇤
⇥ ⇤
(b) A + B = aij + bij = aij + bij = [aij ] + bij = A + B.
(c) cA = [caij ] = [c · aij ] = c [aij ] = cA.
P
(d) Let C = AB; then cij = k aik bkj , so that
"
#
"
#
X
X
⇥
⇤
AB ij = C ij =
aik bkj
=
aik bik
= A · B ij ,
k
k
ij
ij
so that AB = A · B.
T
(e) The ij entry of A is aji . The ij entry of AT is aji , so its conjugate, which is the ij entry of
AT , is aji . So the two matrices have the same entries and thus are equal.
6. These facts follow directly from the facts in Exercise 5.
⇣ ⌘ ⇤ ⇣ ⌘ T ✓ T ◆T
T
⇤ ⇤
(a) (A ) = AT = AT
= A
= AT
= A.
(b) (A + B)⇤ = (A + B)T = AT + B T = A⇤ + B ⇤ .
(c) (cA)⇤ = (cA)T = cAT = cAT = cA⇤ .
(d) (AB)⇤ = (AB)T = B T AT = B T · AT = B ⇤ A⇤ .
2 3
v1
7
⇥
⇤6
P
6 v2 7
7. We have u · v =
ui vi = u1 u2 · · · un 6 . 7 = uT v = u⇤ v.
4 .. 5
vn
8. If A is Hermitian, then aij = aji for all i and j. In particular, when i = j, this gives aii = aii , so that
aii is equal to its own conjugate and therefore must be real.
9. (a) A is not Hermitian since it does not have real diagonal entries.
(b) A is not Hermitian since a12 = 2
3i and a21 = 2
(c) A is not Hermitian since a12 = 1
5i and a21 =
3i = 2 + 3i; the two are not equal.
1
5i; the two are not equal.
(d) A is Hermitian, since every entry is the conjugate of the same entry in its transpose.
(e) A is not Hermitian. Since it is a real matrix, it is equal to its conjugate, but AT 6= A since A is
not symmetric.
(f ) A is Hermitian. Since it is a real matrix, it is equal to its conjugate, and it is symmetric, so that
T
A = AT = A.
10. Suppose that
is an eigenvalue of a Hermitian matrix A, with corresponding eigenvector v. Now,
v⇤ A = v⇤ A⇤ = (Av)⇤ = ( v)⇤
since A is Hermitian. But by Exercise 6(c), ( v)⇤ = v⇤ . Then
(v⇤ v) = v⇤ ( v) = v⇤ (Av) = (v⇤ A)v =
2
v⇤ v = (v⇤ v).
Then since v 6= 0, we have by Exercise 7 v⇤ v = v · v = kvk 6= 0. So we can divide the above equation
through by v⇤ v to get = , so that is real.
Exploration: Vectors and Matrices with Complex Entries
549
11. Following the hint, adapt the proof of Theorem 5.19. Suppose that 1 6= 2 are eigenvalues of a
Hermitian matrix A corresponding to eigenvectors v1 and v2 respectively. Then
· v2 ) = (
1 (v1
1 v1 )
· v2 = (Av1 ) · v2 = (Av1 )⇤ v2 = (v1⇤ A⇤ )v2 .
But A is Hermitian, so A⇤ = A and we continue:
(v1⇤ A⇤ )v2 = (v1⇤ A)v2 = v1⇤ (Av2 ) = v1⇤ (
Thus 1 (v1 · v2 ) =
have v1 · v2 = 0.
2 (v1
· v2 ), so that (
2 )(v1
1
2 v2 )
=
· v2 ) = 0. But
⇤
12. (a) Multiplying, we get AA = I, so that A is unitary and A
⇤
(b) Multiplying, we get AA =

4
0
⇤
2 (v1 v2 )
1
1
6=
=
2
"
pi
2
pi
2
⇤
"
3
5
4
5
=A =
(c) Multiplying, we get AA = I, so that A is unitary and A
(d) Multiplying, we get AA⇤ = I, so that A is unitary and A
1
1
=A =
· v2 ).
by assumption, so we must
⇤
0
6= I, so that A is not unitary.
4
⇤
2 (v1
2
1p i
6 6
= A⇤ = 6
4 0
p2
6
pi
2
pi
2
4i
5
3i
5
0
1
0
#
#
.
.
3
1+i
p
3 7
0 7
5.
p1
3
13. First show (a) , (b). Let ui denote the ith column of a matrix U . Then the ith row of U ⇤ is ui . Then
by the definition of matrix multiplication,
(U ⇤ U )ij = ui · uj .
Then the columns of U form an orthonormal set if and only if
(
0 i 6= j
ui · uj =
,
1 i=j
which holds if and only if
⇤
(U U )ij =
(
0
1
i 6= j
,
i=j
which is to say that U ⇤ U = I. Thus the columns of U are orthogonal if and only if U is unitary.
Now apply this to the matrix U ⇤ . We get that U ⇤ is unitary if and only if the columns of U ⇤ are
orthogonal. But U ⇤ is unitary if and only if U is, since
(U ⇤ )⇤ U ⇤ = U U ⇤ ,
and the columns of U ⇤ are just the conjugates of the rows of U , so that (letting Ui be the ith row of
U)
⇤
Ui · Uj = Ui Uj = (Ui )T Uj = Uj · Ui .
So the columns of U ⇤ are orthogonal if and only if the rows of U are orthogonal. This proves (a) ,
(c).
Next, we show (a) ) (e) ) (d) ) (e) ) (a). To prove (a) ) (e), assume that U is unitary. Then
U ⇤ U = I, so that
⇤
U x · U y = (U x) U y = x⇤ U ⇤ U y = x⇤ y = x · y.
550
CHAPTER 7. DISTANCE AND APPROXIMATION
Next,
for (e) ) (d), assume that (e) holds. Then Qx · Qx = x · x, so that kQxk =
p
x · x = kxk. Next, we show (d) ) (e). Let ui denote the ith column of U . Then
p
Qx · Qx =
⌘
1⇣
2
2
kx + yk
kx yk
4
⌘
1⇣
2
2
=
kU (x + y)k
kU (x y)k
4
⌘
1⇣
2
2
=
kU x + U yk
kU x U yk
4
= U x · U y.
x·y =
Thus (d) ) (e). Finally, to see that (e) ) (a), let ei be the ith standard basis vector. Then ui = U ei ,
so that since (e) holds,
(
0 i 6= j
u i · u j = U ei · U ej = ei · ej =
1 i = j.
This shows that the columns of U form an orthonormal set, so that U is unitary.
14. (a) Since
"
pi
2
pi
2
"
# "
·
pi
2
pi
2
# "
pi
2
pi
2
pi
2
pi
2
#
#
=
i
p ·
2
✓
i
p
2
◆
i
i
p ·p =0
2
2
i
i
i
i
p ·p
p ·p =1
2
2
2
2
"
# "
#
✓
◆
pi
pi
i
i
i
2 ·
2 = pi ·
p
p · p = 1,
i
i
p
p
2
2
2
2
·
2
=
2
the columns of the matrix are orthonormal, so the matrix is unitary.
(b) The dot product of the first column with itself is
1 + i(1 + i) + 1
i(1
i) = (1
i)(1 + i) + (1 + i)(1
so that the matrix is not unitary.
(c) Since
" # "
4
5
3i
5
3
5
4i
5
3
5
4i
5
3
5
4i
5
"
·
#
" # " #
4
5
3i
5
·
# "
·
4
5
3i
5
#
3
= ·
5
=
=
✓
4
5
◆
4i 3i
·
=0
5 5
3 3
·
5 5
4i 4i
·
=1
5 5
✓
◆
4
4
3i 3i
·
·
= 1,
5
5
5 5
the columns of the matrix are orthonormal, so the matrix is unitary.
i) = 4 6= 1,
Exploration: Vectors and Matrices with Complex Entries
551
(d) Clearly column 2 is a unit vector, and is orthogonal to each of the other two columns, so we need
only show the first and third columns are unit vectors and are orthogonal. But
2 1+i 3 2 2 3
p
2
p
6
6
6
7 6 7 1 i 2
4 0 5·4 0 5= p · p +0+
6
6
1 i
1
p
1+i
p
6
p
3
3 2
3
1+i
p
6
1+i 1
2 2i
p
·p =
+
6
3
3
3
6
7 6
7 1 i 1+i
4 0 5·4 0 5= p · p +0+
6
6
1 i
1 i
p
p
3
2
p2
6
3 2
3
p2
6
1+i
p
·
3
1+i
=0
3
1 i
2 2
p
= + =1
6 3
3
3
2
2
1
1
4 1
6 7 6 7
4 0 5 · 4 0 5 = p · p + 0 + p · p = + = 1,
6 3
6
6
3
3
1
1
p
3
p
3
and we are done.
15. (a) With A as given, the characteristic polynomial is (
2)2 i · ( i) = 2 4 + 3 = (
Thus the eigenvalues are 1 = 1 and 2 = 3. Corresponding eigenvectors are


1
1+i
=1:
,
=3:
.
i
1 i
1)(
3).
Since these eigenvectors correspond to di↵erent eigenvalues, they are already orthogonal. Their
norms are


q
p
p
1
1+i
2
2
= 1 · 1 + ( i) · i = 2,
= |1 + i| + |1 i| = 2.
i
1 i
So defining a matrix U whose columns are the normalized eigenvectors gives
"
#
U=
p1
2
pi
2
which is unitary. Then
U ⇤ AU =
1+i
2
1 i
2

1
0
,
0
.
3
(b) With A as given, the characteristic polynomial is 2 + 1. Thus the eigenvalues are
i. Corresponding eigenvectors are
2 =


i
i
=1:
,
=3:
.
1
1
1
= i and
Since these eigenvectors correspond to di↵erent eigenvalues, they are already orthogonal. Their
norms are


q
q
p
p
i
i
2
2
2
2
= |i| + |1| = 2,
= | i| + |1| = 2.
1
1
So defining a matrix U whose columns are the normalized eigenvectors gives
"
#
U=
pi
2
p1
2
which is unitary. Then
U ⇤ AU =
pi
2
p1
2

i
0
,
0
.
i
552
CHAPTER 7. DISTANCE AND APPROXIMATION
(c) With A as given, the characteristic polynomial is ( +1) (1+i)(1 i) = 2 + 2 = ( +2)(
Thus the eigenvalues are 1 = 1 and 2 = 2. Corresponding eigenvectors are


1+i
1+i
=1:
,
=3:
.
2
1
1).
Since these eigenvectors correspond to di↵erent eigenvalues, they are already orthogonal. Their
norms are


q
q
p
p
1+i
1+i
2
2
2
2
= |1 + i| + |2| = 6,
= |1 + i| + |1| = 3.
2
1
So defining a matrix U whose columns are the normalized eigenvectors gives
"
#
U=
1+i
p
6
p2
6
which is unitary. Then
⇤
U AU =

1+i
p
3
p1
3
1
0
,
0
.
2
(d) With A as given, the characteristic polynomial is
(
1) ((
2)(
Thus the eigenvalues are
3)
(1
1)
2
5 +4 =(
1)2 (
4)
= 4. Corresponding eigenvectors are
3 2 3
2
3
0
1
0
= 1 : 4 1 + i5 , 405 ,
= 4 : 41 i5 .
1
0
2
1
= 1 and
2
i)(1 + i)) = (
2
The eigenvectors corresponding to di↵erent eigenvalues are already orthogonal, and the two eigenvectors corresponding to = 1 are also orthogonal, so these vectors form an orthogonal set. Their
norms are
2
3
2 3
q
0
1
p
2
2
2
4 1 + i5 = |0| + | 1 + i| + |1| = 3,
405 = 1,
1
0
2
3
q
0
p
41 i5 = |0|2 + |1 i|2 + |2|2 = 6.
2
So defining a matrix U whose columns are the normalized eigenvectors gives
2
3
0
1
0
6 1+i
7
p
0 1p6i 7
U =6
4 3
5,
1
2
p
0 p6
3
which is unitary. Then
2
1
U ⇤ AU = 40
0
0
1
0
3
0
05 .
4
16. If A is Hermitian, then A⇤ = A, so that A⇤ A = AA = AA⇤ . Next, if U is unitary, then U U ⇤ = I. But if
U is unitary, then its rows are orthonormal. Since the columns of U ⇤ are the conjugates of the rows of
U , they are orthonormal as well and therefore U ⇤ is also unitary. Hence U U ⇤ = U ⇤ U = I and therefore
U is normal. Finally, if A is skew-Hermitian, then A⇤ = A; then A⇤ A = AA = A( A) = AA⇤ . So
all three of these types of matrices are normal.
17. By the result quoted just before Exercise 16, if A is unitarily diagonalizable, then A⇤ A = AA⇤ ; this is
the definition of A being normal.
Exploration: Geometric Inequalities and Optimization Problems
553
Exploration: Geometric Inequalities and Optimization Problems
1. We have
p p
p
p p
x· y+ y· x=2 x y
q
p 2
p 2 p
kuk = kvk =
x + ( y) = x + y,
|u · v| =
p
p p
p
p
so that 2 x y  x + y · x + y = x + y, and thus
p
xy 
x+y
.
2
p
p 2
2. (a) Since
x
y is a square, it is nonnegative. Expanding then gives x
p
that 2 xy  x + y and thus
x+y
p
xy 
.
2
p p
2 x y+y
0, so
(b) Let AD = a and BD = b. Then by the Pythagorean Theorem,
CD2 = a2
x 2 = b2
y2 ,
p
so that 2CD2 = (a2 + b2 ) x2 y 2 = (x + y)2 (x2 + y 2 ) = 2xy, so that CD = xy. But CD
is clearly at most equal to a radius of the circle, since it extends from a diameter to the circle.
Since the radius is x+y
2 , we again get
p
xy 
x+y
.
2
p
3. Since the area is xy = 100, we have 10 = xy  x+y
2 , so that 40  2(x + y), which is the perimeter.
By the AMGM, equality holds when x = y; that is, when x = y = 10, so the square of side 10 is the
rectangle with smallest perimeter.
4. Let
1
x
play the role of y in the AMGM. Then
r
Equality holds if and only if x =
for x > 0.
1
x
p
1
, or 2 1 = 2  x + .
x
x·
x+
1

x
2
1
x,
i.e., when x = 1. Thus 1 +
1
1
= 2 is the minimum value for f (x)
5. If the dimensions of the cut-out square are x ⇥ x, then the base of the box will be (10
so its volume will be x(10 2x)2 . Then
p
3
V 
with equality holding if x = 10
10
10
possible volume is 10
3 ⇥ 3 ⇥ 3 .
6. We have
x + (10
2x) + (10
3
2x. Thus x =
10
3 ,
2x)
=
so that 10
20
3x
3
2x =
2x) ⇥ (10
2x),
,
10
3
p
(x + y) + (y + z) + (z + x)
2
3
f=
= (x + y + z),
3
3
as well. Then the largest
with equality holding if x + y = y + z = z + x. But this implies that x = y = z; since xyz = 1, we
must have x = y = z = 1, so that f (x, y, z) = (1 + 1)(1 + 1)(1 + 1) = 8. This is the minimum value of
f subject to the constraint xyz = 1.
554
CHAPTER 7. DISTANCE AND APPROXIMATION
7. Use the substitution u = x
y. Then the original expression becomes
u+y+
8
.
uy
Using the AMGM for three variables, we get
r
3
u·y·
u+y+
8

uy
3
8
uy
p
8
3
, or 3 8 = 6  u + y +
,
uy
8
with equality holding if and only if u = y = uy
; this implies that u = y = 2, so that x = u + y = 4. So
8
the minimum values is 4 + 2·2 = 6, which occurs when x = 4 and y = 2.
2
3
x
p
8. Follow the method of Example 7.11. x2 + 2y 2 + z 2 is the square of the norm of v = 4 2 y 5, so that
z
2 3
1
p
x + 2y + 4z is the dot product of that vector with u = 4 25. Then the component-wise form of the
4
Cauchy-Schwarz Inequality gives
✓
◆
⇣p ⌘2
p p
2
2
2
2
(1·x+ 2· 2 y +4·z) = (x+2y +4z)  1 +
2 + 4 (x2 +2y 2 +z 2 ) = 19(x2 +2y 2 +z 2 ) = 19.
So x + 2y + 4z 
p
19, and the maximum value is
p
19.
3
x
9. Follow the method of Example 7.11. Here f (x, y, z) is the square of the norm of v = 4 y 5, so set
2
2
3
pz
2
1
u = 4p1 5. Then the component-wise form of the Cauchy-Schwarz Inequality gives
2
02
3 2 3 12
✓
◆
x
1
⇣p ⌘2 ◆ ✓
1 2
2
2
@4 1 5 · 4 y 5A = (x + y + z)2  12 + 12 +
2
x
+
y
+
z
.
p
2
pz
2
2
Since x + y + z = 10, we get
✓
◆
1 2
1
100le4 x + y + z , or 25  x2 + y 2 + z 2 .
2
2
2
2
Thus the minimum value of the function is 25.


1
sin ✓
2
10. Follow the method of Example 7.11 with u =
and v =
. (This works because kvk =
1
cos ✓
sin2 ✓ + cos2 ✓ = 1.) Then the component-wise form of the Cauchy-Schwarz Inequality gives
(u · v)2 = (sin ✓ + cos ✓)2  (12 + 12 )(sin2 ✓ + cos2 ✓) = 2.
p
p
Therefore sin ✓ + cos ✓  2, so the maximum value of the expression is 2.

x
1
11. We want to minimize x + y subject to the constraint x + 2y = 5. Thus set v =
and u =
.
y
2
Then the component-wise form of the Cauchy-Schwarz Inequality gives
2
2
(x + 2y)2  (12 + 22 )(x2 + y 2 ).

Exploration: Geometric Inequalities and Optimization Problems
555
Since x + 2y = 5, we get
25  5(x2 + y 2 ), or 5  x2 + y 2 .
Thus the minimum value is 5, which occurs at equality. To find the point where equality occurs,
substitution 5 2y for x to get
2y)2 + y 2 = 25
5 = (5
20y + 5y 2 , so that y 2
4y + 4 = 0.
This has the solution y = 2; then x = 5 2y = 1. So the point on x + 2y = 5 closest to the origin is
(1, 2).
p
2
12. To show that xy 1/x+1/y
, apply the AMGM to x0 = x1 and y 0 = y1 , giving
p
x0 + y 0
x0 y 0 
, or
2
r
1
x0 y 0
2
p
, or xy
x0 + y 0
2
.
1/x + 1/y
Since this inequality results from the AMGM, we have equality when x0 = y 0 , so when x = y. For the
first inequality, note first that (x y)2 = x2 2xy + y 2 0, so that x2 + y 2 2xy. Then
r
✓
◆2
x2 + y 2
x2 + y 2 + (x2 + y 2 )
x2 + 2xy + y 2
x+y
x2 + y 2
x+y
=
=
, so that
.
2
4
4
2
2
2
Equality occurs when x2 + y 2 = 2xy, which is the same as (x
y)2 = 0, so that x = y.
13. The area of the rectangle shown is 2xy, where x2 + y 2 = r2 . From Exercise 12, we have
r
p
p
p
p
p
x2 + y 2
p
xy )
x2 + y 2
2xy ) r2 = r
2xy = A ) A  r2 .
2
The maximum occurs at equality, and the maximum area is A = r2 . By Exercise 12, this happens
when x = y.
14. Following the hint, we have
(x + y)2
= (x + y)
xy
✓
1
1
+
x y
◆
.
So from Exercise 12, we have
x+y
2
2
) (x + y)
1/x + 1/y
✓
1
1
+
x y
◆
4.
Thus the minimum value is 4; this minimum occurs when x = y, so when
✓
◆
1
1
4x
(x + x)
+
=
= 4,
x x
x
or x = y = 1.
15. Squaring the inequalities in Exercise 12 gives
x2 + y 2
2
Substituting x +
1
x
for x and y +
x+
1
y
1 2
x
✓
x+y
2
◆2
✓
2
1/x + 1/y
◆2
.
for y gives from the first inequality
⇣
⌘2
+ y + y1
2
0
@
x+
1
x
⇣
⌘ 12
+ y + y1
A .
2
556
CHAPTER 7. DISTANCE AND APPROXIMATION
Since the constraint we are working with is x + y = 1, we get
0
@
x+
1
x
⇣
⌘ 12 0
⇣
⌘ 12
+ y + y1
(x + y) + x1 + y1
A =@
A =
2
2
1
+
2
1
x
+
2
1
y
!2
✓
1
2
+
2 x+y
◆
,
using the second and fourth terms in Exercise 12 for the final inequality. But now again since x+y = 1,
we get
✓
◆ 2 ✓ ◆2
1
2
5
25
+
=
=
,
2 x+y
2
4
so that f (x, y) = 2 ·
x = y = 12 .
7.2
25
4
=
25
2
is the minimum when x + y = 1. This equality holds when x = y, so that
Norms and Distance Functions
1. We have
kukE =
p
p
( 1)2 + 42 + ( 5)2 = 42
kuks = | 1| + |4| + | 5| = 10
kukm = max (| 1| , |4| , | 5|) = 5.
2. We have
kvkE =
p
p
22 + ( 2)2 + 02 = 2 2
kvks = |2| + | 2| + |0| = 4
kvkm = max (|2| , | 2| , |0|) = 2.
3. u
2
3
3
v = 4 65, so
5
dE (u, v) =
p
p
( 3)2 + 62 + ( 5)2 = 70
ds (u, v) = | 3| + |6| + | 5| = 14
dm (u, v) = max (| 3| , |6| , | 5|) = 6.
4. (a) ds (u, v) measures the total di↵erence in magnitude between corresponding components.
(b) dm (u, v) measure the greatest di↵erence between any pair of corresponding components.
5. kukH = w(u), which is the number of 1s in u, or 4. kvkH = w(v), which is the number of 1s in v, or
5.
6. dH (u, v) = w(u
1s in u v.
v) = w(u) + w(v) less twice the overlap, or 4 + 5
2 · 2 = 5; this is the number of
pP
P 2
2
7. (a) If kvkE =
vi2 = max{|vi |} = kvkm , then squaring both sides gives
vi = max{|vi | } =
2
2
max{vi }. Since vi
0 for all i, equality holds if and only if exactly one of the vi is nonzero.
P
(b) Suppose kvks =
|vi | = max{|vi |} = kvkm . Since |vi | 0 for all i, equality holds if and only if
exactly one of the vi is nonzero.
(c) By parts (a) and (b), kvkE = kvkm = kvks if and only if exactly one of the components of v is
nonzero.
7.2. NORMS AND DISTANCE FUNCTIONS
557
8. (a) Since
ku + vkE =
X
(ui + vi )2 =
X
u2i +
X
vi2 + 2
X
ui vi = kukE + kvkE + 2u · v,
we see that ku + vkE = kukE + kvkE if and only if u · v = 0; that is, if and only if u and v are
orthogonal.
(b) We have
ku + vks =
X
|ui + vi | ,
kuks + kvks =
X
(|ui | + |vi |).
For a given i, if the signs of ui and vi are the same, then |ui + vi | = |ui | + |vi |, while if they are
di↵erent, then |ui + vi | < |ui | + |vi |. Thus the two norms are equal exactly when ui and vi have
the same sign for each i, which is to say that ui vi 0 for all i.
(c) We have
ku + vks = max{|ui + vi |},
kuks + kvks = max{|ui |} + max{|vi |}.
Suppose that |ur | = max{|ui |} and |vt | = max{|vi |}. Then for any j, we know that
|uj + vj |  |uj | + |vj |  |ur | + |vt | ,
where the first inequality is the triangle inequality and the second comes from the condition on r
and t. Choose j such that |uj + vj | = ku + vks . If the two norms are equal, then
|uj + vj | = ku + vks = kuks + kvks = |ur | + |vt | .
Thus both of the inequalities above must be equalities. The first is an equality if uj and vj have
the same sign. For the second, since |uj |  |ur | and |vj |  |vt |, equality means that |uj | = |ur | and
|vj | = |vt |. Putting this together gives the following (somewhat complicated) condition for equality
to hold in the triangle inequality for the max norm: equality holds if there is some component j
such that uj and vj have the same sign and such that |uj | is a component of maximal magnitude
of u and |vj | is a component of maximal magnitude of v.
9. Let vk be a component of v with maximum magnitude. Then the result follows from the string of
equalities and inequalities
qX
q
kvkm = |vk | = vk2 
vi2 = kvkE .
P 2
P
2
2
2
10. We have kvkE =
vi  ( |vi |)  kvks . Taking square roots gives the desired result since norms
are always nonnegative.
11. We have kvks =
12. We have kvkE =
P
|vi | 
pP
P
vi2 
max{|vi |} = n max{|vi |} = n kvkm .
q
P
2
(max{vi }) =
q
p
p
2
n max{|vi | } = n max{|vi |} = n kvkm .
13. The unit circle in R2 relative to the sum norm is
{(x, y) : |x| + |y| = 1} = {(x, y) : x + y = 1 or x
y = 1 or
The unit circle in R2 relative to the max norm is
{(x, y) : max{|x| , |y|} = 1}.
x + y = 1 or
x
y = 1}.
558
CHAPTER 7. DISTANCE AND APPROXIMATION
These unit circles are shown below:
!1.0
1.0
1.0
0.5
0.5
1.0
0.5
!0.5
!1.0
0.5
!0.5
!0.5
!0.5
!1.0
!1.0
1.0
14. For example, in R2 , let u = h2, 1i and v = h1, 2i. Then kuk = kvk = 2 + 1 = 3. Also, ku + vk =
kh3, 3ik = 6 and ku vk = kh1, 1ik = 2. But then
2
1
1
2
ku + vk + ku
2
2
2
kuk + kvk = 32 + 32 = 18 6=
2
vk = 18 + 2 = 20.

a
0, and since max{|2a| , |3b|} = 0 if and only if both a and b are zero, we see that the
b
norm is zero only for the zero vector. Thus property (1) is satisfied. For property 2,
15. Clearly

a
c
b
Finally,


=

a
c
+
b
d
ca
cb
= max{|2ac| , |3bc|} = |c| max{|2a| , |3b|}.
=

a+c
b+d
= max{|2(a + c)| , |3(b + d)|}.
Now use the triangle inequality:
max{|2(a + c)| , |3(b + d)|}  max{|2a| + |2c| , |3b| + |3d|}  max{|2a| , |3b|} + max{|2c| , |3d|}.
But the last expression is just
16. Clearly max{|aij |}

a
b
+

c
d
, and we are done.
0, and is zero only when all of the aij are zero, so when A = O. Thus property
i,j
(1) is satisfied. For property (2),
kcAk = max{|caij |} = max{|c| |aij |} = |c| max{|aij |} = |c| kAk .
i,j
i,j
i,j
For property (3), using the triangle inequality for absolute values gives
kA + Bk = max{|aij + bij |}  max{|aij | + |bij |}  max{|aij |} + max{|bij |} = kAk + kBk .
i,j
i,j
i,j
i,j
7.2. NORMS AND DISTANCE FUNCTIONS
559
17. Clearly kf k 0, since |f | 0. Since f is continuous, |f | is also continuous, so its integral is zero if
and only if it is identically zero. So property (1) holds. For property (2), use standard properties of
integrals:
Z 1
Z 1
Z 1
kcf k =
|cf (x)| dx =
|c| · |f (x)| dx = |c|
|f (x)| dx = |c| kf k .
0
0
0
Finally, property (3) relies on the triangle inequality for absolute value:
Z 1
Z 1
Z 1
Z
kf + gk =
|f (x) + g(x)| dx 
(|f (x)| + |g(x)|) dx =
|f (x)| dx +
0
0
0
1
0
|g(x)| dx = kf k + kgk .
18. Since the norm is the maximum of a set of nonnegative numbers, it is always nonnegative, and is zero
exactly when all of the numbers are zero, which means that f is the zero function on [0, 1]. So property
(1) holds. For property (2), we have
kcf k = max {|cf (x)|} = max {|c| |f (x)|} = |c| max {|f (x)|} = |c| kf k .
0x1
0x1
0x1
Finally, property (3) relies on the triangle inequality for absolute values:
kf + gk = max {|f (x) + g(x)|}  max {|f (x)|+|g(x)|}  max {|f (x)|}+ max {|g(x)|} = kf k+kgk .
0x1
0x1
0x1
0x1
19. By property (2) of the definition of a norm,
d(u, v) = ku
20.
21.
22.
23.
24.
25.
vk = | 1| ku vk = k( 1)(u v)k = kv uk = d(v, u).
p
p
The Frobenius norm is kAkF = 22 + 32 + 42 + 12 = 30. From Theorem 7.7 and the comment
following, kAk1 is the largest column sum of |A|, which is 2 + 4 = 6, while kAk1 is the largest row
sum of |A|, which is 2 + 3 = 4 + 1 = 5.
p
p
The Frobenius norm is kAkF = 02 + ( 1)2 + ( 3)2 + 32 = 19. From Theorem 7.7 and the comment following, kAk1 is the largest column sum of |A|, which is | 1| + 3 = 4, while kAk1 is the largest
row sum, which is | 3| + 3 = 6.
p
p
The Frobenius norm is kAkF = 12 + 52 + ( 2)2 + ( 1)2 = 31. From Theorem 7.7 and the comment following, kAk1 is the largest column sum of |A|, which is 5 + | 1| = 6, while kAk1 is the largest
row sum, which is 1 + 5 = 6.
p
p
The Frobenius norm is kAkF = 22 + 12 + 12 + 12 + 32 + 22 + 12 + 12 + 32 = 31. From Theorem
7.7 and the comment following, kAk1 is the largest column sum of |A|, which is 1 + 2 + 3 = 6, while
kAk1 is the largest row sum, which is 1 + 3 + 2 = 6.
p
p
The Frobenius norm is kAkF =
02 + ( 5)2 + 22 + 32 + 12 + ( 3)2 + ( 4)2 + ( 4)2 + 32 = 89.
From Theorem 7.7 and the comment following, kAk1 is the largest column sum of |A|, which is | 5| +
1 + | 4| = 10, while kAk1 is the largest row sum, which is | 4| + | 4| + 3 = 11.
p
p
2
2
2
2
2
2
2
2
2
The
p Frobenius norm is kAkF = 4 + ( 2) + ( 1) + 0 + ( 1) + 2 + 3 + ( 3) + 0 = 44 =
2 11. From Theorem 7.7 and the comment following, kAk1 is the largest column sum of |A|, which is
4 + 0 + 3 = 7, while kAk1 is the largest row sum, which is 4 + | 2| + | 1| = 7.
26. Since kAk1 = 6, which is the sum of the first column, such a vector x is e1 , since




2 3 1
2
2
=
, and
= 6.
4 1 0
4
4 s

1
Since kAk1 = 5, which is the sum of (say) the first row, let y =
; then
1




2 3 1
5
5
=
, and
= 5.
4 1 1
5
5 m
560
CHAPTER 7. DISTANCE AND APPROXIMATION
27. Since kAk1 = 4, which is the sum of the magnitudes of the second column, such a vector x is e2 , since




0
1 0
1
1
=
, and
= | 1| + 3 = 4.
3
3 1
3
3 s
Since kAk1 = 6, which is the sum of the magnitudes of the second row, let y =
column in that row is negative; then



0
1
1
1
=
, and
3
3
1
6

1
6

1
since the first
1
= 6.
m
28. Since kAk1 = 6, which is the sum of the magnitudes of the second column, such a vector x is e2 , since




1
5 0
5
5
=
, and
= 5 + | 1| = 6.
2
1 1
1
1 s

1
Since kAk1 = 6, which is the sum of the first row, let y =
; then
1




1
5 1
6
6
=
, and
= 6.
2
1 1
3
3 m
29. Since kAk1 = 6, which is the sum
2
2
41
1
Since kAk1
of the magnitudes of the third column, such a vector x is e3 , since
32 3 2 3
2 3
1 1
0
1
1
3 25 405 = 425 , and 425 = 6.
1 3
1
3
3 s
2 3
1
= 6, which is the sum of the second row, let y = 415; then
1
2
32 3 2 3
2 3
2 1 1
1
4
4
41 3 25 415 = 465 , and 465
= 6.
1 1 3
1
5
5 m
30. Since kAk1 = 10, which is the sum of the magnitudes of the second column, such a vector x is e2 , since
2
32 3 2 3
2 3
0
5
2 0
5
5
4 3
1
35 415 = 4 15 , and 4 15 = 10.
4
4
3 0
4
4 s
2 3
1
Since kAk1 = 11, which is the sum of the magnitudes of the third row, let y = 4 15 since the entries
1
in the first two columns are negative; then
2
32 3 2 3
2 3
0
5
2
1
7
7
4 3
1
35 4 15 = 4 75 , and 4 75
= 11.
4
4
3
1
11
11 m
31. Since kAk1 = 7, which is the sum
2
4
40
3
of the magnitudes of the first column, such a vector x is e1 , since
32 3 2 3
2 3
2
1 1
4
4
1
25 405 = 405 , and 405 = 7.
3
0 0
3
3 s
7.2. NORMS AND DISTANCE FUNCTIONS
561
2
3
1
Since kAk1 = 7, which is the sum of the magnitudes of the first row, let y = 4 15 since the entries
1
in the last two columns are negative; then
2
32 3 2 3
2 3
4
2
1
1
7
7
40
1
25 4 15 = 4 15 , and 4 15
= 7.
3
3
0
1
6
6 m
32. Let Ai be the ith row of A, let M =
max {kAj ks } be the maximum absolute row sum, and let x
j=1,2,...,n
be any vector with kxkm = 1 (which means that |xi |  1 for all i). Then
kAxkm = max{|Ai x|}
i
= max{|ai1 x1 + ai2 x2 + · · · + ain xn |}
i
 max{|ai1 | |x1 | + |ai2 | |x2 | + · · · + |ai1 | |xn |}
i
 max{|ai1 | + |ai2 | + · · · + |ain |}
i
= max{kAi ks } = M.
i
Thus kAxkm is at most equal to the maximum absolute row sum. Next, we find a specific unit vector
x so that it is actually equal to M . Let row k be the row of A that gives the maximum sum norm;
that is, such that kAk ks = M . Let
(
1
Akj 0
yj =
, j = 1, 2, . . . , n.
1 Akj < 0
⇥
⇤T
Then with y = y1 · · · yn , we have kykm = 1, and we have corrected any negative signs in row
k, so that akj yj 0 for all j and therefore the k th entry of Ay is
X
X
X
X
akj yj =
|akj yj | =
|akj | |yj | =
|akj | = kAk ks = M.
j
j
j
j
So we have shown that the 1 norm of A is bounded by the maximum row sum, and then found a
vector x for which kAxk  kAk1 is equal to the maximum row sum. Thus kAk1 is equal to the
maximum row sum, as desired.
33. (a) If k·k is an operator norm, then kIk = max kI x̂k = max kx̂k = 1.
kx̂k=1
kx̂k=1
pP
p
(b) No, there is not, for n > 1, since in the Frobenius norm, kIkF =
12 = n 6= 1. When n = 1,
the Frobenius norm is the same as the absolute value norm on the real number line.
34. Let
be an eigenvalue, and v a corresponding unit vector. Normalize v to v̂. Then Av̂ = v̂, so that
kAk = max kAx̂k
kAv̂k = k v̂k = | | kv̂k = | | .
kx̂k=1
35. Computing the inverse gives
A
1
=
"
1
2
1
2
3
2
#
.
Then
cond1 (A) = kAk1 A
cond1 (A) = kAk1 A
The matrix is well-conditioned.
1
1
1
= (3 + 4) (1 + | 2|) = 21
✓
◆
3
=
(4
+
2)
|
2|
+
= 21.
1
2
562
CHAPTER 7. DISTANCE AND APPROXIMATION
36. Since det A = 0, by definition cond1 (A) = cond1 (A) = 1, so this matrix is ill-conditioned.
37. Computing the inverse gives
1
A
=
Then

1
cond1 (A) = kAk1 A
99
.
100
= (1 + 1)(100 + | 100|) = 400
1
1
cond1 (A) = kAk1 A
100
100
1
= (1 + 1)(100 + | 100|) = 400.
The matrix is ill-conditioned.
38. Computing the inverse gives
1
A
=
"
#
2
2001
50
3001
100
3
2
.
Then
1
cond1 (A) = kAk1 A
◆
2001 3001
+
= 294266
50
100
✓
◆
2001
= (3001 + 4002)
+ 2 = 294266.
50
= (200 + 4002)
1
1
cond1 (A) = kAk1 A
1
✓
The matrix is ill-conditioned.
39. Computing the inverse gives
1
A
Then
cond1 (A) = kAk1 A
2
0
=4 6
5
1
= (1 + 5 + 1)(0 + 6 + 5) = 77
1
1
cond1 (A) = kAk1 A
3
1
15 .
0
0
1
1
1
= (5 + 5 + 6)(6 + 1 + 1) = 128.
The matrix is moderately ill-conditioned.
40. Computing the inverse gives
A
2
9
= 4 36
30
1
Then
1
cond1 (A) = kAk1 A
1
cond1 (A) = kAk1 A
=
1
1
3
30
1805 .
180
36
192
180
◆
1 1
1+ +
(36 + 192 + 180) = 748
2 3
✓
◆
1 1
= 1+ +
(36 + 192 + 180) = 748.
2 3
✓
The matrix is ill-conditioned.
41. (a) Computing the inverse gives
A
Then
cond1 (A) = kAk1 A
1
1
1
=
1
1
k

1
1
k
.
1
= max{|k| + 1, 2} max
⇢
k
k
1
+
1
k
1
,
2
k
1
.
7.2. NORMS AND DISTANCE FUNCTIONS
563
(b) As k ! 1, since the second factor goes to 1, cond1 (A) ! 1. This makes sense since when
k = 1, A is not invertible.
42. Since the matrix norm is compatible, we have from the definition that kAxk  kAk kxk for any x, so
that
1
kAk
kbk = AA 1 b  kAk A 1 b , so that

.
kA 1 bk
kbk
Then
A 1 b
A 1 k bk
k xk
kAk
=


A
kxk
kA 1 bk
kA 1 bk
kbk
43. (a) Since
1
A
=
"
1
k bk = cond(A)
k bk
.
kbk
#
1
,
1
9
10
1
we have cond1 (A) = (10 + 10)(1 + 1) = 40.
(b) Here

10
A=
10
Therefore

10
11
10
10

10
0
=
9
0
0
2
) k Ak = 2.
k xkm
k Ak1
2
 cond1 (A)
= 40 · ,
kx0 km
kAk1
20
so that this can produce at most a change of a factor of 4, or a 400% relative change.
(c) Row-reducing gives
⇥
⇥
Thus
and
k xkm
kxkm
A
A0


⇤
10
b =
10

⇤
10
b =
10
9
x=
,
1
10
9
100
99
!
10
11
100
99
!

11
x =
, so that
1
0


1
0
0
1
9
1
1
0
0
1
11
.
1
x=x
0
x=

2
,
2
= 29 , for approximately a 22% relative error.
(d) Using Exercise 42,
b=

100
101


100
0
=
,
99
2
so that k bkm = 2 and
k xkm
k bkm
2
 cond1 (A)
= 40 ·
= 0.8,
kxkm
kbkm
100
so that at most an 80% change can result.
⇥
⇤
(e) Row-reducing A b gives the same result as above, while


⇥
⇤
10 10 100
1
A b0 =
!
10 9 101
0
Thus
x=
and again
k xkm
kxkm

11
1

9
=
1

2
,
2
= 29 , for about a 22% actual relative error.
0
1
11
.
1
564
CHAPTER 7. DISTANCE AND APPROXIMATION
44. (a) Since
A
1
2
10
6
=4 4
7
3
1
2
we have cond1 (A) = (1 + 5 + | 1|)(| 10| + 4 + 7) = 147.
(b) Here
2
1
A = 42
1
Therefore
3
1 1
5 05
1 2
2
3 2
1 1
0
5 05 = 41
1 2
0
1
41
1
3
5
7
25 ,
3
0
0
0
3
0
05 ) k Ak1 = 1.
0
k xks
k Ak1
1
 cond1 (A)
= 147 · = 21,
kx0 ks
kAk1
7
so that this can produce at most a change of a factor of 21, or a 2100% relative change.
(c) Row-reducing gives
h
h
Thus
and
k xks
kxks
A
A0
2
=
33
21
2
1
6
b = 42
1
2
1
i
6
b = 41
1
i
3
11
6 7
x = 4 45 ,
6
1 1
5 0
1 2
1 1
5 0
1 2
2
6
x0 = 4
3
11
2
3 7
,
2 5
3
1
7
25
3
3
1
7
25
3
!
!
so that
2
1
6
40
0
2
1
6
40
0
0
1
0
0
0
0
0
1
0
0
0
1
x = x0
5
3
11
7
45
6
3
11
2
3 7
.
2 5
5
2
6
x=4
⇡ 1.57, for approximately a 157% actual relative error.
(d) Using Exercise 42,
so that k bks = 1 and
2 3
1
b = 415
3
3
33
2
11 7
,
2 5
11
2 3 2 3
1
0
425 = 4 15 ,
3
0
k xks
k bks
1
 cond1 (A)
= 147 · = 24.5,
kxks
kbks
6
so that at most a 2450% relative change can result.
⇥
⇤
(e) Row-reducing A b gives the same result as above, while
2
3
2
1
1 1 1
1
⇥
⇤
A b0 = 41
5 0 15 ! 4 0
1
1 2 3
0
Thus
and
k xks
kxks
2
=
30
21
3
4
x = 4 15
4
0
1
0
2
3 2
3
11
15
4 45 = 4 55
6
10
⇡ 1.43, for about a 143% actual relative error.
0
0
1
3
4
15 .
4
7.2. NORMS AND DISTANCE FUNCTIONS
1
45. From Exercise 33(a), 1 = kIk = A
565
1
A  A
kAk = cond(A).
46. From Exercise 33(a),
1
cond(AB) = (AB)
kABk = B
1
A
47. From Theorem 4.18 (b) in Chapter 4, if
Exercise 34, kAk
|
1|
and A
n
1
1
n
1
kAk B
1
kBk = cond(A) cond(B).
is an eigenvalue of A, then
1
is an eigenvalue of A
48. With

A=
we have
M=
D
"
(L + U ) =
so that kM k1 = 15 . Then
" #
6
b=
)
4
#
0
5
7
0
c=D
1
n
1
. By
, so that
1
cond(A) = A
1
1
kABk  A
b=
1
"
"
|
|
kAk
1|
n|
7
1
1
,
5
0
1
# "
1
1
7
=
0
0
1
7
0
0
1
5
#"
.
#"
0 0
1
1
5
# " #
6
6
= 74
4
5
k
)
# "
1
0
= 1
0
5
kckm =
1
7
0
#
,
6
.
7
k
Note that x0 = 0, so that x1 = c and then kM k1 kx1 x0 km = kM k1 kckm < 0.0005 means that
4+log (kck /5)
k > log 101/kMm
. In this case, with kM k1 = 15 and kckm = 67 , we get
k1 )
10 (
4 + log10 (6/35)
⇡ 4.6, so that k 5.
log10 5

0.999
= M xk + c gives x5 =
, as compared with an actual solution of
0.999
k>
Computing x5 using xk+1

1
x=
.
1
49. With
A=
we have
M=
D
1
(L + U ) =
so that kM k1 = 27 . Then
" #
1
b=
)
1
"
4.5
0
c=D
#
0
3.5
1
b=

1
"
4.5
1
0.5
,
3.5
"
# "
2
0.5
9
=
0
0
0
1
2
9
0
0
2
7
#"
# " #
2
1
= 92
1
7
k
#"
0 0
2
1
7
)
k
# "
0.5
0
= 2
0
7
kckm =
1
9
0
#
,
2
.
7
Note that x0 = 0, so that x1 = c and then kM k1 kx1 x0 km = kM k1 kckm < 0.0005 means that
4+log (kck /5)
k > log 101/kMm
. In this case, with kM k1 = 27 and kckm = 27 , we get
k1 )
10 (
k>
4 + log10 (2/35)
⇡ 5.06, so that k
log10 (7/2)
6.
566
CHAPTER 7. DISTANCE AND APPROXIMATION

0.2622
Computing x6 using xk+1 = M xk + c gives x6 =
, as compared with an actual solution of
0.3606

0.2623
x=
.
0.3606
50. With
2
20
A=4 1
1
we have
M=
D
1
2
20
6
40
0
(L + U ) =
2
so that kM k1 = 10
= 15 . Then
2 3
17
6 7
b = 4135 ) c = D
18
1
3
0
7
05
10
0
10
0
2
6
b=4
1
1
20
0
1
10
1
10
3
1
15 ,
10
1
10
1
2
0
6
4 1
1
1
0
1
32
3
1
17
20
1 76 7
10 5 4135
0
1
10
0
k
18
3 2
1
0
7 61
15 = 4 10
1
0
10
2
6
=4
3
17
20
13 7
20 5
9
5
3
1
20
1
20
1 7
10 5
0
1
10
)
0
kckm =
9
.
5
k
Note that x0 = 0, so that x1 = c and then kM k1 kx1 x0 km = kM k1 kckm < 0.0005 means that
4+log (kck /5)
k > log 101/kMm
. In this case, with kM k1 = 15 and kckm = 95 , we get
k1 )
10 (
4 + log10 (9/25)
⇡ 5.08, so that k 6.
log10 5
2
3
0.999
= M xk + c gives x6 = 4 1.0005, as compared with an actual solution of
1.999
k>
Computing x6 using xk+1
2
3
1
x = 4 15 .
2
51. With
2
we have
M=
D
1
3
6
40
0
(L + U ) =
= 12 . Then
2 3
1
6 7
b = 415 ) c = D
1
so that kM k1 =
2
3
A = 41
0
2
4
1
0
4
0
2
0
6
b = 4 14
0
3
0
15 ,
3
1
4
1
3
0
7
05
3
1
3
0
1
3
k
1
2
0
6
41
0
1
0
1
3 2
0
0
7 6 1
15 = 4 4
0
0
32 3 2 3
1
1
3
617
17 6 7
=
1
5
4
5
4
4
45
1
0
1
3
1
3
0
1
3
0
17
45
0
0
)
k
3
kckm =
1
.
3
Note that x0 = 0, so that x1 = c and then kM k1 kx1 x0 km = kM k1 kckm < 0.0005 means that
4+log (kck /5)
k > log 101/kMm
. In this case, with kM k1 = 12 and kckm = 13 , we get
k1 )
10 (
k>
4 + log10 (1/15)
⇡ 9.3, so that k
log10 2
10.
7.2. NORMS AND DISTANCE FUNCTIONS
567
Computing x10 using xk+1 = M xk + c gives x10
2
3
0.3
x = 40.15.
0.3
2
3
0.299
= 40.0995, as compared with an actual solution of
0.299
52. The sum and max norms are operator norms induced from the sum and max norms on Rn , so they
are matrix norms.
(a) If C and D are any two matrices and k·k is either the sum or max norm, then kCDk  kCk kDk.
Now assume A is such that kAk = < 1. Then
n
kAn k  kAk =
Thus limn!1 kAn k = limn!1
An ! O.
n
n
.
= 0. So by property (1) in the definition of matrix norms,
(b) A computation shows that
(I
A)(I + A + · · · + An ) = I
An+1 ,
since all the other terms cancel in pairs. Thus
(I
A)(I + A + A2 + A3 + · · · ) = (I
A) lim (I + A + · · · + An )
n!1
= lim ((I
n!1
= lim I
n!1
A)(I + A + · · · + An ))
An+1
= I.
It follows that I
A is invertible and that its inverse is I + A + A2 + A3 + · · · .
(c) Recall that a consumption matrix C is a matrix such that all entries are nonnegative and the sum
of the entries in each column does not exceed 1. A consumption matrix C is productive if I C
is invertible and (I C) 1 O; that is, if it has only nonnegative entries.
To prove Corollary 3.35, use k·k1 , the max norm. Note that since C O, the sum of the absolute
values of the entries of a row is the same as the sum of the entries of the row. If the sum of each
row of C is less than 1, then kCk1 < 1, so by part (b), I C is invertible. Since C
O, it is
clear that each entry of (I C) 1 = I + C + C 2 + C 3 + · · · is nonnegative, so that (I C) 1 O.
Thus C is productive.
The proof of Corollary 3.36 is similar, but it uses k·k1 , the sum norm. If the sum of each column
of C is less than 1, then kCk1 < 1. The rest of the argument is identical to that in the preceding
paragraph.
568
7.3
CHAPTER 7. DISTANCE AND APPROXIMATION
Least Squares Approximation
1. The corresponding data points predicted by the line y =
least squares error is
p
(0 0)2 + (2 1)2 + (4
2 + 2x are (1, 0), (2, 2), and (3, 4), so the
5)2 =
p
2.
A plot of the points and the line on the same set of axes is
6
4
2
0.5
1.
2.
1.5
3.
2.5
3.5
!2
2. The corresponding data points predicted by the line y = x are (1, 1), (2, 2), and (3, 3), so the least
squares error is
p
p
(1 0)2 + (2 1)2 + (3 5)2 = 6.
A plot of the points and the line on the same set of axes is
5
4
3
2
1
1
2
3
4
!1
3. The corresponding data points predicted by the line y = 3 + 52 x are 1,
least squares error is
s✓
p
◆2
✓
◆2
1
9
6
2
0 + (2 1) +
5 =
.
2
2
2
A plot of the points and the line on the same set of axes is
4
2
0.5
!2
1.
1.5
2.
2.5
3.
3.5
1
2
, (2, 2), and 3, 29 , so the
7.3. LEAST SQUARES APPROXIMATION
569
4. The corresponding data points predicted by the line y = 3 13 x are
✓
◆ ✓
◆
✓
◆
1
14
1
5, 3
( 5) =
5,
,
0, 3
· 0 = (0, 3),
3
3
3
✓
◆ ✓
◆
✓
◆ ✓
◆
4
1
1
1
5, 3
· 5 = 5,
,
10, 3
· 10 = 10,
,
3
3
3
3
so the least squares error is
s✓
◆2
14
3
+ (3
3
3)2
✓
+ 2
4
3
◆2
✓
✓
+ 0
1
3
◆◆2
=
r
25
4 1
+0+ + =
9
9 9
p
30
.
3
A plot of the points and the line on the same set of axes is
5
4
3
2
1
10
5
!5
5. The corresponding data points predicted by the line y = 3 13 x are
✓
◆
✓
◆
✓
◆
✓
◆
5
5
5
5
5,
,
0,
,
5,
,
10,
,
2
2
2
2
so the least squares error is
s✓
◆2 ✓
5
3
+ 3
2
5
2
◆2
✓
+ 2
5
2
◆2
✓
+ 0
✓ ◆◆2 r
5
1 1 1 25 p
=
+ + +
= 7.
2
4 4 4
4
A plot of the points and the line on the same set of axes is
5
4
3
2
1
6. The corresponding data points predicted by the line y = 2
( 5, 3) ,
so the least squares error is
q
(3
10
5
!5
2
3) + (3
(0, 2) ,
2
2) + (2
(5, 1) ,
2
1
5x
are
(10, 0) ,
1) + (0
0) 62 =
p
2.
570
CHAPTER 7. DISTANCE AND APPROXIMATION
A plot of the points and the line on the same set of axes is
3.
2.5
2.
1.5
1.
0.5
10
5
!5
7. We are looking for a line y = a+bx that best fits the given points. Using the least squares approximation
theorem, we want a solution to Ax = b, where
3
2 3
2
1 x1
y1
2
3
2 3

1 1
0
61 x2 7
6 y2 7
a
6
7 4
6 7
A = 6.
= 1 25 , x =
, b = 6 . 7 = 4 15 .
7
.
.. 5
b
4 ..
4 .. 5
1 3
5
1 xn
yn
Since
"
1
A A=
1
T
2
# 1
1 6
41
3
1
1
2
we get
x = (AT A)
1
3
"
1
3
7
25 =
6
3
AT b =
"
Thus the least squares approximation is y =
Then kek =
q
1 2
2
1 2
2
+ ( 1)2 +
=
p
#"
1 1
1
1
2
7
3
1
1
2
1
=
"
7
3
1
1
1
2
#
,
2 3
# 0
" #
1 6 7
3
.
415 =
5
3
2
5
3 + 52 x. To find the least squares error, compute
2 3
0
6 7
Ax = 415
5
e=b
#
6
) (AT A)
14
2
1
6
41
1
3
2 3
1
1 " #
2
3
7
6 7
=
25
15 .
4
5
1
2
3
2
6
2 .
8. We are looking for a line y = a+bx that best fits the given points. Using the least squares approximation
theorem, we want a solution to Ax = b, where
2
3
2 3
1 x1
y1
2
3
2 3

1 1
6
61 x2 7
6 y2 7
a
6
7 4
6 7 4 5
5
1
2
3 .
A = 6.
=
,
x
=
,
b
=
=
6 .. 7
.. 7
b
4 ..
4.5
. 5
1 3
1
1 xn
yn
Since
"
1
A A=
1
T
1
2
2
# 1
1 6
41
3
1
3
"
1
3
7
25 =
6
3
#
6
) (AT A)
14
1
=
"
7
3
1
1
1
2
#
,
7.3. LEAST SQUARES APPROXIMATION
571
we get
"
7
3
Thus the least squares approximation is y =
25
3
T
x = (A A)
1
A b=
q
1 2
6
+
1 2
3
1 2
6
+
1
2 3
6
6 7
Ax = 435
1
e=b
Then kek =
T
=
#"
1 1
1
1
2
p
5
2 x.
2
25
3
5
2
#
.
To find the least squares error, compute
3
1 "
7
25
3
1
6
41
1
1
2
2 3
# 6
"
1 6 7
4 35 =
3
1
25
3
5
2
#
2
6
=4
3
1
6
17
.
35
1
6
6
6 .
9. We are looking for a line y = a+bx that best fits the given points. Using the least squares approximation
theorem, we want a solution to Ax = b, where
2
3
2 3
1 x1
y1
2
3
2 3

1 0
4
61 x2 7
6 y2 7
a
6
7 4
6 7 4 5
5
1
1
1 .
A = 6.
=
,
x
=
,
b
=
=
7
6
7
..
.
b
4 ..
4 .. 5
. 5
1 2
0
1 xn
yn
Since
"
1
A A=
0
T
2
# 1
1 6
41
2
1
1
1
we get
3
"
0
3
7
15 =
3
2
"
5
6
1
2
Thus the least squares approximation is y =
11
3
x = (AT A)
1
AT b =
2 3
4
6 7
Ax = 415
0
e=b
Then kek =
q
1 2
3
+
2 2
3
1 2
3
+
=
p
#
3
) (AT A)
5
1
2
1
2
#"
1
0
1
1
1
=
"
5
6
1
2
1
2
1
2
#
,
2 3
# 4
" #
11
1 6 7
415 = 3 .
2
2
0
2x. To find the least squares error, compute
2
1
6
41
1
3
2
0 " 11 #
7
6
15 3 = 4
2
2
3
1
3
27
.
35
1
3
6
3 .
10. We are looking for a line y = a+bx that best fits the given points. Using the least squares approximation
theorem, we want a solution to Ax = b, where
2
3
2 3
1 x1
y1
2
3
2 3

1 0
3
61 x2 7
6 y2 7
a
6
7 4
6
7
5
4
1
1
A = 6.
, x=
, b = 6 . 7 = 35 .
.. 7 =
b
4 ..
4 .. 5
. 5
1 2
5
1 xn
yn
Since
"
1
A A=
0
T
1
1
2
# 1
1 6
41
2
1
3
"
0
3
7
15 =
3
2
#
3
) (AT A)
5
1
=
"
5
6
1
2
1
2
1
2
#
,
572
CHAPTER 7. DISTANCE AND APPROXIMATION
we get
x = (AT A)
1
AT b =
"
5
6
1
2
1
2
1
2
#"
1
0
2 3
# 3
" #
8
1 6 7
435 = 3 .
2
1
5
1
1
Thus the least squares approximation is y = 83 + x. To find the least squares error, compute
2 3 2
3
2 3
1
3
1 0 "8#
3
6 7 6
7 3
6 27
e = b Ax = 435 41 15
= 4 35 .
1
1
5
1 2
3
q
p
2
1 2
2 2
Then kek =
+
+ 13 = 36 .
3
3
11. We are looking for a line y = a+bx that best fits the given points. Using the least squares approximation
theorem, we want a solution to Ax = b, where
2
3 2
2 3 2 3
3
1 x1
y1
1
5
1

61 x2 7 6
6 y2 7 6 7
7
1
0
a
6
7 6
6 7 6 17
7
A = 6.
, x=
, b = 6 . 7 = 4 5.
.. 7 = 41
55
b
2
4 ..
4 .. 5
. 5
1
10
4
1 xn
yn
Since
T
A A=
"
1
5
1
0
1
5
we get
x = (AT A)
1
2
# 1
1 6
61
6
10 41
1
AT b =
"
3
5
"
07
4
7
7=
55
10
#
10
) (AT A)
150
1
=
"
3
10
1
50
1
50
1
125
#
,
10
3
10
1
50
1
50
1
125
#"
1
5
1
0
1
5
2 3
1
#
" #
6
7
1 6 17
7
.
6 7 = 10
8
10 4 25
25
4
7
8
Thus the least squares approximation is y = 10
+ 25
x. To find the least squares error, compute
2 3 2
3
2 13
1
1
5 " #
10
6 17 61
6 37
7
07
6 7 6
7 10
6 10 7
e = b Ax = 6 7 6
7 8 = 6 3 7.
4 25 41
4 10 5
55 25
1
4
1 10
10
q
p
5
1 2
3 2
3 2
1 2
Then kek =
+
+
+
=
10
10
10
10
5 .
12. We are looking for a line y = a+bx that best fits the given points. Using the least squares approximation
theorem, we want a solution to Ax = b, where
2
3 2
2 3 2 3
3
1 x1
y1
1
5
3

61 x2 7 6
6 y2 7 6 7
7
07
a
6
7 61
6 7 637
A = 6.
, x=
, b = 6 . 7 = 4 5.
.. 7 = 41
55
b
2
4 ..
4 .. 5
. 5
1
10
0
1 xn
yn
Since
T
A A=
"
1
5
1
0
1
5
2
# 1
1 6
61
6
10 41
1
3
5
"
07
4
7
7=
55
10
10
#
10
) (AT A)
150
1
=
"
3
10
1
50
1
50
1
125
#
,
7.3. LEAST SQUARES APPROXIMATION
573
we get
T
1
x = (A A)
T
A b=
"
3
10
1
50
1
50
1
125
#"
1
5
1
0
2 3
# 3
"
7
1 6
637
6 7=
10 425
1
5
5
2
1
5
#
.
0
5
2
Thus the least squares approximation is y =
2 3
3
637
6 7
e = b Ax = 6 7
425
0
Then kek =
q
1 2
2
+
1 2
2
+
1 2
2
1 2
2
+
1
5 x.
2
1
61
6
6
41
1
To find the least
3
2
5 " #
6
5
07
7
2 =6
7
6
1
4
55
5
10
squares error, compute
3
1
2
17
27 .
17
5
2
1
2
= 1.
13. We are looking for a line y = a+bx that best fits the given points. Using the least squares approximation
theorem, we want a solution to Ax = b, where
3
2
3 2
2 3 2 3
1 1
1
1 x1
y1
7
637

1
2
61 x2 7 6
6
7
7
a
6
7 6
6 y2 7 6 7
1 37
47
A = 6.
, x=
, b=6 . 7=6
.. 7 = 6
6
7
7.
.
b
4.
4 .. 5 6
. 5 41 45
455
1 xn
yn
1 5
7
Since
"
1
A A=
1
T
1
2
1
3
1
4
we get
T
x = (A A)
1
T
2
1
#6
61
1 6
61
5 6
6
41
1
A b=
"
3
1
7
27 "
7
5
37
7 = 15
7
45
5
11
10
3
10
Thus the least squares approximation is y = 15
2 3
1
6 7
637
6 7
7
e = b Ax = 6
647
6 7
455
7
q
2
1 2
2 2
Then kek =
+ 25 + 02 +
+ 15
5
5
3
10
1
10
#
15
) (AT A)
55
#"
1
1
1
2
1
3
1
4
1
=
"
11
10
3
10
2 3
1
7
#6
6 37 "
7
1 6
6 47 =
6
5 6 7
7
4 55
7
3
10
1
10
1
5
7
5
#
#
,
.
+ 75 x. To find the least squares error, compute
2
3
2 3
1
1 1
6
7 " # 6 52 7
6 1 27
6 57
1
6
7
6 7
5 =6
6 1 37
7
6
7
6 07 .
7
6
7
6
27
5
4 1 45
4 55
1
1 5
5
2
=
p
10
5 .
14. We are looking for a line y = a+bx that best fits the given points. Using the least squares approximation
theorem, we want a solution to Ax = b, where
3
2
3 2
2 3 2 3
1 1
10
1 x1
y1
7
687

1
2
61 x2 7 6
6
7
7
a
6
7 6
6 y2 7 6 7
1 37
57
A = 6.
, x=
, b=6 . 7=6
.. 7 = 6
6
7
7.
.
b
4.
4 .. 5 6
. 5 41 45
435
1 xn
yn
1 5
0
574
CHAPTER 7. DISTANCE AND APPROXIMATION
Since
"
1
A A=
1
T
1
2
1
3
1
4
we get
x = (AT A)
1
2
3
1
7
#
27 "
7
5 15
7
) (AT A)
37 =
15 55
7
45
5
1
#6
61
1 6
61
5 6
6
41
1
AT b =
"
11
10
3
10
3
10
1
10
Thus the least squares approximation is y = 127
10
2 3
10
6 7
687
6 7
7
e = b Ax = 6
657
6 7
435
0
Then kek =
q
1 2
5
+
3 2
10
+
1 2
5
+
3 2
10
#"
1
1
5
2 x.
2
1
6
61
6
61
6
6
41
1
1
2
1
4
=
"
11
10
3
10
3
10
1
10
#
,
2
3
10
7
#6
6 8 7 " 127 #
7
1 6
6 5 7 = 10 .
6
5
5 6 7
7
2
435
0
To find the least squares error, compute
3
2 3
1
1
7 " # 6 35 7
6 10 7
27 127
7
6 17
10
7
7
=6
37
6 57 .
5
7
6 37
2
45
4 10 5
1
5
5
1 2
5
+
1
3
1
p
30
10 .
=
15. Following the method of Example 7.28, we substitute the given values of x into the quadratic equation
a + bx + cx2 = y to get
a+ b+
which in matrix form is
2
1
61
6
41
1
c=
1
a + 2b + 4c =
2
a + 3b + 9c =
3
a + 4b + 16c =
4,
1
2
3
4
3
2
1 2 3
a
6
47
7 4b5 = 6
5
4
9
c
16
3
1
27
7.
35
4
Thus we want the least squares approximation to Ax = b, where A is the 4 ⇥ 3 matrix on the left and
b is the matrix on the right in the above equation. Since
2
3
2
3 1 1 1
2
3
2
3
31
27
5
1 1 1 1 6
4 10 30
7
4
4
4
6
7 61 2 4 7 6
7
6
129
57
AT A = 4 1 2 3 4 5 6
,
7 = 410 30 1005 ) (AT A) 1 = 4 27
4
20
45
41 3 9 5
5
5
1
1 4 9 16
30 100 354
4
4
4
1 4 16
we get
x = (AT A)
1
2
6
AT b = 4
31
4
27
4
5
4
27
4
129
20
5
4
Thus the least squares approximation is y = 3
32
5
1
4
57 6
1
5
4
4
1
1
4
18
5 x
+ x2 .
1
2
4
1
3
9
3
2
1 6
76
4 56
4
16
3
2
3
1
3
7
27 6 18 7
7 = 4 5 5.
35
1
4
7.3. LEAST SQUARES APPROXIMATION
575
16. Following the method of Example 7.28, we substitute the given values of x into the quadratic equation
a + bx + cx2 = y to get
a+ b+
c=6
a + 2b + 4c = 0
a + 3b + 9c = 0
a + 4b + 16c = 2,
which in matrix form is
2
1
61
6
41
1
1
2
3
4
3
2 3
1 2 3
6
a
607
47
7 4b5 = 6 7 .
405
95
c
16
2
Thus we want the least squares approximation to Ax = b, where A is the 4 ⇥ 3 matrix on the left and
b is the matrix on the right in the above equation. Since
2
3
2
3 1 1 1
2
3
2
3
31
27
5
1 1 1 1 6
4 10 30
7
4
4
4
6
7 61 2 4 7 6
7
6
129
57
AT A = 4 1 2 3 4 5 6
,
7 = 410 30 1005 ) (AT A) 1 = 4 27
4
20
45
41 3 9 5
5
5
1
1 4 9 16
30 100 354
4
4
4
1 4 16
we get
x = (AT A)
1
2
6
AT b = 4
31
4
27
4
5
4
27
4
129
20
5
4
Thus the least squares approximation is y = 15
32
5
1
4
57 6
4 5 41
1
1
4
56
5 x
1
2
4
1
3
9
+ 2x2 .
2 3
3 6
2
3
1 6 7
15
7 607 6
7
.
4 5 6 7 = 4 56
5 5
405
16
2
2
17. Following the method of Example 7.28, we substitute the given values of x into the quadratic equation
a + bx + cx2 = y to get
which in matrix form is
2
1
61
6
61
6
41
1
a
2b + 4c =
4
a
b+ c=
7
a
=
3
a+ b+ c=
0
a + 2b + 4c =
1,
3
2
2 4 2 3
6
1 17
7 a
6
4b5 = 6
0 07
7
6
4
1 15 c
2 4
3
4
77
7
37
7.
05
1
Thus we want the least squares approximation to Ax = b, where A is the 5 ⇥ 3 matrix on the left and
b is the matrix on the right in the above equation. Since
2
3
1
2 4
2
36
3
2
3
7 2
17
1
1
1 1 1 1 61
1 17
5 0 10
0
35
7
6
7
6
7
6
7
6
7
T
1
1
AT A = 4 2
1 0 1 25 6
0 07
05 ,
61
7 = 4 0 10 0 5 ) (A A) = 4 0 10
6
7
1
1
4
1 0 1 4 41
1 15
10 0 34
0 14
7
1
2 4
576
CHAPTER 7. DISTANCE AND APPROXIMATION
we get
x = (AT A)
1
2
17
35
6
AT b = 4 0
1
7
0
1
76
05 4 2
1
4
14
1
10
1
7
32
0
Thus the least squares approximation is y =
18
5
17
10 x
2
36
1 1 1 1 6
76
1 0 1 25 6
6
6
1 0 1 4 4
1 2
2x .
3
4
7 2
77
7 6
37
7=4
7
05
1
3
18
5
17 7
.
10 5
1
2
18. Following the method of Example 7.28, we substitute the given values of x into the quadratic equation
a + bx + cx2 = y to get
which in matrix form is
a
2b + 4c =
0
a
b+ c=
11
a
=
10
a+ b+ c=
9
a + 2b + 4c =
8,
2
1
61
6
61
6
41
1
3
2
3
4 2 3
0
6 117
17
7 a
6
7
7
7
4
5
07 b = 6
6 107 .
5
4
1
c
95
4
8
2
1
0
1
2
Thus we want the least squares approximation to Ax = b, where A is the 5 ⇥ 3 matrix on the left and
b is the matrix on the right in the above equation. Since
2
3
1
2 4
2
36
3
2
3
7 2
17
1
1
1 1 1 1 61
1 17
5 0 10
0
35
7
7 6
6
76
7
6
7
T
1
1
AT A = 4 2
1 0 1 25 6
0 07
05 ,
61
7 = 4 0 10 0 5 ) (A A) = 4 0 10
6
7
1
1
4
1 0 1 4 41
1 15
10 0 34
0 14
7
1
2 4
we get
x = (AT A)
1
2
17
35
6
AT b = 4 0
1
7
0
1
76
05 4 2
1
4
14
1
10
1
7
32
0
62
5
Thus the least squares approximation is y =
36
1 1 1 1 6
76
1 0 1 25 6
6
6
1 0 1 4 4
+ 95 x + 4x2 .
19. The normal equation is AT Ax = AT b. We have

3
A A=
1
T
AT b =

1
1
1
2
so that the normal form of this system is

11
6
2
2
3
1 4
1
2
1
3

1
11 6
15 =
6 6
2
2 3

1
3 2 4 5
5
1 =
,
2 1
4
1

6
5
x=
.
6
4
3
0
7 2
117
7 6
107
7=4
7
95
8
3
62
5
97
.
55
4
7.3. LEAST SQUARES APPROXIMATION
Thus
"
#
11 6
x=
6 6
577
1
" # "
5
=
4
1
5
1
5
1
5
11
30
#" # " #
1
5
= 57
4
15
is the least squares solution.
20. The normal equation is AT Ax = AT b. We have
AT A =

1
2
3 2
2 1
AT b =

1
2
3 2
2 1
so that the normal form of this system is

Thus
x=
"
#
6
9
14
6
2
3

1
2
14
43
25 =
6
2
1
2 3

1
6
415 =
,
3
1
14
6
6
x=
9
"
# "
1
6
= 10
1
3
15
1

6
9
6
.
3
1
15
7
45
#"
# "
6
=
3
2
15
1
15
#
is the least squares solution.
21. The normal equation is AT Ax = AT b. We have
AT A =

1
2
AT b =

1
2
2
so that the normal form of this system is

Thus
"
14 8
x=
8 38
#
1
"
3
1
2

0 2 3 6
37
60
7 = 14 8
3 5 0 42
55
8 38
3
0
2 3
4

7
0 2 3 6
12
6 17 =
,
3 5 0 4 25
21
4
14
8
8
x=
38
# "
12
=
21

19
234
2
117
12
.
21
2
117
7
234
#"
# "
12
=
21
is the least squares solution.
22. The normal equation is AT Ax = AT b. We have
AT A =

AT b =

1
0
1
0
2
1
2
1
2
1 0 6
6
1 2 4
2
1 0 6
6
1 2 4
3
1
0

2
17
6
7=
1
15
3
0
2
3
1

57
7 = 12 ,
15
2
2
3
6
4
3
5
6
#
578
CHAPTER 7. DISTANCE AND APPROXIMATION
so that the normal form of this system is

Thus
x=
"
#
3
6
6
3
6
3
1
3
x=
6
"

# "
2
12
= 91
2
9
12
.
2
1
9
2
9
#"
# " #
22
12
= 98
2
9
is the least squares solution.
⇥
⇤
23. Instead of trying to invert AT A, we instead use an alternative method, row-reducing AT A AT b .
We will see that AT A does not reduce to the identity matrix, so that the system does not have a unique
solution. We get
2
32
3 2
3
1 1
0
1 1
1 0 0
3
0
2
1
61 0
6
6
1
17
0 1 17
3
2
17
7 61
7 = 60
7
AT A = 6
40 1
1
15 40
1 1 15 42
2
3
25
0 1
1
0 1
1 1 0
1
1
2
2
2
32 3 2 3
1 1
0
1
1
2
6
7
6
7
6
7
1
0
1
1
3
7 6 7 = 6 57 .
AT b = 6
40 1
1
15 4 25 4 35
0 1
1
0
4
1
Row-reducing, we have
2
3
60
6
42
1
0
3
2
1
2
2
3
2
1
1
2
2
3
2
2
1
60
57
7!6
40
35
1
0
So AT A is not invertible, and therefore there is not
2
3 2
4+t
6 5 t7 6
6
7 6
4 5 2t5 = 4
t
0
1
0
0
a unique
3
2
4
6
57
7 + t6
5
4
5
0
0
0
1
0
1
1
2
0
3
4
57
7.
55
0
solution. The general solution is
3
1
17
7.
25
1
⇥
⇤
24. Instead of trying to invert AT A, we instead use an alternative method, row-reducing AT A AT b .
We will see that AT A does not reduce to the identity matrix, so that the system does not have a unique
solution. We get
2
32
3 2
3
0
1 1 1 0
1 1
0
3 0 3 0
61
6
6
7
1 0 17
1 1
17
7 61
7 = 6 0 3 1 27
AT A = 6
41
5
4
5
4
1 1 1 1
0 1
0
3 1 4 05
0
1 0 1
1
1 1
1
0 2 0 2
2
32 3 2 3
0
1 1 1
5
3
61
7 6 37 6 37
1
0
1
T
76 7 = 6 7.
A b=6
41
1 1 15 4 15 4 85
0
1 0 1
1
2
Row-reducing, we have
2
3
60
6
43
0
0
3
1
2
3
1
4
0
0
2
0
2
3
2
3
1 0 0
60 1 0
37
7!6
40 0 1
85
2
0 0 0
1
1
1
0
3
5
17
7.
65
0
7.3. LEAST SQUARES APPROXIMATION
579
So AT A is not invertible, and therefore there is not a unique
2
3 2 3
2
5 t
5
6 1 t7 6 17
6
6
7 6 7
6
4 6 + t 5 = 4 65 + t 4
t
0
solution. The general solution is
3
1
17
7.
15
1
25. Solving these equations corresponds to solving Ax = b where
2
3
2 3
1
1
1
2
6 0
7
667
1
2
7
7
A=6
b=6
4 3
4115 .
2
15
1
0
1
0
⇥ T
⇤
To solve, we row-reduce A A AT b :
2
3
2
3
2
1
1
1
1
0
3
1 6
11
7
0
1
27 4
1
2
05 6
7
AT A = 4 1
=
4 3
2
15
1
2
1
1
5
1
0
1
2 3
2
3 2
2 3
1
0
3
1 6 7
35
67 4 5
1
2
05 6
18 .
AT b = 4 1
=
4115
1
2
1
1
1
0
Row-reducing, we have
2
11
6
4 7
5
7
6
5
5
5
7
Thus the best approximation is
3
2
35
1
7
6
185 ! 40
1
0
0
1
0
2 3 2 3
42
x
11
6 7 6 19 7
x = 4y 5 = 4 11 5 .
42
z
11
0
0
1
3
42
11
19 7
.
11 5
42
11
26. Solving these equations corresponds to solving Ax = b where
2
3
2 3
2 3
1
21
6 1 1
7
677
1
7
7
A=6
b=6
4 1 1
4145 .
15
0 2
1
0
⇥ T
⇤
To solve, we row-reduce A A AT b :
2
3
2
3
2
2 3
1
2 1
1 0 6
6
7
1 1
17 4
1 25 6
6
AT A = 4 3 1
=
4 1 1
15
1 1
1 1
4
0 2
1
2 3
2
3 21
2 3
2 1
1 0 6 7
35
77 4 5
1 25 6
84 .
AT b = 4 3 1
=
4145
1 1
1 1
14
0
Row-reducing, we have
2
6
6
46
4
6
15
5
4
5
4
3
2
35
1
7
6
845 ! 40
14
0
0
1
0
0
0
1
7
6
5
3
469
66
224 7
.
33 5
133
11
6
15
5
3
4
55
4
3
5
55
7
580
CHAPTER 7. DISTANCE AND APPROXIMATION
Thus the best approximation is
2 3 2
x
6 7 6
x = 4y 5 = 4
z
27. With Ax = QRx = b, we get
Rx = QT b,
"
or
3
0
3
469
66
224 7
.
33 5
133
11
#
"
2
1
x = 31
1
3
Back-substitution gives the solution
"
x=
5
3
2
2
3
2
3
#
1
3
2
3
#
2
3
" #
2
3
6 7
.
4 35 =
2
1
.
28. With Ax = QRx = b, we get
Rx = QT b,
"p
or
6
0
p
6
2
p1
2
#
x=
Back-substitution gives the solution
x=
"
" #
4
3
2
p1
6
p1
2
p2
6
0
2 3
# 1
" #
p1
p2
6 7
6 6 7=
6 .
p
1
4 5
p1
2
2
1
.
29. We are looking for a line y = a+bx that best fits the given points. Using the least squares approximation
theorem, we want a solution to Ax = b, where
2
3
2
3
1 20
14.5
2
3
2 3
1 x1
y1
61 40 7
6 31 7
7
7

61 x2 7 6
6 y2 7 6
6
7
a
36 7
6
7 61 48 7
6 7 6
6
7.
A = 6.
=
,
x
=
,
b
=
=
6 .. 7 6
.. 7 61 60 7
b
4 ..
4 . 5 645.57
. 5 6
7
7
41 80 5
4 59 5
1 xn
yn
1 100
73.5
Since
2
"
1
A A=
20
T
1
40
1
48
1
60
1
80
we get
T
x = (A A)
1
T
A b=
"
1
6
6
# 61
1 6
61
100 6
61
6
41
1
1519
1545
29
2060
29
2060
1
4120
3
20
7
40 7 "
7
48 7
7= 6
60 7
348
7
7
80 5
100
#"
1
20
#
348
) (AT A)
24304
2
1
40
1
48
Thus the least squares approximation is y = 0.918 + 0.730x.
1
60
1
80
1
=
"
1519
1545
29
2060
29
2060
1
4120
3
14.5
6
7
31 7 "
#6
#
6
7
7
36
0.918
1 6
6
7=
.
7
100 6
0.730
645.57
6
7
4 59 5
73.5
#
,
7.3. LEAST SQUARES APPROXIMATION
581
30. (a) We are looking for a line L = a + bF that best fits the given points. Using the least squares
approximation theorem, we want a solution to Ax = b, where
2
3 2
2 3 2
3
3
1 x1
y1
1 2
7.4

61 x2 7 6
6 y2 7 6
6
7 61 47
7 6 9.6 7
7, x = a , b = 6
7.
A = 6.
=4
7
6
7=
.
.
5
.. 5
1 6
b
4 ..
4 .. 5 411.55
1 8
13.6
1 xn
yn
Since
"
1
A A=
2
T
1
4
2
# 1
1 6
61
6
8 41
1
1
6
we get
x = (AT A)
1
AT b =
3
2
47
7
7
65
"
4
=
20
#
20
) (AT A)
120
1
=
"
3
2
1
4
1
4
1
20
#
,
8
"
3
2
1
4
#"
1
1
4
1
2
20
1
4
3
7.4
"
#
7
1 6
5.4
6 9.6 7
.
6
7=
8 411.55
1.025
13.6
#
1
6
2
Thus the least squares approximation is L = 5.4 + 1.025F . a = 5.4 represents the length of the
spring when no force is applied (F = 0).
(b) When a weight of 5 ounces is attached, the spring length will be about 5.4 + 1.025 · 5 = 10.525
inches.
31. (a) We are looking for a line y = a + bx that best fits the given points. Letting x = 0 correspond to
1920 and using the least squares approximation theorem, we want a solution to Ax = b, where
2
3
2
3
1 0
54.1
7
7
2
3 6
2 3 6
61 107
659.77
1 x1
y1
61 207
662.97
7
7

61 x2 7 6
6 y2 7 6
1 307
a
68.27
6
7 6
6 7 6
6
7
6
A = 6.
.. 7 = 61 407 , x = b , b = 6 .. 7 = 669.77
7.
4 ..
4.5 6
. 5 6
7
7
61 507
670.87
1 xn
yn
6
7
6
7
41 605
473.75
1 70
75.4
Since
"
8
A A=
280
T
#
280
) (AT A)
14000
1
=
"
5
12
1
120
1
120
1
4200
#
,
we get
2
T
x = (A A)
1
T
A b=
"
5
12
1
120
1
120
1
4200
#"
1
0
1
10
1
20
1
30
1
40
1
50
1
60
3
54.1
6
7
659.77
6
7
6
7
# 662.97 "
#
6
1 668.27
56.63
7
.
6
7=
7
70 6
69.7
0.291
6
7
670.87
6
7
6
7
473.75
75.4
Thus the least squares approximation is y = 56.63 + 0.291x. The model predicts that the life
expectancy of someone born in 2000 is y(80) = 56.63 + 0.291 · 80 = 79.91 years.
582
CHAPTER 7. DISTANCE AND APPROXIMATION
(b) To see how good the model is, compute the least squares error:
2
3 2
3
2
3
54.1
1 0
2.53
659.77 61 107
6 0.16 7
6
7 6
7
6
7
662.97 61 207
6 0.45 7
6
7
6
7
7 6
668.27 61 307 56.63
6 2.84 7
6
7
6
7
6
7 , so kek ⇡ 4.43.
e = b Ax = 6
=6
7 6
7
7
669.77 61 407 0.291
6 1.43 7
670.87 61 507
6 0.387
6
7 6
7
6
7
473.75 41 605
4 0.395
75.4
1 70
1.6
The model is not bad, with an error of under 10% of the initial value.
32. (a) Following the method of Example 7.28, we substitute the given values of x into the equation
a + bx + cx2 = y to get
a + 0.5b + 0.25c = 11
a+
b+
c = 17
a + 1.5b + 2.25c = 21
which in matrix form is
2
1
61
6
61
6
41
1
a+
2b +
4c = 23
a+
3b +
9c = 18,
0.5
1
1.5
2
3
3
2 3
0.25 2 3
11
6177
1 7
a
7
6 7
4 5 6 7
2.257
7 b = 6217 .
4235
4 5 c
9
18
Thus we want the least squares approximation to Ax = b, where A is the 5 ⇥ 3 matrix on the left
and b is the matrix on the right in the above equation. Since
2
1
AT A = 4 0.5
0.25
1
1
1 1.5
1 2.25
2
3 1
1 1 6
61
2 35 6
61
4 9 41
1
0.5
1
1.5
2
3
3
0.25
2
1 7
5
7
4 8
2.257
=
7
4 5
16.5
9
8
16.5
39.5
(AT A)
1
we get
x = (AT A)
1
2
3.330
AT b = 4 4.082
1.031
4.082
5.735
1.543
32
1.031
1
1.5435 4 0.5
0.436
0.25
2
3
16.5
39.5 5 )
103.125
3.330
= 4 4.082
1.031
4.082
5.735
1.543
3
1.031
1.5435 ,
0.436
2 3
3 11
2
3
7
1
1
1 1 6
1.918
6177
7 4
5
1 1.5 2 35 6
6217 = 20.306 .
1 2.25 4 9 4235
4.972
18
Thus the least squares approximation is y = 1.918 + 20.306t
4.972t2 .
(b) The height at which the object was released is y(0) ⇡ 1.918 m. Its initial velocity was y 0 (0) ⇡
2
20.306 m/sec, and its acceleration due to gravity is y 00 (t) ⇡ 9.944 m/s .
(c) The object hits the ground when y(t) = 0. Solving the quadratic using the quadratic formula
gives y ⇡ 0.092 and y ⇡ 4.176 seconds. So the object hits the ground after about 4.176 seconds.
7.3. LEAST SQUARES APPROXIMATION
583
33. (a) If p(t) = cekt , then p(0) = 150 means that c = 150, so that p(t) = 150ekt . Taking logs gives
ln p(t) = ln 150 + kt ⇡ 5.011 + kt. We wish to find the value of k which best fits the data. Let
t = 0 correspond to 1950 and t = 1 to 1960. Then kt = ln p(t) ln 150 = ln p(t)
150 , so
Since
2 3
1
6 7
6 27
6 7
7
A=6
6 37 ,
6 7
4 45
5
2
3 2
3
ln 179
0.177
150
6 203 7 6
7
6ln 150 7 60.3037
6 227 7 6
7
7 6
7
b=6
6ln 150 7 ⇡ 60.4147 .
6 250 7 6
7
4ln 150 5 40.5115
ln 281
0.628
150
h i
x= k ,
⇥ ⇤
AT A = 55 and thus (AT A)
we get
x = (AT A)
1
AT b =
⇥
1
55
⇤⇥
1
2
3
4
Thus the least squares approximation is p(t) = 150e0.131t .
1
=
2
⇥
1
55
⇤
,
3
0.177
6
7
⇤ 60.3037
6
5 60.4147
7 = 0.131.
40.5115
0.628
(b) 2010 corresponds to t = 6, so the predicted population is p(6) = 150e0.131·6 ⇡ 329.19 million.
34. Let t = 0 correspond to 1970.
(a) For a quadratic approximation, follow the method of Example 7.28, and substitute the given
values of x into the equation a + bx + cx2 = y to get
a
a + 5b +
=
29.3
25c =
44.7
a + 10b + 100c = 143.8
a + 15b + 225c = 371.6
a + 20b + 400c = 597.5
a + 25b + 625c = 1110.8
a + 30b + 900c = 1895.6
a + 35b + 1225c = 2476.6,
which in matrix form is
2
1
61
6
61
6
61
6
61
6
61
6
41
1
0
5
10
15
20
25
30
35
3
2
3
0
29.3
6 44.7 7
25 7
7
6
7
7
100 7 2 3 6
143.8 7
6
7
a
6
7
225 7
7 4 b 5 = 6 371.6 7 .
7
6
7
400 7
6 597.5 7
c
7
6
7
625 7
61110.87
41895.65
900 5
1225
2476.6
Thus we want the least squares approximation to Ax = b, where A is the 8 ⇥ 3 matrix on the left
and b is the matrix on the right in the above equation. Since
2
3
2
3
17
3
1
8
140
3500
24
40
600
6
7
6 3
53
1 7
AT A = 4 140
,
3500
98000 5 ) (AT A) 1 = 4 40
4200
3000 5
1
1
1
3500 98000 2922500
600
3000
105000
584
CHAPTER 7. DISTANCE AND APPROXIMATION
we get
2
6
x = (AT A) 1 AT b = 4
2
17
24
3
40
1
600
1
6
61
6
3 61
1
6
61
600
7
6
1
5
3000 6
61
1
6
105000 6
61
6
41
1
3
40
53
4200
1
3000
Thus the least squares approximation is y = 57.063
0
5
10
15
20
25
30
35
3T
0
7
25 7
7
100 7
7
225 7
7
7
400 7
7
625 7
7
7
900 5
1225
2
3
29.3
6
7
6 44.7 7
6
7
6 143.8 7 2
3
6
7
57.063
6 371.6 7
6
7 6
7
6
7 = 4 20.3355 .
6 597.5 7
6
7
2.589
61110.87
6
7
6
7
41895.65
2476.6
20.335t + 2.589t2 .
(b) With y(t) = cekt , since y(0) = 29.3 we have c = 29.3 so that p(t) = 29.3ekt . Taking logs gives
p(t)
ln p(t) = ln 29.3 + kt, or kt = ln 29.3
. Then
2
3
0
6 7
657
6 7
6107
6 7
6157
6 7
A = 6 7,
6207
6 7
6257
6 7
6 7
4305
35
2
3 2
3
ln 44.7
0.422
29.3
6 ln 143.8 7 61.5917
6
6
7
29.3 7
6 371.6
7 6
7
6 ln 29.3 7 62.5407
6 597.5 7 6
7
7 6
7
b=6
6 ln 29.3 7 ⇡ 63.0157 .
6 1110.8 7 6
7
6ln 29.3 7 63.6357
6 1895.6 7 6
7
4ln 29.3 5 44.1705
ln 2476.6
4.437
29.3
⇥
⇤
⇥
⇤
⇥
⇤
⇥
⇤
Then AT A = 3500 and AT b = 487.696 , so that x = 487.696
⇡ 0.1393 . Thus y(t) =
3500
29.3e0.1393t .
(c) The models produce the following predictions:
Year
1970
1975
1980
1985
1990
1995
2000
2005
Actual
29.3
44.7
143.8
371.6
597.5
1110.8
1895.6
2476.6
Quadratic
57.1
20.1
112.6
334.6
686.0
1166.8
1777.1
2516.9
Exponential
29.3
58.8
118.0
236.8
475.1
953.5
1913.3
3839.4
The quadratic appears to be a better fit overall, even though it does a poor job in the first 15
years or so. The exponential model simply grows too fast.
(d) Using the quadratic approximation, we get for 2010 and 2015
y(40) = 57.063
y(45) = 57.063
20.335 · 40 + 2.589 · 402 = 3386.1
20.335 · 45 + 2.589 · 452 = 4384.7.
35. Using Example 7.29, taking logs gives
2 3 2 3
t1
30
6 7 6 7
A = 4t2 5 = 4605 ,
t3
90
2
3 2
3
kt1
ln 172
200
6 7 6
7
b = 4kt2 5 = 4ln 148
.
200 5
128
kt2
ln 200
7.3. LEAST SQUARES APPROXIMATION
585
⇥
⇤
⇥
⇤
⇥
62.8 , so that x ⇡
Then AT A = 12600 and AT b ⇡
p(t) = ce 0.005t . To compute the half-life, we want to solve
ln 2
ln 2; then t = 0.005
⇡ 139 days.
62.8
12600
1
2c =
⇤
⇡ 0.005. Thus k = 0.005 and
ce 0.005t , so that 0.005t = ln 12 =
36. We want to find the best fit z = a + bx + cy to the given data; the data give the equations
a
4c =
0
=
0
a + 4b
c=
1
a+ b
3c =
1
a
5c =
2,
a + 5b
which in matrix form is
2
1
61
6
61
6
41
1
b
3
2
4 2 3
6
07
7 a
6
7
4
5
17 b = 6
6
4
35 c
5
0
5
4
1
1
3
0
07
7
17
7.
15
2
Thus we want the least squares approximation to Ax = b, where A is the 5 ⇥ 3 matrix on the left and
b is the matrix on the right in the above equation. Since
2
3
2
3
2189
433
541
5
9
13
15
15
15
6
7
6
86
107 7
AT A = 4 9 43
,
25 ) (AT A) 1 = 4 433
15
15
15 5
541
107
134
13
2
51
15
15
15
we get
x = (AT A)
1
2
6
AT b = 4
2189
15
433
15
541
15
433
15
86
15
107
15
32
541
15
107 7 6
15 5 4
134
15
Thus the least squares approximation is y = 13 (43
37. We have
A=

1
1
)
⇥
AT = 1
⇤
1
1
0
4
1
5
0
1
4
1
⇥ ⇤
AT A = 2 , (AT A)
So the standard matrix of the orthogonal projection onto W is
" #
"
i
1
1 h1i h
T
1 T
2
P = A(A A) A =
=
1
1
1
1 2
2
projW (v) = P v =
38. We have
A=

1
2
)
⇥
AT = 1
2
⇤
36
1 6
76
15 6
6
6
5 4
1
1
3
8x + 11y).
)
Then
2
"
1
2
1
2
)
1
2
1
2
1
2
1
2
#
1
=
3
0
7 2
07
7 6
17
7=4
7
15
2
⇥1⇤
2
.
.
#" # " #
7
3
= 27 .
4
2
⇥ ⇤
AT A = 5 , (AT A)
So the standard matrix of the orthogonal projection onto W is
" #
"
h ih
i
1
1
P = A(AT A) 1 AT =
1
2 =
2 5
1
5
2
5
2
5
4
5
1
#
.
=
⇥1⇤
5
.
3
43
3
87
.
35
11
3
586
CHAPTER 7. DISTANCE AND APPROXIMATION
Then
projW (v) = P v =
39. We have
2 3
1
A = 415
1
)
⇥
AT = 1
"
⇤
1
1
2
5
4
5
1
5
2
5
#" # "
1
=
1
1
5
2
5
#
.
⇥ ⇤
AT A = 3 , (AT A)
)
1
=
⇥1⇤
3
.
So the standard matrix of the orthogonal projection onto W is
2 3
1 h ih
6 7 1
T
1 T
P = A(A A) A = 415 3 1
1
Then
projW (v) = P v =
40. We have
2
3
2
A = 4 25
1
)
⇥
AT = 2
2
1
3
61
43
1
3
1
1
1
3
17 6 7
3 5 425
1
3
3
3
1
3
1
3
1
3
1
3
32 3
1
3
1
3
1
3
⇤
1
2
2
1
i
3
61
1 = 43
1
3
17
.
35
1
3
2 3
2
6 7
= 425 .
2
⇥ ⇤
AT A = 9 , (AT A)
)
1
⇥1⇤
=
9
.
So the standard matrix of the orthogonal projection onto W is
2
3
2 h ih
6 7
P = A(AT A) 1 AT = 4 25 19 2
1
Then
2
4
9
4
9
2
9
6
projW (v) = P v = 4
2
i
6
1 =4
2
32 3
4
9
4
9
2
9
2
1
9
27 6 7
05
5
4
9
1
0
9
4
9
4
9
2
9
3
4
9
4
9
2
9
2
2
9
27
.
95
1
9
3
4
9
47
.
95
2
9
6
=4
41. We have
2
1
A=4 0
1
3
1
15
1
)
AT =

1
1
0
1
1
1
)
AT A =

2
0
0
, (AT A)
3
1
=
1
So the standard matrix of the orthogonal projection onto W is
2
1
6
P = A(AT A) 1 AT = 4 0
1
Then
3
1 "1
7
15 2
0
1
2
6
projW (v) = P v = 4
5
6
1
3
1
6
#"
0 1
1
1
3
1
3
1
3
1
3
0
1
2
#
1
6
=4
1
32 3
1
1
6
17 6 7
3 5 405
5
0
6
2
6
=4
5
6
1
3
1
6
3
5
6
17
.
35
1
6
1
3
1
3
1
3
3
1
6
17
.
35
5
6
2
0
0
1
3
.
7.3. LEAST SQUARES APPROXIMATION
42. We have
2
1
A=4 2
1
3
1
05
1
587

1
A =
1
T
)
2
0
So the standard matrix of the orthogonal
2
1
6
T
1 T
P = A(A A) A = 4 2
1
Then
1
1
2
#
0
1
1
3
6
A A=
0
)
projection onto W is
3
#"
1 "1
2
7 6 0 1
05
1
0 2 1
0
1
2
3
1
3
1
3
6
projW (v) = P v = 4
43. With the new basis, we have
2
3
"
1 1
1
6
7
T
A = 41 35 ) A =
1
0 1

T
1
1
3
17 6 7
3 5 425
2
3
3
"
2
A A=
4
T
)
2
#
1
6
=4
1
32 3
1
3
2
3
1
3
0
, (AT A)
2
2
3
1
3
1
3
1
5
6
1
6
1
3
.
1
2
3
3
1
6 7
= 4 05 .
1
So the standard matrix of the orthogonal projection onto W is
2
3
2
#"
#
1 1 " 11
2
1 1 0
6
7
6
3
P = A(AT A) 1 AT = 41 35 62
=4
1
1
3
1
3
3
0 1
0
0
1
3
17
.
35
2
3
2
1
6
=
1
3
2
3
1
3
#
4
, (AT A)
11
1
=
"
11
6
2
3
2
3
1
3
#
.
3
1
6
5
6
1
3
1
3
17
.
35
1
3
This agrees with the matrix found in Example 7.31.
44. (a) To see that P is symmetric, compute its transpose:
⇣
P T = A AT A
1
AT
⌘T
1
= AT
T
⇣
⌘
1 T
AT A
AT = A
⇣
AT A
T
⌘
1
AT
= A AT A
1

1
2
)
So the pseudoinverse of A is
⇥
AT = 1
A+ = (AT A)
46. We have
2
3
1
A = 4 15
2
1
AT = P.
AT
= A AT A
A=
AT = P,
T
recalling that B T
= B 1 for any invertible matrix B.
(b) To show that P is idempotent, we must show that P 2 = P P = P :
⇣
⌘⇣
⌘
⇣
⌘
1 T
1 T
1
P P = A AT A
A
A AT A
A
= A AT A
AT A
AT A
45. We have
1
)
So the pseudoinverse of A is
⇥
AT = 1
A+ = (AT A)
1
2
1
⇤
AT =
⇥1⇤ ⇥
⇤
2
1
AT =
⇥ ⇤
AT A = 5 , (AT A)
)
⇥1⇤ ⇥
6
1
5
1
)
1
⇤ ⇥
2 = 15
2
5
⇤
1
=
5
.
.
⇥ ⇤
AT A = 6 , (AT A)
⇤ ⇥
2 = 16
⇥1⇤
1
6
1
3
⇤
.
1
=
⇥1⇤
6
.
588
CHAPTER 7. DISTANCE AND APPROXIMATION
47. We have
2
3
3
7
15
2
1
6
A=4 1
0
"
#
1 0
1 2
1
A =
3
T
)
"
2
A A=
2
T
)
#
2
, (AT A)
14
1
=
"
#
.
7
12
1
12
1
12
1
12
#
.
So the pseudoinverse of A is
+
T
A = (A A)
48. We have
2
3
3
7
15
2
1
6
A = 43
2
1
T
A =
"
1
A =
3
T
)
"
7
12
1
12
1
12
1
12
#
2
2
3
1
#"
# "
1
1 0
= 31
1 2
6
1
3
"
14
A A=
10
T
)
2
3
1
6
1
6
1
6
#
10
, (AT A)
14
1
=
"
7
48
5
48
5
48
7
48
#
.
So the pseudoinverse of A is
A+ = (AT A)
49. We have
A=

1
0
1
1
1
AT =

AT =
)
"
1
1
7
48
5
48
0
1
5
48
7
48
)
#"
1
3
# "
2
=
2
3
1
AT A =


1
1
1
1
1
6
1
3
1
3
1
6
1
, (AT A)
2
#
1
12
1
12
1
=

.
2
1
1
.
1
So the pseudoinverse of A is
+
T
1
A = (A A)
Note that this is the inverse A
50. We have
"
1
A=
3
#
2
4
)
1
T
A =

2
1
1
1

0
1
=
1
0
1
.
1
of A.
"
#
3
4
1
AT =
2
"
#
14
, (AT A)
20
10
AT A =
14
)
1
=
"
#
7
2
7
2
5
2
2
3
1
3
2
3
1
3
5
3
4
3
5
.
So the pseudoinverse of A is
A+ = (AT A)
Note that this is the inverse A
51. We have
2
1
61
6
A=6
40
1
0
0
1
1
3
2
0
1
7
17
6
T
7 ) A = 40
15
0
1
1
1
AT =
"
5
7
2
7
2
5
2
#"
1
2
# "
3
2
=
3
4
2
1
1
2
#
.
of A.
1
0
1
0
1
1
3
2
1
3
7
6
T
15 ) A A = 41
1
2
1
2
2
3
2
7
25 , (AT A)
3
1
2
6
=4
So the pseudoinverse of A is
A+ = (AT A)
1
2
6
AT = 4
2
3
1
3
2
3
1
3
5
3
4
3
32
2
1
3
47 6
3 5 40
5
0
3
1
0
1
0
1
1
3 2
1
7 6
15 = 4
1
2
3
1
3
2
3
0
1
1
1
3
1
3
1
3
3
1
3
27
.
35
1
3
3
2
3
47
.
35
5
3
7.3. LEAST SQUARES APPROXIMATION
52. We have
2
1
60
A=6
41
0
so that
3
2
0
1
17
7 ) AT = 4 2
25
0
2
2
1
1
0
2
2
6
T
A A=4 3
2
589
3
2
7
35 , (AT A)
9
3
6
3
So the pseudoinverse of A is
A+ = (AT A)
1
2
15
7
1
7
1
6
AT = 4 1
1
76
05 42
1
0
7
2
3
1
7
32
0
0
1
1
3
1 0
1 05 ,
2 2
0
1
1
1
2
15
7
6
=4 1
1
7
1
7
1
3
7
05 .
2
3
1
7
0
3 2
1
1 0
7
7 61
1 05 = 4 3
1
2 2
7
8
7
2
3
1
7
6
7
1
3
1
7
2
7
3
7
05 .
2
7
53. (a) Since A is square with linearly independent columns, it is invertible, so that
1
A+ = AT A
AT = A
1
1
AT
AT = A
1
.
(b) Let the columns of A be qi ; then qi · qj = 0 if i 6= j and 1 if i = j. Then
2 T3
q1
⇤
6 .. 7 ⇥
T
A A = 4 . 5 q 1 · · · q n = In ,
qTn
1
so that A+ = AT A
AT = In AT = AT . (Note that this makes sense, for if A were square, it
would be an orthogonal matrix and then part (a) would hold and AT = A 1 .)
⇣
⌘
1 T
1
54. A+ AA+ = AT A
A AA+ = AT A
AT A A+ = A+ .
1
55. We have A+ A = AT A
AT A = AT A
1
AT A = I, which is certainly a symmetric matrix.
1
1
1
56. (a) (cA)+ = cAT cA
cAT = c12 AT A
cAT = 1c AT A
AT = 1c AT .
(b) By Exercise 53(a), if A is square, then A+ = A 1 . Also A 1 is square with linearly independent
+
1
+
+
columns, so that A 1 = A 1
= A. Then(A+ ) = A 1 = A.
(c) Since A is square, Exercise 53(a) applies to both A and AT , so that
AT
+
= AT
1
= A
1 T
= A+
T
.
57. The fact that not all the points lie on the same vertical line means that there is at least one pair i, j
such that xi 6= xj . Constructing the matrix A gives
2
3
1 x1
6 ..
.. 7
6.
. 7
6
7
61 xi 7
6
7
6
.. 7 .
A = 6 ...
7
.
6
7
61 xj 7
6
7
6.
.. 7
4 ..
. 5
1 xn
Since xi 6= xj , the columns are not multiples of one another, so must be linearly independent. Since
the columns of A are linearly independent, the points have a unique least squares approximating line,
by part (b) of the Least Squares Theorem.
590
CHAPTER 7. DISTANCE AND APPROXIMATION
58. Constructing the matrix A gives
2
1
6
A = 4 ...
1
x1
..
.
x21
..
.
xn
x2n
···
..
.
···
3
xk1
.. 7 .
. 5
xkn
The solution(s) to Ax = b are the least squares approximating polynomials of degree k; by the Least
Squares Theorem, there is a unique solution when the columns of A are linearly independent. Now
suppose that
a0 a1 + a1 a2 + · · · + ak ak+1 = 0.
This means that a0 + a1 xi + · · · + ak xki = 0 for all i. But since at least k + 1 of the xi are distinct, the
polynomial a0 + a1 x + · · · + ak xk has at least k + 1 distinct roots, so it must be the zero polynomial by
Exercise 62 in Section 6.2. Thus all of the ai are zero, so the columns of A are linearly independent.
7.4
The Singular Value Decomposition
1. We have

2
A A=
0
T

0
3

0
4
=
3
0
2
0
0
.
9
Then the singular values of A are the square roots of the eigenvalues ofpthis matrix which,
p since the
matrix is diagonal, are 4 and 9. Thus the singular values of A are 1 = 9 = 3 and 2 = 4 = 2.
2. We have

2
A A=
1
T

1
2

1
5
=
2
4
2
1
4
.
5
Then the singular values of A are the square roots of the eigenvalues of this matrix.
det(AT A
I) =
5
4
4
)2
= (5
5
2
16 =
10 + 9 = (
so that the eigenvalues are 9 and 1. Thus the singular values of A are
3. We have
AT A =

1
1

0
0

1
1
=
0
1
1
0
1
=
p
9)(
9 = 3 and
1),
2
=
p
1
.
1
Then the singular values of A are the square roots of the eigenvalues of this matrix.
det(AT A
I) =
1
so that the singular values of A are
4. We have
AT A =
1
1
1
1
=
p
2
1
p
2 and
p0
2
)2
= (1
2
=
p
2
0
p
1=
2
2 = (
2),
0 = 0.

p1 = p2
2
2
p
2
.
3
Then the singular values of A are the square roots of the eigenvalues of this matrix.
p
2p
2
T
det(A A
I) =
= (2
)(3
) 2= 2 5 +4=(
4)(
2 3
so that the singular values of A are
1
=
p
4 = 2 and
2
=
p
1 = 1.
1),
1 = 1.
7.4. THE SINGULAR VALUE DECOMPOSITION
5. We have
591
⇥
AT A = 3
4

⇤ 3
⇥ ⇤
= 25 .
4
Then the singular values of A are the square roots of the eigenvalues of this matrix.
det(AT A
so that the singular value of A is
p
=
1
I) = 25
= 25
,
25 = 5.
6. We have
AT A =

3 ⇥
3
4

⇤
4 =
9 12
.
12 16
Then the singular values of A are the square roots of the eigenvalues of this matrix.
det(AT A
9
I) =
12
16
12
so that the singular values of A are
AT A =
p
=
1
7. We have
= (9

0
0
)(16
25 = 5 and
2
2
0
2 4
0
0
2
0
3
)
=
p
2
144 =
25 = (
25),
0 = 0.
3

0
4
35 =
0
0
0
.
9
Then the singular values of A are the square roots of the eigenvalues ofpthis matrix. Sincepit is diagonal,
its eigenvalues are 9 and 4, so that the singular values of A are 1 = 9 = 3 and 2 = 4 = 2.
8. We have

0
AT A =
1
1
0
2
0
2 4
1
2
2
3

1
5
5
0 =
4
2
4
.
5
Then the singular values of A are the square roots of the eigenvalues of this matrix.
det(AT A
I) =
5
4
4
= (5
5
)2
16 =
2
10 + 9 = (
so that the eigenvalues are 9 and 1. Thus the singular values of A are
9. We have
2
3
0 
2
25
0
0
2
AT A = 4 0
1
0
2
2
4
1
= 40
0
2
0
4
0
1
=
p
9)(
1),
9 = 3 and
2
=
p
1 = 1.
3
2
05
1
Then the singular values of A are the square roots of the eigenvalues of this matrix.
4
det(AT A
I) =
0
0
2
2
0
4
0
= (4
) ((4
)(1
)
4) =
so that
p the eigenvalues are 5, 4, and 0. Thus the singular values of A are
0 = 0.
3 =
10. We have
2
1
AT A = 4 0
1
(
5)(
4),
1
32
0 1 1
3 05 40
0 1 1
3 2
0 1
2
3 05 = 40
0 1
2
0
9
0
3
2
05
2
1
=
p
5,
2
=
p
4 = 2, and
592
CHAPTER 7. DISTANCE AND APPROXIMATION
Then the singular values of A are the square roots of the eigenvalues of this matrix.
2
det(AT A
I) =
0
0
2
2
0
9
0
= (9
)2
) (2
4 =
(
9)(
4)
2
so that the
peigenvalues are 9, 4, and 0. Thus the singular values of A are
and 3 = 0 = 0.
11. From Exercise 3,
AT A =
with eigenvalues 2 and 0, so that the singular
corresponding eigenspaces, row-reduce:

⇥ T
⇤
1
A A 2I 0 =
1

⇥ T
⇤
1 1
A A 0I 0 =
1 1

1
1
1
p
9 = 3,
=
p
=
2
=
p
4 = 2,
1
,
1
values of A are
1
=
p
2 and
2
0 = 0. To find the


0
1
1 0
1
!
) x1 =
0
0
0 0
1


0
1 1 0
1
!
) x2 =
.
0
0 0 0
1
1
1
These vectors are already orthogonal since they correspond to distinct eigenvalues. Normalizing each
of them gives
" #
"
#
"
#
h
i
p1
p1
p1
p1
2
2
2
2
v1 = 1 ,
v2 =
, so that V = v1 v2 = 1
.
1
1
p
p
2
p
2
2
p
2
The first column of the matrix U is
"
1 1
u1 =
Av1 = p
2 0
1
1
1
0
#"
p1
2
p1
2
#
=
" #
1
0
.
We do not use v2 since it corresponds to an eigenvalue of zero. We must extend this to an orthonormal
basis of R2 by adjoining a second vector, u2 , to get an orthogonal matrix. Clearly e2 is such a vector.
Thus
"
# "p
#"
#
p1
1 0
2 0 p12
T
2
A = U ⌃V =
p1
0 1
0 0 p12
2
is a singular value decomposition of A.
12. We have

1
A A=
1
T
1
1

1
1

1
2
=
1
2
= (2
)2
2
.
2
The characteristic polynomial of AT A is
det(AT A
I) =
2
2
2
2
4=
2
4 = (
4).
Thus the eigenvalues are 1 = 4 and 2 = 0 (and so the singular values are 1 = 2 and
find the corresponding eigenspaces, row-reduce:



⇥ T
⇤
2
2 0
1
1 0
1
A A 4I 0 =
!
) x1 =
2
2 0
0
0 0
1



⇥ T
⇤
2 2 0
1 1 0
1
A A 0I 0 =
!
) x2 =
.
2 2 0
0 0 0
1
2
= 0). To
7.4. THE SINGULAR VALUE DECOMPOSITION
593
To find the matrix V , we must normalize each of these vectors (they are already orthogonal since they
correspond to distinct eigenvalues). This gives
v1 =
"
p1
2
p1
2
#
,
"
v2 =
p1
2
p1
2
#
,
h
so that
"
i
V = v1
v2 =
p1
2
p1
2
p1
2
p1
2
#
.
The first column of the matrix U is
"
1 1
u1 =
Av1 =
2 1
1
#" #
1 p12
1
p1
2
1
=
"
p1
2
p1
2
#
.
We do not use v2 since it corresponds to an eigenvalue of zero. We must extend this to an orthonormal
basis of R2 by adjoining a second
" vector,
# u2 , to get an orthogonal matrix. By looking at the matrix
p1
2
p1
2
V , we see that such a vector is
A = U ⌃V
. Thus
=
"

0
2
T
p1
2
p1
2
p1
2
p1
2
#"
#"
0 p12
2
0
p1
2
p1
2
p1
2
0
#
is a singular value decomposition of A.
13. We have
T
A A=

3
0

2
9
=
0
0
0
3
0
.
4
This matrix is diagonal, so its eigenvalues are
p its diagonal entries,
p which are 1 = 9 and 2 = 4.
Therefore the singular values of A are 1 = 9 = 3 and 2 = 4 = 2. To find the corresponding
eigenspaces, row-reduce:
⇥
⇥
AT A
9I
AT A
4I

⇤
0
0 =
0

⇤
4
0 =
0


0
0 1 0
1
!
) x1 =
0
0 0 0
0


0
1 0 0
0
!
) x2 =
.
0
0 0 0
1
0
5
0
0
These vectors are already orthonormal, so we set
v1 = x1 ,
v2 = x2 ,
so that
⇥
V = v1

⇤
1
v2 =
0
Then U is found by
1
u1 =
Av1 =
3
1
1
Av2 =
1
2
A = U ⌃V T =

u2 =
1
2
Thus
is a singular value decomposition of A.
"
"
0
1
#" #
1
0
2
3
0 0
#" #
2 0
0
3
0
1
0

1
3
0
0
2
=
"
=
"

1
0
#
0
1
#
1
0
0
1
.
0
.
1
594
CHAPTER 7. DISTANCE AND APPROXIMATION
14. We have
AT A =

1
1
1
1

1
1

1
2
=
1
0
0
.
2
This matrix is diagonal, so its eigenvalues
are its diagonal entries. So the only eigenvalue is 2 and the
p
singular values of A are 1 = 2 = 2. To find the corresponding eigenspace, row-reduce:



⇥ T
⇤
0 0 0
1
0
A A 2I 0 =
) x1 =
, x2 =
.
0 0 0
0
1
These vectors are already orthonormal, so we set
v1 = x1 ,
v2 = x2 ,
so that
⇥
V = v1
v2 =
#" #
1 1
"
Then U is found by
"
1 1
u1 =
Av1 = p
2 1
1
"
1
1 1
u2 =
Av2 = p
2 1
2
1
Thus
A = U ⌃V
T
=
"
p1
2
p1
2
1 0
#" #
1 0
1
p1
2
p1
2
⇤
1
# "p
2
0
=
=
"
p1
2
p1
2
#
p1
2
p1
2
#"
0
1
p
2 0

#
1
0
0
.
1
.
#
0
1
is a singular value decomposition of A.
15. From Exercise 5,
⇥ ⇤
AT A = 25 ,
so that the only eigenvalue is 25 and therefore the only singular value is
corresponding eigenspace, row-reduce:
⇥
AT A
25I
⇤ ⇥
0 = 0
1
=
p
25 = 5. To find the
⇤
⇥ ⇤
0 ) x1 = 1 .
This single vector forms an orthonormal set; let v1 = x1 , so that
Then U is found by
⇥ ⇤ ⇥ ⇤
V = v1 = 1 .
" #
" #
3
1 3 h i
u1 =
Av1 =
1 = 54 .
5 4
1
5
1
We extend this to an orthonormal basis for R2 by adjoining the orthogonal unit vector
"
4
5
3
5
#
. Since
⌃ should be 2 ⇥ 1 and there is only one singular value, we adjoin a zero to the singular value. Thus
"
#" #
3
4
5 h i
5
A = U ⌃V T = 54
1
3
0
5
5
is a singular value decomposition of A.
7.4. THE SINGULAR VALUE DECOMPOSITION
16. From Exercise 6,
AT A =
595

9
12
12
,
16
p
with eigenvalues 25 and 0, so that the singular values of A are 1 = 25 = 5 and
find the corresponding eigenspaces, row-reduce:



3
⇥ T
⇤
16 12 0
3
1
0
4
A A 25I 0 =
!
) x1 =
12
9 0
4
0
0 0



4
⇥ T
⇤
9 12 0
4
1 3 0
A A 0I 0 =
!
) x2 =
.
12 16 0
3
0 0 0
2
=
p
0 = 0. To
To find the matrix V , we must normalize each of these vectors (they are already orthogonal since they
correspond to distinct eigenvalues). This gives
" #
" #
"
#
3
4
3
4
h
i
5
5
5
5
v1 = 4 ,
v2 =
, so that V = v1 v2 = 4
.
3
3
5
5
5
5
The first column of the matrix U is
1h
u1 =
Av1 =
3
5
1
1
i
4
" #
3
5
4
5
h i
= 1 .
Then U is the 1 ⇥ 1 matrix containing 1 as its entry. Since ⌃ should be 1 ⇥ 2 and there is only one
nonzero singular value, we adjoin a zero to the singular value, so that
"
#
3
4
h ih
i
T
5
5
A = U ⌃V = 1 5 0
4
3
5
5
is a singular value decomposition of A.
17. From Exercise 7,
AT A =

4
0
0
,
9
with eigenvalues 9 and 4, so with singular values 3 and 2. To find the corresponding eigenspaces,
row-reduce:



⇥ T
⇤
5 0 0
1 0 0
0
A A 9I 0 =
!
) x1 =
0 0 0
0 0 0
1



⇥ T
⇤
0 0 0
0 1 0
1
A A 4I 0 =
!
) x2 =
.
0 5 0
0 0 0
0
These vectors are already orthonormal, so we set
v1 = x1 ,
v2 = x2 ,
so that
Then U is found by
2
14
u1 =
Av1 =
3
1
2
1
14
u2 =
Av2 =
2
2
1
0
0
2
0
0
2
⇥
V = v1
⇤
v2 =

3
2 3
0 
0
0
35
= 4 15
1
0
0
3
2 3

0
0
1
35
= 4 05 .
0
0
1
0
1
1
.
0
596
CHAPTER 7. DISTANCE AND APPROXIMATION
2 3
1
We must extend u1 and u2 to an orthonormal basis for R3 . Clearly u3 = e1 = 405 works. Then U is
0
the matrix whose columns are the ui . Now, ⌃ should be 3 ⇥ 2, so we add a row of zeros at the bottom,
giving
2
32
3
0
0 1 3 0 
0 1
0 05 40 25
A = U ⌃V T = 41
1 0
0
1 0 0 0
for a singular value decomposition of A.
18. From Exercise 8,
AT A =
with eigenvalues 9 and 1 and singular values
eigenspaces, row-reduce:

⇥ T
⇤
4
A A 9I 0 =
4

⇥ T
⇤
4
A A I 0 =
4
1

=
p
5
4
4
,
5
9 = 3 and
2
=
p
1 = 1. To find the corresponding


0
1 1 0
1
!
) x1 =
0
0 0 0
1


0
1
1 0
1
!
) x2 =
.
0
0
0 0
1
4
4
4
4
These vectors are already orthogonal, but we must normalize them:
"
#
" #
h
p1
p1
x1
x2
2
2 ,
v1 =
=
,
v
=
=
so
that
V
=
v1
2
kx1 k
kx2 k
p1
p1
2
2
i
v2 =
"
p1
2
p1
2
p1
2
p1
2
#
Then U is found by
2
0
16
61
34
1
2
2
0
1
16
u2 =
Av2 = 6
1
14
2
2
u1 =
1
Av1 =
3
2
3
1
p
1 " 1 #
3
2
p
7
6 1 7
2 =6
7
07
5
4 3p2 5
1
p
4
2
p
2
3 2
3
2 3
p1
1 " 1 #
7 p
6 27
2
6 1 7
07
5 p1 = 4 p2 5 .
2
2
0
3
We must
pextend u1 and
pu2 to an orthonormal basis for R . To keep computations simpler, we multiply
u1 by 3 2 and v1 by 2. Then adjoin e1 , e2 , and e3 to these vectors and row-reduce:
2
3
2
3
1
1 1 1 0 0
1 0 0
0
4
6
7
6
17
.
1
4 1 1 0 1 05 ! 4 0 1 0
45
1
4 0 0 0 1
0 0 1
1
2
Then a basis for the column space, which is R3 , is given by the columns of the original matrix corresponding to leading 1’s, which is u1 , u2 , and x3 = e1 . Finally, we must turn this into an orthonormal
basis:
2 3
2 3 2 3
2 3
2 3 2 3
4
1
1
1
1
1
9
1
·
1
+
1
·
0
+
4
·
0
1
·
1
+
1
·
0
+
0
·
0
1
1
6
7
6
7
6
7
6
7
6
7
6
7
x03 = x3
4 15
415 = 405 + 4 15
415 = 4 49 5 .
1 · ( 1) + 1 · 1 + 4 · 4
1·1+1·1+0·0
18
2
2
4
0
0
4
0
9
Normalizing this vector gives
u3 =
x03
kx3 k
2
6
=4
3
2
3
27
.
35
1
3
7.4. THE SINGULAR VALUE DECOMPOSITION
597
Then U is the matrix whose columns are the ui . Now, ⌃ should be 3 ⇥ 2, so we add a row of zeros at
the bottom, giving
2
32
3
1
2
p
p1
"
#
3
0
3
2
p1
p1
6 312
7
6
7
T
1
27 6
2
2
7
A = U ⌃V = 6
3 5 4 0 15
4 3p2 p2
p1
p1
4
1
2
2
p
0 0
0
3
3 2
for a singular value decomposition of A.
19. From Exercise 9,
2
4
AT A = 4 0
2
3
2
05 ,
1
0
4
0
p
with eigenvalues 5, 4, and 0, so the singular values are 1 = 5, 2 = 2, and 3 = 0. To find the
corresponding eigenspaces, row-reduce:
2
3
2
3
2 3
1
0
2 0
1 0
2 0
2
⇥ T
⇤
1
0 05 ! 40 1 0 05 ) x1 = 405
A A 5I 0 = 4 0
2
0
4 0
0 0 0 0
1
2
3
2
3
2 3
0 0
2 0
1 0 0 0
0
⇥ T
⇤
0 05 ! 40 0 1 05 ) x2 = 415
A A 4I 0 = 40 0
2 0
3 0
0 0 0 0
0
2
3
2
3
2 3
1
4 0 2 0
1
1 0 2 0
⇥ T
⇤
A A 0I 0 = 40 4 0 05 ! 40 1 0 05 ) x3 = 4 05 .
2 0 1 0
2
0 0 0 0
These vectors are already orthogonal, but we must normalize them:
2 3
2 3
2
3
p2
p1
0
57
6 57
6 7
6
x3
x1
x2
v1 =
=6
, v2 =
=6
, v3 =
=6
07
17
0 7
4
5
4
5
4
5,
kx1 k
kx2 k
kx3 k
1
2
p
p
0
5
5
so that
⇥
V = v1
Then U is found by
u1 =
v2
2 p2
⇤
5
v3 = 4 0
p1
5
"
1 2
Av1 = p
5 0
1
1
"
1 2
u2 =
Av2 =
2 0
2
1
0
2
0
2
3
p1
5
0
1
0
0 5
p2
5
2 3
# p2
" #
57
1 6
6 0 7= 1
0 4 5
0
p1
5
2 3
" #
# 0
7
1 6
617 = 0 .
0 4 5
1
0
We do not use v3 since it corresponds to a zero singular value. Then u1 and u2 already form a basis
for R2 . Now, ⌃ should be 3 ⇥ 2, so we add a column of zeros on the right, giving
2
3
"
# "p
# p2
p1
0
5
57
1 0
5 0 0 6
6 0
A = U ⌃V T =
1 07
4
5
0 1
0 2 0
p1
p2
0
5
5
for a singular value decomposition of A.
598
CHAPTER 7. DISTANCE AND APPROXIMATION
20. We have
2
3
2
3
1 1 
2 2 2
1 1 1
AT A = 4 1 1 5
= 42 2 25
1 1 1
1 1
2 2 2
Then the singular values of A are the square roots of the eigenvalues of this matrix.
2
det(AT A
2
2
2
I) =
2
2
2
2
=
2
(
6),
2
p
so that the eigenvalues are 6 and 0. Thus the singular values of A are 1 = 6 and 2
find the corresponding eigenspaces, row-reduce:
2
3
2
3
2 3
4
2
2 0
1 0
1 0
1
⇥ T
⇤
4
2 05 ! 40 1
1 05 ) x 1 = 4 15
A A 6I 0 = 4 2
2
2
4 0
0 0 0 0
1
2
3
2
3
2 3
2
2 2 2 0
1 1 1 0
1
⇥ T
⇤
A A 0I 0 = 42 2 2 05 ! 40 0 0 05 ) x2 = 4 15 , x3 = 4
2 2 2 0
0 0 0 0
0
We must orthogonalize x2 and x3 . Use Gram-Schmidt:
2 3
1
x2 · x3
6 7
x03 = x3
x2 = 4 05
x2 · x2
1
Multiply through by 2 to clear fractions, giving
x1
=
kx1 k
p1
6 13 7
6p 7 ,
4 35
p1
3
v2 =
h
V = v1
0 = 0. To
3
1
05 .
1
2
3 2 3
1
1
1 6 7 6 21 7
4 15 = 4 2 5
2
0
1
1
1
2
2
3
p1
2
6
7
x2
=6
0 7
4
5,
kx2 k
v2
i
v3 =
Then U is found by
u1 =
p
3
1
x003 = 4 15 .
2
Finally, we normalize the vectors:
2 3
v1 =
=
"
1 1
Av1 = p
6 1
2
p1
2
p1
6 13
6p
4 3
p1
3
1
1
v3 =
2
6
x003
=6
00
kx3 k 4
3
p1
2
p1
67
p1 7 .
65
p2
6
0
p1
2
3
p1
67
p1 7
65
p2
6
)
2 3
# p1
" #
37
p1
1 6
1
2 .
6p 7 =
p1
1 4 35
p1
3
2
We do not use v2 or v3 since they correspond to a zero singular "
value. # Then we extend u1 to an
orthonormal basis of R2 by adding the orthogonal unit vector u2 =
so we adjoin zeros as required:
A = U ⌃V T =
"
p1
2
p1
2
for a singular value decomposition of A.
p1
2
p1
2
# "p
6
0
2
# p1
3
0 0 6
6 p1
0 0 4 12
p
6
p1
2
p1
2
p1
3
0
p1
6
. Finally, ⌃ must be 2 ⇥ 3,
3
p1
37
p1 7
25
p2
6
7.4. THE SINGULAR VALUE DECOMPOSITION
599
21. Using the singular values and the matrices U and V with columns ui and vi from Exercise 11, we get

 h
i
i
p 1 h 1
0 p1
1
T
T
p
p
p1
A = 1 u1 v1 + 2 u2 v2 = 2
+0
.
2
2
2
2
0
1
22. Using the singular values and the matrices U and V with columns ui and vi from Exercise 14, we get
" #
"
#
i p
h
i
p1
p p1 h
T
T
2
2
A = 1 u1 v1 + 2 u2 v2 = 2 1
.
1 0 + 2
0
1
1
p
p
2
2
23. Using the singular values and the matrices U and V with columns ui and vi from Exercise 17, we get
2 3
2 3
0 h
0
i
i
6 7
6 7h
T
T
6
7
6
A = 1 u1 v1 + 2 u2 v2 = 3 415 0 1 + 2 4 07
.
1
0
5
0
1
24. Using the singular values and the matrices U and V with columns ui and vi from Exercise 19, we get
" #
" #
i
i
p 1 h
0 h
T
T
2
1
p
p
A = 1 u1 v1 + 2 u2 v2 = 5
+
2
.
0
0
1
0
5
5
0
1
25. For example, in Exercise 11, we extended u1 to a basis for R2 by adjoining e2 ; we could instead have
adjoined e2 , giving a di↵erent SVD:
"
# "p
#"
#
p1
1
0
2 0 p12
T
2 .
A =
Download