Uploaded by pedromourame

linear-algebra

advertisement
David Poole
A
M 0 'D E R N
I N T R 0 D U, C T I 0 N
4th edition
Fourth edition
David Poole
Trent University
�..
�
•-
CENGAGE
Learning·
Australia• Brazil• Mexico• Singapore•United l<ingdom •United States
�,.,,
#
••
CENGAGE
Learning·
Linear Algebra
A Modern Introduction, 4th Edition
David Poole
Product Director: Liz Covello
Product Team Manager: Richard Stratton
© 2015, 2011, 2006 Cengage Learning
WCN: 02-200-201
ALL RIGHTS RESERVED. No part of this work covered by the copyright
herein may be reproduced, transmitted, stored, or used in any form
or by any means graphic, electronic, or mechanical, including but not
limited to photocopying, recording, scanning, digitizing, taping, web
Content Developer: Laura Wheel
distribution, information networks, or information storage and retrieval
Product Assistant: Danielle Hallock
systems, except as permitted under Section 107 or 108 of the 1976
Media Developer: Andrew Coppola
United States Copyright Act, without the prior written permission of
the publisher.
Content Project Manager: Alison Eigel Zade
Senior Art Director: Linda May
For product information and technology assistance, contact us at
Cengage Learning Customer & Sales Support, 1-800-354-9706
Manufacturing Planner: Doug Bertke
For permission to use material from this text or product,
submit all requests online at www-cengage.com/permissions­
Rights Acquisition Specialist: Shalice
Shah-Caldwell
Further permissions questions can be emailed to
permissionrequest@cengage.com.
Production Service & Compositor:
MPS Limited
Text Designer: Leonard Massiglia
Cover Designer: Chris Miller
Cover & Interior design Image: Image
Source/Getty Images
Library of Congress Control Number: 2013944173
ISBN-13: 978-1-285-46324-7
ISBN-10: 1-285-46324-2
Cengage Learning
200 First Stamford Place, 4th Floor
Stamford, CT 06902
USA
Cengage Learning is a leading provider of customized learning solutions
with office locations around the globe, including Singapore, the United
international.cengage.com/region.
Kingdom, Australia, Mexico, Brazil and japan. Locate your local office at
Cengage Learning products are represented in Canada by Nelson
Education, Ltd.
For your course and learning solutions, visit www.cengage.com.
preferred online store www.cengagebrain.com.
Purchase any of our products at your local college store or at our
Instructors:
Please visit login.cengage.com and log in to access
instructor-specific resources.
Printed in the United States of America
1 2 3 4 5 6 7 17 16 15 14 13
Dedicated to the memory of
Peter Hilton, who was an
exemplary mathematician,
educator, and citizen-a unit
vector in every sense.
Contents
vii
Preface
To the Instructor
xvii
To the Student
xxiii
Chapter 1
Vectors
1 .0
1.1
1 .2
1 .3
1 .4
Chapter 2
Introduction: The Racetrack Game
The Geometry and Algebra of Vectors
Length and Angle: The Dot Product
Exploration: Vectors and Geometry
Lines and Planes
34
2.3
2.4
2.5
3
18
32
Exploration: The Cross Product
48
Writing Project: The Origins of the Dot Product and Cross Product
50
Applications
Force Vectors
50
55
Chapter Review
Systems of Linear Equations
2.0
2.1
2.2
iv
1
57
Introduction: Triviality
57
Introduction to Systems o f Linear Equations
Direct Methods for Solving Linear Systems
58
64
Writing Project: A History of Gaussian Elimination
Explorations: Lies My Computer Told Me
83
84
Partial Pivoting
Counting Operations: An Introduction to the
Analysis of Algorithms
85
88
Spanning Sets and Linear Independence
99
Applications
Allocation of Resources
99
101
Balancing Chemical Equations
Network Analysis
1 02
1 04
Electrical Networks
Linear Economic Models
1 07
Finite Linear Games
1 09
Vignette: The Global Positioning System
121
Iterative Methods for Solving Linear Systems
134
Chapter Review
1 24
82
49
Chapter 3
Chapter 4
Contents
V
Writing Project: Which Came First: The Matrix or the Determinant?
Vignette: Lewis Carroll's Condensation Method
284
Exploration: Geometric Applications of Determinants
286
Eigenvalues and Eigenvectors of n X n Matrices
292
Writing Project: The History of Eigenvalues
30 1
283
Matrices
136
3.0
3. 1
3.2
3.3
3.4
3.5
3.6
136
Introduction: Matrices in Action
138
Matrix Operations
1 54
Matrix Algebra
1 63
The Inverse of a Matrix
The LU Factorization
1 80
Subspaces, Basis, Dimension, and Rank
Introduction to Linear Transformations
3.7
Applications
230
Markov Chains
230
Linear Economic Models
235
Population Growth
239
Graphs and Digraphs
24 1
251
Chapter Review
Vignette: Robotics
226
Eigenvalues and Eigenvectors
4.0
4. 1
4.2
4.3
4.4
4.5
4.6
191
211
253
Introduction: A Dynamical System on Graphs
Introduction to Eigenvalues and Eigenvectors
263
Determinants
253
254
30 1
Similarity and Diagonalization
31 1
Iterative Methods for Computing Eigenvalues
Applications and the Perron-Frobenius Theorem
325
325
Markov Chains
Population Growth
330
The Perron-Frobenius Theorem
332
Linear Recurrence Relations
335
Systems of Linear Differential Equations
340
Discrete Linear Dynamical Systems
348
356
Vignette: Ranking Sports Teams and Searching the Internet
Chapter Review
Chapter 5
Orthogonality
5.0
5.1
5.2
5.3
5.4
5.5
364
366
Introduction: Shadows on a Wall
366
Orthogonality in IR "
368
Orthogonal Complements and Orthogonal Projections
The Gram-Schmidt Process and the QR Factorization
Explorations: The Modified QR Factorization
396
Approximating Eigenvalues with the QR Algorithm
Orthogonal Diagonalization of Symmetric Matrices
408
Applications
Quadratic Forms
408
415
Graphing Quadratic Equations
425
Chapter Review
378
388
398
400
Vi
Contents
Chapter 6
Vector Spaces
6.0
6. 1
Introduction: Fibonacci in (Vector) Space
Vector Spaces and Subspaces
429
6.2
Linear Independence, Basis, and Dimension
6.3
6.4
6.5
6.6
6.7
Chapter 7
427
Writing Project: The Rise of Vector Spaces
Exploration: Magic Squares
443
460
443
Change of Basis
463
472
Linear Transformations
The Kernel and Range of a Linear Transformation
The Matrix of a Linear Transformation
497
48 1
Exploration: Tilings, Lattices, and the Crystallographic Restriction
518
Applications
Homogeneous Linear Differential Equations
Chapter Review
527
Distance and Approximation
7.0
7. 1
Chapter 8
427
518
529
Introduction: Taxicab Geometry
531
Inner Product Spaces
529
Explorations: Vectors and Matrices with Complex Entries
Geometric Inequalities and Optimization Problems
7.2
7.3
7.4
Norms and Distance Functions
552
Least Squares Approximation
568
The Singular Value Decomposition
590
7.5
Applications
610
Approximation o f Functions
618
Chapter Review
Vignette: Digital Image Compression
607
610
Online only
Codes
8. 1
Code Vectors
620
8.2
8.3
8.4
8.5
627
Error-Correcting Codes
632
Dual Codes
639
Linear Codes
The Minimum Distance of a Code
Vignette: The Codabar System
APPENDIX A
APPENDIXB
APPENDIXC
APPENDIXD
APPENDIX£
543
547
620
626
644
Mathematical Notation and Methods of Proof Al
Mathematical Induction
B1
Complex Numbers
Cl
Polynomials
D1
Online only
Technology Bytes
Answers to Selected Odd-Numbered Exercises
II
Index
ANS I
515
Preface
1he last thing one knows when writing a
book is what to put first.
1670
-Blaise Pascal
Pensees,
For more on the recommendations
of the Linear Algebra Curriculum
Study Group, see 1he College
41-46.
Mathematics Journal 24
(1993),
The fourth edition of Linear Algebra: A Modern Introduction preserves the approach
and features that users found to be strengths of the previous editions. However, I have
streamlined the text somewhat, added numerous clarifications, and freshened up the
exercises.
I want students to see linear algebra as an exciting subject and to appreciate its
tremendous usefulness. At the same time, I want to help them master the basic con cepts and techniques of linear algebra that they will need in other courses, both in
mathematics and in other disciplines. I also want students to appreciate the interplay
of theoretical, applied, and numerical mathematics that pervades the subject.
This book is designed for use in an introductory one- or two-semester course
sequence in linear algebra. First and foremost, it is intended for students, and I have
tried my best to write the book so that students not only will find it readable but also
will want to read it. As in the first three editions, I have taken into account the reality
that students taking introductory linear algebra are likely to come from a variety of
disciplines. In addition to mathematics majors, there are apt to be majors from
engineering, physics, chemistry, computer science, biology, environmental science,
geography, economics, psychology, business, and education, as well as other students
taking the course as an elective or to fulfill degree requirements. Accordingly, the book
balances theory and applications, is written in a conversational style yet is fully rigorous,
and combines a traditional presentation with concern for student-centered learning.
There is no such thing as a universally best learning style. In any class, there will be
some students who work well independently and others who work best in groups;
some who prefer lecture-based learning and others who thrive in a workshop setting,
doing explorations; some who enjoy algebraic manipulations, some who are adept at
numerical calculations (with and without a computer), and some who exhibit strong
geometric intuition. In this edition, I continue to present material in a variety of
ways-algebraically, geometrically, numerically, and verbally-so that all types oflearn­
ers can find a path to follow. I have also attempted to present the theoretical, computa­
tional, and applied topics in a flexible yet integrated way. In doing so, it is my hope that
all students will be exposed to the many sides of linear algebra.
This book is compatible with the recommendations ofthe Linear Algebra Curriculum
Study Group. From a pedagogical point of view, there is no doubt that for most students
Vii
Viii
Preface
concrete examples should precede abstraction. I have taken this approach here. I also
believe strongly that linear algebra is essentially about vectors and that students need to
see vectors first (in a concrete setting) in order to gain some geometric insight. Moreover,
introducing vectors early allows students to see how systems of linear equations arise
naturally from geometric problems. Matrices then arise equally naturally as coefficient
matrices oflinear systems and as agents of change (linear transformations). This sets the
stage for eigenvectors and orthogonal projections, both of which are best understood
geometrically. The dart that appears on the cover of this book symbolizes a vector and
reflects my conviction that geometric understanding should precede computational
techniques.
I have tried to limit the number of theorems in the text. For the most part, results
labeled as theorems either will be used later in the text or summarize preceding work.
Interesting results that are not central to the book have been included as exercises or
explorations. For example, the cross product of vectors is discussed only in explo­
rations (in Chapters 1 and 4). Unlike most linear algebra textbooks, this book has no
chapter on determinants. The essential results are all in Section 4.2, with other inter­
esting material contained in an exploration. The book is, however, comprehensive for
an introductory text. Wherever possible, I have included elementary and accessible
proofs of theorems in order to avoid having to say, "The proof of this result is beyond
the scope of this text:' The result is, I hope, a work that is self-contained.
I have not been stingy with the applications: There are many more in the book than
can be covered in a single course. However, it is important that students see the impressive
range of problems to which linear algebra can be applied. I have included some modern
material on finite linear algebra and coding theory that is not normally found in an intro­
ductory linear algebra text. There are also several impressive real-world applications of
linear algebra and one item of historical, if not practical, interest; these applications are
presented as self-contained "vignettes:'
I hope that instructors will enjoy teaching from this book. More important, I hope
that students using the book will come away with an appreciation of the beauty, power,
and tremendous utility of linear algebra and that they will have fun along the way.
W h a t's New i n t h e Fo u rt h Edition
Th e overall structure and style o f Linear Algebra: A Modern Introduction remain the
same in the fourth edition.
Here is a summary of what is new:
The applications to coding theory have been moved to the new online Chapter 8.
To further engage students, five writing projects have been added to the exer­
cise sets. These projects give students a chance to research and write about aspects of
the history and development oflinear algebra. The explorations, vignettes, and many
of the applications provide additional material for student projects.
•
There are over 200 new or revised exercises. In response to reviewers' com­
ments, there is now a full proof of the Cauchy-Schwarz Inequality in Chapter 1 in the
form of a guided exercise.
•
I have made numerous small changes in wording to improve the clarity or
accuracy of the exposition. Also, several definitions have been made more explicit by
giving them their own definition boxes and a few results have been highlighted by
labeling them as theorems.
•
All existing ancillaries have been updated.
•
•
See pages 49, 82, 283,
301, 443
Preface
ix
Features
Clear Writi ng Stvle
The text is written is a simple, direct, conversational style. As much as possible, I have
used "mathematical English" rather than relying excessively on mathematical nota­
tion. However, all proofs that are given are fully rigorous, and Appendix A contains
an introduction to mathematical notation for those who wish to streamline their own
writing. Concrete examples almost always precede theorems, which are then followed
by further examples and applications. This flow-from specific to general and back
again-is consistent throughout the book.
Kev concepts Introduced Early
Many students encounter difficulty in linear algebra when the course moves from the
computational (solving systems of linear equations, manipulating vectors and matri­
ces) to the theoretical (spanning sets, linear independence, subspaces, basis, and
dimension) . This book introduces all of the key concepts of linear algebra early, in a
concrete setting, before revisiting them in full generality. Vector concepts such as dot
product, length, orthogonality, and projection are first discussed in Chapter 1 in the
concrete setting of IR 2 and IR 3 before the more general notions of inner product, norm,
and orthogonal projection appear in Chapters 5 and 7. Similarly, spanning sets and
linear independence are given a concrete treatment in Chapter 2 prior to their gener­
alization to vector spaces in Chapter 6. The fundamental concepts of subspace, basis,
and dimension appear first in Chapter 3 when the row, column, and null spaces of a
matrix are introduced; it is not until Chapter 6 that these ideas are given a general
treatment. In Chapter 4, eigenvalues and eigenvectors are introduced and explored
for 2 X 2 matrices before their n X n counterparts appear. By the beginning of Chap­
ter 4, all of the key concepts of linear algebra have been introduced, with concrete,
computational examples to support them. When these ideas appear in full generality
later in the book, students have had time to get used to them and, hence, are not so
intimidated by them.
Emphasis on Vectors and Geometry
In keeping with the philosophy that linear algebra is primarily about vectors, this
book stresses geometric intuition. Accordingly, the first chapter is about vectors, and
it develops many concepts that will appear repeatedly throughout the text. Concepts
such as orthogonality, projection, and linear combination are all found in Chapter 1 ,
as is a comprehensive treatment o f lines and planes in IR 3 that provides essential
insight into the solution of systems of linear equations. This emphasis on vectors,
geometry, and visualization is found throughout the text. Linear transformations are
introduced as matrix transformations in Chapter 3, with many geometric examples,
before general linear transformations are covered in Chapter 6. In Chapter 4, eigen­
values are introduced with "eigenpictures" as a visual aid. The proof of Perron's
Theorem is given first heuristically and then formally, in both cases using a geometric
argument. The geometry of linear dynamical systems reinforces and summarizes the
material on eigenvalues and eigenvectors. In Chapter 5, orthogonal projections, or­
thogonal complements of subspaces, and the Gram-Schmidt Process are all presented
in the concrete setting of IR 3 before being generalized to IR " and, in Chapter 7, to inner
X
Preface
product spaces. The nature of the singular value decomposition is also explained in­
formally in Chapter 7 via a geometric argument. Of the more than 300 figures in the
text, over 200 are devoted to fostering a geometric understanding of linear algebra.
Exploralions
See pages 1,
136, 427, 529
See pages 32, 286, 460, 515, 543, 547
See pages 83,
84, 85, 396, 398
The introduction to each chapter is a guided exploration (Section O) in which stu­
dents are invited to discover, individually or in groups, some aspect of the upcoming
chapter. For example, "The Racetrack Game" introduces vectors, "Matrices in Action"
introduces matrix multiplication and linear transformations, "Fibonacci in (Vector)
Space" touches on vector space concepts, and "Taxicab Geometry" sets up general­
ized norms and distance functions. Additional explorations found throughout the
book include applications of vectors and determinants to geometry, an investigation
of 3 X 3 magic squares, a study of symmetry via the tilings of M. C. Escher, an intro­
duction to complex linear algebra, and optimization problems using geometric
inequalities. There are also explorations that introduce important numerical consid­
erations and the analysis of algorithms. Having students do some of these explo­
rations is one way of encouraging them to become active learners and to give them
"ownership" over a small part of the course.
APPlicalions
See pages 623,
641
See pages 121,
226, 356, 607, 626
The book contains an abundant selection of applications chosen from a broad range
of disciplines, including mathematics, computer science, physics, chemistry, engi­
neering, biology, business, economics, psychology, geography, and sociology. Note­
worthy among these is a strong treatment of coding theory, from error-detecting
codes (such as International Standard Book Numbers) to sophisticated error­
correcting codes (such as the Reed-Muller code that was used to transmit satellite
photos from space). Additionally, there are five "vignettes" that briefly showcase some
very modern applications of linear algebra: the Global Positioning System (GPS), ro­
botics, Internet search engines, digital image compression, and the Codabar System.
Examples and Exercises
See pages 248,
359, 526, 588
There are over 400 examples in this book, most worked in greater detail than is cus­
tomary in an introductory linear algebra textbook. This level of detail is in keeping
with the philosophy that students should want (and be able) to read a textbook.
Accordingly, it is not intended that all of these examples be covered in class; many can
be assigned for individual or group study, possibly as part of a project. Most examples
have at least one counterpart exercise so that students can try out the skills covered in
the example before exploring generalizations.
There are over 2000 exercises, more than in most textbooks at a similar level.
Answers to most of the computational odd-numbered exercises can be found in the
back of the book. Instructors will find an abundance of exercises from which to select
homework assignments. The exercises in each section are graduated, progressing from
the routine to the challenging. Exercises range from those intended for hand computa­
tion to those requiring the use of a calculator or computer algebra system, and from
theoretical and numerical exercises to conceptual exercises. Many of the examples and
exercises use actual data compiled from real-world situations. For example, there are
problems on modeling the growth of caribou and seal populations, radiocarbon dating
Preface
Xi
of the Stonehenge monument, and predicting major league baseball players' salaries.
Working such problems reinforces the fact that linear algebra is a valuable tool for mod­
eling real-life problems.
Additional exercises appear in the form of a review after each chapter. In each set,
there are 10 true/false questions designed to test conceptual understanding, followed
by 1 9 computational and theoretical exercises that summarize the main concepts and
techniques of that chapter.
Biographical Sketches and Etvmological Notes
See page 34
It is important that students learn something about the history of mathematics and
come to see it as a social and cultural endeavor as well as a scientific one. Accord­
ingly, the text contains short biographical sketches about many of the mathemati­
cians who contributed to the development of linear algebra. I hope that these will
help to put a human face on the subject and give students another way of relating to
the material.
I have found that many students feel alienated from mathematics because the
terminology makes no sense to them-it is simply a collection of words to be learned.
To help overcome this problem, I have included short etymological notes that give
the origins of many of the terms used in linear algebra. (For example, why do we
use the word normal to refer to a vector that is perpendicular to a plane?)
Margin Icons
The margins of the book contain several icons whose purpose is to alert the reader in
various ways. Calculus is not a prerequisite for this book, but linear algebra has many
interesting and important applications to calculus. The � icon denotes an example or
exercise that requires calculus. (This material can be omitted if not everyone in the
class has had at least one semester of calculus. Alternatively, this material can be as­
signed as projects.) The � icon denotes an example or exercise involving complex
numbers. (For students unfamiliar with complex numbers, Appendix C contains all
the background material that is needed.) The cAs icon indicates that a computer algebra
system (such as Maple, Mathematica, or MATLAB) or a calculator with matrix capa­
bilities (such as almost any graphing calculator) is required-or at least very useful­
for solving the example or exercise.
In an effort to help students learn how to read and use this textbook most ef­
fectively, I have noted various places where the reader is advised to pause. These
may be places where a calculation is needed, part of a proof must be supplied, a
claim should be verified, or some extra thought is required. The _.... icon appears
in the margin at such places; the message is "Slow down. Get out your pencil.
Think about this:'
Technology
This book can be used successfully whether or not students have access to technol­
ogy. However, calculators with matrix capabilities and computer algebra systems
are now commonplace and, properly used, can enrich the learning experience as
well as help with tedious calculations. In this text, I take the point of view that stu­
dents need to master all of the basic techniques of linear algebra by solving by hand
examples that are not too computationally difficult. Technology may then be used
Xii
Preface
(in whole or in part) to solve subsequent examples and applications and to apply
techniques that rely on earlier ones. For example, when systems of linear equations
are first introduced, detailed solutions are provided; later, solutions are simply
given, and the reader is expected to verify them. This is a good place to use some
form of technology. Likewise, when applications use data that make hand calcula­
tion impractical, use technology. All of the numerical methods that are discussed
depend on the use of technology.
With the aid of technology, students can explore linear algebra in some exciting
ways and discover much for themselves. For example, if one of the coefficients of a
linear system is replaced by a parameter, how much variability is there in the solu­
tions? How does changing a single entry of a matrix affect its eigenvalues? This book
is not a tutorial on technology, and in places where technology can be used, I have not
specified a particular type of technology. The student companion website that
accompanies this book offers an online appendix called Technology Bytes that gives
instructions for solving a selection of examples from each chapter using Maple, Math­
ematica, and MATLAB. By imitating these examples, students can do further calcula­
tions and explorations using whichever CAS they have and exploit the power of these
systems to help with the exercises throughout the book, particularly those marked
with the cAs icon. The website also contains data sets and computer code in Maple,
Mathematica, and MATLAB formats keyed to many exercises and examples in the
text. Students and instructors can import these directly into their CAS to save typing
and eliminate errors.
Finite and Numerical linear Algebra
See pages 83, 84, 124,
555, 561, 568, 590
See pages 319,
180, 311, 392,
563, 600
The text covers two aspects of linear algebra that are scarcely ever mentioned to­
gether: finite linear algebra and numerical linear algebra. By introducing modular
arithmetic early, I have been able to make finite linear algebra (more properly, "linear
algebra over finite fields;' although I do not use that phrase) a recurring theme
throughout the book. This approach provides access to the material on coding theory
in Chapter 8 (online) . There is also an application to finite linear games in Section 2.4
that students really enjoy. In addition to being exposed to the applications of finite
linear algebra, mathematics majors will benefit from seeing the material on finite
fields, because they are likely to encounter it in such courses as discrete mathematics,
abstract algebra, and number theory.
All students should be aware that in practice, it is impossible to arrive at exact
solutions of large-scale problems in linear algebra. Exposure to some of the tech­
niques of numerical linear algebra will provide an indication of how to obtain
highly accurate approximate solutions. Some of the numerical topics included in
the book are roundoff error and partial pivoting, iterative methods for solving
linear systems and computing eigenvalues, the LU and QR factorizations, matrix
norms and condition numbers, least squares approximation, and the singular
value decomposition. The inclusion of numerical linear algebra also brings up
some interesting and important issues that are completely absent from the theory
of linear algebra, such as pivoting strategies, the condition of a linear system, and
the convergence of iterative methods. This book not only raises these questions
but also shows how one might approach them. Gerschgorin disks, matrix norms,
and the singular values of a matrix, discussed in Chapters 4 and 7, are useful in
this regard.
Preface
Xiii
Appendices
Appendix A contains an overview of mathematical notation and methods of proof,
and Appendix B discusses mathematical induction. All students will benefit from
these sections, but those with a mathematically oriented major may wish to pay
particular attention to them. Some of the examples in these appendices are uncom­
mon (for instance, Example B.6 in Appendix B) and underscore the power of the
methods. Appendix C is an introduction to complex numbers. For students familiar
with these results, this appendix can serve as a useful reference; for others, this sec­
tion contains everything they need to know for those parts of the text that use com­
plex numbers. Appendix D is about polynomials. I have found that many students
require a refresher about these facts. Most students will be unfamiliar with Descartes's
Rule of Signs; it is used in Chapter 4 to explain the behavior of the eigenvalues of
Leslie matrices. Exercises to accompany the four appendices can be found on the
book's website.
Short answers to most of the odd-numbered computational exercises are given at
the end of the book. Exercise sets to accompany Appendixes A, B, C, and D are avail­
able on the companion website, along with their odd-numbered answers.
Ancillaries
For 1ns1ruc1ors
Enhanced WebAssign® �ebAssign
Printed Access Card: 978- 1 -285-85829-6
Online Access Code: 978- 1 -285-85827-2
Exclusively from Cengage Learning, Enhanced WebAssign combines the exceptional
mathematics content that you know and love with the most powerful online home­
work solution, WebAssign. Enhanced WebAssign engages students with immediate
feedback, rich tutorial content, and interactive, fully customizable eBooks (YouBook),
helping students to develop a deeper conceptual understanding of their subject
matter. Flexible assignment options give instructors the ability to release assignments
conditionally based on students' prerequisite assignment scores. Visit us at www.
cengage.com/ewa to learn more.
Cengage Learning Testing Powered by Cognero
Cengage Learning Testing Powered by Cognero is a flexible, online system that allows
you to author, edit, and manage test bank content from multiple Cengage Learning
solutions; create multiple test versions in an instant; and deliver tests from your LMS,
your classroom, or wherever you want.
Complete Solutions Manual
The Complete Solutions Manual provides detailed solutions to all exercises in the
text, including Exploration and Chapter Review exercises. The Complete Solutions
Manual is available online.
Instructor's Guide
This online guide enhances the text with valuable teaching resources such as group
work projects, teaching tips, interesting exam questions, examples and extra
xiv
Preface
material for lectures, and other items designed to reduce the instructor's prepara­
tion time and make linear algebra class an exciting and interactive experience. For
each section of the text, the Instructor's Guide includes suggested time and empha­
sis, points to stress, questions for discussion, lecture materials and examples, tech­
nology tips, student projects, group work with solutions, sample assignments, and
suggested test questions.
Solution Builder
www.cengage.com/solutionbuilder
Solution Builder provides full instructor solutions to all exercises in the text, includ­
ing those in the explorations and chapter reviews, in a convenient online format.
Solution Builder allows instructors to create customized, secure PDF printouts of
solutions matched exactly to the exercises assigned for class.
*Access Cognero and additional instructor resources online at login.cengage.com.
For Students
Student Solutions Manual (ISBN- 13: 978 - 1 -285-84 195-3)
The Student Solutions Manual and Study Guide includes detailed solutions to all odd­
numbered exercises and selected even-numbered exercises; section and chapter
summaries of symbols, definitions, and theorems; and study tips and hints. Complex
exercises are explored through a question-and-answer format designed to deepen
understanding. Challenging and entertaining problems that further explore selected
exercises are also included.
Enhanced WebAssign®
�
WebAssign
Printed Access Card: 978- 1 -285-85829-6
Online Access Code: 978- 1 -285-85827-2
Enhanced Web Assign (assigned by the instructor) provides you with instant feedback
on homework assignments. This online homework system is easy to use and includes
helpful links to textbook sections, video examples, and problem-specific tutorials.
CengageBrain.com
To access additional course materials and companion resources, please visit www.
cengagebrain.com. At the CengageBrain.com home page, search for the ISBN of your
title (from the back cover of your book) using the search box at the top of the page.
This will take you to the product page where free companion resources can be found.
Acknowledgments
The reviewers of the previous edition of this text contributed valuable and often in­
sightful comments about the book. I am grateful for the time each of them took to do
this. Their judgement and helpful suggestions have contributed greatly to the devel­
opment and success of this book, and I would like to thank them personally:
Jamey Bass, City College of San Francisco; Olga Brezhneva, Miami University; Karen
Clark, The College of New Jersey; Marek Elzanowski, Portland State University;
Christopher Francisco, Oklahoma State University; Brian Jue, California State
University, Stanislaus; Alexander Kheyfits, Bronx Community College/CUNY;
Henry Krieger, Harvey Mudd College; Rosanna Pearlstein, Michigan State
To the Instructor
See page 284
See page 286
See pages 325,
330
See page 348
xix
course" in determinants contains all the essential material students need, including
an optional but elementary proof of the Laplace Expansion Theorem. The vignette
"Lewis Carroll's Condensation Method" presents a historically interesting, alternative
method of calculating determinants that students may find appealing. The explo­
ration "Geometric Applications of Determinants" makes a nice project that contains
several interesting and useful results. (Alternatively, instructors who wish to give
more detailed coverage to determinants may choose to cover some of this exploration
in class.) The basic theory of eigenvalues and eigenvectors is found in Section 4.3, and
Section 4.4 deals with the important topic of diagonalization. Example 4.29 on powers
of matrices is worth covering in class. The power method and its variants, discussed
in Section 4.5, are optional, but all students should be aware of the method, and an
applied course should cover it in detail. Gerschgorin's Disk Theorem can be covered
independently of the rest of Section 4.5. Markov chains and the Leslie model of pop­
ulation growth reappear in Section 4.6. Although the proof of Perron's Theorem is
optional, the theorem itself (like the stronger Perron-Frobenius Theorem) should at
least be mentioned because it explains why we should expect a unique positive eigen­
value with a corresponding positive eigenvector in these applications. The applica­
tions on recurrence relations and differential equations connect linear algebra to dis­
crete mathematics and calculus, respectively. The matrix exponential can be covered
if your class has a good calculus background. The final topic of discrete linear dynamical systems revisits and summarizes many of the ideas in Chapter 4, looking at them
in a new, geometric light. Students will enjoy reading how eigenvectors can be used to
help rank sports teams and websites. This vignette can easily be extended to a project
or enrichment activity.
Chapter 5: onhouonalilv
See page 366
See pages 396,
398
See pages 408,
415
The introductory exploration, "Shadows on a Wall;' is mathematics at its best: it takes
a known concept (projection of a vector onto another vector) and generalizes it in a
useful way (projection of a vector onto a subspace-a plane), while uncovering some
previously unobserved properties. Section 5 . 1 contains the basic results about or­
thogonal and orthonormal sets of vectors that will be used repeatedly from here on.
In particular, orthogonal matrices should be stressed. In Section 5.2, two concepts
from Chapter 1 are generalized: the orthogonal complement of a subspace and the
orthogonal projection of a vector onto a subspace. The Orthogonal Decomposition
Theorem is important here and helps to set up the Gram-Schmidt Process. Also note
the quick proof of the Rank Theorem. The Gram-Schmidt Process is detailed in
Section 5.3, along with the extremely important QR factorization. The two explo­
rations that follow outline how the QR factorization is computed in practice and how
it can be used to approximate eigenvalues. Section 5.4 on orthogonal diagonalization
of (real) symmetric matrices is needed for the applications that follow. It also contains
the Spectral Theorem, one of the highlights of the theory of linear algebra. The appli­
cations in Section 5.5 are quadratic forms and graphing quadratic equations. I always
include at least the second of these in my course because it extends what students al­
ready know about conic sections.
Chapter 6: vector Spaces
See page 427
The Fibonacci sequence reappears in Section 6.0, although it is not important that
students have seen it before (Section 4.6). The purpose of this exploration is to show
Preface
XV
University; William Sullivan, Portland State University; Matthias Weber, Indiana
University.
I am indebted to a great many people who have, over the years, influenced my
views about linear algebra and the teaching of mathematics in general. First, I would
like to thank collectively the participants in the education and special linear algebra
sessions at meetings of the Mathematical Association of America and the Canadian
Mathematical Society. I have also learned much from participation in the Canadian
Mathematics Education Study Group and the Canadian Mathematics Education
Forum.
I especially want to thank Ed Barbeau, Bill Higginson, Richard Hoshino, John
Grant McLaughlin, Eric Muller, Morris Orzech, Bill Ralph, Pat Rogers, Peter Taylor,
and Walter Whiteley, whose advice and inspiration contributed greatly to the
philosophy and style of this book. My gratitude as well to Robert Rogers, who devel­
oped the student and instructor solutions, as well as the excellent study guide content.
Special thanks go to Jim Stewart for his ongoing support and advice. Joe Rotman and
his lovely book A First Course in Abstract Algebra inspired the etymological notes in
this book, and I relied heavily on Steven Schwartzman's The Words of Mathematics
when compiling these notes. I thank Art Benjamin for introducing me to the Codabar
system and Joe Grear for clarifying aspects of the history of Gaussian elimination. My
colleagues Marcus Pivato and Reem Yassawi provided useful information about dy­
namical systems. As always, I am grateful to my students for asking good questions
and providing me with the feedback necessary to becoming a better teacher.
I sincerely thank all of the people who have been involved in the production of
this book. Jitendra Kumar and the team at MPS Limited did an amazing job produc­
ing the fourth edition. I thank Christine Sabooni for doing a thorough copyedit. Most
of all, it has been a delight to work with the entire editorial, marketing, and produc­
tion teams at Cengage Learning: Richard Stratton, Molly Taylor, Laura Wheel, Cynthia
Ashton, Danielle Hallock, Andrew Coppola, Alison Eigel Zade, and Janay Pryor. They
offered sound advice about changes and additions, provided assistance when I needed
it, but let me write the book I wanted to write. I am fortunate to have worked with
them, as well as the staffs on the first through third editions.
As always, I thank my family for their love, support, and understanding. Without
them, this book would not have been possible.
David Poole
dpoole@trentu.ca
To the
Instructor
"Would you tell me, please,
which way I ought to go from here?"
"That depends a good deal on where
you want to get to," said the Cat.
-Lewis Carroll
1865
Alice's Adventures in
Wonderland,
This text was written with flexibility in mind. It is intended for use in a one- or
two-semester course with 36 lectures per semester. The range of topics and applica­
tions makes it suitable for a variety of audiences and types of courses. However,
there is more material in the book than can be covered in class, even in a two­
semester course. After the following overview of the text are some brief suggestions
for ways to use the book.
An overview of t h e Text
Chanler 1: vec1ors
See page 1
See page 32
See page 48
The racetrack game in Section 1 .0 serves to introduce vectors in an informal way. (It's also
quite a lot of fun to play!) Vectors are then formally introduced from both algebraic and
geometric points of view. The operations of addition and scalar multiplication and their
properties are first developed in the concrete settings of !R 2 and IR3 before being general­
ized to !R n . Modular arithmetic and finite linear algebra are also introduced. Section 1 .2
defines the dot product of vectors and the related notions of length, angle, and orthogo­
nality. The very important concept of (orthogonal) projection is developed here; it will
reappear in Chapters 5 and 7. The exploration "Vectors and Geometry" shows how vec­
tor methods can be used to prove certain results in Euclidean geometry. Section 1 .3 is a
basic but thorough introduction to lines and planes in IR 2 and IR3. This section is crucial
for understanding the geometric significance of the solution of linear systems in Chap­
ter 2. Note that the cross product of vectors in IR3 is left as an exploration. The chapter
concludes with an application to force vectors.
Chapter 2: svstems of linear Equations
See page 57
The introduction to this chapter serves to illustrate that there is more than one way to
think of the solution to a system of linear equations. Sections 2 . 1 and 2.2 develop the
xv i i
xv i i i
To the Instructor
See pages 72, 205, 386, 486
See page 121
See pages 83,
84, 85
main computational tool for solving linear systems: row reduction of matrices (Gaus­
sian and Gauss-Jordan elimination) . Nearly all subsequent computational methods in
the book depend on this. The Rank Theorem appears here for the first time; it shows
up again, in more generality, in Chapters 3, 5, and 6. Section 2.3 is very important; it
introduces the fundamental notions of spanning sets and linear independence of vec­
tors. Do not rush through this material. Section 2.4 contains six applications from
which instructors can choose depending on the time available and the interests of the
class. The vignette on the Global Positioning System provides another application
that students will enjoy. The iterative methods in Section 2.5 will be optional for many
courses but are essential for a course with an applied/numerical focus. The three ex­
plorations in this chapter are related in that they all deal with aspects of the use of
computers to solve linear systems. All students should at least be made aware of these
issues.
Chanler 3: Malrices
See page 136
See pages 172,
206, 296, 512, 605
See page 226
See pages 230,
239
This chapter contains some of the most important ideas in the book. It is a long
chapter, but the early material can be covered fairly quickly, with extra time allowed
for the crucial material in Section 3.5. Section 3.0 is an exploration that introduces
the notion of a linear transformation: the idea that matrices are not just static objects
but rather a type of function, transforming vectors into other vectors. All of the basic
facts about matrices, matrix operations, and their properties are found in the first two
sections. The material on partitioned matrices and the multiple representations of the
matrix product is worth stressing, because it is used repeatedly in subsequent sections.
The Fundamental Theorem of Invertible Matrices in Section 3.3 is very important
and will appear several more times as new characterizations of invertibility are pre­
sented. Section 3.4 discusses the very important LU factorization of a matrix. If this
topic is not covered in class, it is worth assigning as a project or discussing in a work­
shop. The point of Section 3.5 is to present many of the key concepts of linear algebra
(subspace, basis, dimension, and rank) in the concrete setting of matrices before stu dents see them in full generality. Although the examples in this section are all famil­
iar, it is important that students get used to the new terminology and, in particular,
understand what the notion of a basis means. The geometric treatment of linear
transformations in Section 3.6 is intended to smooth the transition to general linear
transformations in Chapter 6. The example of a projection is particularly important
because it will reappear in Chapter 5. The vignette on robotic arms is a concrete
demonstration of composition of linear (and affine) transformations. There are four
applications from which to choose in Section 3.7. Either Markov chains or the Leslie
model of population growth should be covered so that they can be used again in
Chapter 4, where their behavior will be explained.
Chanler 4: Eigenvalues and Eigenveclors
See page 253
The introduction Section 4.0 presents an interesting dynamical system involving
graphs. This exploration introduces the notion of an eigenvector and foreshadows the
power method in Section 4.5. In keeping with the geometric emphasis of the book,
Section 4. 1 contains the novel feature of "eigenpictures" as a way of visualizing the
eigenvectors of 2 X 2 matrices. Determinants appear in Section 4.2, motivated by
their use in finding the characteristic polynomials of small matrices. This "crash
XX
To the Instructor
See page 515
that familiar vector space concepts (Section 3.5) can be used fruitfully in a new
setting. Because all of the main ideas of vector spaces have already been introduced in
Chapters 1 -3, students should find Sections 6. 1 and 6.2 fairly familiar. The emphasis
here should be on using the vector space axioms to prove properties rather than rely­
ing on computational techniques. When discussing change of basis in Section 6.3, it
is helpful to show students how to use the notation to remember how the construc­
tion works. Ultimately, the Gauss-Jordan method is the most efficient here. Sec­
tions 6.4 and 6.5 on linear transformations are important. The examples are related to
previous results on matrices (and matrix transformations) . In particular, it is impor­
tant to stress that the kernel and range of a linear transformation generalize the null
space and column space of a matrix. Section 6.6 puts forth the notion that (almost) all
linear transformations are essentially matrix transformations. This builds on the
information in Section 3.6, so students should not find it terribly surprising. However,
the examples should be worked carefully. The connection between change of basis
and similarity of matrices is noteworthy. The exploration "Tilings, Lattices, and the
Crystallographic Restriction" is an impressive application of change of basis. The con­
nection with the artwork of M. C. Escher makes it all the more interesting. The appli­
cations in Section 6.7 build on previous ones and can be included as time and interest
permit.
Chapter 1: Distance and Approximation
See page 529
See page 543
See page 547
See page 607
Section 7 .0 opens with the entertaining "Taxicab Geometry" exploration. Its
purpose is to set up the material on generalized norms and distance functions
(metrics) that follows. Inner product spaces are discussed in Section 7 . 1 ; the em­
phasis here should be on the examples and using the axioms. The exploration "Vec­
tors and Matrices with Complex Entries" shows how the concepts of dot product,
symmetric matrix, orthogonal matrix, and orthogonal diagonalization can be ex­
tended from real to complex vector spaces. The following exploration, "Geometric
Inequalities and Optimization Problems:' is one that students typically enjoy. (They
will have fun seeing how many "calculus" problems can be solved without using
calculus at all!) Section 7.2 covers generalized vector and matrix norms and shows
how the condition number of a matrix is related to the notion of ill-conditioned
linear systems explored in Chapter 2. Least squares approximation (Section 7.3) is
an important application of linear algebra in many other disciplines. The Best Ap­
proximation Theorem and the Least Squares Theorem are important, but their
proofs are intuitively clear. Spend time here on the examples-a few should suffice.
Section 7.4 presents the singular value decomposition, one of the most impressive
applications of linear algebra. If your course gets this far, you will be amply re­
warded. Not only does the SVD tie together many notions discussed previously; it
also affords some new (and quite powerful) applications. If a CAS is available, the
vignette on digital image compression is worth presenting; it is a visually impres­
sive display of the power of linear algebra and a fitting culmination to the course.
The further applications in Section 7.5 can be chosen according to the time avail­
able and the interests of the class.
Chapter 8: Codes
This online chapter contains applications of linear algebra to the theory of codes.
Section 8 . 1 begins with a discussion of how vectors can be used to design
To the Instructor
See page 626
XXi
error- detecting codes such as the familiar Universal Product Code (UPC) and
International Standard Book Number (ISBN). This topic only requires knowl­
edge of Chapter 1 . The vignette on the Codabar system used in credit and bank
cards is an excellent classroom presentation that can even be used to introduce
Section 8 . 1 . Once students are familiar with matrix operations, Section 8.2 de­
scribes how codes can be designed to correct as well as detect errors. The
Hamming codes introduced here are perhaps the most famous examples of such
error- correcting codes. Dual codes, discussed in Section 8.3, are an important
way of constructing new codes from old ones. The notion of orthogonal comple­
ment, introduced in Chapter 5, is the prerequisite concept here. The most
important, and most widely used, class of codes is the class of linear codes that is
defined in Section 8.4. The notions of subspace, basis, and dimension are key
here. The powerful Reed-Muller codes used by NASA spacecraft are important
examples of linear codes. Our discussion of codes concludes in Section 8.5 with
the definition of the minimum distance of a code and the role it plays in deter­
mining the error-correcting capability of the code.
H o w to u s e t h e B o o k
Students find the book easy to read, s o I usually have them read a section before I
cover the material in class. That way, I can spend class time highlighting the most
important concepts, dealing with topics students find difficult, working examples,
and discussing applications. I do not attempt to cover all of the material from the
assigned reading in class. This approach enables me to keep the pace of the course
fairly brisk, slowing down for those sections that students typically find
challenging.
In a two-semester course, it is possible to cover the entire book, including a rea­
sonable selection of applications. For extra flexibility, you might omit some of the
topics (for example, give only a brief treatment of numerical linear algebra), thereby
freeing up time for more in-depth coverage of the remaining topics, more applica­
tions, or some of the explorations. In an honors mathematics course that emphasizes
proofs, much of the material in Chapters 1 -3 can be covered quickly. Chapter 6 can
then be covered in conjunction with Sections 3.5 and 3.6, and Chapter 7 can be in­
tegrated into Chapter 5. I would be sure to assign the explorations in Chapters 1, 4,
6, and 7 for such a class.
For a one-semester course, the nature of the course and the audience will deter­
mine which topics to include. Three possible courses are described below and on the
following page. The basic course, described first, has fewer than 36 hours suggested,
allowing time for extra topics, in-class review, and tests. The other two courses build
on the basic course but are still quite flexible.
A Basic Course
A course designed for mathematics majors and students from other disciplines is
outlined on the next page. This course does not mention general vector spaces at all
(all concepts are treated in a concrete setting) and is very light on proofs. Still, it is a
thorough introduction to linear algebra.
XXii
To the Instructor
Section
1.1
1.2
1.3
2.1
2.2
2.3
3.1
3.2
3.3
3.5
Number of Lectures
1 - 1 .5
1 - 1 .5
0.5- 1
1 -2
1 -2
1-2
2
2
Section
Number of Lectures
3.6
4.1
4.2
4.3
4.4
5.1
5.2
5.3
5.4
7.3
1 -2
1
2
1
1 -2
1 - 1 .5
1 - 1 .5
0.5
1
2
Total: 23-30 lectures
Because the students in a course such as this one represent a wide variety of dis­
ciplines, I would suggest using much of the remaining lecture time for applications.
In my course, I do code vectors in Section 8. 1 , which students really seem to like, and
at least one application from each of Chapters 2-5. Other applications can be as­
signed as projects, along with as many of the explorations as desired. There is also
sufficient lecture time available to cover some of the theory in detail.
A Course with a Comoulalional Emphasis
For a course with a computational emphasis, the basic course outlined on the previous
page can be supplemented with the sections of the text dealing with numerical linear
algebra. In such a course, I would cover part or all of Sections 2.5, 3.4, 4.5, 5.3, 7.2, and
7.4, ending with the singular value decomposition. The explorations in Chapters 2
and 5 are particularly well suited to such a course, as are almost any of the applications.
A course tor Sludenls Who Have Already
SIUdied Some linear Algebra
Some courses will be aimed at students who have already encountered the basic prin­
ciples of linear algebra in other courses. For example, a college algebra course will
often include an introduction to systems of linear equations, matrices, and deter­
minants; a multivariable calculus course will almost certainly contain material on
vectors, lines, and planes. For students who have seen such topics already, much early
material can be omitted and replaced with a quick review. Depending on the back­
ground of the class, it may be possible to skim over the material in the basic course up
to Section 3.3 in about six lectures. If the class has a significant number of mathemat­
ics majors (and especially if this is the only linear algebra course they will take),
I would be sure to cover Sections 6. 1 - 6.5, 7. 1 , and 7.4 and as many applications as time
permits. If the course has science majors (but not mathematics majors), I would cover
Sections 6. 1 and 7. 1 and a broader selection of applications, being sure to include the
material on differential equations and approximation of functions. If computer sci­
ence students or engineers are prominently represented, I would try to do as much of
the material on codes and numerical linear algebra as I could.
There are many other types of courses that can successfully use this text. I hope
that you find it useful for your course and that you enjoy using it.
To the
Student
"Where shall I begin, please your
Majesty?" he asked.
"Begin at the beginning," the King
said, gravely, "and go on till you come
to the end: then stop."
-Lewis Carroll
1865
Alice's Adventures in
Wonderland,
Linear algebra is an exciting subject. It is full of interesting results, applications to
other disciplines, and connections to other areas of mathematics. The Student Solu­
tions Manual and Study Guide contains detailed advice on how best to use this book;
following are some general suggestions.
Linear algebra has several sides: There are computational techniques, concepts, and
applications. One of the goals of this book is to help you master all of these facets of
the subject and to see the interplay among them. Consequently, it is important that
you read and understand each section of the text before you attempt the exercises in
that section. If you read only examples that are related to exercises that have been
assigned as homework, you will miss much. Make sure you understand the defini­
tions of terms and the meaning of theorems. Don't worry if you have to read some­
thing more than once before you understand it. Have a pencil and calculator with you
as you read. Stop to work out examples for yourself or to fill in missing calculations.
The � icon in the margin indicates a place where you should pause and think over
what you have read so far.
Answers to most odd-numbered computational exercises are in the back of the
book. Resist the temptation to look up an answer before you have completed a ques­
tion. And remember that even if your answer differs from the one in the back, you
may still be right; there is more than one correct way to express some of the solutions.
For example, a value of l / v2 can also be expressed as v2/2 and the set of all scalar
multiples of the vector
[� ]
[� ] .
is the same as the set of all scalar multiples of
1 2
As you encounter new concepts, try to relate them to examples that you know.
Write out proofs and solutions to exercises in a logical, connected way, using com plete sentences. Read back what you have written to see whether it makes sense.
Better yet, if you can, have a friend in the class read what you have written. If it doesn't
make sense to another person, chances are that it doesn't make sense, period.
You will find that a calculator with matrix capabilities or a computer algebra sys­
tem is useful. These tools can help you to check your own hand calculations and are
indispensable for some problems involving tedious computations. Technology also
xx i i i
XXiV
To the Student
enables you to explore aspects of linear algebra on your own. You can play "what if?"
games: What if I change one of the entries in this vector? What if this matrix is of a
different size? Can I force the solution to be what I would like it to be by changing
something? To signal places in the text or exercises where the use of technology is
recommended, I have placed the icon cAs in the margin. The companion website that
accompanies this book contains computer code working out selected exercises from
the book using Maple, Mathematica, and MATLAB, as well as Technology Bytes, an
appendix providing much additional advice about the use of technology in linear
algebra.
You are about to embark on a journey through linear algebra. Think of this book
as your travel guide. Are you ready? Let's go!
Vectors
Here they come pouring out of the blue.
Little arrowsfor me andfor you.
-Albert Hammond and
Mike Hazelwood
1968
Little Arrows
Dutchess Music/BM!,
1.0
I n t ro d u ctio n : T h e R a cetrack G a m e
Many measurable quantities, such as length, area, volume, mass, and temperature,
can be completely described by specifying their magnitude. Other quantities, such
as velocity, force, and acceleration, require both a magnitude and a direction for
their description. These quantities are vectors. For example, wind velocity is a vector
consisting of wind speed and direction, such as 1 0 km/h southwest. Geometrically,
vectors are often represented as arrows or directed line segments.
Although the idea of a vector was introduced in the 1 9th century, its usefulness
in applications, particularly those in the physical sciences, was not realized until the
20th century. More recently, vectors have found applications in computer science,
statistics, economics, and the life and social sciences. We will consider some of these
many applications throughout this book.
This chapter introduces vectors and begins to consider some of their geometric
and algebraic properties. We begin, though, with a simple game that introduces some
of the key ideas. [You may even wish to play it with a friend during those (very rare!)
dull moments in linear algebra class.]
The game is played on graph paper. A track, with a starting line and a finish line,
is drawn on the paper. The track can be of any length and shape, so long as it is wide
enough to accommodate all of the players. For this example, we will have two players
(let's call them Ann and Bert) who use different colored pens to represent their cars
or bicycles or whatever they are going to race around the track. (Let's think of Ann
and Bert as cyclists.)
Ann and Bert each begin by drawing a dot on the starting line at a grid point on
the graph paper. They take turns moving to a new grid point, subject to the following
rules:
1. Each new grid point and the line segment connecting it to the previous grid point
must lie entirely within the track.
2. No two players may occupy the same grid point on the same turn. (This is the
"no collisions" rule.)
3. Each new move is related to the previous move as follows: If a player moves
a units horizontally and b units vertically on one move, then on the next move
he or she must move between a 1 and a + 1 units horizontally and between
-
2
Chapter
1
Vectors
b - 1 and b + 1 units vertically. In other words, if the second move is c units
horizontally and d units vertically, then l a - c l :::::: 1 and l b - d i :::::: 1 . (This is the
"acceleration/deceleration" rule.) Note that this rule forces the first move to be
1 unit vertically and/or 1 unit horizontally.
:r:
r§
(1805-1865)
A player who collides with another player or leaves the track is eliminated. The
winner is the first player to cross the finish line. If more than one player crosses
the finish line on the same turn, the one who goes farthest past the finish line is the
winner.
In the sample game shown in Figure 1 . 1 , Ann was the winner. Bert accelerated too
quickly and had difficulty negotiating the turn at the top of the track.
To understand rule 3, consider Ann's third and fourth moves. On her third move,
she went 1 unit horizontally and 3 units vertically. On her fourth move, her options
were to move 0 to 2 units horizontally and 2 to 4 units vertically. (Notice that some
of these combinations would have placed her outside the track.) She chose to move
2 units in each direction.
The Irish mathematician William
Rowan Hamilton
used vector concepts in his study
of complex numbers and their
generalization, the quaternions.
I
A
I
B
Figure 1 . 1
r
r
l
l
A sample game of racetrack
Problem 2
Problem 1
of moves?
Play a few games of racetrack.
Is it possible for Bert to win this race by choosing a different sequence
Problem 3 Use the notation [a, b] to denote a move that is a units horizontally
and b units vertically. (Either a or b or both may be negative.) If move [3, 4] has just
been made, draw on graph paper all the grid points that could possibly be reached
on the next move.
Problem 4 What is the net effect of two successive moves? In other words, if you
move [a, b] and then [c, d] , how far horizontally and vertically will you have moved
altogether?
Section
1.1
The Geometry and Algebra of Vectors
3
Problem 5 Write out Ann's sequence of moves using the [a, b] notation. Suppose
she begins at the origin (O, 0) on the coordinate axes. Explain how you can find the
coordinates of the grid point corresponding to each of her moves without looking at
the graph paper. If the axes were drawn differently, so that Ann's starting point was
not the origin but the point (2, 3), what would the coordinates of her final point be?
Although simple, this game introduces several ideas that will be useful in our
study of vectors. The next three sections consider vectors from geometric and alge­
braic viewpoints, beginning, as in the racetrack game, in the plane.
T h e G e o m etrv a n d A l g e b ra o f vectors
Vectors in the Plane
The Cartesian plane is named
after the French philosopher and
mathematician Rene Descartes
whose introduction
of coordinates allowed geometric
problems to be handled using
algebraic techniques.
(1596-1650),
The word vector comes from the
Latin root meaning "to carrY:' A
vector is formed when a point is
displaced-or "carried off" -a given
distance in a given direction. Viewed
another way, vectors "carry" two
pieces of information: their length
and their direction.
When writing vectors by hand,
it is difficult to indicate boldface.
Some people prefer to write v for
the vector denoted in print by v,
but in most cases it is fine to use an
ordinary lowercase v. It will usu­
ally be clear from the context when
the letter denotes a vector.
We begin by considering the Cartesian plane with the familiar x- and y-axes.
A vector is a directed line segment that corresponds to a displacement from one point
A to another point B; see Figure 1 .2.
The vector from A to B is denoted by AB ; the point A is called its initial point,
or tail, and the point B is called its terminal point, or head. Often, a vector is simply
denoted by a single boldface, lowercase letter such as v.
The set of all points in the plane corresponds to the set of all vectors whose tails
---->
are at the origin 0. To each point A, there corresponds the vector a = OA; to each
vector a with tail at 0, there corresponds its head A. (Vectors of this form are some­
times called position vectors.)
It is natural to represent such vectors using coordinates. For example, in
---->
Figure 1 .3, A = (3, 2) and we write the vector a = OA = [3, 2] using square brackets.
Similarly, the other vectors in Figure 1 .3 are
[ - 1 , 3 ] and c = [ 2, - 1 ]
The individual coordinates (3 and 2 in the case of a) are called the components of the
vector. A vector is sometimes said to be an ordered pair of real numbers. The order is
important since, for example, [3, 2] � [2, 3] . In general, two vectors are equal if and
only if their corresponding components are equal. Thus, [x, y] = [ l , 5] implies that
x = 1 and y = 5.
b =
[ �]
It is frequently convenient to use column vectors instead of (or in addition to)
row vectors. Another representation of [3, 2] is
. (The important point is that the
y
y
The word component is derived
from the Latin words co, meaning
"together with;' and ponere, mean­
ing "to put:' Thus, a vector is "put
together" out of its components.
A �B
Figure 1 . 2
Figure 1 . 3
4
Chapter
1
Vectors
IR2 is pronounced "r two:'
components are ordered.) In later chapters, you will see that column vectors are some­
what better from a computational point of view; for now, try to get used to both
representations.
It may occur to you that we cannot really draw the vector [O, OJ = 00 from the
origin to itself. Nevertheless, it is a perfectly good vector and has a special name: the
zero vector. The zero vector is denoted by 0.
The set of all vectors with two components is denoted by IR 2 (where IR denotes
the set of real numbers from which the components of vectors in IR 2 are chosen).
Thus, [ - 1, 3.5] , [ \/2, 7f ], and rn, 4 ] are all in IR 2 •
Thinking back to the racetrack game, let's try to connect all of these ideas to vec­
tors whose tails are not at the origin. The etymological origin of the word vector in
the verb "to carry" provides a clue. The vector [3, 2] may be interpreted as follows:
Starting at the origin 0, travel 3 units to the right, then 2 units up, finishing at P. The
same displacement may be applied with other init � oin� igure 1 .4 shows two
equivalent displacements, represented by the vectors AB and CD .
y
c
Figure 1 . 4
When vectors are referred to by
their coordinates, they are being
considered analytically.
We define two vectors as equal if they have the same length and the same direc­
tion. Thus, AB = CD in Figure 1 .4. (Even though they have different initial and ter­
minal points, they represent the same displacement.) Geometrically, two vectors are
equal if one can be obtained by sliding (or translating) the other parallel to itself until
the two vectors coincide. In terms of components, in Figure 1 .4 we have A = ( 3, 1)
and B = (6, 3). Notice that the vector [3, 2] that records the displacement is just the
difference of the respective components:
----->
AB
[ 3, 2 ] = [ 6
-
3, 3
-
l]
CD = r - 1 - ( -4), 1 - ( - 1) J
Similarly,
----->
=
[ 3, 2 ]
---->
and thus AB = CD , as expected.
A vector such as oP with its tail at the origin is said to be in standard position.
The foregoing discussion shows that every vector can be drawn as a vector in stan­
dard position. Conversely, a vector in standard position can be redrawn (by transla­
tion) so that its tail is at any point in the plane.
Exa m p l e 1 . 1
If A = ( - 1 , 2 ) and B = ( 3, 4), find AB and redraw it (a) in standard position and
(b) with its tail at the point C = (2, - 1).
----->
----->
We compute AB = [3 - ( - 1), 4 - 2] = [4, 2] . If AB is then translated
to CD, where C = ( 2, - 1 ), then we must have D = ( 2 + 4, - 1 + 2) = (6, 1). (See
Figure 1 .5. )
Solulion
1.1
Section
The Geometry and Algebra of Vectors
5
y
D(6, 1 )
Figure 1 . 5
New Veclors from Old
As in the racetrack game, we often want to "follow" one vector by another. This leads
to the notion of vector addition, the first basic vector operation.
If we follow u by v, we can visualize the total displacement as a third vector,
denoted by u + v. In Figure 1 .6, u = [ 1, 2] and v = [ 2, 2 ] , so the net effect of follow­
ing u by v is
[ 1 + 2, 2 + 2 ]
which gives u + v. In general, if u
is the vector
=
=
[3, 4]
[u 1 , u 2 ] and v = [ v 1 , v2 ] , then their s u m u + v
It is helpful to visualize u + v geometrically. The following rule is the geometric
version of the foregoing discussion.
y
/
Figure 1 . 6
Vector addition
�(/lj2
:
2
/_ )
U
_
I
I
2_ _
4
_ _ _ _ _ _ _ _ _
3
_n
6
Chapter
1
Vectors
The Head-to-Tail R u l e
Given vectors u and v in IR 2 , translate v s o that its tail coincides with the head
of u. The s u m u + v of u and v is the vector from the tail of u to the head of v.
(See Figure 1 .7. )
Figure 1 . 1
The head-to-tail rule
Figure 1 . 8
The parallelogram
determined by u and v
The Parallelogram Rule
By translating u and v parallel to themselves, we obtain a parallelogram, as
shown in Figure 1 .8. This parallelogram is called the parallelogram determined by u
and v. It leads to an equivalent version of the head-to-tail rule for vectors in standard
position.
Given vectors u and v in IR 2 (in standard position), their s u m u + v is the vector
in standard position along the diagonal of the parallelogram determined by u and
v. (See Figure 1 .9. )
y
Figure 1 . 9
The parallelogram rule
Exa m p l e 1 . 2
If u = [3, - 1 ] and v = [ l , 4], compute and draw u + v.
We compute u + v = [3 + 1, - 1 + 4] = [4, 3] . This vector is drawn
using the head-to-tail rule in Figure l . l O(a) and using the parallelogram rule in
Figure l . l O(b).
Solulion
Section
1.1
The Geometry and Algebra of Vectors
y
1
y
u
(b)
(a)
Figure 1 . 1 0
The second basic vector operation is scalar multiplication. Given a vector v and
a real number c, the scalar multiple c v is the vector obtained by multiplying each
component of v by c. For example, 3 [ - 2, 4] = [ - 6, 12] . In general,
Geometrically, cv is a "scaled" version of v.
Exa m p l e 1 . 3
If v = [ - 2, 4], compute and draw 2v, tv, and - 2v.
Solution
We calculate as follows:
2v
tv
- 2v
=
=
=
[ 2 ( - 2 ) , 2 ( 4 ) ] = [ - 4, 8 J
[t ( - 2 ) , t ( 4 ) ] = [ - 1 , 2 ]
[ - 2 ( - 2 ) , - 2 ( 4 ) ] = [ 4, - 8 ]
These vectors are shown in Figure 1 . 1 1 .
y
2v
- 2v
Figure 1 . 1 1
Chapter
8
1
Vectors
/./
2v
u
- 2v
Figure 1 . 1 3
u
+
( - v)
Vector subtraction
Figure 1 . 1 2
The term scalar comes from the
Latin word scala, meaning "lad­
dd' The equally spaced rungs on
a ladder suggest a scale, and in vec­
tor arithmetic, multiplication by a
constant changes only the scale (or
length) of a vector. Thus, constants
became known as scalars.
Observe that cv has the same direction as v if c > 0 and the opposite direction if
< 0. We also see that cv is le I times as long as v. For this reason, in the context of
vectors, constants (i.e., real numbers) are referred to as scalars. As Figure 1 . 1 2 shows,
when translation of vectors is taken into account, two vectors are scalar multiples of
each other if and only if they are parallel.
A special case of a scalar multiple is ( - l )v, which is written as -v and is called
the negative ofv. We can use it to define vector subtraction: The difference of u and
v is the vector u - v defined by
c
u - v = u + ( - v)
Figure 1 . 1 3 shows that u - v corresponds to the "other" diagonal of the parallelo­
gram determined by u and v.
Exa m p l e 1 . 4
y
A
Figure 1 . 1 4
�
If u = [ 1 , 2] and v = [ - 3, 1 ] , then u - v = [ 1
-
( - 3), 2 - 1 ]
=
[ 4, l ] .
The definition of subtraction in Example 1 .4 also agrees with the way we cal­
culate a vector such as AB .---If the
points A and B correspond to the vectors a and b
->
=
in standard position, then AB b - a, as shown in Figure 1 . 14. [Observe that the
head-to-tail rule applied to this diagram gives the equation a + (b - a ) = b. If we
had accidentally drawn b - a with its head at A instead of at B, the diagram would
have read b + (b - a) = a, which is clearly wrong! More will be said about algebraic
expressions involving vectors later in this section.]
vec1ors in � 3
Everything we have just done extends easily to three dimensions. The set of all or­
dered triples of real numbers is denoted by IR 3 . Points and vectors are located using
three mutually perpendicular coordinate axes that meet at the origin 0. A point such
as A = ( 1 , 2, 3) can be located as follows: First travel 1 unit along the x-axis, then
move 2 units parallel to the y-axis, and finally move 3 units parallel to the z-axis. The
corresponding vector a = [ l , 2, 3] is then OA, as shown in Figure 1 . 15.
Another way to visualize vector a in IR 3 is to construct a box whose six sides are de­
termined by the three coordinate planes (the xy-, xz-, and yz-planes) and by three planes
through the point ( 1 , 2, 3 ) parallel to the coordinate planes. The vector [ 1 , 2, 3] then corre­
sponds to the diagonal from the origin to the opposite corner of the box (see Figure 1 . 16).
Section
1.1
The Geometry and Algebra of Vectors
z
z
A ( l , 2,
3
•
I
I
I
I
3)
,.
:3
-
I
I
I
-- 1
2 -
x
Figure 1 . 1 5
9
�
y
x
Figure 1 . 1 6
The "componentwise" definitions of vector addition and scalar multiplication are
extended to IR 3 in an obvious way.
Vectors in � n
In general, we define !R n as the set of all ordered n-tuples of real numbers written as
row or column vectors. Thus, a vector v in !R n is of the form
�
Figure 1 . 1 1
u+v=v+u
The individual entries of v are its components; V; is called the ith component.
We extend the definitions of vector addition and scalar multiplication to !R n in
the obvious way: If u = [u 1 , u 2 , • • • , u n l and v = [ v 1 , v2 , . . . , vn ] , the ith component of
u + v is U ; + V; and the ith component of cv is just C V;.
Since in !R n we can no longer draw pictures of vectors, it is important to be able to
calculate with vectors. We must be careful not to assume that vector arithmetic will be
similar to the arithmetic of real numbers. Often it is, and the algebraic calculations we
do with vectors are similar to those we would do with scalars. But, in later sections,
we will encounter situations where vector algebra is quite unlike our previous experi­
ence with real numbers. So it is important to verify any algebraic properties before
attempting to use them.
One such property is commutativity of addition: u + v = v + u for vectors u and
v. This is certainly true in IR 2 • Geometrically, the head-to-tail rule shows that both
u + v and v + u are the main diagonals of the parallelogram determined by u and v.
(The parallelogram rule also reflects this symmetry; see Figure 1 . 1 7.)
Note that Figure 1 . 1 7 is simply an illustration of the property u + v = v + u. It
is not a proof, since it does not cover every possible case. For example, we must also
include the cases where u = v, u = - v, and u = 0. (What would diagrams for these
cases look like?) For this reason, an algebraic proof is needed. However, it is just as
easy to give a proof that is valid in !R n as to give one that is valid in IR 2 .
The following theorem summarizes the algebraic properties of vector addition
and scalar multiplication in !R n . The proofs follow from the corresponding properties
of real numbers.
Chapter
10
1
Vectors
Theorem 1 . 1
Algebraic Properties o f Vectors i n !R n
Let u, v, and w be vectors in !R n and let c and d be scalars. Then
a.
b.
c.
d.
e.
f.
g.
h.
u+v=v+u
(u + v) + w = u + (v + w)
u+0=u
u + ( -u) = 0
c (u + v) = cu + cv
(c + d)u = cu + du
c (du) = (cd)u
lu = u
Commutativity
Associativity
Distributivity
Distributivity
R e m a rks
The word theorem is derived from
the Greek word theorema, which
in turn comes from a word mean­
ing "to look af' Thus, a theorem
is based on the insights we have
when we look at examples and
extract from them properties that
we try to prove hold in general.
Similarly, when we understand
something in mathematics-the
proof of a theorem, for example­
we often say, "I see:'
•
Properties (c) and (d) together with the commutativity property (a) imply
that 0 + u = u and - u + u = 0 as well.
•
If we read the distributivity properties (e) and (f) from right to left, they say
that we can factor a common scalar or a common vector from a sum.
We prove properties (a) and (b) and leave the proofs of the remain­
ing properties as exercises. Let u = [u 1 , u 2 , . . . , u n ] , v = [ v 1 , v2 , . . . , vn ] , and w =
Proof
[ W1 , Wz , . . . , Wn ] .
(a)
U
+ = [U1, U z , . . . , U n ] + [V1, Vz , . . . , Vn ]
= [U1 + V1 , U z + Vz , . . . , U n + Vn ]
= [ v 1 + u 1 , v2 + u 2 , , v n + u n ]
= [v 1 , Vz , . . . , Vn ] + [u 1 , U z , . . . , U n ]
=v+u
V
•
.
.
The second and fourth equalities are by the definition of vector addition, and the
third equality is by the commutativity of addition of real numbers.
(b) Figure 1 . 1 8 illustrates associativity in IR 2 . Algebraically, we have
[ (u 1 + v 1 ) + w 1 , (u 2 + vz ) + w2 , . . . , (u n + vn ) + wn ]
[ u 1 + (v l + W1 ), U 2 + ( Vz + Wz ), . . . , U n + (vn + Wn ) J
[ u 1 , u 2 , . . . , u n ] + ( [ v 1 , v2 , . . . , vn ] + [ w 1 , w2 , . . . , w n ] )
= u + (v + w)
u
Figure 1 . 1 8
The fourth equality is by the associativity of addition of real numbers. Note the care­
ful use of parentheses.
Section
1.1
The Geometry and Algebra of Vectors
11
By property (b) of Theorem 1 . 1 , we may unambiguously write u + v + w without
parentheses, since we may group the summands in whichever way we please. By (a),
we may also rearrange the summands-for example, as w + u + v-if we choose.
Likewise, sums of four or more vectors can be calculated without regard to order or
grouping. In general, if v1 , v2 , , vk are vectors in !R n , we will write such sums with­
out parentheses:
•
•
.
The next example illustrates the use of Theorem 1 . 1 in performing algebraic
calculations with vectors.
Exa m p l e 1 . 5
Let a, b, and x denote vectors in !R n .
(a) Simplify 3a + (Sb - 2a) + 2(b - a).
(b) If Sx - a = 2(a + 2x), solve for x in terms of a.
We will give both solutions in detail, with reference to all of the properties
in Theorem 1 . 1 that we use. It is good practice to justify all steps the first few times
you do this type of calculation. Once you are comfortable with the vector properties,
though, it is acceptable to leave out some of the intermediate steps to save time and
space.
Solution
(a) We begin by inserting parentheses.
3a + (Sb - 2a) + 2(b - a) = (3a + (Sb - 2a)) + 2(b - a)
(3a + ( - 2a + Sb)) + (2b - 2a)
((3a + ( - 2a)) + Sb) + (2b - 2a)
((3 + ( - 2))a + Sb) + (2b - 2a)
( la + Sb) + (2b - 2a)
((a + Sb) + 2b) - 2a
(a + (Sb + 2b)) - 2a
(a + (S + 2)b) - 2a
(7b + a) - 2a
= 7b + (a - 2a)
= 7b + ( 1 - 2)a
= 7b + ( - l )a
= 7b - a
(a), (e)
(b)
(f)
(b ), (h)
(b)
(f)
(a)
(b)
(f),
(h)
You can see why we will agree to omit some of these steps! In practice, it is acceptable to
simplify this sequence of steps as
3a + (Sb - 2a) + 2(b - a) = 3a + Sb - 2a + 2b - 2a
(3a - 2a - 2a) + (Sb + 2b)
= - a + 7b
or even to do most of the calculation mentally.
12
Chapter
1
Vectors
(b) In detail, we have
Sx - a
Sx - a
Sx - a
Sx - a
( Sx - a ) - 4x
( - a + Sx) - 4x
- a + ( Sx - 4x)
- a + ( 5 - 4)x
- a + ( l )x
a + ( - a + x)
( a + ( - a )) + x
0+x
x
=
=
=
=
=
=
=
=
=
=
=
=
=
2 ( a + 2x)
2a + 2 ( 2x)
2a + ( 2 · 2 ) x
2a + 4x
( 2a + 4x) - 4x
2a + ( 4x - 4x)
2a + 0
2a
2a
a + 2a
( 1 + 2)a
3a
3a
(e)
(g)
(a), (b)
(b), (d)
(f), (c)
(h)
(b), (f)
(d)
(c)
Again, we will usually omit most of these steps.
Linear Combin alions and Coordin a1es
A vector that is a sum of scalar multiples of other vectors is said to be a linear combi­
nation of those vectors. The formal definition follows.
D e f i n i l i O D A vector v is a linear combination of vectors v1 , v2 , . . . , vk if
there are scalars c 1 , c2 , . . . , c k such that v = c 1 v1 + c2v2 +
+ ckvk . The scalars
c 1 , c2 , . . . , ck are called the coefficients of the linear combination.
·
·
·
Exa m p l e 1 . 6
Remark
Determining whether a given vector is a linear combination of other
vectors is a problem we will address in Chapter 2.
In IR 2 , it is possible to depict linear combinations of two (nonparallel) vectors
quite conveniently.
Exa m p l e 1 . 1
Let u =
[�]
and v =
way that e 1 =
[�]
[�] .
We can use u and v to locate a new set of axes (in the same
and e2 =
[�]
locate the standard coordinate axes) . We can use
Section
1.1
The Geometry and Algebra of Vectors
13
y
I
u
I
Figure 1 . 1 9
I-
I
I
these new axes to determine a coordinate grid that will let us easily locate linear
combinations of u and v.
As Figure 1 . 1 9 shows, w can be located by starting at the origin and traveling
- u followed by 2v. That is,
w = -u + 2v
We say that the coordinates of w with respect to u and v are - 1 and 2 . (Note that
this is just another way of thinking of the coefficients of the linear combination.)
It follows that
(Observe that - 1 and 3 are the coordinates of w with respect to e 1 and e2 . )
Switching from the standard coordinate axes to alternative ones is a useful idea. It
has applications in chemistry and geology, since molecular and crystalline structures
often do not fall onto a rectangular grid. It is an idea that we will encounter repeatedly
in this book.
Binarv vec1ors and Modular Arilhmelic
We will also encounter a type of vector that has no geometric interpretation-at least
not using Euclidean geometry. Computers represent data in terms of Os and ls (which
can be interpreted as off/on, closed/open, false/true, or no/yes) . Binary vectors are
vectors each of whose components is a 0 or a 1 . As we will see in Chapter 8, such
vectors arise naturally in the study of many types of codes.
In this setting, the usual rules of arithmetic must be modified, since the result of
each calculation involving scalars must be a 0 or a 1 . The modified rules for addition
and multiplication are given below.
The only curiosity here is the rule that 1 + 1 = 0. This is not as strange as it appears;
if we replace 0 with the word "even" and 1 with the word "odd," these tables simply
14
Chapter
1
Vectors
summarize the familiar parity rules fo r the addition and multiplication of even and
odd integers. For example, 1 + 1 = 0 expresses the fact that the sum of two odd inte­
gers is an even integer. With these rules, our set of scalars {O, l } is denoted by 22 and
is called the set of integers modulo 2.
Exa m p l e 1 . 8
We are using the term length dif­
ferently from the way we used it in
!FR". This should not be confusing,
since there is no geometric notion
of length for binary vectors.
Exa m p l e 1 . 9
Exa m p l e 1 . 1 0
In 22, 1 + 1 + 0 + 1 = 1 and 1 + 1 + 1 + 1 = 0. (These calculations illustrate
the parity rules again: The sum of three odds and an even is odd; the sum of four
odds is even.)
.+
With 22 as our set of scalars, we now extend the above rules to vectors. The set of
all n-tuples of Os and l s (with all arithmetic performed modulo 2) is denoted by 2� .
The vectors in 2� are called binary vectors of length n.
The vectors in 2� are [O, OJ , [O, l ] , [ l , OJ , and [ l , l J . (How many vectors does 2�
contain, in general?)
Let u = [l, 1, 0, 1, OJ and v = [O, 1, 1, 1, OJ be two binary vectors oflength 5. Find u +
Solulion
v.
The calculation of u + v takes place over 22 , so we have
u
+
v =
=
=
[ 1 , i , o, i , o ] + [o, I , I , i , o ]
[ 1 + o, I + I , o + I , I + I , o + o ]
[ 1 , 0, 1 , 0, 0 ]
It is possible to generalize what we have just done for binary vectors to vectors whose
components are taken from a finite set {O, 1, 2, . . . , k} for k 2: 2. To do so, we must
first extend the idea of binary arithmetic.
Exa m p l e 1 . 1 1
The integers modulo 3 is the set 2 3
by the following tables:
=
{O, 1, 2} with addition and multiplication given
2
0 1 2
0 0 0 0
0 0 1 2
1 2 0
2
1 0
2 0 2 1
2 2 0
Observe that the result of each addition and multiplication belongs to the set
{O, 1, 2}; we say that 2 3 is closed with respect to the operations of addition and multi­
plication. It is perhaps easiest to think of this set in terms of a 3-hour clock with 0, 1 ,
and 2 o n its face, as shown i n Figure 1 .20.
The calculation 1 + 2 = 0 translates as follows: 2 hours after 1 o'clock, it is
0 o'clock. Just as 24:00 and 1 2:00 are the same on a 1 2-hour clock, so 3 and 0 are
equivalent on this 3-hour clock. Likewise, all multiples of 3-positive and negative­
are equivalent to 0 here; 1 is equivalent to any number that is 1 more than a multiple
of 3 (such as - 2, 4, and 7); and 2 is equivalent to any number that is 2 more than a
+ 0
Section
1.1
The Geometry and Algebra of Vectors
15
multiple of 3 (such as - 1 , 5, and 8). We can visualize the number line as wrapping
around a circle, as shown in Figure 1 .2 1 .
0
. . . , - 3 , 0, 3, . . .
2
. . . , 1 , 2, 5 , . . .
Figure 1 . 2 0
Arithmetic modulo
Exa m p l e 1 . 1 2
3
. . . , - 2, 1 , 4, . . .
Figure 1 . 2 1
To what is 3548 equivalent in Z/
Solution This is the same as asking where 3548 lies on our 3-hour clock. The key is
to calculate how far this number is from the nearest (smaller) multiple of 3; that is,
we need to know the remainder when 3548 is divided by 3. By long division, we find that
3548 = 3 · 1 1 82 + 2, so the remainder is 2. Therefore, 3548 is equivalent to 2 in l'.. 3 •
4
In courses in abstract algebra and number theory, which explore this concept in
greater detail, the above equivalence is often written as 3548 = 2 (mod 3) or 3548 = 2
(mod 3), where = is read "is congruent to." We will not use this notation or termi­
nology here.
Exa m p l e 1 . 1 3
In l'.. 3 , calculate 2 + 2 + 1 + 2.
We use the same ideas as in Example 1 . 12. The ordinary sum is 2 + 2 +
1 + 2 = 7, which is 1 more than 6, so division by 3 leaves a remainder of 1 . Thus, 2 +
2 + 1 + 2 = 1 in l'.. 3 .
Solution 1
Solution 2
A better way to perform this calculation is to do it step by step entirely in l'.. 3 •
2+2+ 1 +2=
=
=
=
=
(2 + 2) + 1 + 2
1 + 1 +2
(1 + 1) + 2
2+2
1
Here we have used parentheses to group the terms we have chosen to combine. We could
speed things up by simultaneously combining the first two and the last two terms:
(2 + 2) + ( 1 + 2) = 1 + 0
= 1
Chapter
16
1
Vectors
Repeated multiplication can be handled similarly. The idea is to use the addition and
multiplication tables to reduce the result of each calculation to 0, 1 , or 2.
4
Extending these ideas to vectors is straightforward.
m
-
m
2
-
1
Exa m p l e 1 . 1 4
In Z�, let u = [2, 2, 0, 1 , 2] and v = [ l , 2, 2, 2, l ] . Then
0
+ v = [ 2, 2, 0, 1, 2 ] + [ 1, 2, 2, 2, l ]
[ 2 + 1, 2 + 2, 0 + 2, 1 + 2, 2 + l ]
[ 0, 1 , 2, 0, 0 ]
Vectors in Z� are referred to as ternary vectors of length 5.
_.,---._
u
2
3
In general, we have the set Z m = {O, 1, 2, . . . , m - l } of integers modulo m (cor­
responding to an m-hour clock, as shown in Figure 1 .22). A vector of length n whose
entries are in Z m is called an m-ary vector of length n. The set of all m-ary vectors of
length n is denoted by z::i.
Figure 1.22
Arithmetic modulo m
..
I
Exercises 1 . 1
1 . Draw the following vectors in standard position
in IR 2 :
(a) a =
[]
3
0
(b) b =
(d) d =
[]
[]
2
3
3
-2
2 . Draw the vectors in Exercise 1 with their tails at the
point (2, - 3).
3. Draw the following vectors in standard position in IR 3 :
(a) a = [O, 2, OJ
(b) b = [ 3 , 2, l ]
(c) c = [ l , - 2, l ]
(d) d = [ - 1, - 1 , - 2]
4. If the vectors in Exercise 3 are translated so that their
heads are at the point (3, 2, 1), find the points that
correspond to their tails.
5. For each of the following pairs of points, draw the
---->
---->
vector AB. Then compute and redraw AB as a vector
in standard position.
(a) A = ( 1, - 1 ), B = (4, 2)
(b) A = (O, - 2), B = ( 2, - 1 )
(c) A = (2, f), B = (t, 3 )
(d) A = (t, t), B = (i, t)
6. A hiker walks 4 km north and then 5 km northeast.
Draw displacement vectors representing the hiker's
trip and draw a vector that represents the hiker's net
displacement from the starting point.
Exercises 7- 1 0 refer to the vectors in Exercise 1. Compute
the indicated vectors and also show how the results can be
obtained geometrically.
8. b - c
7. a + b
10. a + d
9. d - c
Exercises 1 1 and 12 refer to the vectors in Exercise 3.
Compute the indicated vectors.
1 1 . 2a + 3c
12. 3b - 2c + d
13. Find the components of the vectors u, v, u + v, and
u - v, where u and v are as shown in Figure 1 .23.
14. In Figure 1 . 2 4 , A, B, C, D, E, and F are the vertices of a
regular hexagon centered at the origin.
Express each of the following vectors in terms of
---->
---->
a = OA and b = OB :
(a) Ai
(c) Ai5
(b) BC
(d) a
(e) xc
(f) BC
+ ill + PX
Section
y
21. u =
22. u =
1.1
[- �],
[ -�],
The Geometry and Algebra of Vectors
v=
v=
11
[ � ] , [�]
[ �] , [ � ]
w=
w=
23. Draw diagrams to illustrate properties (d) and (e) of
Theorem 1 . 1 .
24. Give algebraic proofs of properties ( d) through (g) of
Theorem 1 . 1 .
In Exercises 25-28, u and v are binary vectors. Find u + v
in each case.
25. u =
Figure 1 . 2 3
y
c
B
E
F
Figure 1 . 2 4
[ � ] ,v [ � ]
=
27. u = [ 1 , 0, 1 , 1 ] , v = [ 1, 1 , 1, 1 ]
28. u = [ 1 , 1, 0, 1, 0 ] , v = [ O, 1, 1 , 1, 0 ]
29. Write out the addition and multiplication tables for Z 4 .
30. Write out the addition and multiplication tables for Zs.
In Exercises 3 1 -43, perform the indicated calculations.
31. 2 + 2 + 2 in Z3
32. 2 · 2 · 2 in Z3
(
)
33. 2 2 + 1 + 2 in Z3 34. 3 + 1 + 2 + 3 in Z4
35. 2 · 3 · 2 in Z4
36. 3 ( 3 + 3 + 2 ) in Z4
37. 2 + 1 + 2 + 2 + 1 in Z3 , Z4 , and Zs
38. ( 3 + 4 ) ( 3 + 2 + 4 + 2 ) in Zs
39. 8 ( 6 + 4 + 3 ) in Z9 40. 2 100 in Z 1 1
41. [ 2 , 1 , 2 ] + [ 2, 0, 1 ] in Z� 42. 2 [ 2, 2, 1 ] in Z�
43. 2 ( [ 3, l , 1 , 2 ] + [ 3, 3, 2, l ] ) in Z! and Z�
In Exercises 1 5 and 1 6, simplify the given vector expression.
Indicate which properties in Theorem 1 . 1 you use.
15. 2 (a - 3b) + 3 (2b + a)
16. - 3(a - c) + 2(a + 2b) + 3(c - b)
In Exercises 1 7 and 1 8, solve for the vector x in terms of the
vectors a and b.
17. x - a = 2(x - 2a)
18. x + 2a - b = 3 (x + a) - 2 (2a - b)
In Exercises 19 and 20, draw the coordinate axes relative to
u and v and locate w.
[ � ] [ �l w
[ -�l [ �l
19. u = _ , v =
20. u =
v= _
= 2 u + 3v
w = - u - 2v
In Exercises 21 and 22, draw the standard coordinate axes
on the same diagram as the axes relative to u and v. Use
these to find w as a linear combination of u and v.
In Exercises 44-55, solve the given equation or indicate that
there is no solution.
45. x + 5 = 1 in Z6
44. x + 3 = 2 in Zs
46. 2x = 1 in Z3
47. 2x = 1 in Z4
48. 2x = 1 in Zs
49. 3x = 4 in Zs
51. 6x = 5 in Zs
50. 3x = 4 in Z6
52. Bx = 9 in Z 1 1
53. 2x + 3 = 2 in Z5
55. 6x + 3 = 1 in Zs
54. 4x + 5 = 2 in Z6
56. (a) For which values of a does x + a = 0 have a solu­
tion in Zs?
(b) For which values of a and b does x + a = b have a
solution in Z6?
(c) For which values of a, b, and m does x + a = b
have a solution in Z m ?
57. (a) For which values of a does ax = 1 have a solution
in Zs?
(b) For which values of a does ax = 1 have a solution
in Z6?
(c) For which values of a and m does ax = 1 have a
solution in z m ?
18
Chapter
1
Vectors
Length a n d A n g l e : T h e D o t P ro d u ct
It is quite easy to reformulate the familiar geometric concepts of length, distance,
and angle in terms of vectors. Doing so will allow us to use these important and
powerful ideas in settings more general than IR 2 and IR 3 • In subsequent chapters,
these simple geometric tools will be used to solve a wide variety of problems arising
in applications-even when there is no geometry apparent at all!
The Doi Producl
The vector versions of length, distance, and angle can all be described using the
notion of the dot product of two vectors.
Definilion
If
then the dot product u · v of u and v is defined by
In words, u · v is the sum of the products of the corresponding components of u
and v. It is important to note a couple of things about this "product" that we have just
defined: First, u and v must have the same number of components. Second, the dot
product u · v is a number, not another vector. (This is why u · v is sometimes called
the scalar product of u and v.) The dot product of vectors in !R n is a special and im­
portant case of the more general notion of inner product, which we will explore in
Chapter 7.
Exa m p l e 1 . 1 5
Solution
u · v = l · ( - 3) + 2 · 5 + ( - 3) · 2 = 1
Notice that if we had calculated v · u in Example 1 . 15, we would have computed
v · u = ( - 3) · 1 + 5 · 2 + 2 · ( - 3) = 1
That u · v = v · u in general is clear, since the individual products of the components
commute. This commutativity property is one of the properties of the dot product
that we will use repeatedly. The main properties of the dot product are summarized
in Theorem 1 .2.
Section
Theorem 1 . 2
1.2
Length and Angle: The Dot Product
19
Let u, v, and w be vectors in � n and let c be a scalar. Then
a.
b.
c.
d.
Commutativity
u·v = v·u
u . ( v + w) = u . v + u . w
Distributivity
( cu) . v = c ( u . v)
u · u 2: 0 and u · u = 0 if and only if u = 0
We prove (a) and (c) and leave proof of the remaining properties for the
exercises.
Proof
(a) Applying the definition of dot product to u · v and v · u, we obtain
u · v = U 1 V 1 + U z Vz + . . + U n V n
= V 1 U 1 + Vz U z + . . . + Vn U n
= v·u
·
where the middle equality follows from the fact that multiplication of real numbers
is commutative.
(c) Using the definitions of scalar multiplication and dot product, we have
(cu) · v = [ cu 1 , cu 2 , . . . , cu " ] • [ v 1 , v2 , , v" ]
+ CU n Vn
CU 1 V 1 + CU z V2 +
+ U n Vn )
c( U 1 V 1 + U z Vz +
c(u · v)
•
·
·
·
·
·
·
.
.
R e m a rlls
Property (b) can be read from right to left, in which case it says that we can
factor out a common vector u from a sum of dot products. This property also has
a "right-handed" analogue that follows from properties (b) and (a) together:
(v + w) · u = v · u + w · u.
•
Property (c) can be extended to give u · (cv) = c(u · v) (Exercise 58). This
extended version of ( c) essentially says that in taking a scalar multiple of a dot product
of vectors, the scalar can first be combined with whichever vector is more convenient.
For example,
•
(� [ - 1, -3, 2 ] ) · [ 6, - 4, 0 ] = [ - 1 , -3, 2 ] · (� [6, - 4, 0 ] ) = [ - 1, -3, 2 ] · [3, - 2, 0] = 3
With this approach we avoid introducing fractions into the vectors, as the original
grouping would have.
•
The second part of ( d) uses the logical connective if and only if. Appendix A dis­
cusses this phrase in more detail, but for the moment let us just note that the wording
signals a double i mplication- namely,
if u = 0, then u · u = 0
and
if u · u = 0, then u = 0
Theorem 1 .2 shows that aspects of the algebra of vectors resemble the algebra of
numbers. The next example shows that we can sometimes find vector analogues of
familiar identities.
Chapter
20
1
Vectors
Exa m p l e 1 . 1 6
Prove that (u + v) · (u + v) = u · u + 2(u · v) + v · v for all vectors u and v in !R n .
Solulion
(u + v) · (u + v) =
=
=
=
( u + v) · u + ( u + v) · v
u·u + v·u + u·v + v·v
u·u + u·v + u·v + v·v
u · u + 2(u · v) + v · v
(Identify the parts of Theorem 1 .2 that were used at each step.)
y
b
v =
[�]
Length
To see how the dot product plays a role in the calculation oflengths, recall how lengths
are computed in the plane. The Theorem of Pythagoras is all we need.
In IR 2 , the length of the vector v =
a
[�]
is the distance from the origin to the point
(a, b), which, by Pythagoras' Theorem, is given by Va 2 + b 2 , as in Figure 1 .25.
Observe that a 2 + b 2 = v · v. This leads to the following definition.
Figure 1 . 2 5
Definition
The length (or norm) of a vector v =
tive scalar ll v ll defined by
[ �: ]
·
in !R n is the nonnega-
vn
ll v ll = VV:V = Vvi + v� +
·
·
·
+ v�
In words, the length of a vector is the square root of the sum of the squares of its
components. Note that the square root of v · v is always defined, since v · v 2': 0 by
Theorem l .2 (d). Note also that the definition can be rewritten to give ll v ll 2 = v · v,
which will be useful in proving further properties of the dot product and lengths of
vectors.
vT3
11 [2, 3 J 11 = v2 2 + 3 2 =
Exa m p l e 1 . 1 1
Theorem 1 . 3 lists some of the main properties of vector length.
Theorem 1 . 3
Let v be a vector in IR " and let c be a scalar. Then
a. ll v ll = 0 if and only ifv = 0
b. ll cv ll = l c l ll v ll
Proof
Property (a) follows immediately from Theorem l .2(d). To show (b), we have
ll cv ll 2 = (cv) · (cv) = c 2 (v · v) = c 2 ll v ll 2
using Theorem l .2 (c). Taking square roots ofboth sides, using the fact that W = I c l
for any real number c, gives the result.
Section
1.2
Length and Angle: The Dot Product
21
A vector of length 1 is called a unit vector. In IR 2 , the set of all unit vectors can
be identified with the unit circle, the circle of radius 1 centered at the origin (see
Figure 1 .26). Given any nonzero vector v, we can always find a unit vector in the
same direction as v by dividing v by its own length (or, equivalently, multiplying by
1/ ll v ll ) . We can show this algebraically by using property (b) of Theorem 1.3 above:
If u = ( 1 / ll v ll )v, then
ll u ll
=
11 ( 1/ ll v ll )v ll
=
l l / ll v ll I ll v ll
=
( 1/ ll v ll ) ll v ll
=
1
and u is in the same direction as v, since 1 / I v I is a positive scalar. Finding a unit vec­
tor in the same direction is often referred to as normalizing a vector (see Figure 1 .27).
y
v
>:;:.
�
rr�1r
Figure 1 . 2 7
Figure 1 . 2 6
Normalizing a vector
Unit vectors in lffi 2
Exa m p l e 1 . 1 8
In IR 2 , let e1
/
=
[�]
and e2
=
[�] .
Then e1 and e2 are unit vectors, since the sum of the
squares of their components is 1 in each case. Similarly, in IR 3 , we can construct unit
vectors
Observe in Figure 1 .28 that these vectors serve to locate the positive coordinate axes
in IR 2 and IR 3 .
t
.--+
z
y
Figure 1 . 2 8
Standard unit vectors in lffi 2 and !ffi 3
x
y
Chapter
22
1
Vectors
In general, in !R n , we define unit vectors e1, e2 ,
, en , where e; has 1 in its ith
component and zeros elsewhere. These vectors arise repeatedly in linear algebra and
are called the standard unit vectors.
•
Exa m p l e 1 . 1 9
Nmmalfae the vectm v
�
[ -n
V2 2 + ( - 1 ) 2 + 3 2
tion as v is given by
Solulion
ll v ll
=
u
�
( 1/ ll v ll l v �
=
.
.
\/14, so a unit vector in the same direc­
{ :J [
( 1/ v'14 -
2/ \/14
- 1/ \/14
3/ \/14
]
Since property (b) of Theorem 1 .3 describes how length behaves with respect to
scalar multiplication, natural curiosity suggests that we ask whether length and vec­
tor addition are compatible. It would be nice if we had an identity such as II u + v II =
ll u 1 + ll v ll , but for almost any choice of vectors u and v this turns out to be false. [See
Exercise 52(a).] However, all is not lost, for it turns out that if we replace the = sign by
:s , the resulting inequality is true. The proof of this famous and important result-the
Triangle Inequality-relies on another important inequality-the Cauchy-Schwarz
Inequality-which we will prove and discuss in more detail in Chapter 7.
Theorem 1 . 4
The Cauchy-Schwarz Inequality
For all vectors u and v in !R n ,
l u · vl
-
u
Figure 1 . 2 9
The Triangle Inequality
Theorem 1 . 5
:s
ll n ll ll v ll
See Exercises 7 1 and 72 for algebraic and geometric approaches to the proof of this
inequality.
In IR 2 or IR 3 , where we can use geometry, it is clear from a diagram such as
Figure 1 .29 that ll u + v ii :s ll n ll + ll v ll for all vectors u and v. We now show that
this is true more generally.
The Triangle Inequality
For all vectors u and v in !R n ,
ll u + v ii
:s
ll u ll + ll v ll
Section 1 . 2 Length and Angle: The Dot Product
�
23
Proof
Since both sides of the inequality are nonnegative, showing that the square of
the left-hand side is less than or equal to the square of the right-hand side is equiva­
lent to proving the theorem. (Why?) We compute
ll u + v ll 2 = (u + v) - (u + v)
= u · u + 2(u · v) + v · v
By Example 1.9
2
2
l
l
ll
ll
ll
ll
::; u + 2 u · v + v
::; ll n ll 2 + 2 ll n ll ll v ll + ll v ll 2 By Cauchy-Schwarz
= ( ll n ll + ll v ll ) 2
as required.
Distance
The distance between two vectors is the direct analogue of the distance between two
points on the real number line or two points in the Cartesian plane. On the number
line (Figure 1 .30), the distance between the numbers a and b is given by l a - bl. (Tak­
ing the absolute value ensures that we do not need to know which of a or b is larger.)
This distance is also equal to V ( a - b ) 2 , and its two-dimensional generalization is
the familiar formula for the distance d between points (a 1 , a 2 ) and (b 1 , b 2 )-namely,
d = V ( a 1 - b 1 ) 2 + ( a 2 - bi ) 2 •
4
I
I
I
a
+
-2
I
0
I
Figure 1 . 3 0
d
= la
In terms of vectors, if a =
- bl
[ :: ]
= l-2
-
and b =
31 =
[ �:],
b
+
3
I
I
�
5
then d is just the length of a - b,
as shown in Figure 1 . 3 1 . This is the basis for the next definition.
I
I
I
I
: a 2 - b2
d
_ _ _ _ _ _ _ _ _ _ _
a1 - b1
I
I
I
I
_f]
Figure 1 . 3 1
d
= V( a ,
- b,)2
Definition
+ ( a2
- b )2
2
=
I l a - b ll
The distance d(u, v) between vectors u and v in u;g n is defined by
d ( U, v) = I u - v II
24
Chapter
1
Vectors
Exa m p l e 1 . 2 0
�
Find the di,tance between u
Solution
We rnmpute u - v
d ( u , v)
=
ll u
•nd v
�
·
�
v ii
-
[ 1] [ _ � l
[ �J '"
=
V( \/2) 2 + ( - 1 ) 2 + 1 2
= V4 =
2
Angles
The dot product can also be used to calculate the angle between a pair of vectors.
In IR 2 or IR 3 , the angle between the nonzero vectors u and v will refer to the angle (}
determined by these vectors that satisfies 0 :::::: (} :::::: 1 80° (see Figure 1 .3 2).
�
v
v4
v
u
(}
0
u
Figure 1 . 3 2
The angle between u and v
'D
. /7
u
u
In Figure 1 .33, consider the triangle with sides u, v, and u - v, where (} is the angle
between u and v. Applying the law of cosines to this triangle yields
v
ll u - v ll 2
ll u ll 2 + ll v ll 2 - 2 ll u ll ll v ll cos (}
Expanding the left-hand side and using ll v ll 2 = v · v several times, we obtain
Figure 1 . 3 3
=
ll u ll 2 - 2 ( u · v) + ll v ll 2 = ll u ll 2 + ll v ll 2 - 2 ll u ll ll v ll cos (}
which, after simplification, leaves us with u · v = ll u ll ll v ll cos (} . From this we obtain
the following formula for the cosine of the angle (} between nonzero vectors u and v.
We state it as a definition.
Definition
Exa m p l e 1 . 2 1
For nonzero vectors u and v in !R n ,
u·v
cos (} =
u
ll ll ll v ll
Compute the angle between the vectors u = [2, 1 , - 2] and v = [1, 1 , 1 ] .
Section
1.2
Length and Angle: The Dot Product
25
We calculate u·v = 2 · 1 + l · l + ( - 2 ) · 1 = 1 , JJ u JJ = V2 2 + 1 2 + (- 2 ) 2 =
V9 = 3, and JJ v JJ = Vl 2 + 1 2 + 1 2 = v3. Therefore, cos e = 1 / 3 v3, so
1
e = cos - ( 1 /3 v3) = 1 .377 radians, or 78.9°.
Solution
Exa m p l e 1 . 2 2
...._+
Compute the angle between the diagonals on two adjacent faces of a cube.
The dimensions of the cube do not matter, so we will work with a cube
with sides of length 1 . Orient the cube relative to the coordinate axes in IR 3 , as shown
in Figure 1 .34, and take the two side diagonals to be the vectors [ 1 , 0, 1 ] and [O, 1 , 1 ] .
Then angle e between these vectors satisfies
Solution
cos e =
l·O + O·l + l·l
2
\/2 v'2
from which it follows that the required angle is n /3 radians, or 60°.
-----
z
[O,
1, l]
y
x
Figure 1 . 3 4
(Actually, we don't need to do any calculations at all to get this answer. If we draw
a third side diagonal joining the vertices at ( 1 , 0, 1 ) and (O, 1 , 1 ) , we get an equilateral
triangle, since all of the side diagonals are of equal length. The angle we want is one of
the angles of this triangle and therefore measures 60°. Sometimes, a little insight can
save a lot of calculation; in this case, it gives a nice check on our work!)
R e m a rks
As this discussion shows, we usually will have to settle for an approximation
to the angle between two vectors. However, when the angle is one of the so-called
special angles (0°, 30°, 45°, 60°, 90°, or an integer multiple of these), we should be able
to recognize its cosine (Table 1 . 1 ) and thus give the corresponding angle exactly. In
all other cases, we will use a calculator or computer to approximate the desired angle
by means of the inverse cosine function.
•
e
Ta b l e 1 . 1
cos e
cosines or Special Angles
30°
V4
-= 1
2
v3
2
45°
v'2
2
60°
1
v'2
Vi
2
90°
2
Vo
-=O
2
26
Chapter
1
Vectors
The derivation of the formula for the cosine of the angle between two vectors
is valid only in IR 2 or IR 3 , since it depends on a geometric fact: the law of cosines. In
IR n , for n > 3, the formula can be taken as a definition instead. This makes sense, since
u·v
u·v
the Cauc�y-Schwarz ! � equality �mplies that
ranges from
:::::: 1 , so I
I
u
u
ll ll v ll
ll ll v ll
- 1 to 1, J USt as the cosme funct10n does.
•
I I
Orthogonal Veclors
The word orthogonal is derived
from the Greek words orthos, mean­
ing "upright;' and gonia, meaning
"angle:' Hence, orthogonal literally
means "right-angled'.' The Latin
equivalent is rectangular.
The concept of perpendicularity is fundamental to geometry. Anyone studying
geometry quickly realizes the importance and usefulness of right angles. We now gen­
eralize the idea of perpendicularity to vectors in !R n , where it is called orthogonality.
In IR 2 or IR 3 , two nonzero vectors u and v are perpendicular if the angle (J between
u·v
= cos 90° = 0,
them is a right angle-that is, if (J = 1T /2 radians, or 90°. Thus, I
ll u ll v ll
and it follows that u · v = 0. This motivates the following definition.
DefiniliOD
Since 0 · v
vector.
Exa m p l e 1 . 2 3
Two vectors u and v in !R n are orthogonal to each other if u ·v = 0.
=
0 for every vector v in !R n , the zero vector is orthogonal to every
In IR 3 , u = [ l , 1 , - 2 ] and v = [3, 1 , 2] are orthogonal, since u · v
3+1-4
=
0.
Using the notion of orthogonality, we get an easy proof of Pythagoras' Theorem,
valid in !R n .
Theorem 1 . 6
Pythagoras' 1heoreITI
For all vectors u and v in lR n , ll u + v ll 2
orthogonal.
=
ll u ll 2 + ll v ll 2 if and only if u and v are
From Example 1 . 16, we have ll u + v ll 2 = ll u ll 2 + 2 ( u · v) + ll v ll 2 for all
vectors u and v in !R n . It follows immediately that ll u + v ll 2 = ll u ll 2 + ll v ll 2 if and
only if u · v = 0. See Figure 1 .35.
Proof
Figure 1 . 3 5
The concept of orthogonality is one of the most important and useful in linear
algebra, and it often arises in surprising ways. Chapter 5 contains a detailed treatment
of the topic, but we will encounter it many times before then. One problem in which
it clearly plays a role is finding the distance from a point to a line, where "dropping a
perpendicular" is a familiar step.
Section
1.2
Length and Angle: The Dot Product
21
Proieclions
We now consider the problem of finding the distance from a point to a line in the
context of vectors. As you will see, this technique leads to an important concept: the
projection of a vector onto another vector.
As Figure 1 .36 shows, the problem of finding the distance from a point B to a
line € (in IR 2 or IR 3 ) reduces to the problem of finding the length of the perpendicular
line segment PB or, equivalently, the length of the vector PB. If we choose a point
A
"---->
on €, then, in the----"
right-angled
triangle
fl.APB,
the
other
two
vectors
are
the
leg
AP
and
----->
-----;
the hypotenuse AB. AP is called the projection of AB onto the line €. We will now look
at this situation in terms of vectors.
B
B
e
A
Figure 1 . 3 6
The distance from a point to a line
_..,.,.-figure 1 . 3 1
The projection of v onto u
u
vu= l Puv.l u, u l u )u u v,
=
= v
u. u·v
= u l v1 .
= l v l C1 uil · �l vl l ) (ilf)u
u·v )u
= (�
= (�)u
u·u
Consider two nonzero vectors and Let p be the vector obtained by dropping
a perpendicular from the head of onto and let e be the angle between and as
( I / I is the unit vector in
shown in Figure 1 .37. Then clearly p
where
I II cos e , and
the direction of Moreover, elementary trigonometry gives I p I
Thus, after substitution, we obtain
we know that cos e II II
p
This is the formula we want, and it is the basis of the following definition for vec­
tors in !R n .
v u
u v
u * 0,
u·v
= (-u·u) u
If and are vectors in
onto is the vector proju(v) defined by
Definition
proj0 ( v )
!R n
and
then the projection of
An alternative way to derive this formula is described in Exercise 73.
28
Chapter
1
Vectors
R e m a rks
The term projection comes from the idea of projecting an image onto a wall
(with a slide projector, for example) . Imagine a beam oflight with rays parallel to each
other and perpendicular to u shining down on v. The projection of v onto u is just the
shadow cast, or projected, by v onto u.
•
It may be helpful to think of proj0(v) as a function with variable v. Then the
variable v occurs only once on the right-hand side of the definition. Also, it is helpful
to remember Figure 1 .38, which reminds us that proj0(v) is a scalar multiple of the
vector u (not v) .
•
Although in our derivation of the definition of proj0(v) we required v as well
as u to be nonzero (why?), it is clear from the geometry that the projection of the
0
zero vector onto u is 0. The definition is in agreement with this, since "
u=
u·u
Ou = 0.
•
If the angle between u and v is obtuse, as in Figure 1 .38, then proj0(v) will be in
the opposite direction from u; that is, proj0(v) will be a negative scalar multiple of u.
•
If u is a unit vector then proj0(v) = ( u · v) u. (Why?)
•
proju(v)
u
Figure 1 . 3 8
(
Exa m p l e 1 . 2 4
.
Find the projection of v onto u in each case.
(a) v =
(c) v =
Solution
[ - �] [ �]
[ � ] [ ��� ]
[ �] [ - �]
and u =
and u =
1 / \/2
3
·
(a) We compute u · v =
proj u ( v) =
(b) Since e3 is a unit vector,
(
[ �] [ �]
[] [ ]
= 1 and u · u =
)
·
2 /5
u·v
1 2
=
=
"
u·u
1 /5
s 1
(c) We see that ll u ll = V� + � + t = 1 . Thus,
proj0 ( v) = ( u · v) u =
�
(�
3(I
= 5, so
+ 1 +
�) [ ��� ]
[ �l
: V2J
1 / \/2
=
3( 1
� \/2)
[ ��� ]
1 / \/2
)
J
Section
= [ 1 . 1 2, - 3.25, 2.07, - 1 .83 ] ,
v = [ - 2.29, 1 .72, 4.33, - 1 .54 ]
U
In Exercises 7-12, find II u II for the given exercise, and give
a unit vector in the direction of u.
GAs
7. Exercise 1
10. Exercise 4
8. Exercise 2
1 1 . Exercise 5
GAs
9. Exercise 3
12. Exercise 6
In Exercises 13-16, find the distance d(u, v) between u and
v in the given exercise.
14. Exercise 2
13. Exercise 1
GAs
15. Exercise 3
16. Exercise 4
1 7. If u, v, and w are vectors in ll�r, n 2: 2, and c is a
scalar, explain why the following expressions make
no sense:
(b) u · v + w
(a) ll u · v ll
(
)
(c) u · v · w
(d) c . ( u + w )
In Exercises 18-23, determine whether the angle between
u and v is acute, obtuse, or a right angle.
20. u =
GAS 2 1 . U =
22. u =
23. u =
[4, 3, - 1 ] , v = [ l , - 1 , l ]
[0.9, 2 . 1 , 1 .2] , V = [ - 4.5, 2.6, - 0.8]
[ 1 , 2, 3, 4], v = [ - 3, 1 , 2, - 2 ]
[ 1 , 2, 3, 4], v = [5, 6, 7, 8]
In Exercises 24-29, find the angle between u and v in the
given exercise.
24. Exercise 1 8
29
..
5. u = [ 1 , \/2, \/3, o ] , v = [ 4, - \/2, 0, - 5 ]
6.
Length and Angle: The Dot Product
Exercises 1 . 2
In Exercises 1-6, find u · v.
GAS
1.2
25. Exercise 1 9
GAs
27. Exercise 2 1
26. Exercise 20
GAs
GAs 28. Exercise 22
29. Exercise 23
30. Let A = ( - 3, 2), B = ( 1 , 0), and C = (4, 6). Prove that
MBC is a right-angled triangle.
31. Let A = ( 1 , 1, - 1 ), B = ( - 3, 2, - 2), and C = (2, 2, - 4).
Prove that �ABC is a right-angled triangle.
GAs 32. Find the angle between a diagonal of a cube and an ad­
jacent edge.
33. A cube has four diagonals. Show that no two of them
are perpendicular.
34. A parallelogram has diagonals determined by the
vectors
d,
�
[H
and d,
�
[ - �-
Show that the parallelogram is a rhombus (all sides of
equal length) and determine the side length.
35. The rectangle ABCD has vertices at A = ( 1 , 2, 3),
B = (3, 6, - 2), and C = (O, 5, - 4) . Determine the
coordinates of vertex D.
36. An airplane heading due east has a velocity of
200 miles per hour. A wind is blowing from the north
at 40 miles per hour. What is the resultant velocity of
the airplane?
37. A boat heads north across a river at a rate of 4 miles
per hour. If the current is flowing east at a rate of
3 miles per hour, find the resultant velocity of
the boat.
38. Ann is driving a motorboat across a river that is 2 km
wide. The boat has a speed of 20 km/h in still water, and
the current in the river is flowing at 5 km/h. Ann heads
out from one bank of the river for a dock directly across
from her on the opposite bank. She drives the boat in a
direction perpendicular to the current.
(a) How far downstream from the dock will
Ann land?
(b) How long will it take Ann to cross the river?
39. Bert can swim at a rate of 2 miles per hour in still
water. The current in a river is flowing at a rate of
1 mile per hour. If Bert wants to swim across the river
to a point directly opposite, at what angle to the bank
of the river must he swim?
30
Chapter
1
Vectors
In Exercises 40-45, find the projection of v onto u. Draw a
sketch in Exercises 40 and 41.
In Exercises 48 and 49, find all values of the scalar k for
which the two vectors are orthogonal.
48. u
�
[ : Jv [ � � : l
�
50. Describe all vectors v =
[
]
to u =
]
[
1 .34
3.01
CAS 45. U =
4.25
- 0.33 V =
- 1 .66
2.52
Figure 1 .39 suggests two ways in which vectors
may be used to compute the area of a triangle.
The area A of
,
51. Describe all vectors v =
to u =
[�] .
�
[ -J [ _fl
�
[;]
that are orthogonal
[;]
that are orthogonal
52. Under what conditions are the following true for
vectors u and v in IR 2 or IR 3 ?
(a) ll u + v ii = ll u ll + ll v ll (b) ll u + v ii = ll u ll - ll v ll
53. Prove Theorem 1 .2 (b).
54. Prove Theorem 1 .2 (d) .
In Exercises 55-57, prove the stated property of distance
between vectors.
55. d(u, v) = d(v, u) for all vectors u and v
56. d(u, w) :s d(u, v) + d(v, w) for all vectors u, v, and w
57. d(u, v) = 0 if and only if u = v
58. Prove that u · c v = c( u · v) for all vectors u and v in !R n
and all scalars c.
59. Prove that ll u - v ii 2': ll u ll - ll v ll for all vectors u and
v in !R n . [Hint: Replace u by u - v in the Triangle
(a)
(b)
[�] .
49. u
u
Figure 1 . 3 9
the triangle in part (a) is given by t ll u ll ll v - proj u ( v ) II ,
and part (b) suggests the trigonometric form of the
area of a triangle: A = t I u I I v II sin e (We can use the
identity sin e = v 1 - cos 2 e to find sin e.)
In Exercises 46 and 47, compute the area of the triangle
with the given vertices using both methods.
46. A = ( 1 , - 1 ), B = (2, 2), C = (4, O)
47. A = (3, - 1 , 4), B = (4, - 2 , 6), C = (5, 0, 2)
Inequality.]
60. Suppose we know that u · v = u · w. Does it follow that
v = w? If it does, give a proof that is valid in !R n ;
otherwise, give a counterexample (i.e., a specific set of
vectors u, v, and w for which u · v = u · w but v -=F w) .
61. Prove that (u + v) · (u - v) = ll u ll 2 - ll v ll 2 for all vec­
tors u and v in !R n .
62. (a) Prove that ll u + v ll 2 + ll u - v ll 2 = 2 ll u ll 2 + 2 ll v ll 2
for all vectors u and v in !R n .
(h) Draw a diagram showing u, v, u + v, and u - v
in IR 2 and use (a) to deduce a result about
parallelograms.
1
1
63. Prove that u · v = - ll u + v ll 2 - - ll u - v ll 2 for all
4
4
vectors u and v in !R n .
Section
64. (a) Prove that ll u + v ii = ll u - v ii if and only if u and
v are orthogonal.
(b) Draw a diagram showing u, v, u + v, and u - v
in IR 2 and use (a) to deduce a result about
parallelograms.
65. ( a) Prove that u + v and u - v are orthogonal in !R n if
and only if ll u ll = ll v ll .
(b) Draw a diagram showing u, v, u + v, and u - v
in IR 2 and use (a) to deduce a result about
parallelograms.
1.2
Length and Angle: The Dot Product
31
72. Figure 1 .40 shows that, in IR 2 or IR 3 ,
ll proju ( v ) I :::::: ll v ll ·
(a) Prove that this inequality is true in general. [Hint:
Prove that proju(v) is orthogonal to v - proju(v)
and use Pythagoras' Theorem.]
(b) Prove that the inequality ll proju ( v ) I :::::: ll v ll is
equivalent to the Cauchy-Schwarz Inequality.
66. If ll u ll = 2, ll v ll = v'3 , and u · v = 1, find ll u + v ii .
67. Show that there are no vectors u and v such that ll u ll = 1,
ll v ll = 2, and u · v = 3.
68. ( a) Prove that if u is orthogonal to both v and w, then
u is orthogonal to v + w.
(b) Prove that if u is orthogonal to both v and w, then
u is orthogonal to sv + tw for all scalars s and t.
69. Prove that u is orthogonal to v - proju(v) for all
vectors u and v in !R n , where u * 0.
70. ( a) Prove that proju(proju(v)) = proju(v) ·
(b) Prove that proju(v - proju(v)) = 0.
(c )
proju(v)
u
Figure 1 . 4 0
73. Use the fact that proju(v) = c u for some scalar c, to­
gether with Figure 1 . 4 1 , to find c and thereby derive
the formula for proju(v) .
Explain (a) and (b) geometrically.
71. The Cauchy-Schwarz Inequality I u · v i :::::: ll u ll ll v ll is
v
equivalent to the inequality we get by squaring both
sides: (u · v) 2 :::::: ll u ll 2 ll v ll 2 .
( a) In IR 2 , with u =
[ �: ]
and v =
[ :J
this becomes
Prove this algebraically. [Hint: Subtract the left-hand
side from the right-hand side and show that the
difference must necessarily be nonnegative.]
(b) Prove the analogue of (a) in IR 3 .
-
cu
Figure 1 . 4 1
cu
u
74. Using mathematical induction, prove the following
generalization of the Triangle Inequality:
ll v1 + Vz + · · · + vn ll ::=::: ll v1 ll + ll vz ll + · · · + ll vn ll
for all n 2: 1.
Exp loration
Ve ctors and G e o m etry
Many results i n plane Euclidean geometry can b e proved using vector techniques.
For example, in Example 1 .24, we used vectors to prove Pythagoras' Theorem. In
this exploration, we will use vectors to develop proofs for some other theorems from
Euclidean geometry.
As an introduction to the notation and the basic approach, consider the following
easy example.
Exa m p l e 1 . 2 5
Give a vector description of the midpoint M of a line segment AB.
We first convert everything to vector notation. If 0 denotes the origin and
P----is-> a point,
let---->
p be the vector GP. In this situation, a = GA, b = GB, m = oM, and
------>
AB = OB - OA = b - a (Figure 1.42).
Now, since M is the midpoint of AB, we have
Solution
A
B
m
so
Figure 1 . 4 2
The midpoint of AB
m
-a
=
= AM =
a + t (b - a)
c
.... B
----------
Figure 1 . 4 3
32
=
t (a + b)
1 . Give a vector description of the point P that is one-third of the way from A to
B on the line segment AB. Generalize.
A
fAB = t ( b - a )
2. Prove that the line segment joining the midpoints of two sides of a triangle is
parallel to the third side and half as long. (In vector notation, prove that PQ = t AB in
Figure 1 .43.)
3. Prove that the quadrilateral PQRS (Figure 1 .44), whose vertices are the mid­
points of the sides of an arbitrary quadrilateral ABCD, is a parallelogram.
4. A median of a triangle is a line segment from a vertex to the midpoint of
the opposite side (Figure 1 .45). Prove that the three medians of any triangle are con­
current (i.e., they have a common point of intersection) at a point G that is two­
thirds of the distance from each vertex to the midpoint of the opposite side. [ Hint: In
Figure 1 .46, show that the point that is two-thirds of the distance from A to P is given
by t ( a + b + c). Then show that t ( a + b + c ) is two-thirds of the distance from B
to Q and two-thirds of the distance from C to R.] The point G in Figure 1 .46 is called
the centroid of the triangle.
A
A
B
B
Figure 1 . 4 4
c
c
Figure 1 . 4 5
Figure 1 . 4 6
A median
The centroid
5. An altitude of a triangle is a line segment from a vertex that is perpendicu­
lar to the opposite side (Figure 1 .47) . Prove that the three altitudes of a triangle are
concurrent. [Hint: Let H be the point of intersection of the altitudes from A and B in
Figure 1 .48. Prove that cH is orthogonal to AB .] The point H in Figure 1 .48 is called
the orthocenter of the triangle.
6. A perpendicular bisector of a line segment is a line through the midpoint of
the segment, perpendicular to the segment (Figure 1 .49). Prove that the perpendicular
bisectors of the three sides of a triangle are concurrent. [Hint: Let K be the point of in­
tersection of t� erpendicular bisectors of AC and BC in Figure 1 .50. Prove that RK is
orthogonal to AB .] The point K in Figure 1 .50 is called the circumcenter of the triangle.
c
A
B
A --�-*��������--=e B
Figure 1 . 41
Figure 1 . 4 8
Figure 1 . 4 9
An altitude
The orthocenter
A perpendicular bisector
7. Let A and B be the endpoints of a diameter of a circle. If C is any point on the
circle, prove that LACE is a right angle. [Hint: In Figure 1 . 5 1 , let 0 be the center of the
circle. Express everything in terms of a and c and show that AC is orthogonal to BC. J
8. Prove that the line segments joining the midpoints of opposite sides of a
quadrilateral bisect each other (Figure 1 .52).
c
c
0
D
R
Figure 1 . 5 0
The circumcenter
Figure 1 . 5 1
Figure 1 . 5 2
33
34
Chapter
1
Vectors
lines a n d P l a n es
We are all familiar with the equation of a line in the Cartesian plane. We now want
to consider lines in IR 2 from a vector point of view. The insights we obtain from this
approach will allow us to generalize to lines in IR 3 and then to planes in IR 3 . Much of
the linear algebra we will consider in later chapters has its origins in the simple geom­
etry of lines and planes; the ability to visualize these and to think geometrically about
a problem will serve you well.
Lines in � 2 and � 3
In the xy-plane, the general form of the equation of a line is ax + by = c. If b * 0, then
the equation can be rewritten as y = - ( a/b )x + c/b, which has the form y = mx + k.
[This is the slope-intercept form; m is the slope of the line, and the point with coordi­
nates (O, k) is its y-intercept.J To get vectors into the picture, let's consider an example.
Exa m p l e 1 . 2 6
The line C with equation 2x + y = 0 is shown in Figure 1 .53. It is a line with slope - 2
passing through the origin. The left-hand side of the equation is in the form of a dot
product; in fact, if we let n
=
[ �]
and x
=
[; ]
, then the equation becomes n · x = 0.
The vector n is perpendicular to the line-that is, it is orthogonal to any vector x that
is parallel to the line (Figure 1 .54)-and it is called a normal vector to the line. The
equation n x = 0 is the normal form of the equation of e.
Another way to think about this line is to imagine a particle moving along the
line. Suppose the particle is initially at the origin at time t = 0 and it moves along
the line in such a way that its x-coordinate changes 1 unit per second. Then at t = 1
the particle is at ( 1 , - 2 ), at t = 1 .5 it is at ( 1 .5, - 3 ), and, if we allow negative values of t
(i.e., we consider where the particle was in the past), at t = - 2 it is (or was) at ( - 2, 4).
.
y
y
The Latin word norma refers to a
carpenter's square, used for draw­
ing right angles. Thus, a normal
vector is one that is perpendicular
to something else, usually a plane.
2x + 0
Figure 1 . 5 3
The line
y =
Figure 1 . 5 4
A normal vector n
Section
1. 3
Lines and Planes
35
This movement is illustrated in Figure 1 .55. In general, if x = t, then y = -2t, and we
may write this relationship in vector form as
What is the significance of the vector d
=
[ �] ?
_
It is a particular vector parallel
to e, called a direction vector for the line. As shown in Figure 1 .56, we may write the
equation of e as x = td. This is the vector form of the equation of the line.
If the line does not pass through the origin, then we must modify things
slightly.
y
y
e
Figure 1 . 5 6
A direction vector d
Figure 1 . 5 5
Exa m p l e 1 . 2 1
Consider the line C with equation 2x + y = 5 (Figure 1 .57). This is just the line from
Example 1 .26 shifted upward 5 units. It also has slope - 2, but its y-intercept is the
point (O, 5). It is clear that the vectors d and n from Example 1 .26 are, respectively, a
direction vector and a normal vector for this line too.
Thus, n is orthogonal to every vector that is parallel to e. The point P = ( 1, 3)
---+
is on C. If X = ( x, y) represents a general point on C, then the vector PX = x p is
parallel to e and n · (x p ) = 0 (see Figure 1 .58 ) . Simplified, we have n · x = n · p .
As a check, we compute
-
-
n·x
=
[ �] [;]
·
=
2x + y and
n
·p
=
[ �] [ �]
·
=
5
Thus, the normal form n · x = n · p is just a different representation of the general
form of the equation of the line. (Note that in Example 1 .26, p was the zero vector, so
n · p = 0 gave the right-hand side of the equation.)
36
Chapter
1
Vectors
y
y
n
-t--j--+--t--t--f--\-t-jl-t-... x
2x + 5
figure 1 . 5 1
The line
y =
(x
- p)
figure 1 . 5 8
n •
=
0
These results lead to the following definition.
Definition
The normal form of the equation of a line e in IR 2 is
n · ( x - p ) = 0 or n · x = n · p
where p is a specific point on e and n =F 0 is a normal vector for e.
The general form of the equation of e is ax +
= where n =
by
normal vector for e.
C,
[ ab]
i· s a
Continuing with Example 1 .27, let us now find the vector form of the equation
of e. Note that, for each choice of X, x - p must be parallel to-and thus a multiple
of-the direction vector d. That is, x - p = td or x = p + td for some scalar t. In
terms of components, we have
or
The word parameter and the cor­
responding adjective parametric
come from the Greek words para,
meaning "alongside;' and metron,
meaning "measure:' Mathemati­
cally speaking, a parameter is a
variable in terms of which other
variables are expressed-a new
"measure" placed alongside
old ones.
[;] [ �] [ - � ]
y +t
(1)
x= 1 + t
= 3
2t
(2)
Equation ( 1) is the vector form of the equation of €, and the componentwise Equa­
tions ( 2) are called parametric equations of the line. The variable t is called a parameter.
How does all of this generalize to IR 3 ? Observe that the vector and parametric
forms of the equations of a line carry over perfectly. The notion of the slope of a line
in IR 2 -which is difficult to generalize to three dimensions-is replaced by the more
convenient notion of a direction vector, leading to the following definition.
The vector form of the equation of a line e in IR 2 or IR 3 is
x = p + td
where p is a specific point on e and d =F 0 is a direction vector for e.
The equations corresponding to the components of the vector form of the
equation are called parametric equations of e.
Definition
Section
1.3
Lines and Planes
31
We will often abbreviate this terminology slightly, referring simply to the general,
normal, vector, and parametric equations of a line or plane.
Exa m p l e 1 . 2 8
[ -:J
Find vector and parametric equations of the line in IR 3 through the point P = ( 1, 2, - 1 ),
pornllel to the mtm d
Solution
�
The vector equation x =
The parametric form is
p
+ td is
x = 1 + St
y= 2 - t
z = - 1 + 3t
R e m a rks
The vector and parametric forms of the equation of a given line e are not
unique-in fact, there are infinitely many, since we may use any point on e to de­
termine p and any direction vector for e. However, all direction vectors are clearly
multiples of each other.
1o
In Example 1 .28, (6, 1 , 2) is another point on the line (take t = 1), and - 2 is
another direction vector. Therefore,
6
•
[ l
gives a different (but equivalent) vector equation for the line. The relationship between
the two parameters s and t can be found by comparing the parametric equations: For
a given point ( x, y, z) on e, we have
x = 1 + St = 6 + 10s
y = 2 - t = 1 - 2s
z = - 1 + 3t = 2 + 6s
implying that
- 10s + St = S
2s - t = -1
- 6s + 3t = 3
Each of these equations reduces to t = 1 + 2s.
38
Chapter
1
Vectors
3
Intuitively, we know that a line is a one-dimensional object. The idea of
"dimension" will be clarified in Chapters and 6, but for the moment observe that
this idea appears to agree with the fact that the vector form of the equation of a line
requires one parameter.
•
Exa m p l e 1 . 2 9
One often hears the expression "two points determine a line:' Find a vector equation
of the line f in IR 3 determined by the points P = ( - 1 , 5, o) and Q = ( 2, 1 , 1 ) .
Solulion
fine) .
3[ ]
We may choose any point on f for p, so we will use P ( Q would also be
A convenient direction vector is d = PQ =
Thus, we obtain
-4
1
(or any scalar multiple of this).
Planes i n lR 3
n is orthogonal to infinitely many
vectors
Figure 1 . 5 9
n
p
n · (x
- p)
Figure 1 . 6 0
=
0
The next question we should ask ourselves is, How does the general form of the equa­
tion of a line generalize to IR 3 ? We might reasonably guess that if ax + by = c is the
general form of the equation of a line in IR 2 , then ax + by + cz = d might represent a
line in IR 3 . In normal form, this equation would be n · x = n · p, where n is a normal
vector to the line and p corresponds to a point on the line.
To see if this is a reasonable hypothesis, let's think about the special case of the
equation ax + by +
cz �
0. In normal furn, it becomes n • x � 0, wh"e n
�
[�l
·
However, the set of all vectors x that satisfy this equation is the set of all vectors or­
thogonal to n. As shown in Figure 1 .59, vectors in infinitely many directions have
this property, determining a family of parallel planes. So our guess was incorrect: It
appears that ax + by + cz = d is the equation of a plane-not a line-in IR 3 .
Let's make this finding more precise. Every plane <!]' in IR 3 can be determined by
specifying a point p on <!I' and a nonzero vector n normal to <!I' (Figure 1 .60) . Thus,
if x represents an arbitrary point on <!I', we have n · (x - p) = 0 or n · x = n · p. If
n
�
[�]
and x
�
[� l
then, in teems of rnmponents, the equation bernmes
ax + by + cz = d (where d = n · p ) .
The normal form of the equation of a plane <!I' in IR 3 is
n · (x - p) = 0 or n · x = n · p
where p is a specific point on <!I' and n * 0 is a normal vector for <!I' .
Definition
The general form of the equation of <!I' i s ax + by +
is a normal vector for <!I' .
cz
= d , where n =
Section
1.3
Lines and Planes
39
Note that any scalar multiple of a normal vector for a plane is another normal
vector.
Exa m p l e 1 . 3 0
Find the normal and general forms of the equation of the plane that contains the
point P � (6, O, 1 ) and har normal vodor n �
- With p �
[fl
and x �
[� ].
[H
we havr n · p � 1 · 6 + 2 ' 0 � 3 · 1 � 9, 'o
the normal equation n · x = n · p becomes the general equation x + 2y + 3z = 9 .
.+
Geometrically, it is clear that parallel planes have the same normal vector(s) .
Thus, their general equations have left-hand sides that are multiples of each other. So, for
example, 2x + 4y + 6z = 10 is the general equation of a plane that is parallel to the
plane in Example 1 .30, since we may rewrite the equation as x + 2y + 3z = 5-from
which we see that the two planes have the same normal vector n. (Note that the planes
do not coincide, since the right-hand sides of their equations are distinct.)
We may also express the equation of a plane in vector or parametric form. To do
so, we observe that a plane can also be determined by specifying one of its points
P (by the vector p ) and two direction vectors u and v parallel to the plane (but not
parallel to each other). As Figure 1 .6 1 shows, given any point X in the plane (located
,<= I
x
tv
-
p = su
+ tv
-x
SU
figure 1 . 6 1
X
-
p
= SU + t v
by x), we can always find appropriate multiples s u and tv of the direction vectors such
that x - p = su + tv or x = p + s u + tv. If we write this equation componentwise,
we obtain parametric equations for the plane.
The vector form of the equation of a plane <lP in IR 3 is
X = p + SU + tv
where p i s a point o n <lP and u and v are direction vectors fo r <lP ( u and v are non zero and parallel to <lP , but not parallel to each other) .
The equations corresponding to the components of the vector form of the
equation are called parametric equations of <lP .
Definition
Chapter
40
1
Vectors
Exa m p l e 1 . 3 1
Find vector and parametric equations for the plane in Example 1 .30.
We need to find two direction vectors. We have one point P = (6, 0, 1 ) in
the plane; if we can find two other points Q and R in <!J', then the vectors PQ and PR
can serve as direction vectors (unless by bad luck they happen to be parallel!). By
trial and error, we observe that Q = (9, 0, O) and R = (3, 3, O) both satisfy the general
equation x + 2y + 3z = 9 and so lie in the plane. Then we compute
Solulion
Figure 1 . 6 2
\
which, since they are not scalar multiples of each other, will serve as direction vectors.
Therefore, we have the vector equation of <!J',
Two normals determine a line
and the corresponding parametric equations,
<;JP I
Figure 1 . 6 3
The intersection of
two planes is a line
x = 6 + 3s - 3t
�
y = 3t
z= l - s- t
[What would have happened had we chosen R = (O, 0, 3)?]
R e m a rks
A plane is a two-dimensional object, and its equation, in vector or parametric
form, requires two parameters.
•
As Figure 1 .59 shows, given a point P and a nonzero vector n in IR 3 , there are
infinitely many lines through P with n as a normal vector. However, P and two non­
parallel normal vectors n 1 and n2 do serve to locate a line e uniquely, since e must
then be the line through P that is perpendicular to the plane with equation x = p +
sn 1 + tn2 (Figure 1 .62) . Thus, a line in IR 3 can also be specified by a pair of equations
•
a 1 x + b1y + c 1 z = d 1
a 1X + b 2y + C 2 Z = d1
one corresponding to each normal vector. But since these equations correspond to a
pair of nonparallel planes (why nonparallel?), this is just the description of a line as
the intersection of two nonparallel planes (Figure 1 .63). Algebraically, the line con­
sists of all points (x, y, z) that simultaneously satisfy both equations. We will explore
this concept further in Chapter 2 when we discuss the solution of systems of linear
equations.
Tables 1 .2 and 1 . 3 summarize the information presented so far about the equa­
tions of lines and planes.
Observe once again that a single (general) equation describes a line in IR 2 but
a plane in IR 3 . [In higher dimensions, an object (line, plane, etc.) determined by a
single equation of this type is usually called a hyperplane.] The relationship among
Section
Ta b l e 1 . 2
Normal Form
Lines and Planes
41
Equations or lines in !R 2
Vector Form
General Form
n·x=n·p
Ta b l e 1 . 3
1. 3
ax + by =
x=p
c
+ td
{ yx == P 1 ++ tdtd1
Parametric Form
P2
2
lines and Planes in !R 3
Lines
{n1 · x = n1 · p1
Planes
n·x = n·p
Normal Form
n2 • x = n2 • P2
{a 1X + b 1y + C1Z = d1
Vector Form
General Form
a zX + b1Y + C2 Z = dz
ax + by + cz = d
x=p
+ td
X = p + SU + tv
Parametric Form
{ ;:
{;:
P1 + td 1
P2 + td2
z = p 3 + td3
p 1 + su 1 + tv 1
p2 + su 2 + tv2
z = p 3 + S U 3 + tv3
the dimension of the object, the number of equations required, and the dimension of
the space is given by the "balancing formula'':
(dimension of the object) + (number of general equations) = dimension of the space
The higher the dimension of the object, the fewer equations it needs. For
example, a plane in IR 3 is two-dimensional, requires one general equation, and lives
in a three-dimensional space: 2 + 1 = 3. A line in IR 3 is one-dimensional and so
needs 3 1 = 2 equations. Note that the dimension of the object also agrees with
the number of parameters in its vector or parametric form. Notions of "dimension"
will be clarified in Chapters 3 and 6, but for the time being, these intuitive observa­
tions will serve us well.
We can now find the distance from a point to a line or a plane by combining the
results of Section 1 .2 with the results from this section.
-
Exa m p l e 1 . 3 2
Find the distance from the point B
A � (3, ! , I ) with dimhon vedoc d
�
=
( 1 , 0, 2) to the line € through the point
[ -u
--+
As we have already determined, we need to calculate the length of ---->
PB,
where P is the point on e at the foot of the perpendicular from B. If we label v = AB,
then AP = projd(v) and PB = v projd ( v) (see Figure 1 .64) . We do the necessary
calculations in several steps.
Solution
-
42
Chapter
1
Vectors
B
v - profa(v)
___...--- e
d
A
= llv - projd ( v ) II
Figure 1 . 6 4
d ( B, € )
Step 2: The projection of v onto d is
projd ( v ) =
=
(:��)
d
(( - 1 ) · ( (--21) ) 1·1 ( - 10) 0· 1 ) [ - �
ol
+
2+
+
+
Step 3: The vector we want is
Step 4:
The distance d(B, f) from B to e is
Using Theorem
1.3(b)
ll v - prnj , ( v ) I
�
[ =! ]
[ =� -
to simplify the calculation, we have
ll v - prnj , ( v ) II � l
=
=
t v9 + 9 + 4
t v22
Nole
•
In terms of our earlier notation, d(B, f) = d(v, projd(v)).
1.3
Section
Lines and Planes
43
In the case where the line £ is in IR 2 and its equation has the general form
ax + by = c, the distance d(B, €) from B = (x0, y0 ) is given by the formula
d ( B, € ) =
l ax0 + by0 c l
Va z + b 2
-
(3)
You are invited to prove this formula in Exercise 39.
Exa m p l e 1 . 3 3
Find the distance from the point B = ( 1 , 0, 2) to the plane <fP whose general equation
is x + y z = 1 .
-
<fP
In this case, we need to calculate the length of PB, where P is the point on
at the foot of the perpendicular from B. As Figure 1 .65 shows, if A is any point on
'ii'
and we 'ituate thr nurnrnl vectu' n
Solution
�
[_:]
of 'ii'
'°
that it; tail [, at A, then we
need to find the length of the projection of AB onto n. Again we do the necessary
calculations in steps.
n
B
I
I
I
I
I
11 p
Figure 1 . 6 5
d ( B, <!P ) = ll proj n (AB) II
1: By trial and error, we find any point whose coordinates satisfy the equation
z = 1 . A = ( 1 , 0, O) will do.
Step 2: Set
Step
x+y
-
Step 3:
The projection of v onto n is
proj n ( V )
=
=
(�
n n)n
( 1 + 1 + ( - 1 )2 )[
·
1 ·0 + 1·0 - 1•2
-t: l [=! ]
�1
-1
Chapter
44
1
Vectors
Step 4:
The distance d(B, <JP) from B to <JP is
[J
[J
l l r rnj " ( vl ll � H
�i
In general, the distance d(B, <JP) from the point B = (x0, y0 , z0 ) to the plane whose
general equation is ax + by + cz = d is given by the formula
d ( B, <JP )
l ax0 + by0 + CZ0 - d i
Va 2 + b 2 + c 2
(4)
You will be asked to derive this formula in Exercise 40.
..
I
Exercises 1 . 3
In Exercises 1 and 2, write the equation of the line passing
through P with normal vector n in (a) normal form and
(b) general form.
I . P = ( 0, 0 ) , n =
[�]
2. P = ( 1, 2 ) , n =
[ _! ]
In Exercises 3-6, write the equation of the line passing
through P with direction vector d in (a) vector form and
(b) parametric form.
3. P = ( 1, 0 ) , d =
S. P � (0, 0, 0), d �
[ - �]
[ -: l
[�]
_ m
4. p = ( - 4, 4 ) , d =
•.
p � ( 3 . o. 2 ) . d �
In Exercises 7 and 8, write the equation of the plane passing
through P with normal vector n in (a) normal form and
(b) general form.
7. P � ( 0, 1, 0 ) , n �
[;]
8. P � ( 3, 0, - 2 ) , n �
m
In Exercises 9 and 1 0, write the equation of the plane pass­
ing through P with direction vectors u and v in (a) vector
form and (b) parametric form.
9. p � ( o. o. 0 ) . u �
}
[ [ :l
[ l [ - :i
10. p � ( 6, - 4, - 3), u �
�
�
In Exercises 1 1 and 12, give the vector equation of the line
passing through P and Q.
1 1 . P = ( 1, - 2), Q = (3, O)
12. P = (O, 1, - 1), Q = ( - 2, 1, 3)
In Exercises 1 3 and 1 4, give the vector equation of the plane
passing through P, Q, and R.
13. P = ( 1, 1, 1 ) , Q = (4, 0, 2), R = (O, 1, - 1 )
14. P = ( 1, 1, O), Q = ( 1 , 0, 1 ), R = (O, 1 , 1)
15. Find parametric equations and an equation in vector
form for the lines in IR 2 with the following equations:
(a) y = 3x - 1
(b) 3x + 2y = 5
Section
16. Consider the vector equation x = p + t(q - p ) , where
p and q correspond to distinct points P and Q in IR 2
or IR 3 .
(a) Show that this equation describes the line segment
PQ as t varies from 0 to 1 .
(b) For which value of t is x the midpoint of PQ,
and what is x in this case?
(c) Find the midpoint of PQ when P = (2, - 3 ) and
Q = ( 0, 1 ) .
(d) Find the midpoint o f P Q when P = ( 1 , 0, 1 )
and Q = ( 4, 1 , - 2) .
(e) Find the two points that divide P Q i n part ( c ) into
three equal parts.
(f) Find the two points that divide PQ in part (d) into
three equal parts.
17. Suggest a "vector proof" of the fact that, in IR 2 , two
lines with slopes m 1 and m 2 are perpendicular if and
only if m 1 m 2 = - 1 .
18. The line e passes through the point P = ( 1 , - 1 , 1 ) and
h"' dfr,dion mtoc d
�
[ _;l
Foe mh of th,
following planes <if', determine whether e and <if' are
parallel, perpendicular, or neither:
(a) 2x + 3y - z = 1 (h) 4x - y + 5z = 0
(c) x - y - z = 3
(d) 4x + 6y - 2z = 0
19. The plane <if' 1 has the equation 4x - y + 5z = 2. For
each of the planes <if' in Exercise 18, determine whether
<if' 1 and <if' are parallel, perpendicular, or neither.
20. Find the vector form of the equation of the line in IR 2
that passes through P = (2, - 1 ) and is perpendicular
to the line with general equation 2x - 3y = 1 .
21. Find the vector form of the equation of the line in IR 2
that passes through P = (2, - 1 ) and is parallel to the
line with general equation 2x - 3y = 1 .
22. Find the vector form of the equation of the line in IR 3
that passes through P = ( - 1 , 0, 3 ) and is perpendicular
to the plane with general equation x - 3y + 2z = 5.
23. Find the vector form of the equation of the line in IR 3
that passes through P = ( - 1 , 0, 3 ) and is parallel to
the line with parametric equations
passes through P = (O, - 2, 5 ) and is parallel to the
plane with general equation 6x - y + 2z = 3.
Lines and Planes
45
25. A cube has vertices at the eight points (x, y, z) , where
each of x, y, and z is either 0 or 1 . (See Figure 1 .34.)
(a) Find the general equations of the planes that
determine the six faces (sides) of the cube.
(b) Find the general equation of the plane that con­
tains the diagonal from the origin to ( 1, 1, 1 ) and
is perpendicular to the xy-plane.
(c) Find the general equation of the plane that
contains the side diagonals referred to in
Example 1 .22.
26. Find the equation of the set of all points that are
equidistant from the points P = ( 1 , 0, - 2) and
Q = ( 5, 2, 4 ) .
In Exercises 27 and 28, find the distance from the point Q to
the line e.
27. Q = (2, 2), f with equation
[;] [ - � ] [ _ � ]
�[ ] [ : J { - � ]
28. Q � (O, ! , O), M h 'qoation
+ t
=
�
+
In Exercises 29 and 30, find the distance from the point Q to
the plane <if' .
29. Q = (2, 2, 2), <if' with equation x + y - z = 0
30. Q = (O, 0, O), <if' with equation x - 2y + 2z = 1
Figure 1 . 66 suggests a way to use vectors to locate the point
R on e that is closest to Q.
31. Find the point R on f that is closest to Q in Exercise 27.
32. Find the point R on e that is closest to Q in Exercise 28.
Q
I
x= 1 - t
y = 2 + 3t
z = -2 - t
24. Find the normal form of the equation of the plane that
1. 3
Fioure 1 . 6 6
r
----->
= p + PR
46
Chapter
1
Vectors
Figure 1.67 suggests a way to use vectors to locate the point
R on <!J' that is closest to Q.
k --
the angle between <!J' and <!J' z to be either e or 1 80° whichever is an acute angle. (Figure 1.68)
l
e
0
\
r
= p
------>
------>
+ PQ + QR
In Exercises 35 and 36, find the distance between the
parallel lines.
36.
[;] [ � ] [ � ] [;] [ � ] [ � ]
[�] [ _�] {] [;] [:J {]
+s
-
+
and
+ t
and
-
+
In Exercises 37 and 38, find the distance between the
parallel planes.
37. 2x + y 2z = 0 and 2x + y 2z = 5
38. x + y + z = 1 and x + y + z = 3
39. Prove Equation (3) on page 43.
40. Prove Equation (4) on page 44.
41. Prove that, in !R z , the distance between parallel lines
with equations n · x = c1 and n · x = Cz is given by
l c 1 - Cz l
ll n ll .
42. Prove that the distance between parallel planes with
equations n · x = d 1 and n · x = dz is given by
I d , - dz l
ll n ll
-
�
1 80
Figure 1 . 6 8
33. Find the point R on <!J' that is closest to Q in Exercise 29.
34. Find the point R on <!J' that is closest to Q in Exercise 30.
35.
\lJ> I
-
Figure 1 . 6 1
-
If two nonparallel planes <!J' 1 and <!J' z have normal vectors n 1
and nz and e is the angle between n1 and Dz, then we define
e,
e
In Exercises 43-44, find the acute angle between the planes
with the given equations.
43. x + y + z = 0 and 2x + y - 2z = 0
44. 3x - y + 2z = 5 and x + 4y - z = 2
In Exercises 45-46, show that the plane and line with the
given equations intersect, and then find the acute angle of
intersection between them.
45. The plane given by x + y + 2z = 0 and the line
given by x = 2 + t
y = 1 - 2t
z= 3 + t
46. The plane given by 4x - y - z 6 and the line
given by x = t
y = 1 + 2t
z = 2 + 3t
Exercises 47-48 explore one approach to the problem of
finding the projection of a vector onto a plane. As Figure 1.69 shows, if <!J' is a plane through the origin in IR 3 with
normal vector n, and v is a vector in IR 3, then p = proj9p(v)
is a vector in <!J' such that v - en = p for some scalar c.
n
Figure 1 . 6 9
Projection onto a plane
Section
47. Using the fact that n is orthogonal to every vector in <!f'
(and hence to p ), solve for c and thereby find an expres­
sion for p in terms of v and n.
48. Use the method of Exercise 43 to find the projection of
1. 3
Lines and Planes
onto the planes with the following equations:
(b) 3x - y + z = 0
(a) x + y + z = 0
(c) x - 2z = 0
(d) 2x - 3y + z = 0
41
Exp loration
T h e C ro s s Pro duct
It would b e convenient i f we could easily convert the vector form x = p + s u + t v of
the equation of a plane to the normal form n · x = n · p. What we need is a process
that, given two nonparallel vectors u and v, produces a third vector n that is orthogo­
nal to both u and v. One approach is to use a construction known as the cross product
of vectors. Only valid in IR 3 , it is defined as follows:
Definition
defined by
U
X
V =
[U2U3VV31 -- U3U1VV2]3
U 1 V2 - U 2 V 1
A shortcut that can help you remember how to calculate the cross product of
two vectors is illustrated below. Under each complete vector, write the first two com­
ponents of that vector. Ignoring the two components on the top line, consider each
block of four: Subtract the products of the components connected by dashed lines
from the products of the components connected by solid lines. (It helps to notice that
the first component of u X v has no ls as subscripts, the second has no 2s, and the
third has no 3s.)
UU21 VVz1
U3U1 xx V3V1 UU23VV31 -- U3U1 VV23
Uz x Vz U1V2 - U2V1
The following problems briefly explore the cross product.
1 . Compute u X v.
48
u X v
v
Figure 1 . 1 0
2. Show that e 1 X e = e3 , e X e3 = e 1 , and e3 X e 1 = e .
2
2
2
3. Using the definition of a cross product, prove that u X v (as shown in Figure 1 .70)
is orthogonal to u and v.
4. Use the cross product to help find the normal form of the equation of the plane.
(a) The pfane P"'ing thrnugh P � ( 1 . 0, - 2 ), parn!kl to u
�
[:] [ - �]
and v �
(b) The plane passing through P = (0, - 1 , 1 ) , Q = (2, 0, 2), and R = ( 1 , 2, - 1 )
5. Prove the following properties of the cross product:
(a) v X u = - ( u X v) (b) u X 0 = 0
(c) u X u = 0
(d) u X kv = k ( u X v)
(e) u X ku = 0
(f) u X ( v + w) = u X v + u X w
6. Prove the following properties of the cross product:
(a) u · ( v X w) = ( u X v ) · w
(b) u X (v X w) = (u · w)v - (u · v)w
(c) ll u X v ll 2 = ll u ll 2 ll v ll 2 - ( u · v ) 2
7. Redo Problems 2 and 3, this time making use of Problems 5 and 6.
8. Let u and v be vectors in IR 3 and let (} be the angle between u and v.
(a) Prove that ll u X v ii = ll u ll ll v ll sin (}. [Hint: Use Problem 6(c) .]
(b) Prove that the area A of the triangle determined by u and v (as shown in Fig­
ure 1. 7 1 ) is given by
Figure 1 . 1 1
A = H u X v ii
( c) Use the result in part (b) to compute the area of the triangle with vertices
A = ( 1 , 2, 1 ) , B = (2, 1 , O), and C = (5, - 1 , 3).
Writi n g Project
The Origins o f the Dot Product and Cross Product
The notations for dot and cross product that we use today were introduced in the
late 1 9th century by Josiah Willard Gibbs, a professor of mathematical physics at
Yale University. Edwin B. Wilson was a graduate student in Gibbs's class, and he
later wrote up his class notes, expanded upon them, and had them published in
1 90 1 , with Gibbs's blessing, as Vector Analysis: A Text-Book for the Use of Students
of Mathematics and Physics. However, the concepts of dot and cross product arose
earlier and went by various other names and notations.
Write a report on the evolution of the names and notations for the dot product
and cross product.
1 . Florian Cajori, A History of Mathematical Notations (New York: Dover, 1 993).
2. J. Willard Gibbs and Edwin Bidwell Wilson, Vector Analysis: A Text-Book for the
Use of Students of Mathematics and Physics (New York: Charles Scribner's Sons,
1 90 1 ) . Available online at http://archive.org/details/ 1 1 77 1 4283.
3. Ivor Grattan-Guinness, Companion Encyclopedia of the History and Philosophy
of the Mathematical Sciences (London: Routledge, 2013).
49
50
Chapter
1
Vectors
A p p l icati o n s
Force Veclors
Force is defined as the product of
mass and acceleration due to grav­
ity (which, on Earth, is m/s2 ).
Thus, a kg mass exerts a down­
ward force of kg X m/s2 or
kg • m/s2 • This unit of measure­
ment is a newton (N). So the force
exerted by a kg mass is N.
1 1 9.89.8
9.8
1
9.8
We can use vectors to model force. For example, a wind blowing at 30 km/h in a west­
erly direction or the Earth's gravity acting on a 1 kg mass with a force of 9.8 newtons
downward are each best represented by vectors since they each consist of a magnitude
and a direction.
It is often the case that multiple forces act on an object. In such situations, the
net result of all the forces acting together is a single force called the resultant, which
is simply the vector sum of the individual forces (Figure 1 .72). When several forces
act on an object, it is possible that the resultant force is zero. In this case, the object
is clearly not moving in any direction and we say that it is in equilibrium. When an
object is in equilibrium and the force vectors acting on it are arranged head-to-tail,
the result is a closed polygon (Figure 1 .73).
Figure 1 . 1 2
Figure 1 . 1 3
The resultant of two forces
Equilibrium
Exa m p l e 1 . 3 4
Ann and Bert are trying to roll a rock out of the way. Ann pushes with a force of 20 N
in a northerly direction while Bert pushes with a force of 40 N in an easterly direction.
(a) What is the resultant force on the rock?
(b) Carla is trying to prevent Ann and Bert from moving the rock. What force must
Carla apply to keep the rock in equilibrium?
Solution
(a) Figure 1 .74 shows the position of the two forces. Using the paral­
lelogram rule, we add the two forces to get the resultant r as shown. By Pythagoras'
a
Figure 1 . 1 4
a
b
The resultant of two forces
b
Section
1.4
Applications
51
Theorem, we see that ll r ll = V20 2 + 40 2 = V20QO = 44.72 N. For the direc­
tion of r, we calculate the angle e between r and Bert's easterly force. We find that
sine = 20/ ll r ll = 0.447, so e = 26.57°.
(b) If we denote the forces exerted by Ann, Bert, and Carla by a, b, and c, respec­
tively, then we require a + b + c = 0. Therefore c = - ( a + b) = - r, so Carla
needs to exert a force of 44. 72 N in the direction opposite to r.
Often, we are interested in decomposing a force vector into other vectors whose
resultant is the given vector. This process is called resolving a vector into com­
ponents. In two dimensions, we wish to resolve a vector into two components.
However, there are infinitely many ways to do this; the most useful will be to re­
solve the vector into two orthogonal components. (Chapters 5 and 7 explore this
idea more generally.) This is usually done by introducing coordinate axes and by
choosing the components so that one is parallel to the x-axis and the other to the
y-axis. These components are usually referred to as the horizontal and vertical
components, respectively. In Figure 1 .75, f is the given vector and fx and fy are its
horizontal and vertical components.
y
x
Figure 1 . 1 5
Resolving a vector into components
Exa m p l e 1 . 3 5
Ann pulls on the handle of a wagon with a force of 100 N. If the handle makes an
angle of20° with the horizontal, what is the force that tends to pull the wagon forward
and what force tends to lift it off the ground?
Solution
consider.
Figure 1 .76 shows the situation and the vector diagram that we need to
Figure 1 . 1 6
52
Chapter
1
Vectors
We see that
l l fx I I = 11 £ 11 cos 20° and l l fy I = 1 £ 11 sin 20°
1 00 ( 0.9397 ) = 93.97 and ll fy ll = 1 00 ( 0.3420 )
Thus, ll fx ll =
= 34.20. So the
wagon is pulled forward with a force of approximately 93.97 N and it tends to lift off
the ground with a force of approximately 34.20 N.
We solve the next example using two different methods. The first solution considers
a triangle of forces in equilibrium; the second solution uses resolution of forces into
components.
Exa m p l e 1 . 3 6
Figure 1. 77 shows a painting that has been hung from the ceiling by two wires. If the
painting has a mass of 5 kg and if the two wires make angles of 45 and 60 degrees with
the ceiling, determine the tension in each wire.
Figure 1 . 1 1
We assume that the painting is in equilibrium. Then the two wires must
supply enough upward force to balance the downward force of gravity. Gravity
exerts a downward force of 5 X 9.8 = 49 N on the painting, so the two wires must
collectively pull upward with 49 N of force. Let f1 and f2 denote the tensions in the
wires and let r be their resultant (Figure 1 . 78 ). It follows that II r II = 49 since we are
in equilibrium.
Solulion 1
Section
1.4
Applications
53
Using the law of sines, we have
ll f1 ll
sin 45°
so
11 f1 11
sin 30°
ll r ll
sin 1 05°
ll r ll sin 45° = 49 ( 0.7071 ) =
ll r ll sin 30° = 49 ( 0.5 )
35.87 and I f2 II =
sin 1 05°
0.9659
sin 1 05°
0.9659
Therefore, the tensions in the wires are approximately 35.87 N and 25.36 N.
=
=
25.36
We resolve f1 and f2 into horizontal and vertical components, say, f1 =
h1 + v1 and f2 = h2 + v2 , and note that, as above, there is a downward force of 49 N
(Figure 1 .79).
Solution 2
30 °
It follows that
Figure 1 . 1 8
ll vz ll
Vz
60 °
/
=
ll f2 ll sin 45°
=
I�
Since the painting is in equilibrium, the horizontal components must balance, as
must the vertical components. Therefore, I h1 I = ll h2 II and ll v1 II + ll v2 II = 49, from
which it follows that
and
45 °
Substituting the first of these equations into the second equation yields
v'3 ll f2 ll
, ;;:;
VL
+
fil - 49, or ll f2 I - 49 \/2 - 25.36
,v;;:;2
1 + ,v;;:;3
_
_
_
Thus, 11 £1 11 = v2 ll f2 ll = 1 .4 1 42 ( 25.36 ) = 35.87, so the tensions in the hires are
approximately 35.87 N and 25.36 N, as before.
49 N
Figure 1 . 1 9
I
Exercises 1 . 4
Force Vectors
In Exercises 1 -6, determine the resultant of the given
forces.
1 . f1 acting due north with a magnitude of 1 2 N and
f2 acting due east with a magnitude of 5 N
2. f1 acting due west with a magnitude of 1 5 N and
f2 acting due south with a magnitude of 20 N
3. f1 acting with a magnitude of 8 N and f2 acting at an
angle of 60° to f1 with a magnitude of 8 N
4. f1 acting with a magnitude of 4 N and f2 acting at an
angle of 1 35° to f1 with a magnitude of 6 N
54
Chapter
1
Vectors
5. f1 acting due east with a magnitude of 2 N, f2 acting
due west with a magnitude of 6 N, and f3 acting at an
angle of 60° to f1 with a magnitude of 4 N
6. f1 acting due east with a magnitude of 1 0 N, f2 acting
due north with a magnitude of 1 3 N, f3 acting due west
with a magnitude of 5 N, and f4 acting due south with
a magnitude of 8 N
7. Resolve a force of 1 0 N into two forces perpendicular
to each other so that one component makes an angle
of 60° with the 10 N force.
8. A 10 kg block lies on a ramp that is inclined at an angle
of 30° (Figure 1 .80). Assuming there is no friction, what
force, parallel to the ramp, must be applied to keep the
block from sliding down the ramp?
10. A lawn mower has a mass of 30 kg. It is being pushed
with a force of 1 00 N. If the handle of the lawn mower
makes an angle of 45° with the ground, what is the
horizontal component of the force that is causing the
mower to move forward?
1 1 . A sign hanging outside Joe's Diner has a mass of 50 kg
(Figure 1 .82). If the supporting cable makes an angle
of 60° with the wall of the building, determine the
tension in the cable.
Figure 1 . 8 2
12. A sign hanging in the window of Joe's Diner has a
Figure 1 . 8 0
9. A tow truck is towing a car. The tension in the tow
cable is 1 500 N and the cable makes a 45° with the
horizontal, as shown in Figure 1 . 8 1 . What is the verti­
cal force that tends to lift the car off the ground?
mass of 1 kg. If the supporting strings each make an
angle of 45° with the sign and the supporting hooks
are at the same height (Figure 1 .83), find the tension in
each string.
•
_
_
f = 1500 N
,
,
'
_
,
,
•
, ' ,f2
45 °
OPEN FOR BUSINESS
Figure 1 . 8 3
13. A painting with a mass of 1 5 kg is suspended by two
Figure 1 . 8 1
wires from hooks on a ceiling. If the wires have lengths
of 1 5 cm and 20 cm and the distance between the
hooks is 25 cm, find the tension in each wire.
14. A painting with a mass of 20 kg is suspended by two
wires from a ceiling. If the wires make angles of 30°
and 45° with the ceiling, find the tension in each wire.
Chapter Review
Kev Defi nitions and concepts
head-to-tail rule, 6
integers modulo m (Z m ), 14- 1 6
length (norm) of a vector, 2 0
linear combination of
vectors, 1 2
normal vector, 34, 38
m-ary vector 16
orthogonal vectors, 26
parallel vectors, 8
parallelogram rule, 6
algebraic properties of vectors, 10
angle between vectors, 24
binary vector, 1 3
Cauchy-Schwarz Inequality, 22
cross product, 48
direction vector, 3 5
distance between vectors, 2 3
dot product, 1 8
equation o f a line, 3 6
equation o f a plane, 38-39
Review Questions
1 . Mark each of the following statements true or false:
(a) For vectors U, v, and w in rr;r, if u + w = v + w,
(b)
(c)
(d)
(e)
(f)
( g)
(h)
(i)
(j)
then u = v.
For vectors U, v, and w in rr;r, if u . w = v . w, then
u = v.
For vectors u, v, and w in IR 3 , if u is orthogonal
to v, and v is orthogonal to w, then u is orthogonal
to w.
In IR 3 , if a line C is parallel to a plane <:IP, then a di­
rection vector d for e is parallel to a normal vector
n for <:IP .
I n IR 3 , i f a line e i s perpendicular t o a plane <:IP , then
a direction vector d for e is a parallel to a normal
vector n for <:IP .
I n IR 3 , i f two planes are not parallel, then they must
intersect in a line.
In IR 3 , if two lines are not parallel, then they must
intersect in a point.
If v is a binary vector such that v · v = 0, then
v = 0.
In l:'.5, if ab = 0 then either a = 0 or b = 0.
In l:'.6, if ab = 0 then either a = 0 or b = 0.
2. If u =
[ -�l [�l
[ -�l [�l
v=
and the vector 4u + v is drawn
with its tail at the point ( 1 0 , 1 0 ), find the coordinates
of the point at the head of 4u + v.
3. If u =
for x.
v=
and 2x + u 3 (x - v), solve
=
projection o f a vector onto
a vector, 27
Pythagoras' Theorem, 26
scalar multiplication, 7
standard unit vectors, 22
Triangle Inequality, 22
unit vector, 2 1
vector, 3
vector addition, 5
zero vector, 4
4. Let A, B, C, and D be the vertices of a square centered
---->
at the origin 0, labeled in clockwise order. If a = OA
and b = OB , find BC in terms of a and b.
5. Find the angle between the vectors [ - 1 , 1, 2] and
[2, 1, - 1 ] .
6. Hnd ilie prnjection of v �
[:]
m
onto u
�
[ -n
7. Find a unit vector in the xy-plane that is orthogonal
to
8. Find the general equation of the plane through the
point ( 1, 1, 1 ) that is perpendicular to the line with
parametric equations
x= 2- t
y = 3 + 2t
z = -1 + t
9. Find the general equation of the plane through the
-
point (3, 2, 5) that is parallel to the plane whose general
equation is 2 x + 3y z = 0.
10. Find the general equation of the plane through the
points A ( l , 1 , O ) , B ( l , 0, 1 ) , and C ( O, 1 , 2).
11. Find the area of the triangle with vertices A(l, 1 , O),
B ( l , 0, 1 ) , and C(O, 1 , 2).
55
56
Chapter
1
Vectors
12. Find the midpoint of the line segment between
A = (5, 1 , - 2) and B = (3, - 7, 0).
13. Why are there no vectors u and v in !R n such that
ll u ll = 2, ll v ll = 3, and u · v = - 7?
14. Find the distance from the point (3, 2, 5) to the plane
whose general equation is 2x + 3y - z = 0.
15. Find the distance from the point (3, 2, 5) to the line
with parametric equations x = t, y = 1 + t, z = 2 + t.
16. Compute 3 - (2 + 4) 3 (4 + 3) 2 in Z5.
17. If possible, solve 3(x + 2) = 5 in Z7•
18. If possible, solve 3 (x + 2) = 5 in Z9•
19. Compute [2, 1, 3, 3] · [3, 4, 4, 2] in zt.
20. Let u [ l , 1, 1, O] in Zi. How many binary vectors v
=
satisfy u · v = O?
Systems of Lineari
Eq u ations
The world was full of equations . . . .
There must be an answerfor everything,
if only you knew how to setforth
the questions.
-Anne Tyler
1985, p. 235
The Accidental Tourist
Alfred A. Knopf,
2.0
I n t ro d u ctio n : Triv i a l ilV
The word trivial is derived from the Latin root tri ("three") and the Latin word via
("road"). Thus, speaking literally, a triviality is a place where three roads meet. This
common meeting point gives rise to the other, more familiar meaning of trivial­
commonplace, ordinary, or insignificant. In medieval universities, the trivium con­
sisted of the three "common" subjects (grammar, rhetoric, and logic) that were taught
before the quadrivium (arithmetic, geometry, music, and astronomy) . The "three
roads" that made up the trivium were the beginning of the liberal arts.
In this section, we begin to examine systems oflinear equations. The same system
of equations can be viewed in three different, yet equally important, ways-these will
be our three roads, all leading to the same solution. You will need to get used to this
threefold way of viewing systems of linear equations, so that it becomes common­
place (trivial!) for you.
The system of equations we are going to consider is
2x + y = 8
x - 3y = - 3
Problem 1 Draw the two lines represented by these equations. What is their point
of intersection?
Problem 2
Consider the vectors u =
[�]
and v =
[ _ �]
. Draw the coordinate
grid determined by u and v. [Hint: Lightly draw the standard coordinate grid first and
use it as an aid in drawing the new one.]
Problem 3
On the u-v grid, find the coordinates of w =
[ _: ]
.
Problem 4 Another way to state Problem 3 is to ask for the coefficients x and y
for which xu + yv = w. Write out the two equations to which this vector equation is
equivalent (one for each component) . What do you observe?
Problem 5 Return now to the lines you drew for Problem 1 . We will refer to the
line whose equation is 2x + y = 8 as line 1 and the line whose equation is x - 3y = - 3
as line 2. Plot the point (O, O) on your graph from Problem 1 and label it P0 . Draw a
51
58
Chapter
2
Systems of Linear Equations
Ta b l e 2 . 1
Point
Po
P1
P2
P3
P4
Ps
p6
x
0
y
0
horizontal line segment from P0 to line 1 and label this new point P1 . Next draw a
vertical line segment from P1 to line 2 and label this point P2 • Now draw a horizontal
line segment from P2 to line 1, obtaining point P3 • Continue in this fashion, drawing
vertical segments to line 2 followed by horizontal segments to line 1 . What appears to
be happening?
Problem 6 Using a calculator with two-decimal-place accuracy, find the (approxi­
mate) coordinates of the points P 1 , P2 , P3 , . . . , P6• (You will find it helpful to first
solve the first equation for x in terms of y and the second equation for y in terms
of x.) Record your results in Table 2 . 1 , writing the x- and y-coordinates of each point
separately.
The results of these problems show that the task of "solving" a system of linear
equations may be viewed in several ways. Repeat the process described in the prob­
lems with the following systems of equations:
( a ) 4x - 2y = 0 ( b ) 3x + 2y = 9 ( c ) x + y = S ( d ) x + 2y = 4
x + 2y = 5
x + 3y = 1 0
x-y=3
2x - y = 3
Are all of your observations from Problems 1-6 still valid for these examples? Note
any similarities or differences. In this chapter, we will explore these ideas in more detail.
I n t ro d u ct i o n to svste m s o t l i n e a r E q u a t i o n s
Recall that the general equation o f a line in IR 2 is of the form
ax + by =
c
and that the general equation of a plane in IR 3 is of the form
ax
+
by
+
cz
=
d
Equations of this form are called linear equations.
D e f i n i t i o n A linear equation in the n variables x 1 , x2 , • • • , xn is an equation
that can be written in the form
where the coefficients a 1 , a 2 , . • . , a n and the constant term b are constants.
Exa m p l e 2 . 1
The following equations are linear:
3x - 4y = - 1
\/2x +
r
- ts - lft = 9
:y - (sin :)z =
1
3.2x1 - O.O lx2 = 4 . 6
Observe that the third equation is linear because it can be rewritten in the form x 1 +
5x 2 + x 3 - 2x4 = 3. It is also important to note that, although in these examples (and
in most applications) the coefficients and constant terms are real numbers, in some
examples and applications they will be complex numbers or members of "11.P for some
prime number p.
Section
2. 1
Introduction to Systems of Linear Equations
59
The following equations are not linear:
xy + 2z = 1
xi - x� = 3
x
-+z=2
y
Thus, linear equations do not contain products, reciprocals, or other functions of the
variables; the variables occur only to the first power and are multiplied only by con stants. Pay particular attention to the fourth example in each list: Why is it that the
fourth equation in the first list is linear but the fourth equation in the second list is not?
4
A solution of a linear equation a 1 x 1 + a 2 x 2 +
+ a nxn = b is a vector
, sn l whose components satisfy the equation when we substitute x 1 = s 1 ,
[ s 1 , s2 ,
X2 = S 2> , Xn = Sn ·
•
Exa m p l e 2 . 2
•
•
•
•
·
·
•
(a) [5, 4] is a solution of 3x - 4y = - 1 because, when we substitute x = 5 and y = 4,
the equation is satisfied: 3 (5) - 4(4) = - 1 . [ l , 1] is another solution. In general, the
solutions simply correspond to the points on the line determined by the given equa­
tion. Thus, setting x = t and solving for y, we see that the complete set of solutions
can be written in the parametric form [t, � + �t] . (We could also set y equal to some
parameter-say, s -and solve for x instead; the two parametric solutions would look
different but would be equivalent. Try this.)
(b) The linear equation x 1 - x 2 + 2x 3 = 3 has [3, 0, O ] , [O, 1, 2], and [6, 1 , - 1 ]
as specific solutions. The complete set of solutions corresponds to the set of points
in the plane determined by the given equation. If we set x 2 = s and x 3 = t, then a
parametric solution is given by [3 + s - 2t, s, t] . (Which values of s and t produce the
three specific solutions above?)
A system of linear equations is a finite set of linear equations, each with the same
variables. A solution of a system of linear equations is a vector that is simultaneously
a solution of each equation in the system. The solution set of a system of linear equa­
tions is the set of all solutions of the system. We will refer to the process of finding the
solution set of a system of linear equations as "solving the system:'
Exa m p l e 2 . 3
The system
2x - y = 3
x + 3y 5
has [2, 1 ] as a solution, since it is a solution of both equations. On the other hand,
[ 1 , - 1 ] is not a solution of the system, since it satisfies only the first equation.
=
Exa m p l e 2 . 4
Solve the following systems of linear equations:
(a) x - y = 1
x+y=3
(b) x - y = 2
2x - 2y = 4
(c) x - y = 1
x-y= 3
60
Chapter
2
Systems of Linear Equations
Solution
(a) Adding the two equations together gives 2x = 4, so x = 2, from which we find
that y = 1 . A quick check confirms that [2, 1 ] is indeed a solution of both equations.
That this is the only solution can be seen by observing that this solution corresponds
to the (unique) point of intersection ( 2, 1 ) of the lines with equations x - y = 1 and
x + y = 3, as shown in Figure 2 . l (a). Thus, [2, 1 ] is a unique solution.
(b) The second equation in this system is just twice the first, so the solutions are the
solutions of the first equation alone-namely, the points on the line x - y = 2. These
can be represented parametrically as [2 + t, t] . Thus, this system has infinitely many
solutions [Figure 2 . 1 (b) ] .
(c) Two numbers x and y cannot simultaneously have a difference of 1 and 3 . Hence,
this system has no solutions. (A more algebraic approach might be to subtract the second
equation from the first, yielding the absurd conclusion 0 = - 2.) As Figure 2.l (c) shows,
the lines for the equations are parallel in this case .
.J
y
y
(b)
(c)
4
( a)
Figure 2 . 1
A system oflinear equations is called consistent if it has at least one solution. A sys­
tem with no solutions is called inconsistent. Even though they are small, the three sys­
tems in Example 2.4 illustrate the only three possibilities for the number of solutions of
a system of linear equations with real coefficients. We will prove later that these same
three possibilities hold for any system of linear equations over the real numbers.
A
system of linear equations with real coefficients has either
(a) a unique solution (a consistent system) or
(b) infinitely many solutions (a consistent system) or
( c) no solutions (an inconsistent system) .
Solving a svstem or Linear Equations
�
Two linear systems are called equivalent if they have the same solution sets. For
example,
and
x-y= l
x-y= l
x+y=3
y= 1
are equivalent, since they both have the unique solution [2, l ] . (Check this.)
Section
2. 1
Introduction to Systems of Linear Equations
61
Our approach to solving a system of linear equations is to transform the given
system into an equivalent one that is easier to solve. The triangular pattern of the
second example above (in which the second equation has one less variable than the
first) is what we will aim for.
Exa m p l e 2 . 5
Solve the system
x-y- z= 2
y + 3z = 5
5z = 1 0
Starting from the last equation and working backward, we find successively
that z = 2, y = 5 - 3(2) = - 1 , and x = 2 + ( - 1 ) + 2 = 3. So the unique solution is
[3, - 1 , 2] .
Solution
The procedure used to solve Example 2.5 is called back substitution.
We now turn to the general strategy for transforming a given system into an
equivalent one that can be solved easily by back substitution. This process will be
described in greater detail in the next section; for now, we will simply observe it in
action in a single example.
Exa m p l e 2 . 6
Solve the system
x- y- z= 2
3x - 3y + 2z = 1 6
2x - y + z = 9
To transform this system into one that exhibits the triangular structure
of Example 2.5, we first need to eliminate the variable x from Equations 2 and 3.
Observe that subtracting appropriate multiples of equation 1 from Equations 2 and
3 will do the trick. Next, observe that we are operating on the coefficients, not on
the variables, so we can save ourselves some writing if we record the coefficients and
constant terms in the matrix
-1 -1 2
-3 2 16
1 9
-1
Solution
The word matrix is derived from
the Latin word mater, meaning
"mother:' When the suffix - ix
is added, the meaning becomes
"womb:' Just as a womb surrounds
a fetus, the brackets of a matrix
surround its entries, and just as
the womb gives rise to a baby, a
matrix gives rise to certain types of
functions called linear transforma­
tions. A matrix with m rows and
n columns is called an m X n
matrix (pronounced "m by n").
The plural of matrix is matrices,
not "matrixes:'
[:
]
where the first three columns contain the coefficients of the variables in order, the final
column contains the constant terms, and the vertical bar serves to remind us of the
equal signs in the equations. This matrix is called the augmented matrix of the system.
There are various ways to convert the given system into one with the triangular
pattern we are after. The steps we will use here are closest in spirit to the more general
method described in the next section. We will perform the sequence of operations on
the given system and simultaneously on the corresponding augmented matrix. We
begin by eliminating x from Equations 2 and 3.
x- y- z= 2
3x - 3y + 2z = 1 6
2x - y + z = 9
-1
-3
-1
2
2 16
1 9
-1
]
62
Chapter
2
Systems of Linear Equations
Subtract 3 times the first equation
from the second equation:
Subtract 3 times the first row from the
second row:
x-y- z= 2
5z = 1 0
2x - y + z = 9
Subtract 2 times the first equation
from the third equation:
[:
-1
0
-1
Subtract 2 times the first row from the
third row:
x-y- z= 2
5z = 1 0
y + 3z = 5
Interchange Equations 2 and 3:
[:
-1
0
1
Interchange rows 2 and 3:
x-y- z= 2
y + 3z = 5
5z = 1 0
[i
-1
1
0
-1
5
3
-1
3
5
!�:
lt
This is the same system that we solved using back substitution in Example 2.5, where
we found that the solution was [ 3, - 1 , 2] . This is therefore also the solution to the sys­
tem given in this example. Why? The calculations above show that any solution of the
given system is also a solution of thefinal one. But since the steps we just performed are
reversible, we could recover the original system, starting with the final system. (How?)
So any solution of the final system is also a solution of the given one. Thus, the systems
are equivalent (as are all of the ones obtained in the intermediate steps above). More­
over, we might just as well work with matrices instead of equations, since it is a simple
matter to reinsert the variables before proceeding with the back substitution. (Work­
ing with matrices is the subject of the next section.)
Remark
Calculators with matrix capabilities and computer algebra systems can
facilitate solving systems of linear equations, particularly when the systems are large
or have coefficients that are not "nice;' as is often the case in real-life applications. As
always, though, you should do as many examples as you can with pencil and paper
until you are comfortable with the techniques. Even if a calculator or CAS is called
for, think about how you would do the calculations manually before doing anything.
After you have an answer, be sure to think about whether it is reasonable.
Do not be misled into thinking that technology will always give you the answer
faster or more easily than calculating by hand. Sometimes it may not give you the
answer at all! Roundoff errors associated with the floating-point arithmetic used by
calculators and computers can cause serious problems and lead to wildly wrong an­
swers to some problems. See Exploration: Lies My Computer Told Me for a glimpse
of the problem. (You've been warned! )
Section
1
2. 1
Introduction to Systems of Linear Equations
63
Exercises 2 . 1
In Exercises 1-6, determine which equations are linear
equations in the variables x, y, and z. If any equation is not
linear, explain why not.
1 . x - 'TTY + efsz = 0
2. x 2 + y 2 + z 2 = 1
3. x - 1 + 7y + z = sin
(;)
4. 2x - xy - 5z = 0
6. (cos3)x - 4y + z = v3
5. 3 cos x - 4y + z = v3
_
In Exercises 7-1 0, find a linear equation that has the same
solution set as the given equation (possibly with some
restrictions on the variables).
x2 y2
8. x - y
7. 2x + y = 7 - 3y
1 1 4
9. - + - = 10. log 1 0 x - log 1 0 y = 2
x y xy
--­
In Exercises 1 1 -14, find the solution set of each equation.
12. 2x 1 + 3x 2 = 5
1 1 . 3x - 6y = 0
13. x + 2y + 3 z = 4
In Exercises 15-18, draw graphs corresponding to the given
linear systems. Determine geometrically whether each sys­
tem has a unique solution, infinitely many solutions, or no
solution. Then solve each system algebraically to confirm
your answer.
16. x - 2y = 7
15. x + y = 0
2x + y = 3
3x + y = 7
18. O. lOx - 0.05y = 0.20
17. 3x - 6y = 3
-0.06x + 0.03y = -0.12
-x + 2y = l
In Exercises 1 9-24, solve the given system by back
substitution.
19. x - 2y = 1
20. 2u - 3v = 5
2v = 6
y=3
21. x - y + z = 0
22. X 1 + 2X 2 + 3X 3 = 0
- 5x2 + 2x 3 = 0
2y - z = 1
4X 3 = 0
3z = - 1
23. X 1 + Xz - X3 - X4 = 1 24. x - 3y + z = 5
y - 2z = - 1
X2 + X3 + X4 = 0
X 3 - X4 = 0
X4 = 1
The systems in Exercises 25 and 26 exhibit a "lower trian­
gular" pattern that makes them easy to solve by forward
substitution. (We will encounter forward substitution again
in Chapter 3.) Solve these systems.
= -1
25. X
2 26. X1
!
= -3
2x + y
5
- x1 + x2
�X 1 + 2X 2 + X3 = 7
- 3x - 4y + z = - 10
Find the augmented matrices of the linear systems in
Exercises 27-30.
28. 2X1 + 3X2 - X3 = 1
27. x - y = 0
2x + y = 3
+ X3 = 0
X1
- X 1 + 2X2 - 2X 3 = 0
29. x + Sy = - 1
30. a - 2b
+ d=2
-x + y = - 5
- a + b - c - 3d = 1
2x + 4y = 4
In Exercises 31 and 32, find a system of linear equations
that has the given matrix as its augmented matrix.
31.
32.
[ � =: � :J
[i
�]
-1 0 3
2
-1
0 2 3
For Exercises 33-38, solve the linear systems in the given
exercises.
33. Exercise 27
34. Exercise 28
35. Exercise 29
36. Exercise 3 0
38. Exercise 32
37. Exercise 3 1
39. (a) Find a system of two linear equations in the vari­
ables x and y whose solution set is given by the
parametric equations x = t and y = 3 - 2t.
(b) Find another parametric solution to the system in
part (a) in which the parameter is s and y = s.
40. ( a) Find a system of two linear equations in the vari­
ables x 1 , x 2 , and x 3 whose solution set is given by
the parametric equations x 1 = t, x 2 = 1 + t, and
X 3 = 2 - t.
(b) Find another parametric solution to the system in
part (a) in which the parameter is s and x3 = s.
64
Chapter
2
Systems of Linear Equations
In Exercises 4 1 -44, the systems of equations are nonlinear.
Find substitutions (changes of variables) that convert each
system into a linear system and use this linear system to help
solve the given system.
2 3
41. -x + - = 0
y
-3 + -4 = l
x y
42. x 2 + 2y 2 = 6
x2 - y 2 = 3
43. tan x 2 sin y
2
sin y + cos z = 2
tan x
sin y cos z = - 1
44. - 2a + 2 (3 b ) = 1
3(2a) - 4(3 b ) = 1
-- -
D i rect M e t h o d s f o r S o lvi n g l i n e a r Svst e m s
In this section, we will look at a general, systematic procedure fo r solving a system
of linear equations. This procedure is based on the idea of reducing the augmented
matrix of the given system to a form that can then be solved by back substitution.
The method is direct in the sense that it leads directly to the solution (if one exists) in
a finite number of steps. In Section 2.5, we will consider some indirect methods that
work in a completely different way.
Matrices a n d Echelon Form
There are two important matrices associated with a linear system. The coefficient
matrix contains the coefficients of the variables, and the augmented matrix (which
we have already encountered) is the coefficient matrix augmented by an extra column
containing the constant terms.
For the system
2x + y - z = 3
x
+ Sz = 1
-x + 3y - 2z = O
the coefficient matrix is
and the augmented matrix is
[
-1]
- � � -�
2 1
1 -1 3
0
5 1
3 -2 0
]
Note that if a variable is missing (as y is in the second equation), its coefficient 0 is
entered in the appropriate position in the matrix. If we denote the coefficient matrix
of a linear system by A and the column vector of constant terms by b, then the form
of the augmented matrix is [A I b J .
Section
2.2
Direct Methods for Solving Linear Systems
65
In solving a linear system, it will not always be possible to reduce the coefficient
matrix to triangular form, as we did in Example 2.6. However, we can always achieve
a staircase pattern in the nonzero entries of the final matrix.
The word echelon comes from
the Latin word scala, meaning
"ladder" or "stairs:' The French
word for "ladder:' echelle, is also
derived from this Latin base. A
matrix in echelon form exhibits
a staircase pattern.
Definition
properties:
A matrix is in row echelon form if it satisfies the following
1. Any rows consisting entirely of zeros are at the bottom.
2. In each nonzero row, the first nonzero entry (called the leading entry) is
in a column to the left of any leading entries below it.
Note that these properties guarantee that the leading entries form a staircase pat­
tern. In particular, in any column containing a leading entry, all entries below the
leading entry are zero, as the following examples illustrate.
Exa m p l e 2 . 1
The following matrices are in row echelon form:
[: �] [: :J [i
4
-1
0
0
1
0
1 2
0 1
0 0
�-
l�
1 -1
2
0 -1 1
4
0
0 0
2
0
0
0 0
0
�-4
If a matrix in row echelon form is actually the augmented matrix of a linear sys­
tem, the system is quite easy to solve by back substitution alone.
Exa m p l e 2 . 8
Assuming that each of the matrices in Example 2.7 is an augmented matrix, write out
the corresponding systems of linear equations and solve them.
We first remind ourselves that the last column in an augmented matrix is
the vector of constant terms. The first matrix then corresponds to the system
Solution
2x1
+
4X2 = 1
-x2 = 2
(Notice that we have dropped the last equation 0 = 0, or Ox 1 + Ox 2 = 0, which is
clearly satisfied for any values of x 1 and x 2 .) Back substitution gives x 2 = - 2 and then
2x1 = 1 - 4( - 2) = 9, so x1 = � . The solution is [ �, - 2 ] .
The second matrix has the corresponding system
x, = 1
X2 = 5
0 =4
The last equation represents Ox 1 + Ox2 = 4, which clearly has no solutions. Therefore,
the system has no solutions. Similarly, the system corresponding to the fourth matrix
has no solutions. For the system corresponding to the third matrix, we have
66
Chapter
2
Systems of Linear Equations
X1
+ X2 + 2X3
=
1
X3 = 3
so x1 = 1 - 2 ( 3 ) - x2 = - 5 x2• There are infinitely many solutions, since we may
assign x2 any value t to get the parametric solution [ 5 t, t, 3 ] .
-
-
-
Elemenlarv Row Operalions
We now describe the procedure by which any matrix can be reduced to a matrix
in row echelon form. The allowable operations, called elementary row operations,
correspond to the operations that can be performed on a system of linear equations
to transform it into an equivalent system.
DefiniliOD
matrix:
The following elementary row operations can be performed on a
1 . Interchange two rows.
2. Multiply a row by a nonzero constant.
3. Add a multiple of a row to another row.
Observe that dividing a row by a nonzero constant is implied in the above
definition, since, for example, dividing a row by 2 is the same as multiplying it by ! .
Similarly, subtracting a multiple o f one row from another row i s the same a s adding a
negative multiple of one row to another row.
We will use the following shorthand notation for the three elementary row
operations:
1 . R; � Rj means interchange rows i and j.
2. kR; means multiply row i by k.
3. R; + kRj means add k times row j to row i (and replace row i with the result).
The process of applying elementary row operations to bring a matrix into row
echelon form, called row reduction, is used to reduce a matrix to echelon form.
Exa m p l e 2 . 9
Reduce the following matrix to echelon form:
[ _:
-4 -4
0 0
2
3
1
3
6
1
2
4
�]
We work column by column, from left to right and from top to bottom.
The strategy is to create a leading entry in a column and then use it to create zeros
below it. The entry chosen to become a leading entry is called a pivot, and this phase
of the process is called pivoting. Although not strictly necessary, it is often convenient
to use the second elementary row operation to make each leading entry a 1 .
Solution
Section
2.2
Direct Methods for Solving Linear Systems
5]
55 [
61
-- :5 ]
We begin by introducing zeros into the first column below the leading 1 in the
first row:
2 -4 -4
4 0 0 2
1
3 2
3 6
R, - 2R1
R3 - 2R1
�
R, + R1
l
2 -4 -4
0 0 8 8
0 - 1 10 9
0 3 -1 2
10
The first column i s now a s we want it, s o the next thing t o do i s to create a leading entry
in the second row, aiming for the staircase pattern of echelon form. In this case, we do
this by interchanging rows. (We could also add row 3 or row 4 to row 2.)
]�
2 -4 -4
- 1 10 9
0 8 8 -8
3 - 1 2 10
]
The pivot this time was - 1 . We now create a zero at the bottom o f column 2, using
the leading entry - 1 in row 2:
-�
-5
2 -4 -4
- 1 10 9
0 8 8 -8
0 29 29
Column 2 is now done. Noting that we already have a leading entry in column 3,
we just pivot on the 8 to introduce a zero below it. This is easiest if we first divide
row 3 by 8:
�
� R,
[1
]�
2 -4 -4
- 1 10 9
1 -1
0 1
0 29 29 - 5
]
Now we use the leading 1 in row 3 to create a zero below it:
·�· [ i
-�
2 -4 -4
- 1 10 9
1 -1
0 1
0 0 0 24
With this final step, we have reduced our matrix to echelon form.
R e m arks
The row echelon form of a matrix is not unique. (Find a different row echelon
form for the matrix in Example 2.9. )
•
68
Chapter
2
Systems of Linear Equations
The leading entry in each row is used to create the zeros below it.
The pivots are not necessarily the entries that are originally in the posi­
tions eventually occupied by the leading entries. In Example 2.9, the pivots were
1 , - 1 , 8, and 24. The original matrix had 1 , 4, 2, and 5 in those positions on the
"staircase."
•
Once we have pivoted and introduced zeros below the leading entry in a
column, that column does not change. In other words, the row echelon form emerges
from left to right, top to bottom.
•
•
Elementary row operations are reversible-that is, they can be "undone:' Thus,
if some elementary row operation converts A into B, there is also an elementary row
operation that converts B into A. (See Exercises 1 5 and 16.)
Matrices A and B are row equivalent if there is a sequence of
elementary row operations that converts A into B.
Definition
The matrices in Example 2.9,
2 -4 -4
4 0 0
1
3 2
3 6
�] [�
md
-�]
2 -4 -4
- 1 10 9
1 -1
0 1
0 0 0 24
are row equivalent. In general, though, how can we tell whether two matrices are row
equivalent?
Theorem 2 . 1
Matrices A and B are row equivalent if and only if they can be reduced to the same
row echelon form.
Proof If A and B are row equivalent, then further row operations will reduce B (and
therefore A) to the (same) row echelon form.
Conversely, if A and B have the same row echelon form R, then, via elementary
row operations, we can convert A into R and B into R. Reversing the latter sequence of
operations, we can convert R into B, and therefore the sequence A � R � B achieves
the desired effect.
Remark In practice, Theorem 2 . 1 is easiest to use if R is the reduced row echelon
form of A and B, as defined on page 73. See Exercises 17 and 18.
Gaussian Elimination
When row reduction is applied to the augmented matrix of a system of linear
equations, we create an equivalent system that can be solved by back substitution.
The entire process is known as Gaussian elimination.
Section 2.2 Direct Methods for Solving Linear Systems
G a u s s i a n Elimination
69
1. Write the augmented matrix of the system of linear equations.
2. Use elementary row operations to reduce the augmented matrix to row
echelon form.
3. Using back substitution, solve the equivalent system that corresponds to the
row-reduced matrix.
Remark When performed by hand, step 2 of Gaussian elimination allows quite a
bit of choice. Here are some useful guidelines:
(a) Locate the leftmost column that is not all zeros.
(b) Create a leading entry at the top of this column. (It will usually be easiest if you
make this a leading 1 . See Exercise 22.)
(c) Use the leading entry to create zeros below it.
( d) Cover up the row containing the leading entry, and go back to step (a) to repeat the pro­
cedure on the remaining submatrix. Stop when the entire matrix is in row echelon form.
Exa m p l e 2 . 1 0
Solve the system
2x2 + 3x3 = 8
2x1 + 3x2 + x3 = 5
X1 - Xz - 2X3 = 5
-
Solution
The augmented matrix is
[�
2 3
3
- 1 -2
-
�l
We proceed to reduce this matrix to row echelon form, following the guidelines given
for step 2 of the process. The first nonzero column is column 1 . We begin by creating
Carl Friedrich Gauss (1777-1855) is generally considered to be one of the three greatest
mathematicians of all time, along with Archimedes and Newton. He is often called the "prince of
mathematicians;' a nickname that he richly deserves. A child prodigy, Gauss reportedly could do
arithmetic before he could talk. At the age of he corrected an error in his father's calculations for
the company payroll, and as a young student, he found the formula n(n + 1) /2 for the sum of the
first n natural numbers. When he was 19, he proved that a 17-sided polygon could be constructed
using only a straightedge and a compass, and at the age of21, he proved, in his doctoral
dissertation, that every polynomial of degree n with real or complex coefficients has exactly
n zeros, counting multiple zeros-the Fundamental Theorem of Algebra.
Gauss's 1801 publication Disquisitiones Arithmeticae is generally considered to be the
foundation of modern number theory, but he made contributions to nearly every branch of
mathematics as well as to statistics, physics, astronomy, and surveying. Gauss did not publish
all of his findings, probably because he was too critical of his own work. He also did not like
to teach and was often critical of other mathematicians, perhaps because he discovered-but
did not publish-their results before they did.
The method called Gaussian elimination was known to the Chinese in the third century B . c .
and was well known by Gauss's time. The method bears Gauss's name because of his use of it in a
paper in which he solved a system oflinear equations to describe the orbit of an asteroid.
3,
10
Chapter
2
Systems of Linear Equations
a leading entry at the top of this column; interchanging rows 1 and 3 is the best way
to achieve this.
[�
2
3
-1
SJ [
3
s
-2 -s
[
Ri ++ R,
�
l
2
0
-1
3
2
-2 -S
1 s
3 8
i
We now create a second zero in the first column, using the leading 1 :
R, 2R
�
-
1
l - l -2 -S
0 s s lS
0 2 3 8
i
We now cover up the first row and repeat the procedure. The second column is
the first nonzero column of the submatrix. Multiplying row 2 by � will create a
leading 1 .
[i
-1
s
2
-2
s
3
- 5 1 [ i -! ]
I[ 5i
lS
8
tR2
�
-1
-2
2
3
We now need another zero at the bottom of column 2:
�
,, - rn,
0
0
- 1 -2
1
1
0 1
3
2
The augmented matrix is now in row echelon form, and we move to step 3. The cor­
responding system is
X 1 - Xz - 2X3 = - S
Xz + x3 = 3
X3 = 2
and back substitution gives x3 = 2, then x2 = 3 - x3 = 3 - 2 = 1, and
finally x 1 = - S + x2 + 2x3 = - S + 1 + 4 = 0. We write the solution in vector
form as
(We are going to write the vector solutions of linear systems as column vectors from
now on. The reason for this will become clear in Chapter 3.)
Exa m p l e 2 . 1 1
Solve the system
w - x - y + 2z
2w - 2x - y + 3z
-w + x - y
=
=
=
3
-3
Section
Solution
2.2
Direct Methods for Solving Linear Systems
11
The augmented matrix is
u
-1 -1 2
-2 - 1 3
-1 0
which can be row reduced as follows:
u
J [:
[:
-1 -1 2
-2 - 1 3
-1 0
R, - 2R,
R, + R,
-----+
R3 + 2R2
-----+
The associated system is now
w
-x
-y
y
J
-1 -1 2
0
-1
0 -2 2
-1 -1 2
0 1 -1
0 0 0
J
il
+ 2z = 1
- z = 1
which has infinitely many solutions. There is more than one way to assign param eters, but we will proceed to use back substitution, writing the variables correspond­
ing to the leading entries (the leading variables) in terms of the other variables (the
free variables) .
In this case, the leading variables are w and y, and the free variables are x and z.
Thus, y = 1 + z, and from this we obtain
w
If we assign parameters x =
form as
= 1 + x + y - 2z
= 1 + x + (1 + z) - 2z
=2+x-z
s
and z = t, the solution can be written in vector
Example 2. 1 1 highlights a very important property: In a consistent system, the
free variables are just the variables that are not leading variables. Since the number
of leading variables is the number of nonzero rows in the row echelon form of the
coefficient matrix, we can predict the number of free variables (parameters) before
we find the explicit solution using back substitution. In Chapter 3, we will prove that,
although the row echelon form of a matrix is not unique, the number of nonzero rows
is the same in all row echelon forms of a given matrix. Thus, it makes sense to give a
name to this number.
12
Chapter
2
Systems of Linear Equations
Definition
echelon form.
The rank of a matrix is the number of nonzero rows in its row
We will denote the rank of a matrix A by rank(A) . In Example 2 . 1 0, the rank of
the coefficient matrix is 3, and in Example 2. 1 1 , the rank of the coefficient matrix is
2. The observations we have just made justify the following theorem, which we will
prove in more generality in Chapters 3 and 6.
Theorem 2 . 2
The Rank Theorem
-
Let A be the coefficient matrix of a system of linear equations with n variables. If
the system is consistent, then
number of free variables
=
-
n
rank (A )
Thus, in Example 2 . 1 0, we have 3 - 3 = 0 free variables (in other words, a unique
solution), and in Example 2. 1 1 , we have 4 2 = 2 free variables, as we found.
Exa m p l e 2 . 1 2
Solve the system
X 1 - X2 + 2X3 =
X 1 + 2X2 - X3 =
2x2 - 2X3 =
Solulion
When we row reduce the augmented matrix, we have
[i
-1
2
2
2
-1
-2
- : l [i
[i
[I
Ri - R 1
�
!. R2
�
�
� - ,,,
(1842-1899)
Wilhelm Jordan
was
a German professor of geodesy
whose contribution to solving
linear systems was a systematic
method of back substitution
closely related to the method
described here.
3
-3
0
0
-1
2
3 -3
2
-2
-1
1
2
2
-1
-2
-1
1
2
-1
0
0
-;]
�]
-�]
-
-
leading to the impossible equation 0 = 5. (We could also have performed R3 t R2 as the
second elementary row operation, which would have given us the same contradiction
but a different row echelon form.) Thus, the system has no solutions-it is inconsistent.
Gauss-Jordan Eliminalion
4
A modification of Gaussian elimination greatly simplifies the back substitution phase
and is particularly helpful when calculations are being done by hand on a system with
Section
2.2
Direct Methods for Solving Linear Systems
13
infinitely many solutions. This variant, known as Gauss-Jordan elimination, relies
on reducing the augmented matrix even further.
Definition
properties:
A matrix is in reduced row echelon form ifit satisfies the following
1. It is in row echelon form.
2. The leading entry in each nonzero row is a 1 (called a leading 1 ) .
3. Each column containing a leading 1 has zeros everywhere else.
The following matrix is in reduced row echelon form:
1
0
0
0
0
2
0
0
0
0
1 0
0 0 -3
0 4 -1 0
0
3 -2 0
0 0 0 0 1
0 0 0 0 0
For 2 X 2 matrices, the possible reduced row echelon forms are
For a short proof that the reduced
row echelon form of a matrix is
unique, see the article by Thomas
Yuster, "The Reduced Row Echelon
Form of a Matrix Is Unique: A
Simple Proof;' in the March
issue of Mathematics Magazine
(vol. no. pp.
57, 2, 93-94).
1984
G a u ss-Jordan
Elimination
where * can be any number.
It is clear that after a matrix has been reduced to echelon form, further elementary
row operations will bring it to reduced row echelon form. What is not clear (although
intuition may suggest it) is that, unlike the row echelon form, the reduced row ech­
elon form of a matrix is unique.
In Gauss-Jordan elimination, we proceed as in Gaussian elimination but reduce
the augmented matrix to reduced row echelon form.
1 . Write the augmented matrix of the system of linear equations.
2. Use elementary row operations to reduce the augmented matrix to reduced
row echelon form.
3. If the resulting system is consistent, solve for the leading variables in terms
of any remaining free variables.
Exa m p l e 2 . 1 3
Solve the system in Example 2.1 1 by Gauss-Jordan elimination.
SolUtion
[:
The reduction proceeds as it did in Example 2.1 1 until we reach the echelon form:
ll
-1 -1 2
0 1 -1 1
0 0 0 0
14
Chapter
2
Systems of Linear Equations
We now must create a zero above the leading 1 in the second row, third column. We
do this by adding row 2 to row 1 to obtain
[:
-1 0
0
0 0
The system has now been reduced to
w-x
+z=2
y-z= I
It is now much easier to solve for the leading variables:
w=2+x-z
and
y= I +z
If we assign parameters x = s and z = t as before, the solution can be written in vector
form as
Remark
From a computational point of view, it is more efficient (in the sense
that it requires fewer calculations) to first reduce the matrix to row echelon form
and then, working from right to left, make each leading entry a 1 and create zeros
above these leading 1 s. However, for manual calculation, you will find it easier to
just work from left to right and create the leading ls and the zeros in their columns
as you go.
Let's return to the geometry that brought us to this point. Just as systems of linear
equations in two variables correspond to lines in IR 2 , so linear equations in three vari­
ables correspond to planes in IR 3 . In fact, many questions about lines and planes can
be answered by solving an appropriate linear system.
Exa m p l e 2 . 1 4
Find the line of intersection of the planes x + 2y - z = 3 and 2x + 3y + z = 1 .
Solution
First, observe that there will b e a line of intersection, since the normal
vectors of the two planes- [ I, 2, - I ] and [2, 3, 1 ] -are not parallel. The points that
lie in the intersection of the two planes correspond to the points in the solution set
of the system
x + 2y - z = 3
2x + 3y + z = 1
Gauss-Jordan elimination applied to the augmented matrix yields
Section
[� 3
2.2
Direct Methods for Solving Linear Systems
2
1
3
1
-1 3 ]
1- 3 - 7 J
7
-7
2
-5
0
s
Replacing variables, we have
5
+ Sz =
x
z
15
y
3z = 5
We set the free variable z equal to a parameter t and thus obtain the parametric equa­
tions of the line of intersection of the two planes:
x=
- St
y = 5 + 3t
z=
In vector form, the equation is
x
y
Figure 2 . 2
The intersection of two planes
Exa m p l e 2 . 1 5
See Figure 2.2.
Lct p
x
=
p
�
[ _ H m' [:J.
q �
+ tu and x =
u
q
�
md v
�
[J
Dcte;m;ne whcthec the lilles
+ tv intersect and, if so, find their point of intersection.
We need to be careful here. Although t has been used as the parameter
in the equations of both lines, the lines are independent and therefore so are their
parameters. Let's use a different parameter-say, s -for the first line, so its equation
Solution
bewmes
x � p
+ su. If the lines illtmed, then we wmt to find on x
satisfies both equations simultaneously. That is, we want x =
SU
- tv = q
-
p.
p
Substituting the given p, q, u, and v, we obtain the equations
s - 3t =
s+ t=
s+ t=
�
+ su =
[;]
q
thot
+ tv or
-1
2
2
whose solution is easily found to be s = �, t = � . The point of intersection is therefore
Chapter
16
2
Systems of Linear Equations
z
See Figure 2.3. (Check that substituting t
point.)
x
=
i in the other equation gives the same
In IR 3 , it is possible for two lines to intersect in a point, to be parallel, or
to do neither. Nonparallel lines that do not intersect are called skew lines.
Remark
y
Figure 2 . 3
Two intersecting lines
Homogeneous svste ms
We have seen that every system of linear equations has either no solution, a unique
solution, or infinitely many solutions. However, there is one type of system that
always has at least one solution.
A system of linear equations is called homogeneous ifthe constant
term in each equation is zero.
Definition
In other words, a homogeneous system has an augmented matrix of the form
[A I O] . The following system is homogeneous:
2x + 3y - z
-x + Sy + 2z
=
=
0
0
Since a homogeneous system cannot have no solution (forgive the double negative!),
it will have either a unique solution (namely, the zero, or trivial, solution) or infinitely
many solutions. The next theorem says that the latter case must occur if the number
of variables is greater than the number of equations.
Theorem 2 . 3
If [A I O] is a homogeneous system of m linear equations with n variables, where
m < n, then the system has infinitely many solutions.
Since the system has at least the zero solution, it is consistent. Also,
rank(A) :::::: m (why?) . By the Rank Theorem, we have
Proof
�
number of free variables
=
n - rank (A ) 2: n - m > 0
So there is at least one free variable and, hence, there are infinitely many solutions.
Note Theorem 2.3 says nothing about the case where m 2: n. Exercise 44 asks
you to give examples to show that, in this case, there can be either a unique solution
or infinitely many solutions.
Section
IR and ZP are examples offields.
The set of rational numbers Q and
the set of complex numbers C are
other examples. Fields are covered
in detail in courses in abstract
algebra.
2.2
Direct Methods for Solving Linear Systems
linear svs1ems over Z p
Thus far, all o f the linear systems we have encountered have involved real numbers,
and the solutions have accordingly been vectors in some !R n . We have seen how other
number systems arise-notably, "ll_P ' When p is a prime number, "ll_P behaves in many
respects like IR; in particular, we can add, subtract, multiply, and divide (by nonzero
numbers). Thus, we can also solve systems of linear equations when the variables and
coefficients belong to some "ll_P ' In such instances, we refer to solving a system over "ll_P '
For example, the linear equation x1 + x2 + x3 = 1 , when viewed as an equation
over Z 2 , has exactly four solutions:
(where the last solution arises because 1 + 1 + 1
wh,n wni,w th"q"'Hon X ; + x, +
�
Exa m p l e 2 . 1 6
11
X; �
1 in Z 2 ) .
=
1 ov" 2,, thno!utiom
[ :: l
m
(Check these.)
But we need not use trial-and-error methods; row reduction of augmented matri­
ces works just as well over "ll_P as over IR.
Solve the following system of linear equations over Z 3 :
X1
X1
+ 2x2 + x3
+ X3
x2 + 2x3
=
=
=
0
2
1
The first thing to note in examples like this one is that subtraction and
division are not needed; we can accomplish the same effects using addition and mul­
tiplication. (This, however, requires that we be working over "ll_P , where p is a prime;
see Exercise 60 at the end of this section and Exercise 57 in Section 1 . 1 .)
We row reduce the augmented matrix of the system, using calculations modulo 3.
Solution
[i ;] [i ;]
[i :J
2
0
1 2
�
R2 + 2R1
R, + R2
�
R3 + 2R2
2 1
1 0
1 2
0
1 0
0 2
18
Chapter
2
Systems of Linear Equations
R, + R,
�
Thus, the solution is Xi
Exa m p l e 2 . 1 1
=
[
1 Q 0 1
0 1 02
0 0 1 1
l
1, x2 = 2, x3 = 1 .
Solve the following system of linear equations over 2 2 :
X i + Xz + X 3 + X4
X i + X2
X2 + X3
X3 + X4
+ X4
Xi
Solution
=
=
=
=
=
1
1
0
0
1
The row reduction proceeds as follows:
1 1
1
0 0 1
0 1
0 0
1 0
0 0
0 0
R2 + R1
R5 + R,
�
R2 <--> R3
R1 + R2
�
R5 + R2
R2 + R3
�
R4 + R3
1
1
0 0 1 1 0
0 1 1 0 0
0 0 1
0
0
0 0
1
0
0
0
0
0 0 1
0
0
0 1 1
0 0 0
1
0
0
0
0
0 0
1
0
0
0 1 1 0
0 0 0 0
0 0 0 0
X4
X4
X4
=
Therefore, we have
X1
Setting the free variable x4
=
Xz
t yields
+
+
X3 +
=
=
1
0
Q
1
0
0
0
0
Section
2.2
Direct Methods for Solving Linear Systems
0 1,
Since t can take on the two values and
�
1
..
In Exercises 1 -8, determine whether the given matrix is in
row echelon form. If it is, state whether it is also in reduced
row echelon form.
!.
3.
5.
3
3
2.
4.
6.
5
7
.
01 -11 ]
[� 0 0 �
00
[ : 0 :J
01
[: 0 :J
001 01
[ � 0 0 -� ]
3
8
.
there are exactly two solutions:
Remark For linear systems over "11..P , there can never be infinitely many
solutions. (Why not?) Rather, when there is more than one solution, the number
of solutions is finite and is a function of the number of free variables and p. (See
Exercise 59.)
Exercises 2 . 2
0 ]
0
[i �
[ � 01 0 � ]
00 0 - 40
[i 0 :J
20
[ j 01 �:
19
In Exercises 9-14, use elementary row operations to reduce
the given matrix to (a) row echelon form and (b) reduced
row echelon form.
0
•
[: : J
[: -!]
13. [ : -- 21 --- 111 i
9.
II
�]
1 2. [� - 41 - 62 : J
2 4
14. [ -- � --6 1 : ]
2
10.
.
-3
[�
-3
15. Reverse the elementary row operations used in
Example 2.9 to show that we can convert
2- 1 -104 - 4
1
1
0
1
[� 0 0 0 -�]24
9
into
Chapter
80
2
Systems of Linear Equations
2 -4 -4
4 0 0
3 2
3 6
�]
16. In general, what is the elementary row operation that
"undoes" each of the three elementary row operations
R; � �' kR;, and R; + kR/
In Exercises 1 7 and 1 8, show that the given matrices are row
equivalent and find a sequence of elementary row operations
that will convert A into B.
1
3 -1
'B =
17. A =
3 4
1
0
2]
[
18.
]
[
A= [ � � i [ � - � ]
0 -1
-1
1
,B =
3
5
2
19. What is wrong with the following "proof" that every
matrix with at least two rows is row equivalent to a
matrix with a zero row?
Perform R 2 + R 1 and R 1 + R2 • Now rows 1 and 2
are identical. Now perform R 2 - R 1 to obtain a
row of zeros in the second row.
20. What is the net effect of performing the following
sequence of elementary row operations on a matrix
(with at least two rows)?
R2 + R , , R , - R2 , R2 + R , , - R ,
21. Students frequently perform the following type of cal­
culation to introduce a zero into a matrix:
However, 3R 2 - 2R 1 is not an elementary row opera­
tion. Why not? Show how to achieve the same result
using elementary row operations.
22. Consider the matrix A =
[�
!]
. Show that any of
the three types of elementary row operations can be
used to create a leading 1 at the top of the first column.
Which do you prefer and why?
23. What is the rank of each of the matrices in Exercises 1 -8?
24. What are the possible reduced row echelon forms of
3 X 3 matrices?
In Exercises 25-34, solve the given system of equations using
either Gaussian or Gauss-Jordan elimination.
26. x - y + z = 0
25. X 1 + 2X2 - 3X3 = 9
2X 1 - X2 + X3 = 0
-x + 3y + z = 5
4x 1 - X2 + x3 = 4
3x + y + 7z = 2
27. X 1 - 3X2 - 2X3 = 0 28. 2w + 3x - y + 4z = 1
3w - x
-X1 + 2X2 + X3 = 0
+ z=1
3w - 4x + y - z = 2
2x1 + 4x2 + 6x3 = 0
29. 2r + s = 3
4r + s = 7
2r + Ss = - 1
30. -X1 + 3X2 - 2X3 + 4X4 = 0
2X 1 - 6X2 + X3 - 2X4 = - 3
X 1 - 3X2 + 4X3 - 8x4 = 2
2
31. � X 1 + X2 - X3 - 6X4
ix , + � Xz
- 3X4 + X5 = - 1
- 4X5 = 8
32. Vlx + y + 2z =
1
Vly - 3z = - V2
-y + Vlz =
33. w + x + 2y + z = 1
w-x- y+z= 0
x+ y
-1
+z= 2
w+x
34. a + b + c + d = 4
a + 2b + 3c + 4d = 1 0
a + 3b + 6c + lOd = 20
a + 4b + lOc + 20d = 35
In Exercises 35-38, determine by inspection (i.e., without
performing any calculations) whether a linear system with
the given augmented matrix has a unique solution, infinitely
many solutions, or no solution. Justify your answers.
0
-2 0
35.
2 -3
36.
1
0
4 -6
2 3 4 O
2 3 4
6 7 8 0
5 4 3
37.
38.
10 1 1 12 0
7 7 7
39. Show that if ad - be -=!= 0, then the system
ax + by = r
ex + dy = s
[:
[i
[�
J [�
has a unique solution.
Section
In Exercises 40-43,for what value(s) of k, if any, will the
systems have (a) no solution, (b) a unique solution, and
(c) infinitely many solutions?
kx + 2y = 3
x + ky = 1
kx + y = 1
2x - 4y =
x - 2y + 3 z = 2
x + y + kz =
x + ky + z =
x+ y+ z=k
2
kx + y + z = - 2
2x - y + 4z = k
Give examples of homogeneous systems of m linear
equations in n variables with m = n and with m > n
40.
42.
44.
445..
467.
-6
41.
43.
that have (a) infinitely many solutions and (b) a
unique solution.
2.2
(c )
Direct Methods for Solving Linear Systems
6
Give an example of three planes, exactly two of
which are parallel (Figure 2. ) .
figure 2 . 6
(d) Give an example of three planes that intersect in a
single point (Figure 2.7).
In Exercises 45 and 46, find the line of intersection of the
given planes.
3x + 2y + z = - 1 and 2x - y + 4z = 5
4x + y + z = 0 and 2x - y + 3z = 2
(a) Give an example of three planes that have a com­
mon line of intersection (Figure 2.4).
p
q
48.p � [ - Hq � [H u � [ _+ � [ - i l
49.p � [�J q � [: J . u � [ + � m
50. p � m. u � [ J. � m
p
51. u
uu
� [ :: l [ :: l
figure 2 . 1
In Exercises 48 and 49, determine whether the lines
x = + su and x = + tv intersect and, if they do, find
their point of intersection.
Figure 2 . 4
(b) Give an example of three planes that intersect in
pairs but have no common point of intersection
(Figure 2.5 ).
let
ond v
Desaibe
all points Q = ( a, b, c) such that the line through
Q with direction vector v intersects the line with
equation x = + su.
Recall that the cross product of vectors and v is a
vector X v that is orthogonal to both and v. (See
Exploration: The Cross Product in Chapter 1 .) If
Figure 2 . 5
81
U
and
F
82
Chapter
2
Systems of Linear Equations
show that there are infinitely many vectors
that simultaneously satisfy u · x = 0 and v · x = 0
and that all are multiples of
U
X
V
[
U 2 V3 - U 3 V2
= U 3 V 1 - U 1 V3
U 1 V2 - U 2 V 1
]
Show that the lines x = p + su and x = q + tv are
skew lines. Find vector equations of a pair of parallel
planes, one containing each line.
In Exercises 53-58, solve the systems of linear equations
over the indicated ZP.
53. x + 2y = 1 over Z 3
x+ y=2
54. x + y
= 1 over Z 2
y+z=O
x
+z= 1
Writi n g Proiect
55. x + y
= 1 over Z 3
y+z=O
x
+z= 1
56. 3 x + 2y = 1 over Z s
x + 4y = 1
57. 3 x + 2y = 1 over Z 7
x + 4y = 1
58. x 1
+ 4x4 = 1 over Z s
=3
X 1 + 2X2 + 4x3
2X 1 + 2X2
+ X4 = 1
=2
X1
+ 3X3
59. Prove the following corollary to the Rank Theorem:
Let A be an m X n matrix with entries in Zp . Any
consistent system of linear equations with coefficient
matrix A has exactly p n - rank(A) solutions over ZP '
60. When p is not prime, extra care is needed in solving
a linear system (or, indeed, any equation) over ZP '
Using Gaussian elimination, solve the following system
over Z6• What complications arise?
2x + 3y = 4
4x + 3y = 2
A History of Gaussian Elimination
As noted in the biographical sketch of Gauss in this section, Gauss did not actually
"invent" the method known as Gaussian elimination. It was known in some form as
early as the third century B . C . and appears in the mathematical writings of cultures
throughout Europe and Asia.
Write a report on the history of elimination methods for solving systems of
linear equations. What role did Gauss actually play in this history, and why is his
name attached to the method?
1. S. Athloen and R. McLaughlin, Gauss-Jordan reduction: A brief history,
American Mathematical Monthly 94 ( 1 987), pp. 1 30- 142.
2. Joseph F. Grear, Mathematicians of Gaussian Elimination, Notices of the AMS,
Vol. 58, No. 6 (20 1 1 ) , pp. 782-792. (Available online at http://www.ams.org/
notices/20 1 1 06/index.html)
3. Roger Hart, The Chinese Roots of Linear Algebra (Baltimore: Johns Hopkins
University Press, 20 1 1 ) .
4. Victor J. Katz, A History of Mathematics: An Introduction (Third Edition)
(Reading, MA: Addison Wesley Longman, 2008).
cAs
L i e s My C o mp uter Told M e
Computers and calculators store real numbers in floating-point form. For example,
2001 is stored as 0.200 1 X 10 4 , and - 0.00063 is stored as - 0.63 X 10 - 3 . In general,
the floating-point form of a number is :±:.M X lO k , where k is an integer and the
mantissa M is a (decimal) real number that satisfies 0. 1 ::::: M < 1 .
The maximum number of decimal places that can be stored in the mantissa depends
on the computer, calculator, or computer algebra system. If the maximum number of
decimal places that can be stored is d, we say that there are d significant digits. Many
calculators store 8 or 12 significant digits; computers can store more but still are subject
to a limit. Any digits that are not stored are either omitted (in which case we say that the
number has been truncated) or used to round the number to d significant digits.
For example,-rr = 3.141592654, and its floating-point form is 0.3 141592654 X 10 1 .
In a computer that truncates to five significant digits, 1T would be stored as 0.3 1415 X
10 1 (and displayed as 3.1415); a computer that rounds to five significant digits would
store 1T as 0.3 1416 X 10 1 (and display 3.1416). When the dropped digit is a solitary
5, the last remaining digit is rounded so that it becomes even. Thus, rounded to two
significant digits, 0.735 becomes 0.74 while 0.725 becomes 0.72.
Whenever truncation or rounding occurs, a roundoff error is introduced, which
can have a dramatic effect on the calculations. The more operations that are per­
formed, the more the error accumulates. Sometimes, unfortunately, there is nothing
we can do about this. This exploration illustrates this phenomenon with very simple
systems of linear equations.
1 . Solve the following system of linear equations exactly (that is, work with
rational numbers throughout the calculations).
x+ y=O
01 = 1
X + 8sooY
2. As a decimal, ��b = 1 .00 125, so, rounded to five significant digits, the system
becomes
x+
y=O
x + 1 .00 1 2y = 1
Using your calculator or CAS, solve this system, rounding the result of every calcula­
tion to five significant digits.
83
3. Solve the system two more times, rounding first to four significant digits and
then to three significant digits. What happens?
4. Clearly, a very small roundoff error (less than or equal to 0.00 125) can re­
sult in very large errors in the solution. Explain why geometrically. (Think about the
graphs of the various linear systems you solved in Problems 1 - 3 .)
Systems such as the one you just worked with are called ill-conditioned. They
are extremely sensitive to roundoff errors, and there is not much we can do about it.
We will encounter ill-conditioned systems again in Chapters 3 and 7. Here is another
example to experiment with:
4.552x + 7.083y
1 .73 1x + 2.693y
=
=
1 .93 1
2.00 1
Play around with various numbers of significant digits to see what happens, starting
with eight significant digits (if you can).
Partial P ivoting
In Exploration: Lies My Computer Told Me, we saw that ill-conditioned linear sys­
tems can cause trouble when roundoff error occurs. In this exploration, you will dis­
cover another way in which linear systems are sensitive to roundoff error and see that
very small changes in the coefficients can lead to huge inaccuracies in the solution.
Fortunately, there is something that can be done to minimize or even eliminate this
problem (unlike the problem with ill-conditioned systems).
84
1 . (a) Solve the single linear equation 0.0002 lx = 1 for x.
(b) Suppose your calculator can carry only four decimal places. The equa­
tion will be rounded to 0.0002x = 1 . Solve this equation.
The difference between the answers in parts (a) and (b) can be thought of as the
effect of an error of 0.0000 1 on the solution of the given equation.
2. Now extend this idea to a system of linear equations.
(a) With Gaussian elimination, solve the linear system
0.400x + 99.6y = 1 00
75.3x - 45. 3y = 30.0
using three significant digits. Begin by pivoting on 0.400 and take each
calculation to three significant digits. You should obtain the "solution" x =
- 1 .00, y = 1 .0 1 . Check that the actual solution is x = 1 .00, y = 1 .00. This is a
huge error-200% in the x value! Can you discover what caused it?
(b) Solve the system in part (a) again, this time interchanging the two equa­
tions (or, equivalently, the two rows of its augmented matrix) and pivoting
on 75.3. Again, take each calculation to three significant digits. What is the
solution this time?
The moral of the story is that, when using Gaussian or Gauss-Jordan elimination
to obtain a numerical solution to a system of linear equations (i.e., a decimal approxi­
mation), you should choose the pivots with care. Specifically, at each pivoting step,
choose from among all possible pivots in a column the entry with the largest absolute
value. Use row interchanges to bring this element into the correct position and use it to
create zeros where needed in the column. This strategy is known as partial pivoting.
3.
Solve the following systems by Gaussian elimination, first without and then
with partial pivoting. Take each calculation to three significant digits. (The exact
solutions are given.)
(a)
O.- O10.O l2xx 0.1.99050yy - 51.0.000 -l Ox3x - 2.09y7y 6z 73.91
5x - y [5zx l 6[ 0.00
]
.
0
0
5
0
0
y
1.
[;] [ 1.00 ]
z - 1.00
+
+
(b)
=
=
Exact solution:
+
+
+
Exact solution:
=
=
=
=
C o unting O p e r at i o n s : An I n t r o d u c t i o n
to t h e An alys i s of Algorith m s
Gaussian and Gauss-Jordan elimination are examples of algorithms: systematic pro­
cedures designed to implement a particular task-in this case, the row reduction of
the augmented matrix of a system of linear equations. Algorithms are particularly well
suited to computer implementation, but not all algorithms are created equal. Apart
from the speed, memory, and other attributes of the computer system on which they
are running, some algorithms are faster than others. One measure of the so-called
of an algorithm (a measure of its efficiency, or ability to perform its task in a
reasonable number of steps) is the number of basic operations it performs as a func­
tion of the number of variables that are input.
Let's examine this proposition in the case of the two algorithms we have for
solving a linear system: Gaussian and Gauss-Jordan elimination. For our pur­
poses, the basic operations are multiplication and division; we will assume that
all other operations are performed much more rapidly and can be ignored. (This
is a reasonable assumption, but we will not attempt to justify it.) We will consider
only systems of equations with
coefficient matrices, so, if the coefficient
matrix is n X n, the number of input variables is n. Thus, our task is to find the
number of operations performed by Gaussian and Gauss-Jordan elimination as
a function of n. Furthermore, we will not worry about special cases that may
arise, but rather establish the
that can arise-when the algorithm takes
as long as possible. Since this will give us an estimate of the time it will take a
computer to perform the algorithm (if we know how long it takes a computer to
perform a single operation) , we will denote the number of operations performed
by an algorithm by T ( n ) . We will typically be interested in T ( n ) for large values of
n, so comparing this function for different algorithms will allow us to determine
which will take less time to execute.
com­
plexity
0
00
j
780-850)
Abu Ja'far Muhammad ibn Musa
al-Khwarizmi (c.
was
a Persian mathematician whose
book Hisab al-jabr w'al muqabalah
(c.
described the use of Hindu­
Arabic numerals and the rules
of basic arithmetic. The second
word of the book's title gives rise
to the English word algebra, and
the word algorithm is derived from
al-Khwarizmi's name.
825)
9
square
worst case
1.
Consider the augmented matrix
[A l h J
9
4
85
I
Count the number of operations required to bring [A b] to the row echelon
form
4]
[:
2 3
1 -1 0
0 1 1
(By "operation" we mean a multiplication or a division.) Now count the number
of operations needed to complete the back substitution phase of Gaussian elimi­
nation. Record the total number of operations.
2. Count the number of I operations needed to perform Gauss-Jordan
elimination-that is, to reduce [A b ] to its reduced row echelon form
(where the zeros are introduced into each column immediately after the leading 1 is
created in that column) . What do your answers suggest about the relative efficiency
of the two algorithms?
We will now attempt to analyze
I the algorithms in a general, systematic way. Sup­
pose the augmented matrix
[A
b ] arises from a linear system with n equations and
I
n variables; thus, [A b ] is n X (n + 1 ):
I
[A b ]
=
[
all
�
a 1
an ]
We will assume that row interchanges are never needed-that we can always create a
leading 1 from a pivot by dividing by the pivot.
3. (a) Show that n operations are needed to create the first leading 1 :
�
86
l
�
a.,
a 12
aln b,
a2 1
a 22
a2 n
an l
a n2
a nn b n
2
l l
�
l�
:,
l
*
*
a I
a z2
a2 n
an l
an2
a nn b n
(Why don't we need to count an operation for the creation of the leading 1 ?) Now
show that n operations are needed to obtain the first zero in column 1 :
la�,
*
*
*
*
an2
a nn
iJ
....,..
(Why don't we need to count an operation for the creation of the zero itself?) When
the first column has been "swept out;' we have the matrix
[� : : :]
* *
0 *
Show that the total number of operations needed up to this point is n + (n - 1) n.
(b) Show that the total number of operations needed to reach the row
echelon form
is
[!
*
*
* *
0
1 *
'
]
[ n + (n - l)n ] + [ (n - 1) + (n - 2)(n - 1 ) ] + [ (n - 2) + (n - 3)(n - 2 ) ]
+ . . . + [2 + 1 . 2 ] + 1
which simplifies to
n 2 + (n - 1 ) 2 +
·
·
·
+ 22 + 1 2
( c) Show that the number of operations needed to complete the back substi­
tution phase is
1 +2+
·
·
·
+ (n - 1)
(d) Using summation formulas for the sums in parts (b) and (c) (see
Exercises 51 and 52 in Section 2.4 and Appendix B), show that the total number of
operations, T(n), performed by Gaussian elimination is
T(n)
=
tn 3 + n 2 - t n
Since every polynomial function is dominated by its leading term fo r large values of
the variable, we see that T (n) = t n 3 for large values of n.
4. Show that Gauss-Jordan elimination has T(n) = ! n 3 total operations if we
create zeros above and below the leading l s as we go. (This shows that, for large
systems of linear equations, Gaussian elimination is faster than this version of Gauss­
Jordan elimination.)
81
88
Chapter
2
Systems of Linear Equations
�
S p a n n i n g se1s a n d l i n e a r I n d e p e n d e n c e
The second of the three roads in our "trivium" is concerned with linear combina­
tions of vectors. We have seen that we can view solving a system of linear equations
as asking whether a certain vector is a linear combination of certain other vectors.
We explore this idea in more detail in this section. It leads to some very important
concepts, which we will encounter repeatedly in later chapters.
Spanning Sets of Vectors
We can now easily answer the question raised in Section 1 . 1 : When is a given vector
a linear combination of other given vectors?
Exa m p l e 2 . 1 8
(o)
(b)
fa
fa
th, wdoc
m
m
m [ � �}
m [ � �}
' linm rnmbinotion of the vodorn
' linemomb;mtion of the vectorn
Solulion
and
and
(a) We want to find scalars x and y such that
Expanding, we obtain the system
whose augmented matrix is
x- y= 1
y=2
3x - 3y = 3
[� �� �]
(Observe that the columns of the augmented matrix are just the given vectors; notice
the order of the vectors-in particular, which vector is the constant vector.)
The reduced echelon form of this matrix is
[� : �]
�
(Verify this.) So the solution is x = 3, y
combination is
2, and the corresponding linear
Section
2. 3
Spanning Sets and Linear Independence
89
(b) Utilizing our observation in part (a), we obtain a linear system whose augmented
matrix is
which reduces to
m [J
;eve.ling that th, 'Y'"m h"' no ,oJotion. Th°', in thi, ""·
bination of
and
[ �]
4
i' not a linm com-
The notion of a spanning set is intimately connected with the solution of linear
systems. Look back at Example 2 . 1 8. There we saw that a system with augmented
matrix [A I b] has a solution precisely when b is a linear combination of the columns
of A. This is a general fact, summarized in the next theorem.
Theorem 2 . 4
A system of linear equations with augmented matrix [A I b] is consistent if and
only if b is a linear combination of the columns of A.
Let's revisit Example 2.4, interpreting it in light of Theorem 2.4.
(a) The system
x-y= l
x+y=3
has the unique solution x = 2, y = 1 . Thus,
See Figure 2.S(a).
(b) The system
x- y=2
2 x - 2y = 4
has infinitely many solutions of the form x = 2 + t, y = t. This implies that
for all values of t. Geometrically, the vectors
[ � ], [ = � ], [! ]
and
are all parallel and
so all lie along the same line through the origin [see Figure 2.S(b) ] .
Chapter
90
2
Systems of Linear Equations
y
y
y
5
5
5
4
4
-2
2
-1
3
-1
-2
-2
-3
-3
( a)
-3
(b)
(c)
Figure 2 . 8
(c) The system
x-y= I
x-y= 3
has no solutions, so there are no values of x and y that satisfy
In this case,
[�] [=�]
and
are parallel, but
through the origin [see Figure 2.S(c) ] .
[ �]
does not lie along the same line
We will often b e interested in the collection of all linear combinations of a given
set of vectors.
D e f i n i t i o n If S = {v1 , v2 , . . . , vk} is a set of vectors in u;g n , then the set of all
linear combinations of v1 , v2 , . . . , vk is called the span of v1 , v2 , . . . , vk and is de, vk) or span(S ). If span(S ) = W, then S is called a span­
noted by span(v1 , v2 ,
ning set for u;g n .
.
Exa m p l e 2 . 1 9
Show that u;g z = span
Solulion
•
( [ _ � l [ �] ) .
We need to show that an arbitrary vector
combination of
1 [�] [ : ]
=
.
[ _ � ] [� }
and
[:]
can be written as a linear
that is, we must show that the equation x
[ _�] +
can always be solved for x and y (in terms of a and b ), regardless of the
values of a and b.
Section
The augmented matrix is
2. 3
Spanning Sets and Linear Independence
91
[ _ � � I �] '
1] [ 1 ]
and row reduction produces
b
3 b �1 - 1 3
0 7 a + 2b
1 a
at which point it is clear that the system has a (unique) solution. (Why?) If we con­
tinue, we obtain
---'-----+
b R1
] [ I
[ 1
R 1 - 3R, - 1 O (b - 3 a )/ 7
b
3
�
0 1 (a + 2b)/7
0 1 (a + 2b)/7
-1
]
from which we see that x = ( 3a - b)/7 and y = (a + 2b)/7. Thus, for any choice of
a and b, we have
�
(Check this.)
Remark
( [ _�l [�l [�] } [:l
[ _ �] y [�] [ : l
[ _ �] [�]
It is also true that IR 2
find x and y such that x
[�] [:] .
=
+
span
=
If, given
we can
then we also have x
+y
+
In fact, any set of vectors that contains a spanning set for IR 2 will also be
a spanning set for IR 2 (see Exercise 20).
The next example is an important (easy) case of a spanning set. We will encounter
versions of this example many times.
0
Exa m p l e 2 . 2 0
=
Let e , , e, . and e; be the standMd unit vedocs in HI'. Then I°' any mtn<
=
[;] � x[ � ] m m �
+y
+z
xe , + ye, + ze.,
[;J
Thus, IR 3 span(e 1 , e2 , e 3 ) .
You should have n o difficulty seeing that, i n general, !R n = span(e 1 , e2 ,
.
•
.
we haw
, en
4
When the span of a set of vectors in !R n is not all of !R n , it is reasonable to ask for a
description of the vectors' span.
Exa m p l e 2 . 2 1
92
Chapter
2
Systems of Linear Equations
z
m [J
Solulion
and
m [ �: ]
[�l m { ]
m [J
Thinking geometrically, we can see that the set of all linear combinations of
i' iu't the plane thrnugh the migm wiili
and
�
vecto" ( Figme 2. 9). The vectm equotion of this pfane is
which i*t anothe; way of "ying that
Figure 2 . 9
Two nonparallel vectors span a
plane
[� l
is m th, ,pan of
OS
s
dimtion
+
and
Suppose we want to obtain the general equation of this plane. There are several
ways to proceed. One is to use the fact that the equation ax + by + cz = 0 must be
satisfied by the points ( 1 , 0, 3) and ( - 1, 1, - 3) determined by the direction vectors.
Substitution then leads to a system of equations in a, b, and c. (See Exercise 1 7.)
Another method is to use the system of equations arising from the vector equation:
s- t=x
t=y
3s - 3t = z
If we row reduce the augmented matrix, we obtain
N�� we know that this ry'1em is consistent, since
[ _�]
[; ]
is in the span of
by ossumption. So we must have z - 3x � 0 (o< 3x -
z�
[�]
and
0, in mo<e standa<d
form), giving us the general equation we seek.
Remark
product
A normal vector to the plane in this example is also given by the cross
Linear Independence
ill
E"mple 2 18, wdound that 3
m [ �: l m
+2
�
Let's abb<e�ate thi"qua­
tion as 3u + 2v = w. The vector w "depends" on u and v in the sense that it is a
linear combination of them. We say that a set of vectors is linearly dependent if one
Section
2. 3
Spanning Sets and Linear Independence
93
of them can be written as a linear combination of the others. Note that we also have
�v + t w and v = � u + ! w. To get around the question of which vector to
express in terms of the rest, the formal definition is stated as follows:
u=
-
-
Definition
scalars c 1 , c2 ,
A set of vectors v1 , v2 ,
, vk is linearly dependent if there are
, ck, at least one of which is not zero, such that
.
.
•
•
•
.
A set of vectors that is not linearly dependent is called linearly independent.
R e m arks
In the definition of linear dependence, the requirement that at least one of the
scalars c 1 , c2 , , ck must be nonzero allows for the possibility that some may be zero.
In the example above, u, v, and w are linearly dependent, since 3u + 2v - w = 0 and,
in fact, all of the scalars are nonzero. On the other hand,
•
so
.
.
•
[!l [� l [�]
are linearly dependent, since at least one (in fact, two) of the
and
three scalars 1 , - 2, and 0 is nonzero. (Note that the actual dependence arises simply
from the fact that the first two vectors are multiples.) (See Exercise 44.)
•
Since Ov1 + Ov2 + . . . + Ovk = 0 for any vectors v1 , v2 , . . . , vk, linear de­
pendence essentially says that the zero vector can be expressed as a nontrivial linear
combination of v1 , v2 , . . . , vk · Thus, linear independence means that the zero vector
can be expressed as a linear combination of v1 , v2 ,
, vk only in the trivial way: c 1 v1
+ c2v2 + + ckvk = 0 only if c 1 = 0, c2 = 0, . . . , ck = 0.
•
·
·
•
.
·
The relationship between the intuitive notion of dependence and the formal defi­
nition is given in the next theorem. Happily, the two notions are equivalent!
Theorem 2 . 5
Vectors v1 , v2 , , vm in !R n are linearly dependent if and only if at least one of the
vectors can be expressed as a linear combination of the others.
•
.
.
Proof
If one of the vectors-say, v1 -is a linear combination of the others, then
there are scalars c2 , , c m such that v1 = c2v2 + . . . + c mvm . Rearranging, we obtain
v1 c2v2
cmvm = 0, which implies that v1 , v2 , . . . , vm are linearly dependent,
since at least one of the scalars (namely, the coefficient 1 of v1 ) is nonzero.
Conversely, suppose that v1 , v2 ,
, vm are linearly dependent. Then there are
scalars c 1 , c2 , . . . , cm , not all zero, such that c 1 v 1 + c2v2 + . . . + c mvm = 0. Suppose
c 1 -=!= 0. Then
.
-
-
·
·
·
•
•
-
•
.
.
and we may multiply both sides by 1/ c 1 to obtain v1 as a linear combination of the
other vectors:
v1
=
()
- � v2
C1
-
·
·
·
-
( cCm1 )
vm
94
Chapter
2
Systems of Linear Equations
It may appear as if we are cheating a bit in this proof. After all, we cannot
be sure that v1 is a linear combination of the other vectors, nor that c1 is nonzero.
However, the argument is analogous for some other vector V; or for a different scalar
cj . Alternatively, we can just relabel things so that they work out as in the above proof.
In a situation like this, a mathematician might begin by saying, "without loss of gen­
erality, we may assume that v1 is a linear combination of the other vectors" and then
proceed as above.
Note
Exa m p l e 2 . 2 2
Any set of vectors containing the zero vector is linearly dependent. For if 0, v2, , vm
are in IJ�r, then we can find a nontrivial combination of the form c10 + c2v2 + . . . +
Cm Vm = 0 by setting c1 = 1 and c2 = c3 = . . . = C m = 0 .
Exa m p l e 2 . 2 3
Determine whether the following sets of vectors are linearly independent:
.
(a)
(c)
[�] [ �]
[ -i ] . [ J
•
-
and
nl
In answering any question of this type, it is a good idea to see if you can
determine by inspection whether one vector is a linear combination of the others. A
little thought may save a lot of computation!
(a) The only way two vectors can be linearly dependent is if one is a multiple of
the other. (Why?) These two vectors are clearly not multiples, so they are linearly
independent.
(b) There is no obvious dependence relation here, so we try to find scalars c1, c2, c3
such that
Solution
�
and
.
The corresponding linear system is
C1
C1
and the augmented matrix is
+
+ Cz
[i
C3
=0
=O
0 1 O
0 0
1 0
J
Once again, we make the fundamental observation that the columns of the coefficient
matrix are just the vectors in question!
Section
2. 3
Spanning Sets and Linear Independence
OJ
The reduced row echelon form is
�
�
[i
95
0 0
1 0 0
0 1 0
(check this), so c 1 = 0, c2 = 0, c 3 = 0. Thus, the given vectors are linearly independent.
( c) A little reflection reveals that
so the three vectors are linearly dependent. [Set up a linear system as in part (b) to
check this algebraically.]
( d) Once again, we observe no obvious dependence, so we proceed directly to reduce
a homogeneous linear system whose augmented matrix has as its columns the given
vectors:
[ � � : � 1 �I [ �
1 1 0
-1 2 0
0 -1 2 0
0 -1 2 0
If we let the scalars be c 1 , c2 , and c 3 , we have
c1 +
3c 3
c2 - 2c3
=
=
l �: �:
�
_
R
,
[ -� �1
1 0
0 1
0 0
0 0
0
0
from which we see that the system has infinitely many solutions. In particular, there
must be a nonzero solution, so the given vectors are linearly dependent.
If we continue, we can describe these solutions exactly: c 1 = - 3c3 and c2 = 2c3 .
Thus, for any nonzero value of c3 , we have the linear dependence relation
�
(Once again, check that this is correct.)
We summarize this procedure for testing for linear independence as a theorem.
Theorem 2 . 6
Let v 1 , v2 , . . . , vm be (column) vectors in IK£ n and let A be the n X m matrix
[v1 v2 vm l with these vectors as its columns. Then v1 , v2 , , vm are linearly
dependent if and only if the homogeneous linear system with augmented matrix
[A I O ] has a nontrivial solution.
·
·
.
·
[ :: ]
•
•
, vm are linearly dependent ifand only if there are scalars c 1 , c2 , , cm ,
not all zero, such that c 1 v1 + c2v2 + . . . + cmvm = 0. By Theorem 2.4, this is equivalent
Proof
v1 , v2 ,
. . •
to sa'.i� g that the nonzero vector
matnx is [v1 v2 vm I OJ .
.
•
.
.
·
cm
• • .
is a solution of the system whose augmented
96
Chapter
2
Systems of Linear Equations
Exa m p l e 2 . 2 4
The standard unit vectors e 1 , e2 , and e 3 are linearly independent in IR 3 , since the sys­
tem with augmented matrix [ e1 e2 e3 I OJ is already in the reduced row echelon form
[:
0 0 O
1 0 0
0 1 0
J
and so clearly has only the trivial solution. In general, we see that e1 , e2 , . . . , en will be
linearly independent in !R n .
Performing elementary row operations on a matrix constructs linear combina­
tions of the rows. We can use this fact to come up with another way to test vectors for
linear independence.
Exa m p l e 2 . 2 5
Consider the three vectors of Example 2.23(d) as row vectors:
[ 1 , 2, 0] , ( 1 , 1 , - 1 ] , and [ 1 , 4, 2]
We construct a matrix with these vectors as its rows and proceed to reduce it to eche­
lon form. Each time a row changes, we denote the new row by adding a prime symbol:
[:
2
1
4
2
-1
2
From this we see that
O
-1
2
J
�
R; = R; + ZR;
l -�J
[
2
0 -1
0
0
or, in terms of the original vectors,
- 3 ( 1 , 2, 0 ] + 2 [ 1 , 1 , - 1 ] + ( 1 , 4, 2 ] = [ O, O, O ]
[Notice that this approach corresponds to taking c 3 = 1 in the solution to
Example 2.23(d).]
Thus, the rows of a matrix will be linearly dependent if elementary row opera­
tions can be used to create a zero row. We summarize this finding as follows:
Theorem 2 . 1
Let v, , v, . . . . . vm be ( rnw) ve<to" in
these vectors as its rows. Then v1 , v2 ,
rank(A) < m .
and kt A be t he m X n motrix
�"
•
.
•
[J
with
, vm are linearly dependent if and only if
Proof Assume that v1 , v2 ,
, vm are linearly dependent. Then, by Theorem 2.2,
at least one of the vectors can be written as a linear combination of the others.
•
•
.
Section
2. 3
Spanning Sets and Linear Independence
91
We relabel the vectors, if necessary, so that we can write vm = c 1 v1 + c 2v2 +
+
cm - iYm - l · Then the elementary row operations R m - c 1 R 1 , R m - c 2 R2 ,
,
R m - cm - i R m - I applied to A will create a zero row in row m. Thus, rank(A) < m.
Conversely, assume that rank(A) < m. Then there is some sequence of row opera­
tions that will create a zero row. A successive substitution argument analogous to that
used in Example 2.25 can be used to show that 0 is a nontrivial linear combination of
v1 , v2 , , vm . Thus, v1 , v2 , , vm are linearly dependent.
·
•
•
.
.
.
•
·
·
•
•
.
In some situations, we can deduce that a set of vectors is linearly dependent with­
out doing any work. One such situation is when the zero vector is in the set (as in
Example 2.22). Another is when there are "too many" vectors to be independent. The
following theorem summarizes this case. (We will see a sharper version of this result
in Chapter 6.)
Theorem 2 . 8
Any set of m vectors in !R n is linearly dependent if m > n.
Let v1 , v2 ,
, vm be (column) vectors in !R n and let A be the n X m matrix
[v1 v2 vm l with these vectors as its columns. By Theorem 2.6, v1 , v2 , , vm are lin­
early dependent if and only if the homogeneous linear system with augmented matrix
[A I O] has a nontrivial solution. But, according to Theorem 2.6, this will always be the
case if A has more columns than rows; it is the case here, since number of columns m
is greater than number of rows n.
Proof
.
.
Exa m p l e 2 . 2 6
...
I
•
•
•
•
.
The vectors
[ � ], [ !] [ � ]
, and
.
.
are linearly dependent, since there cannot be more
than two linearly independent vectors in IR 2 • (Note that if we want to find the ac­
tual dependence relation among these three vectors, we must solve the homogeneous
system whose coefficient matrix has the given vectors as columns. Do this! )
Exercises 2 . 3
In Exercises 1 -6, determine if the vector v is a linear combi­
nation of the remaining vectors.
� m. u, m. u, [ : J
u, � m
3[ .2 ] [ 0.4 ] [ 31..44 ]
-[ �:� ]
5. v
GAS
6. V =
U3
�
�
=
2.0 , U 1
- 2.6
- 1 .0
1 .0
=
, ll2
48
.
=
-
64
.
,
Chapter
98
2
Systems of Linear Equations
In Exercises 7 and 8, determine if the vector b is in the span
of the columns of the matrix A.
7. A =
8. A
=
[� ! l
[� !
[:J
�], [ :l
([�] [ �])
( [ � l [ �] ).
'P""( [ H [J [ : ] )
'P"" ( [J m. [ _ : ] )
b=
9 10 1 1
b=
9 . Show that IR 2 = span
1 1 . Show thot � ' �
12. Show thot � ' �
12
, _
10. Show that IR 2 = span
[Hint: We know that IR 3 = span(e 1 , e2 , e3 ).]
.
Use the method of Example 2.23 and Theorem 2. 6 to deter­
mine if the sets of vectors in Exercises 22-31 are linearly in­
dependent. If, for any of these, the answer can be determined
by inspection (i.e., without calculation), state why. For any
sets that are linearly dependent, find a dependence relation­
ship among the vectors.
_
::r!fFr ::�fa [:J
m. u i
[ _�irn [ -:i
;�:�;�; �� �rt��� � ��� �;���:� �;::! � ��:
15.
16.
r
,
f
1
n
f
r
i
,
1
form ax + by + cz = 0. Solve for a, b, and c.
18. Prove that u, v, and w are all in span(u, v, w) .
19. Prove that u, v, and w are all in span(u, u + v, u +
� ; ;:�
ve that if u 1 , . . . , um are vectors in !R n , S =
{ U 1 , U2 , . . . , Uk } , an d T = { U 1 , . . . , Uk , Uk + !> . . . ,
um }, then span (S ) � span ( T) . [Hint: Rephrase this
question in terms of linear combinations.]
(b) Deduce that if [R n = span (S), then [R n = span ( T)
also.
2 1 . (a) Suppose that vector w is a linear combination
of vectors u 1 , . . . , uk and that each u; is a linear
combination of vectors v1 , . . . , vm . Prove that w is
a linear combination ofv1 , . . . , vm and therefore
span(u 1 , . . . , uk) � span(v 1 , . . . , vm ) .
20 . a
[ : ] . m. [ - � ]
[ :H�lE
m. [ !J Hl
[ -; i . [ -!J [ : J m [ ;] . [ H [ � :
23.
In Exercises 1 3- 1 6, describe the span of the given vectors
d (b) alg,bm
17.
(b) In part (a), suppose in addition that each vj is also
a linear combination of u 1 , . . . , uk. Prove that
span(u 1 , . . . , uk) = span(v1 , . . . , vm ) .
(c ) Use the result o f part (b) t o prove that
24 •
26.
28
25
27.
r�H �J u1
[ l [ -1 1 [ l [ o l
�[ l [ � l [ � l [ : l
3
�[ l [ - � 1 [ � l , [ -- 3� 1
29 .
30 ·
31.
l
-1
,
1
0
l
1
1
0
,
,
-1
1
0
1
1
-1
0 ' 2 ' 2 ' 2
1
1
1
1
1
l
-
,
,
-l
l
-l
l
In Exercises 32- 41, determine if the sets of vectors in the
given exercise are linearly independent by converting the
Section
vectors to row vectors and using the method of Example 2.25
and Theorem 2.7. For any sets that are linearly dependent,
find a dependence relationship among the vectors.
32. Exercise 22
33. Exercise 23
34. Exercise 24
35. Exercise 25
36. Exercise 26
37. Exercise 27
38. Exercise 28
39. Exercise 29
40. Exercise 30
41. Exercise 3 1
42. (a) If the columns of an n X n matrix A are linearly in­
dependent as vectors in !R n , what is the rank of A?
Explain.
(b) If the rows of an n X n matrix A are linearly inde­
pendent as vectors in !R n , what is the rank of A?
Explain.
43. (a) If vectors u, v, and w are linearly independent, will
u + v, v + w, and u + w also be linearly indepen dent? Justify your answer.
11 f
2.4
Applications
99
(b) If vectors u, v, and w are linearly independent, will
u - v, v - w, and u - w also be linearly indepen­
dent? Justify your answer.
44. Prove that two vectors are linearly dependent if
and only if one is a scalar multiple of the other.
[Hint: Separately consider the case where one of the
vectors is O.]
45. Give a "row vector proof" of Theorem 2.8.
46. Prove that every subset of a linearly independent set is
linearly independent.
47. Suppose that S = {v1 , , vk> v} is a set of vectors in
some !R n and that v is a linear combination of v 1 , . . . ,
vk. If S' = {v1 , . . . , vd, prove that span (S ) = span (S').
[Hint: Exercise 2 l (b) is helpful here.]
48. Let {v1 , . . . , vd be a linearly independent set of vec­
tors in !R n , and let v be a vector in !R n . Suppose that
v = c 1 v1 + c 2 v2 + · · · + ck vk with c 1 * 0. Prove that
{v, v2 , , vd is linearly independent.
.
.
•
•
.
.
A p p l icati o n s
There are too many applications of systems of linear equations to do them justice in a
single section. This section will introduce a few applications, to illustrate the diverse
settings in which they arise.
Allocation o f Resources
A great many applications of systems of linear equations involve allocating limited
resources subject to a set of constraints.
Exa m p l e 2 . 2 1
A biologist has placed three strains of bacteria (denoted I, II, and III) in a test tube,
where they will feed on three different food sources (A, B, and C). Each day 2300 units
of A, 800 units of B, and 1 500 units of C are placed in the test tube, and each bacte­
rium consumes a certain number of units of each food per day, as shown in Table 2.2.
How many bacteria of each strain can coexist in the test tube and consume all of the
food?
Ta b l e 2 . 2
Food A
Food B
Food C
Bacteria
Strain I
2
1
1
Bacteria
Strain II
2
2
3
Bacteria
Strain III
4
0
100
Chapter
2
Systems of Linear Equations
Let x 1 , x 2 , and x3 b e the numbers of bacteria of strains I , II, and III,
respectively. Since each of the x 1 bacteria of strain I consumes 2 units of A per day,
strain I consumes a total of 2x 1 units per day. Similarly, strains II and III consume a
total of 2x 2 and 4x 3 units of food A daily. Since we want to use up all of the 2300 units
of A, we have the equation
2x 1 + 2x2 + 4x3 = 2300
Solulion
Likewise, we obtain equations corresponding to the consumption of B and C:
= 800
X 1 + 2X 2
X 1 + 3X 2 + X3 = 1 500
Thus, we have a system of three linear equations in three variables. Row reduction of
the corresponding augmented matrix gives
[:
2 4 2300
2 0 800
3 1 1 500
l
] [
�
0 0 1 00
0
0 350
0 0 1 350
]
Therefore, x 1 = 1 00, x 2 = 350, and x 3 = 350. The biologist should place 100 bacteria
of strain I and 350 of each of strains II and III in the test tube if she wants all the food
to be consumed.
Exa m p l e 2 . 2 8
Repeat Example 2.27, using the data on daily consumption of food (units per day)
shown in Table 2.3. Assume this time that 1 500 units of A, 3000 units of B, and 4500
units of C are placed in the test tube each day.
Ta b l e 2 . 3
Bacteria
Strain II
Bacteria
Strain I
Food A
Food B
Food C
2
3
Bacteria
Strain III
1
3
5
Let x 1 , x2 , and x3 again be the numbers of bacteria of each type. The aug­
mented matrix for the resulting linear system and the corresponding reduced echelon
form are
Solulion
[:
� ���� i
1
2
3 5 4500
�
�[ � O J
-1
2 1 500
0 0
0
0
We see that in this case we have more than one solution, given by
0
- X3 =
X 2 + 2X 3 = 1 500
Letting x3 = t, we obtain x 1 = t, x2 = 1 500 - 2t, and x3 = t. In any applied problem,
we must be careful to interpret solutions properly. Certainly the number of bacteria
X1
Section
2.4
Applications
101
cannot be negative. Therefore, t 2: 0 and 1 500 2t 2: 0. The latter inequality implies
that t :::::: 750, so we have 0 :::::: t :::::: 750. Presumably the number of bacteria must be a
whole number, so there are exactly 75 1 values of t that satisfy the inequality. Thus, our
75 1 solutions are of the form
-
one for each integer value of t such that 0 :::::: t :::::: 750. (So, although mathematically
this system has infinitely many solutions, physically there are only finitely many.)
Balancing Chemical Equations
.+
When a chemical reaction occurs, certain molecules (the reactants) combine to form
new molecules (the products). A balanced chemical equation is an algebraic equation
that gives the relative numbers of reactants and products in the reaction and has the
same number of atoms of each type on the left- and right-hand sides. The equation is
usually written with the reactants on the left, the products on the right, and an arrow
in between to show the direction of the reaction.
For example, for the reaction in which hydrogen gas (H 2 ) and oxygen (0 2 ) com­
bine to form water (H 2 0), a balanced chemical equation is
2H 2 + 0 2 -----+ 2H 2 0
indicating that two molecules of hydrogen combine with one molecule of oxygen to
form two molecules of water. Observe that the equation is balanced, since there are
four hydrogen atoms and two oxygen atoms on each side. Note that there will never
be a unique balanced equation for a reaction, since any positive integer multiple of
a balanced equation will also be balanced. For example, 6H 2 + 30 2 -----+ 6H 2 0 is also
balanced. Therefore, we usually look for the simplest balanced equation for a given
reaction.
While trial and error will often work in simple examples, the process of balancing
chemical equations really involves solving a homogeneous system oflinear equations,
so we can use the techniques we have developed to remove the guesswork.
Exa m p l e 2 . 2 9
The combustion of ammonia (NH 3 ) in oxygen produces nitrogen (N 2 ) and water.
Find a balanced chemical equation for this reaction.
Solution
If we denote the numbers of molecules of ammonia, oxygen, nitrogen, and
water by w, x, y, and z, respectively, then we are seeking an equation of the form
wNH 3 + x0 2 -----+ yN 2 + zH 2 0
Comparing the numbers of nitrogen, hydrogen, and oxygen atoms in the reactants
and products, we obtain three linear equations:
Nitrogen: w = 2y
Hydrogen: 3w = 2z
Oxygen: 2x = z
Rewriting these equations in standard form gives us a homogeneous system of three
linear equations in four variables. [Notice that Theorem 2.3 guarantees that such a
102
Chapter
2
Systems of Linear Equations
system will have (infinitely many) nontrivial solutions.] We reduce the corresponding
augmented matrix by Gauss-Jordan elimination.
w
3w
- 2y
2x
=
- 2z
=
-
=
z
0O -----+ [
0
1 0
3 0
0 2
-2
0
0
- � �] -----+ [ � �
0
0
0 0 1
-1 0
0]
-t
-! 0
-t 0
Thus, w = t z, x = ! z, and y = t z . The smallest positive value of z that will produce
integer values for all four variables is the least common denominator of the fractions
t, !, and t-namely, 6-which gives w = 4, x = 3, y = 2, and z = 6. Therefore, the
balanced chemical equation is
Network Analysis
j20
f
+2
i 30
Many practical situations give rise to networks: transportation networks, communi­
cations networks, and economic networks, to name a few. Of particular interest are
the possible flows through networks. For example, vehicles flow through a network
of roads, information flows through a data network, and goods and services flow
through an economic network.
For us, a network will consist of a finite number of nodes (also called junctions
or vertices) connected by a series of directed edges known as branches or arcs. Each
branch will be labeled with a flow that represents the amount of some commodity
that can flow along or through that branch in the indicated direction. (Think of cars
traveling along a network of one-way streets.) The fundamental rule governing flow
through a network is conservation offlow:
At each node, the flow in equals the flow out.
Figure 2 . 1 0
Flow at a node: f1 + f2 =
50
Exa m p l e 2 . 3 0
Figure 2 . 1 0 shows a portion of a network, with two branches entering a node and two
leaving. The conservation of flow rule implies that the total incoming flow, f1 + f
2
units, must match the total outgoing flow, 20 + 30 units. Thus, we have the linear
equation f1 + f = 50 corresponding to this node.
2
We can analyze the flow through an entire network by constructing such equa­
tions and solving the resulting system of linear equations.
Describe the possible flows through the network of water pipes shown in Figure 2. 1 1 ,
where flow is measured in liters per minute.
At each node, we write out the equation that represents the conservation
of flow there. We then rewrite each equation with the variables on the left and the
constant on the right, to get a linear system in standard form.
Solution
Node A :
Node B :
Node C:
Node D:
15
+ 14
+ 10
!2 + f3 + 5 = 3 0
f4 + 20 = f3
= !1
!1 = !2
-----+
!1
!1 - !2
!2
+ f4 = 1 5
=
+ f3
=
f, - f4 =
10
25
20
Section
jQ.
20
5!
-
!1
A
J4 !
D
3 0!
Figure 2 . 1 1
] 1
[
Applications
10
c
5
+--
Using Gauss-Jordan elimination, we reduce the augmented matrix:
.......,..
[i
1
0 0
0
-1 0
0
1
0 1 -1
15
10
25
20
�
103
-
B
1d
h
-
2.4
0 0
0
0
0 0 1
0 0 0
1 15
1 5
- 1 20
0 0
]
(Check this.) We see that there is one free variable, f4 , so we have infinitely many
solutions. Settingf4 = t and expressing the leading variables in terms off4 , we obtain
f,
!2
f3
!4
= 15 - t
= 5-t
= 20 + t
=
These equations describe all possible flows and allow us to analyze the network. For
example, we see that if we control the flow on branch AD so that t = 5 L/min, then
the other flows are f1 = 1 O,f2 = 0, and f3 = 25.
We can do even better: We can find the minimum and maximum possible flows
on each branch. Each of the flows must be nonnegative. Examining the first and sec­
ond equations in turn, we see that t :s 15 (otherwise f1 would be negative) and t :s 5
(otherwisef2 would be negative). The second of these inequalities is more restrictive
than the first, so we must use it. The third equation contributes no further restrictions
on our parameter t, so we have deduced that 0 :s t :s 5. Combining this result with
the four equations, we see that
10
0
20
0
:s f,
:S f2
:s f3
:S .fi
:s
:S
:s
:S
15
5
25
5
We now have a complete description of the possible flows through this network.
4
104
Chapter
2
Systems of Linear Equations
Electrical Networks
Electrical networks are a specialized type of network providing information about
power sources, such as batteries, and devices powered by these sources, such as light
bulbs or motors. A power source "forces" a current of electrons to flow through the
network, where it encounters various resistors, each of which requires that a certain
amount of force be applied in order for the current to flow through it.
The fundamental law of electricity is Ohm's law, which states exactly how much
force E is needed to drive a current I through a resistor with resistance R.
force
O h m 's Law
or
=
resistance X current
E
=
RI
Force is measured in volts, resistance in ohms, and current in amperes (or amps, for
short). Thus, in terms of these units, Ohm's law becomes "volts = ohms X amps;' and
it tells us what the "voltage drop" is when a current passes through a resistor-that is,
how much voltage is used up.
Current flows out of the positive terminal of a battery and flows back into the
negative terminal, traveling around one or more closed circuits in the process. In
a diagram of an electrical network, batteries are represented by � I- (where the
positive terminal is the longer vertical bar) and resistors are represented by -'VV'v- .
The following two laws, whose discovery we owe to Kirchhoff, govern electrical net­
works. The first is a "conservation of flow" law at each node; the second is a "balancing
of voltage" law around each circuit.
Kirchh off 's Laws
Current Law (nodes)
The sum of the currents flowing into any node is equal to the sum of the currents
flowing out of that node.
Voltage Law (circuits)
The sum of the voltage drops around any circuit is equal to the total voltage around
the circuit (provided by the batteries) .
Figure 2 . 1 2 illustrates Kirchhoff's laws. I n part (a), the current law gives 11 = 12 + 13
(or 11 - 12 - 13 = 0, as we will write it); part (b) gives 41 = 10, where we have used
Ohm's law to compute the voltage drop 41 at the resistor. Using Kirchhoff's laws, we
can set up a system of linear equations that will allow us to determine the currents in
an electrical network.
Exa m p l e 2 . 3 1
Determine the currents 11 , 12 , and 13 in the electrical network shown in Figure 2 . 1 3 .
This network has two batteries and four resistors. Current 11 flows through
the top branch BCA, current 12 flows across the middle branch AB, and current 13
flows through the bottom branch BDA.
At node A, the current law gives 11 + 13 = 12 , or
11 - 12 + 13 = 0
(Observe that we get the same equation at node B.)
Solution
Section
1
lY',,
11
1
12
2 ohms
10 volts
1
�
h
4 ohms
8 volts
I
11
lz
ohm
D
105
+---
--+
+---
4 ohms
=
c
--+
A
(b) 41
Applications
+---
+---
+---
2. 4
2 ohms
B
h
+---
1 6 volts
10
Figure 2 . 1 3
Figure 2 . 1 2
Next we apply the voltage law for each circuit. For the circuit CABC, the voltage
drops at the resistors are 2Ii , I2 , and 2Ii . Thus, we have the equation
4Ii + I2
Similarly, for the circuit DABD, we obtain
I2 + 4I3
=
=
8
16
(Notice that there i s actually a third circuit, CADBC, if we "go against the flow:' In this
case, we must treat the voltages and resistances on the "reversed" paths as negative.
Doing so gives 2Ii + 2Ii - 4I3 = 8
= - 8 or 4I - 4I3 = - 8, which we observe
i
is just the difference of the voltage equations for the other two circuits. Thus, we can
omit this equation, as it contributes no new information. On the other hand, includ­
ing it does no harm.)
We now have a system of three linear equations in three variables:
- 16
0
16
- 1 1 o 01 o1 � 41
[� � 1! l [ 0 o 1 3 l
1
I1 - I2 + I3
4I1 + I2
I2 + 4I3
Gauss-Jordan elimination produces
=
8
=
�
Hence, the currents are Ii
=
amp, I2
=
4 amps, and I3 = 3 amps.
In some electrical networks, the currents may have fractional values or may
even be negative. A negative value simply means that the current in the correspond­
ing branch flows in the direction opposite that shown on the network diagram.
Remark
CAS
Exa m p l e 2 . 3 2
The network shown in Figure 2 . 1 4 has a single power source A and five resistors. Find
the currents I, Ii , . . . , I5 • This is an example of what is known in electrical engineering
as a Wheatstone bridge circuit.
106
Chapter
2
Systems of Linear Equations
1i
i
13
14
--
B
I
ohm
-1
l ohm
c
2 ohms
!Ii
2 ohms
15
--
!
E
D
2 ohms
A
1 0 volts
-1
Figure 2 . 1 4
A bridge circuit
Solulion
nodes:
Kirchhoff's current law gives the following equations at the four
Node B: I - I1 - I4 = 0
Node C: I1 - I2 - I3 = 0
Node D: I - I2 - Is = 0
Node E : I3 + I4 - Is = 0
For the three basic circuits, the voltage law gives
Circuit ABEDA: I4 + 2Is = 1 0
Circuit BCEB: 2I1 + 2I3 - I4 = 0
Circuit CDEC: I2 - 2Is - 2I3 = 0
(Observe that branch DAB has no resistor and therefore no voltage drop; thus, there
is no I term in the equation for circuit ABEDA. Note also that we had to change signs
three times because we went "against the current:' This poses no problem, since we
will let the sign of the answer determine the direction of current flow.)
We now have a system of seven equations in six variables. Row reduction gives
1
0
1
0
0
0
0
�
-1
1
0
0
0
2
0
0
-1
-1
0
0
0
0
-1
0
-1
0
0
0
2
-2
-1
0
0 0
0 0
-1 0
-1 0
2 10
0 0
-2 0
�
1
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
1
0
0
0
7
0
3
0
4
0 -1
0
4
1
3
0
0
(Use your calculator or CAS to check this.) Thus, the solution (in amps) is I = 7, I1 =
Is = 3, I2 = I4 = 4, and I3 = - 1 . The significance of the negative value here is that
the current through branch CE is flowing in the direction opposite that marked on
the diagram.
-+
There is only one power source in this example, so the single 1 0-volt battery sends a current of 7 amps through the network. If we substitute these values into
Remark
Section
2.4
Applications
101
Ohm's law, E = RI, we get 1 0 = 7R or R = 1?-. Thus, the entire network behaves as if
there were a single 1?--ohm resistor. This value is called the effective resistance (Reff) of
the network.
linear Economic Models
An economy is a very complex system with many interrelationships among the vari­
ous sectors of the economy and the goods and services they produce and consume.
Determining optimal prices and levels of production subject to desired economic
goals requires sophisticated mathematical models. Linear algebra has proven to be a
powerful tool in developing and analyzing such economic models.
In this section, we introduce two models based on the work of Harvard econo­
mist Wassily Leontief in the 1 930s. His methods, often referred to as input-output
analysis, are now standard tools in mathematical economics and are used by cities,
corporations, and entire countries for economic planning and forecasting.
We begin with a simple example.
Exa m p l e 2 . 3 3
The economy of a region consists of three industries, or sectors: service, electricity,
and oil production. For simplicity, we assume that each industry produces a single
commodity (goods or services) in a given year and that income (output) is gener­
ated from the sale of this commodity. Each industry purchases commodities from the
other industries, including itself, in order to generate its output. No commodities are
purchased from outside the region and no output is sold outside the region. Further­
more, for each industry, we assume that production exactly equals consumption (out­
put equals input, income equals expenditure). In this sense, this is a closed economy
that is in equilibrium. Table 2.4 summarizes how much of each industry's output is
consumed by each industry.
Ta b l e 2 . 4
(1906-1999)
Wassily Leontief
was
born in St. Petersburg, Russia. He
studied at the University of Lenin­
grad and received his Ph.D. from the
University of Berlin. He emigrated
to the United States in
teach­
ing at Harvard University and later
at New York University. In
Leontiefbegan compiling data for the
monumental task of conducting an
input-output analysis of the United
States economy, the results of which
were published in
He was also
an early user of computers, which he
needed to solve the large-scale linear
systems in his models. For his pio­
neering work, Leontief was awarded
the Nobel Prize in Economics in
1931,
1932,
1941.
1973.
Consumed by
(input)
Service
Electricity
Oil
Service
Produced by (output)
Electricity
Oil
1/4
1/4
1/2
1/3
1/3
1/3
1 /2
1 /4
1 /4
From the first column of the table, we see that the service industry consumes 1 I4
of its own output, electricity consumes another 1/4, and the oil industry uses 1/2 of
the service industry's output. The other two columns have similar interpretations.
Notice that the sum of each column is 1 , indicating that all of the output of each
industry is consumed.
Let X i , x2 , and x3 denote the annual output (income) of the service, electricity,
and oil industries, respectively, in millions of dollars. Since consumption corresponds
to expenditure, the service industry spends � X i on its own commodity, .!.3 x2 on electricity, and t x3 on oil. This means that the service industry's total annual expenditure is � X i + .!.3 x2 + t x3 . Since the economy is in equilibrium, the service industry's
108
Chapter
2
Systems of Linear Equations
expenditure must equal its annual income x 1 . This gives the first of the following
equations; the other two equations are obtained by analyzing the expenditures of the
electricity and oil industries.
Service:
Electricity:
Oil:
......
"]
"]
Rearranging each equation, we obtain a homogeneous system of linear equations,
which we then solve. (Check this!)
-� x 1 + l3 x2 + t x3 = 0
i x1 - � x2 - i x3 = 0
3
t x1 + l3 x2 - � x3 = 0
---7
n
I
3
2
l3
-3
l
l 0
-i 0
�
[�
0 1 - :4 0
0
0 0
Setting x3 = t, we find that x 1 = t and x2 = � t. Thus, we see that the relative outputs of
the service, electricity, and oil industries need to be in the ratios x 1 : x2 : x3 = 4 : 3 : 4
for the economy to be in equilibrium.
4
R e m a rks
The last example illustrates what is commonly called the Leontief closed model.
Since output corresponds to income, we can also think of x 1 , x2 , and x3 as the
prices of the three commodities.
•
•
We now modify the model in Example 2.33 to accommodate an open economy, one
in which there is an external as well as an internal demand for the commodities that
are produced. Not surprisingly, this version is called the Leontief open model.
Exa m p l e 2 . 3 4
Consider the three industries of Example 2.33 but with consumption given by
Table 2.5. We see that, of the commodities produced by the service industry, 20% are
consumed by the service industry, 40% by the electricity industry, and 1 0% by the oil
industry. Thus, only 70% of the service industry's output is consumed by this econ­
omy. The implication of this calculation is that there is an excess of output (income)
over input (expenditure) for the service industry. We say that the service industry is
productive. Likewise, the oil industry is productive but the electricity industry is non­
productive. (This is reflected in the fact that the sums of the first and third columns
are less than 1 but the sum of the second column is equal to 1 ) . The excess output may
be applied to satisfy an external demand.
Tab l e 2 . 5
Consumed by
(input)
Service
Electricity
Oil
Service
Produced by (output)
Electricity
Oil
0.20
0.40
0. 1 0
0.50
0.20
0.30
0. 1 0
0.20
0.30
Section
2.4
Applications
109
For example, suppose there is an annual external demand (in millions of dollars)
for 10, 1 0, and 30 from the service, electricity, and oil industries, respectively. Then,
equating expenditures (internal demand and external demand) with income (out­
put), we obtain the following equations:
internal demand
output
Service
Xi
=
external demand
0.2X 1 + 0.5X2 + O.lX3
+ 10
+ 10
+ 30
0.4X 1 + 0.2X2 + 0.2X3
= O.lx 1 + 0.3x2 + 0.3x3
Oil
Rearranging, we obtain the following linear system and augmented matrix:
Electricity
X2
X3
=
0.8X 1 - O.SX2 - O. lX3
- 0.4X 1 + 0.8X2 - 0.2X3
- 0. lx 1 - 0.3X2 + 0.7X3
CAS
=
=
=
10
10
30
---+
[
08
- 0.4
-0.l
Row reduction yields
[�
0
1
0
0
0
6 1 .74
63.04
78.70
- 0.5
0.8
- 0.3
-0.1
- 0.2
0.7
10
10
30
]
l
from which we see that the service, electricity, and oil industries must have an an­
nual production of $6 1 . 7 4, $63.04, and $78. 70 (million), respectively, in order to meet
both the internal and external demand for their commodities.
We will revisit these models in Section 3.7.
Finile Linear Games
There are many situations in which we must consider a physical system that has only a
finite number of states. Sometimes these states can be altered by applying certain pro­
cesses, each of which produces finitely many outcomes. For example, a light bulb can be
on or off and a switch can change the state of the light bulb from on to off and vice versa.
Digital systems that arise in computer science are often of this type. More frivolously, many
computer games feature puzzles in which a certain device must be manipulated by various
switches to produce a desired outcome. The finiteness of such situations is perfectly suited
to analysis using modular arithmetic, and often linear systems over some ZP play a role.
Problems involving this type of situation are often called finite linear games.
Exa m p l e 2 . 3 5
A row of five lights is controlled by five switches. Each switch changes the state (on or
off) of the light directly above it and the states of the lights immediately adjacent to
the left and right. For example, if the first and third lights are on, as in Figure 2 . l S (a),
then pushing switch A changes the state of the system to that shown in Figure 2 . l S (b).
If we next push switch C, then the result is the state shown in Figure 2 . l S (c).
110
Chapter
B
2
c
Systems of Linear Equations
D
E
A
B
( a)
D
c
E
'-
A
B
"
c
D
IH
E
(c)
(b)
Figure 2 . 1 5
Suppose that initially all the lights are off. Can we push the switches in some order
so that only the first, third, and fifth lights will be on? Can we push the switches in
some order so that only the first light will be on?
The on/off nature of this problem suggests that binary notation will be helpful
and that we should work with 2 2 • Accordingly, we represent the states of the five lights by
a vector in Z�, where 0 represents off and 1 represents on. Thus, for example, the vector
Solution
0
1
0
0
corresponds to Figure 2 . I S (b).
We may also use vectors in Z� to represent the action of each switch. If a switch
changes the state of a light, the corresponding component is a 1 ; otherwise, it is 0.
With this convention, the actions of the five switches are given by
a =
1
0
0
0
, b
=
1
0
0
, c
=
0
1
1
1
0
0
0
, d
=
,
e =
0
0
0
The situation depicted in Figure 2 . I S(a) corresponds to the initial state
s =
1
0
1
0
0
followed by
a =
1
0
0
0
Section
2.4
Applications
111
It is the vector sum (in Z� )
0
s
+
a
1
1
=
0
0
Observe that this result agrees with Figure 2 . 1 5 (b).
Starting with any initial configuration s, suppose we push the switches in the order
A, C, D, A, C, B. This corresponds to the vector sum s + a + c + d + a + c + b. But
in Z�, addition is commutative, so we have
s
.-
+
a
+
c
a
+ d +
+
c
+
b
= s + 2a + b +
= s + b + d
2c
+ d
where we have used the fact that 2 = 0 in z:'. 2 . Thus, we would achieve the same result
by pushing only B and D-and the order does not matter. (Check that this is correct.)
Hence, in this example, we do not need to push any switch more than once.
So, to see if we can achieve a target configuration t starting from an initial
configuration s, we need to determine whether there are scalars x 1 , , x5 in z:'. 2 such that
s + x 1 a + x2 b + · · · + x5 e = t
. • •
In other words, we need to solve (if possible) the linear system over z:'. 2 that corre­
sponds to the vector equation
x1 a
In this case, s =
x2 b + · · · + x5 e = t s
0 and our first target configuration is
+
-
= t +
s
1
0
t =
1
0
The augmented matrix of this system has the given vectors as columns:
1
1
0
0
1
0
0
1
0
0
0
1
0
0
0
1
0
0
0
0
1
1
1
0
0
0
0
0
0
1
1
We reduce it over z:'. 2 to obtain
0
0
0
1
0
0
0
1
0
0
0
0
0
0
0
1 1
0
112
Chapter
2
Systems of Linear Equations
Thus, x5 i s a free variable. Hence, there are exactly two solutions (corresponding to
x5 = 0 and x5 = 1 ) . Solving for the other variables in terms of x5 , we obtain
X1
X2
X3
X4
=
=
=
=
X5
+ X5
+ X5
So, when x5 = 0 and x5 = 1 , we have the solutions
X1
X2
X3
X4
X5
_...
0
1
0
and
1
0
1
0
X1
X2
X3
X4
X5
respectively. (Check that these both work.)
Similarly, in the second case, we have
1
0
t= 0
0
0
The augmented matrix reduces as follows:
1
0
0
0
1
1
0
0
0 0 0 1
1 0 0 0
1
0 0
1
0
0
0
-----+
1
0
0
0
0
0 0
0
0 1
0 0
0 0
0
0
0
0 0
1 1
0 0
showing that there is no solution in this case; that is, it is impossible to start with all
of the lights off and turn only the first light on.
Example 2.35 shows the power of linear algebra. Even though we might have
found out by trial and error that there was no solution, checking all possible ways to
push the switches would have been extremely tedious. We might also have missed the
fact that no switch need ever be pushed more than once.
Exa m p l e 2 . 3 6
Consider a row with only three lights, each of which can be off, light blue, or dark blue.
Below the lights are three switches, A, B, and C, each of which changes the states of
particular lights to the next state, in the order shown in Figure 2 . 1 6. Switch A changes
the states of the first two lights, switch B all three lights, and switch C the last two
Section
Dark blue
2.4
Applications
113
j
Light blue
�
B
c
Figure 2.11
Figure 2 . 1 6
lights. If all three lights are initially off, is it possible to push the switches in some order
so that the lights are off, light blue, and dark blue, in that order (as in Figure 2 . 1 7) ?
Whereas Example 2 . 3 5 involved "1!.. 2 , this one clearly (is it clear?) involves
"1!.. 3 • Accordingly, the switches correspond to the vectors
Solution
in Zl. and th, final wnfigmation w' "' <riming fod, t
�
[n
( Offo 0, light bl"' i' 1 ,
and dark blue i s 2 . ) We wish t o find scalars X i , x2 , x3 i n "1!.. 3 such that
X 1 a + X 2 b + X3 C
= t
(where X; represents the number of times the ith switch is pushed) . This equation
gives rise to the augmented matrix [a b c I t] , which reduces over "1!.. 3 as follows:
..
I
Hence, there is a unique solution: X i = 2, x2 = 1 , x3 = 1 . In other words, we must push
switch A twice and the other two switches once each. (Check this.)
Exercises 2 . 4
1. Suppose that, in Example 2.27, 400 units of food A,
A l l o c a t i o n of R e s o u rces
600 units of B, and 600 units of C are placed in the test
tube each day and the data on daily food consump­
tion by the bacteria (in units per day) are as shown
in Table 2.6. How many bacteria of each strain can
coexist in the test tube and consume all of the food?
2. Suppose that in Example 2.27, 400 units of food A,
500 units of B, and 600 units of C are placed in
the test tube each day and the data on daily food
Tab l e 2 . 6
Food A
Food B
Food C
Bacteria
Strain I
1
2
1
Bacteria
Strain II
2
1
Bacteria
Strain III
0
1
2
consumption by the bacteria (in units per day) are
as shown in Table 2.7. How many bacteria of each
114
Chapter
2
Systems of Linear Equations
Tab l e 2 . 1
Food A
Food B
Food C
Bacteria
Strain I
2
1
Bacteria
Strain II
2
Bacteria
Strain III
0
3
1
the special blend produces a profit of $ 1 .50, and one bag
of the gourmet blend produces a profit of $2.00. How
many bags of each type should the merchant prepare
if he wants to use up all of the beans and maximize his
profit? What is the maximum profit?
Balancing Chemical Equations
In Exercises 7-14, balance the chemical equation for each
reaction.
7. FeS 2 + 0 2 -----+ Fe2 0 3 + S0 2
8. C0 2 + H 2 0 -----+ C6H 1 2 06 + 0 2 (This reaction takes
strain can coexist in the test tube and consume all
of the food?
place when a green plant converts carbon dioxide and
3. A florist offers three sizes of flower arrangements
water to glucose and oxygen during photosynthesis.)
containing roses, daisies, and chrysanthemums. Each
small arrangement contains one rose, three daisies,
9. C4H 1 0 + 0 2 -----+ C0 2 + H 2 0 (This reaction occurs
and three chrysanthemums. Each medium arrange­
when butane, C 4 H 1 0 , burns in the presence of oxygen
ment contains two roses, four daisies, and six chry­
to form carbon dioxide and water.)
santhemums. Each large arrangement contains four
10. C 7 H60 2 + 0 2 -----+ H 2 0 + C0 2
roses, eight daisies, and six chrysanthemums. One
1 1 . C 5 H 1 1 0H + 0 2 -----+ H 2 0 + C0 2 (This equation rep­
day, the florist noted that she used a total of 24 roses,
resents the combustion of amyl alcohol.)
50 daisies, and 48 chrysanthemums in filling orders
12.
HC10
4 + P 4 0 1 0 ----+ H 3 P0 4 + Cl2 0 7
for these three types of arrangements. How many
arrangements of each type did she make?
13. Na2 C0 3 + C + N2 -----+ NaCN + CO
4. (a) In your pocket you have some nickels, dimes, and
cA5 14. C 2 H 2 Cl4 + Ca(OH) z -----+ C 2 HC13 + CaC12 + H 2 0
quarters. There are 20 coins altogether and exactly
twice as many dimes as nickels. The total value of the
N e t w o r k A n alvsis
coins is $3.00. Find the number of coins of each type.
15. Figure 2 . 1 8 shows a network of water pipes with flows
(b) Find all possible combinations of 20 coins (nickels,
measured in liters per minute.
dimes, and quarters) that will make exactly $3.00.
(a) Set up and solve a system of linear equations to find
5. A coffee merchant sells three blends of coffee. A bag
the possible flows.
of the house blend contains 300 grams of Colombian
(b) If the flow through AB is restricted to 5 L/min, what
beans and 200 grams of French roast beans. A bag of the
will the flows through the other two branches be?
special blend contains 200 grams of Colombian beans,
(c) What are the minimum and maximum possible
200 grams of Kenyan beans, and 100 grams of French
flows through each branch?
roast beans. A bag of the gourmet blend contains
(
d)
We have been assuming that flow is always posi­
100 grams of Colombian beans, 200 grams of Kenyan
tive. What would negative flow mean, assum­
beans, and 200 grams of French roast beans. The mer­
ing we allowed it? Give an illustration for this
chant has on hand 30 kilograms of Colombian beans,
example.
15 kilograms of Kenyan beans, and 25 kilograms of French
roast beans. Ifhe wishes to use up all of the beans, how
many bags of each type of blend can be made?
!1
--+
A
6. Redo Exercise 5, assuming that the house blend contains
300 grams of Colombian beans, 50 grams of Kenyan
beans, and 1 50 grams of French roast beans and the
gourmet blend contains 100 grams of Colombian beans,
c
350 grams of Kenyan beans, and 50 grams of French roast
beans. This time the merchant has on hand 30 kilograms of
Colombian beans, 1 5 kilograms of Kenyan beans, and
B
1 5 kilograms of French roast beans. Suppose one bag of
the house blend produces a profit of $0.50, one bag of
Fioure 2 . 1 8
Section
16. The downtown core of Gotham City consists of
one-way streets, and the traffic flow has been
measured at each intersection. For the city block
shown in Figure 2 . 1 9, the numbers represent the
average numbers of vehicles per minute entering and
leaving intersections A, B, C, and D during business
hours.
(a) Set up and solve a system oflinear equations to find
the possible flows f1 , . . . ,f4 .
(b) If traffic is regulated on CD so that f4 = 1 0 vehi­
cles per minute, what will the average flows on the
other streets be?
(c) What are the minimum and maximum possible
flows on each street?
(d) How would the solution change if all of the direc­
tions were reversed?
!
10
J.Q.
A
h
.£
!
D
f
!1
f4
+---
!
10
B
13
t
c
to be maintained through DB?
( c ) From Figure 2.20 it is clear that DB cannot be
(d)
closed. (Why not?) How does your solution in part
(a) show this?
From your solution in part (a), determine the mini­
mum and maximum flows through DB.
18. (a) Set up and solve a system of linear equations to
find the possible flows in the network shown in
Figure 2.2 1 .
(b) Is it possible forf1 = 1 0 0 andf6 = 1 50? [Answer
this question first with reference to your solution
in part (a) and then directly from Figure 2.2 1 . ]
( c) I ff4 = 0 , what will the range o f flow b e o n each of
the other branches?
is
f
200
13
.£
t
200
f
1 50
f
A
D
Ji
!
--
16
f4
!
1 00
h
f
20o
B
--
E
+---
+---
1 0o
Figure 2 . 1 9
!
fs
h
t
l
c
F
l.QQ..
.l2Q.
1 0o
Figure 2 . 2 1
17. A network of irrigation ditches is shown in Figure 2.20,
with flows measured in thousands of liters per day.
l.QQ.. A
E l e c t r i c a l Networks
For Exercises 19 and 20, determine the currents for the
given electrical networks.
19.
c
/1
/1
+---
+---
8 volts
1 ohm
.l2Q.
A
"
/2
c
Figure 2 . 2 0
/2
--
!
..!22.
--
1 ohm
D
200
115
(b) Suppose DC is closed. What range of flow will need
1 0o
.2....
Applications
(a) Set up and solve a system oflinear equations to find
the possible flows f1 , . . . ,f5 •
20
--
2.4
4 ohms
/3
+---
D
1 3 volts
/3
+---
B
Chapter
116
2
Systems of Linear Equations
c
Ii
20.
-
11
R 1
1
-+­
R 1 Rz
eff
-
5 volts
1 ohm
h
12
--
A
---
--
B
2 ohms
4 ohms
13
-
D
13
E
-
( a)
8 volts
21. (a) Find the currents I, I1 , . . . , I5 in the bridge circuit
in Figure 2.22.
(b) Find the effective resistance of this network.
O?
(c) Can you change the resistance in branch BC (but
leave everything else unchanged) so that the cur­
rent through branch CE becomes
l
Ii
i
B
c
ohm
2 ohms
/4
13
2 ohms
I
-
!
E
A
Ri
12
R2
1
f
--
E
(b)
!h
1 ohm
--
11
--
15
--
14 volts
1 ohm
D
I
-
Figure 2 . 2 2
22. The networks in parts (a) and (b) of Figure 2.23
show two resistors coupled in series and in parallel,
respectively. We wish to find a general formula for the
effective resistance of each network-that is, find Reff
such that E = Refl
(a) Show that the effective resistance Reff of a network
with two resistors coupled in series [Figure 2.23(a) J
is given by
(b) Show that the effective resistance Reff of a net­
work with two resistors coupled in parallel
[Figure 2.23(b)] is given by
Figure 2 . 2 3
Resistors in series and in parallel
l i n e a r E c o n o m i c M o d e ls
23. Consider a simple economy with just two industries:
farming and manufacturing. Farming consumes 1/2 of
the food and 1/3 of the manufactured goods. Manufac­
turing consumes 1/2 of the food and 2/3 of the manu­
factured goods. Assuming the economy is closed and
in equilibrium, find the relative outputs of the farming
and manufacturing industries.
24. Suppose the coal and steel industries form a closed
economy. Every $1 produced by the coal industry
requires $0.30 of coal and $0.70 of steel. Every $ 1
produced by steel requires $0.80 o f coal and $0.20 of
steel. Find the annual production (output) of coal and
steel if the total annual production is $20 million.
25. A painter, a plumber, and an electrician enter into a
cooperative arrangement in which each of them agrees
to work for himself/herself and the other two for a
total of 1 0 hours per week according to the schedule
shown in Table 2.8. For tax purposes, each person must
establish a value for his/her services. They agree to do
this so that they each come out even-that is, so that the
Section
total amount paid out by each person equals the amount
he/she receives. What hourly rate should each person
charge if the rates are all whole numbers between $30
and $60 per hour?
Ta b l e 2 . 8
Supplier
Painter Plumber Electrician
2
4
4
Painter
Consumer Plumber
Electrician
5
1
4
5
4
26. Four neighbors, each with a vegetable garden, agree to
share their produce. One will grow beans (B), one will
grow lettuce (L), one will grow tomatoes (T), and one
will grow zucchini (Z). Table 2.9 shows what fraction
of each crop each neighbor will receive. What prices
should the neighbors charge for their crops if each
person is to break even and the lowest-priced crop has
a value of $50?
2.4
Applications
111
they produce, each department uses a certain amount
of the services produced by the other departments
and itself, as shown in Table 2 . 1 0. Suppose that, dur­
ing the year, other city departments require $ 1 million
in Administrative services, $ 1 .2 million in Health
services, and $0.8 million in Transportation services.
What does the annual dollar value of the services
produced by each department need to be in order to
meet the demands?
Ta b l e 2 . 1 0
Department
H
A
Buy
A
H
T
$0.20
0. 1 0
0.20
0. 1 0
0. 1 0
0.40
T
0.20
0.20
0.30
Finite linear Games
29. (a) In Example 2.35, suppose all the lights are initially
Ta b l e 2 . 9
Consumer
B
L
T
z
B
Producer
L
T
z
0
1 /2
1 /4
1 /4
1/4
1/4
1/4
1/4
1/8
1/4
1/2
1/8
1 /6
1 /6
1/3
1/3
27. Suppose the coal and steel industries form an open
economy. Every $1 produced by the coal industry
requires $0. 1 5 of coal and $0.20 of steel. Every $ 1
produced by steel requires $0.25 of coal and $0. 1 0 of
steel. Suppose that there is an annual outside demand
for $45 million of coal and $ 1 24 million of steel.
(a) How much should each industry produce to satisfy
the demands?
(b) If the demand for coal decreases by $5 million
per year while the demand for steel increases by
$6 million per year, how should the coal and steel
industries adjust their production?
28. In Gotham City, the departments of Administra­
tion (A), Health (H), and Transportation (T) are
interdependent. For every dollar's worth of services
off. Can we push the switches in some order so
that only the second and fourth lights will be on?
(b) Can we push the switches in some order so that
only the second light will be on?
30. (a) In Example 2.35, suppose the fourth light is
initially on and the other four lights are off. Can
we push the switches in some order so that only
the second and fourth lights will be on?
(b) Can we push the switches in some order so that
only the second light will be on?
31. In Example 2.35, describe all possible configurations
of lights that can be obtained if we start with all the
lights off.
32. (a) In Example 2.36, suppose that all of the lights
are initially off. Show that it is possible to
push the switches in some order so that the
lights are off, dark blue, and light blue, in that order.
(b) Show that it is possible to push the switches in
some order so that the lights are light blue, off,
and light blue, in that order.
(c) Prove that any configuration of the three lights
can be achieved.
33. Suppose the lights in Example 2.35 can be off, light
blue, or dark blue and the switches work as described
Chapter
118
2
Systems of Linear Equations
in Example 2.36. (That is, the switches control the same
lights as in Example 2.35 but cycle through the colors as
in Example 2.36.) Show that it is possible to start with
all of the lights off and push the switches in some order
so that the lights are dark blue, light blue, dark blue,
light blue, and dark blue, in that order.
34. For Exercise 33, describe all possible configurations
of lights that can be obtained, starting with all the
lights off.
cAs 35. Nine squares, each one either black or white, are ar­
ranged in a 3 X 3 grid. Figure 2.24 shows one possible
how the state changes work. (Touching the square
whose number is circled causes the states of the
squares marked * to change.) The object of the game
is to turn all nine squares black. [Exercises 35 and 36
are adapted from puzzles that can be found in
the interactive CD-ROM game The Seventh Guest
(Trilobyte Software/Virgin Games, 1 992) .]
(a) If the initial configuration is the one shown in
Figure 2.24, show that the game can be won and
describe a winning sequence of moves.
(b) Prove that the game can always be won, no matter
what the initial configuration.
CAS 36. Consider a variation on the nine squares puzzle. The
game is the same as that described in Exercise 35
except that there are three possible states for each
square: white, gray, or black. The squares change as
shown in Figure 2.25, but now the state changes follow
the cycle white � gray � black � white. Show how
the winning all-black configuration can be achieved
from the initial configuration shown in Figure 2.26.
Figure 2 . 2 4
The nine squares
puzzle
arrangement. When touched, each square changes
its own state and the states of some of its neighbors
(black � white and white � black) . Figure 2.25 shows
CD*
2
4
5
*
*
*
8
7
@*
3
4
5
7
8
3
l
6
9
*
1
2
6
4
5
9
7
8
*
*
*
®*
6
*
Figure 2.26
The nine squares puzzle
with more states
9
Miscellaneous Problems
l
2
3
l
@*
5
6
4
7
8
9
2
*
*
l
4
*
(j)*
5
8
*
*
2
*
3
1
2
3
4
5
®*
G)*
6
7
8
9
7
8
9
3
l
2
3
l
2
3
6
4
5
6
4
5
9
7
®*
9
7
8
*
*
*
Figure 2 . 2 5
State changes for the nine squares puzzle
*
*
*
*
6
*
*
*
®*
In Exercises 37-53, set up and solve an appropriate system
of linear equations to answer the questions.
37. Grace is three times as old as Hans, but in 5 years she
will be twice as old as Hans is then. How old are they
now?
38. The sum of Annie's, Bert's, and Chris's ages is 60.
Annie is older than Bert by the same number of years
that Bert is older than Chris. When Bert is as old as
Annie is now, Annie will be three times as old as Chris
is now. What are their ages?
The preceding two problems are typical of those found in
popular books of mathematical puzzles. However, they have
their origins in antiquity. A Babylonian clay tablet that sur­
vives from about 300 B. c. contains the following problem.
Section
Over 2000 years ago, the Chinese developed methods for
solving systems of linear equations, including a version of
Gaussian elimination that did not become well known in
Europe until the 1 9th century. (There is no evidence that
Gauss was aware of the Chinese methods when he devel­
oped what we now call Gaussian elimination. However, it is
clear that the Chinese knew the essence of the method, even
though they did not justify its use.) The following problem
is taken from the Chinese text Jiuzhang suanshu (Nine
Chapters in the Mathematical Art), written during the early
Han Dynasty, about 200 B . C .
46. Through any three noncollinear points there also
passes a unique circle. Find the circles (whose general
equations are of the form x2 + y2 + ax + by + c = 0)
that pass through the sets of points in Exercise 45. (To
check the validity of your answer, find the center and
radius of each circle and draw a sketch.)
The process of adding rationalfunctions (ratios ofpolyno­
mials) by placing them over a common denominator is
the analogue of adding rational numbers. The reverse
process of taking a rationalfunction apart by writing it as
a sum ofsimpler rationalfunctions is useful in several
areas of mathematics; for example, it arises in calculus
when we need to integrate a rationalfunction and in dis­
crete mathematics when we use generatingfunctions to
solve recurrence relations. The decomposition of a rational
function as a sum ofpartialfractions leads to a system of
linear equations. In Exercises 47-50,find the partial
fraction decomposition of the givenform. (The capital
letters denote constants.)
�
b
c 3 6
d 4 5
42. What conditions on w, x, y, and z will guarantee that
we can find a, b, c, and d so that the following is a valid
addition table?
3x + 1
A
B
+ -=
x + 2x - 3 x - 1 x + 3
C
B
x2 - 3x + 3 A
=48. 3
+
+
2
x + 2x + x x x + 1 (x + 1) 2
x-1
49.
(X + l )(x 2 + l )(x 2 + 4 )
A
Bx + C Dx + E
=
+ 2
+ --x+ 1 x + 1
x2 + 4
x3 + x + 1
A
B
50 •
=-+
x (x - l) (x 2 + x + l) (x 2 + 1 ) 3 x x - 1
Cx + D
Ex + F Gx + H
Ix + J
+
+ x2
+
+
2
2
2
+ x + l x + 1 (x + 1 )
(x 2 + 1 ) 3
47. 2
-
--
--
43. Describe all possible values of a, b, c, d, e, and f
that will make each of the following a valid addition
table.
( a)
+
d
e
f
a b c
3 2 1
5 4 3
4 3
(b)
+
d
e
f
a b c
1 2 3
3 4 5
4 5 6
119
(a) (O, 1), ( - 1, 4), and (2, 1 )
(b) ( - 3, 1), ( - 2, 2), and ( - 1, 5)
40. There are three types of corn. Three bundles of the
first type, two of the second, and one of the third
make 39 measures. Two bundles of the first type, three
of the second, and one of the third make 34 measures.
And one bundle of the first type, two of the second,
and three of the third make 26 measures. How many
measures of corn are contained in one bundle of
each type?
41. Describe all possible values of a, b, c, and d that
will make each of the following a valid addition
table. [Problems 4 1 -44 are based on the article
''An Application of Matrix Theory" by Paul Glaister in
The Mathematics Teacher, 85 ( 1 992), pp. 220-223.]
(b)
Applications
44. Generalizing Exercise 42, find conditions on the en­
tries of a 3 X 3 addition table that will guarantee that
we can solve for a, b, c, d, e, and f as previously.
45. From elementary geometry we know that there
is a unique straight line through any two points
in a plane. Less well known is the fact that there is a
unique parabola through any three noncollinear
points in a plane. For each set of points below, find
a parabola with an equation of the form y = ax 2 +
bx + c that passes through the given points. (Sketch
the resulting parabola to check the validity of your
answer.)
39. There are two fields whose total area is 1 800 square
yards. One field produces grain at the rate of � bushel
per square yard; the other field produces grain at the
rate of t bushel per square yard. If the total yield is
1 1 00 bushels, what is the size of each field?
( a)
2.4
CAS
---
------
--
CAS
--
----
120
Chapter
2
Systems of Linear Equations
Following are two useful formulas for the sums ofpowers of
consecutive natural numbers:
n(n + 1 )
1 + 2 + · · · + n = ---2
and
n(n + 1 ) (2n + 1)
1 2 + 2 2 + · . · + n 2 = ----6
The validity of these formulas for all values of n 2:: 1 (or
even n 2:: OJ can be established using mathematical induc­
tion (see Appendix B) . One way to make an educated guess
as to what the formulas are, though, is to observe that we
can rewrite the two formulas above as
respectively. This leads to the conjecture that the sum ofpth
powers of the first n natural numbers is a poly nomial of
degree p + 1 in the variable n.
5 1 . Assuming that 1 + 2 + · · · + n = an 2 + bn + c,
find a, b, and c by substituting three values for n and
thereby obtaining a system of linear equations in a,
b, and c.
52. Assume that 1 2 + 2 2 + · · · + n 2 = an 3 + bn 2 + en + d.
Find a, b, c, and d. [Hint: It is legitimate to use n = 0.
What is the left-hand side in that case?]
53. Show that 1 3 + 2 3 + · · · + n 3 = (n(n + 1)/2) 2 .
Vi gnette
T h e Glob al Po s it i o n i n g System
This application is based on the
article "An Underdetermined
Linear System for GPS" by
Dan Kalman in The College
33 (2002),
pp. 384-390. For a more in-depth
Mathematics Journal,
treatment of the ideas introduced
here, see G. Strang and K. Borre,
Linear Algebra, Geodesy, and GPS
1997).
(Wellesley-Cambridge Press, MA,
The Global Positioning System (GPS) i s used i n a variety o f situations fo r determin­
ing geographical locations. The military, surveyors, airlines, shipping companies,
and hikers all make use of it. GPS technology is becoming so commonplace that
some automobiles, cellular phones, and various handheld devices are now equipped
with it.
The basic idea of GPS is a variant on three-dimensional triangulation: A point
on Earth's surface is uniquely determined by knowing its distances from three other
points. Here the point we wish to determine is the location of the GPS receiver, the
other points are satellites, and the distances are computed using the travel times of
radio signals from the satellites to the receiver.
We will assume that Earth is a sphere on which we impose an xyz-coordinate
system with Earth centered at the origin and with the positive z-axis running through
the north pole and fixed relative to Earth.
For simplicity, let's take one unit to be equal to the radius of Earth. Thus Earth's
surface becomes the unit sphere with equation x 2 + y 2 + z 2 = 1 . Time will be
measured in hundredths of a second. GPS finds distances by knowing how long it
takes a radio signal to get from one point to another. For this we need to know the
speed of light, which is approximately equal to 0.47 (Earth radii per hundredths of
a second) .
Let's imagine that you are a hiker lost in the woods at point (x, y, z) at some time
t. You don't know where you are, and furthermore, you have no watch, so you don't
know what time it is. However, you have your G PS device, and it receives simultaneous
signals from four satellites, giving their positions and times as shown in Table 2 . 1 1 .
(Distances are measured in Earth radii and time in hundredths of a second past
midnight.)
Ta b l e 2 . 1 1 sa1em1e Dara
Satellite
2
3
4
Position
Time
( 1 . 1 1 , 2.55, 2 . 1 4)
(2.87, 0.00, 1 .43)
(0.00, 1 .08, 2.29)
( 1 .54, 1 . 0 1 , 1 .23)
1 .29
1.31
2.75
4.06
121
Let (x, y, z) be your position, and let t be the time when the signals arrive. The
goal is to solve for x, y, z, and t. Your distance from Satellite 1 can be computed as
follows. The signal, traveling at a speed of 0.47 Earth radii/ 1 0 - 2 sec, was sent at time
1 .29 and arrived at time t, so it took t - 1 .29 hundredths of a second to reach you.
Distance equals velocity multiplied by (elapsed) time, so
d = 0.47(t - 1 .29)
We can also express d in terms of (x, y, z) and the satellite's position ( 1 . 1 1 , 2.55, 2. 14)
using the distance formula:
d = V( x - 1 . 1 1 ) 2 + (y - 2.55 ) 2 + ( z - 2 . 1 4 ) 2
Combining these results leads to the equation
(x - 1 . 1 1 ) 2 + (y - 2.55) 2 + (z - 2 . 1 4) 2 = 0.47 2 (t - 1 .29) 2
Expanding, simplifying, and rearranging, we find that Equation ( 1 ) becomes
2.22x + 5 . l Oy + 4.28z - 0.57t = x 2 + y 2 + z 2 - 0.22t 2 + 1 1 .95
(1)
Similarly, we can derive a corresponding equation for each of the other three satel­
lites. We end up with a system of four equations in x, y, z, and t:
2.22x + 5 . l Oy + 4.28z - 0.57t = x 2 + y 2 + z 2 - 0.22t 2 + 1 1 .95
5.74x
+ 2.86z - 0.58t = x 2 + y 2 + z 2 - 0.22t 2 + 9.90
2 . 1 6y + 4.58z - 1 .2 1 t = x 2 + y 2 + z 2 - 0.22t 2 + 4.74
3.08x + 2.02y + 2.46z - 1 .79t = x 2 + y 2 + z 2 - 0.22t 2 + 1 .26
These are not linear equations, but the nonlinear terms are the same in each equation.
If we subtract the first equation from each of the other three equations, we obtain a
linear system:
3.52x - 5 . l Oy - 1 .42z - O.O l t = 2.05
- 2.22x - 2.94y + 0.30z - 0.64t = 7.2 1
0.86x - 3.08y - 1 .82z - 1 .22t = - 1 0.69
The augmented matrix row reduces as
122
[
3.52
- 2.22
0.86
- 5. 1 0
- 2.94
- 3.08
- 1 .42
0.30
- 1 .82
= �:��1 [� � � �:�: ]
- 0.01
- 0.64
- 1 .22 - 1 0.69
�
0 0
2.97
0.81
1 0.79 5.91
from which we see that
x = 2.97 - 0.36t
y = 0.81 - 0.03t
z = 5.91 - 0.79t
(2)
with t free. Substituting these equations into ( 1 ), we obtain
(2.97 - 0.36t - 1 . 1 1 ) 2 + (0.81 - 0.03t - 2.55) 2
+ (5.9 1 - 0.79t - 2. 14) 2 = 0.47 2 (t - 1 .29) 2
which simplifies to the quadratic equation
0.54t 2 - 6.65t + 20.32 = 0
There are two solutions:
t = 6.74 and t = 5.60
Substituting into (2), we find that the first solution corresponds to (x, y, z) = (0.55,
0.6 1 , 0.56) and the second solution to (x, y, z) = (0.96, 0.65, 1 .46). The second solution
is clearly not on the unit sphere (Earth), so we reject it. The first solution produces
x 2 + y 2 + z 2 = 0.99, so we are satisfied that, within acceptable roundoff error, we
have located your coordinates as (0.55, 0.6 1 , 0.56).
In practice, GPS takes significantly more factors into account, such as the fact
that Earth's surface is not exactly spherical, so additional refinements are needed in­
volving such techniques as least squares approximation (see Chapter 7). In addition,
the results of the GPS calculation are converted from rectangular (Cartesian) coor­
dinates into latitude and longitude, an interesting exercise in itself and one involving
yet other branches of mathematics.
123
124
Chapter
'�
2
Systems of Linear Equations
�
llerative M e l h o d s f o r S o l v i n g l i n e a r svste m s
The direct methods for solving linear systems, using elementary row operations, lead
to exact solutions in many cases but are subject to errors due to roundoff and other
factors, as we have seen. The third road in our "trivium" takes us down quite a different
path indeed. In this section, we explore methods that proceed iteratively by succes­
sively generating sequences of vectors that approach a solution to a linear system. In
many instances (such as when the coefficient matrix is sparse-that is, contains many
zero entries), iterative methods can be faster and more accurate than direct methods.
Also, iterative methods can be stopped whenever the approximate solution they gen­
erate is sufficiently accurate. In addition, iterative methods often benefit from inac­
curacy: Roundoff error can actually accelerate their convergence toward a solution.
We will explore two iterative methods for solving linear systems: Jacobi's method
and a refinement of it, the Gauss-Seidel method. In all examples, we will be consid­
ering linear systems with the same number of variables as equations, and we will
assume that there is a unique solution. Our interest is in finding this solution using
iterative methods.
Exa m p l e 2 . 3 1
Consider the system
7x 1 - x2
3x 1 - 5x2
= 5
= -7
Jacobi's method begins with solving the first equation for x1 and the second equation
for x2, to obtain
5 + Xz
X1 = 7
(1)
7 + 3x 1
Xz =
5
We now need an initial approximation to the solution. It turns out that it does not
matter what this initial approximation is, so we might as well take x1 = 0, x2 = 0. We
use these values in Equations ( 1 ) to get new values of x1 and x2:
(1804-1851)
Carl Gustav Jacobi
was
a German mathematician who made
important contributions to many
fields of mathematics and physics,
including geometry, number theory,
analysis, mechanics, and fluid
dynamics. Although much of his
work was in applied mathematics,
Jacobi believed in the importance of
doing mathematics for its own sake.
A fine teacher, he held positions
at the Universities of Berlin and
Konigsberg and was one of the most
famous mathematicians in Europe.
-- = - =
Xz =
5 + 0 5
0.714
7
7
7 + 3.0 7
1 .400
5
5
Now we substitute these values into ( 1 ) to get
X1 =
Xz =
5 + 1 .4
= 0.9 14
7
7 + 3 2---7 = 1 .829
5
---
.
(written to three decimal places) . We repeat this process (using the old values of x2
and x1 to get the new values of x1 and x2 ) , producing the sequence of approximations
given in Table 2 . 1 2 .
Section
2. 5
1
0
Ta b l e 2 . 1 2
2
4
3
125
5
6
00 0.1.47001 4 0.1.982914 0.1.994769 0.1.999385 0.1.999698 0.1.999999
4,
[::]
[ 01.·999385 ] .
[�l
n
The successive vectors
are called iterates, so, for example, when n
the fourth iterate is
.......
Iterative Methods for Solving Linear Systems
approaching
=
We can see that the iterates in this example are
which is the exact solution of the given system. (Check this.)
4
We say in this case that Jacobi's method converges.
2. 1 3.
Jacobi's method calculates the successive iterates in a two-variable system accord­
ing to the crisscross pattern shown in Table
1
0
Ta b l e 2 . 1 3
n
The Gauss-Seidel method is named
after C. F. Gauss and Philipp Ludwig
von Seidel
Seidel
worked in analysis, probability
theory, astronomy, and optics.
Unfortunately, he suffered from
eye problems and retired at a young
age. The paper in which he described
the method now known as Gauss­
Seidel was published in
Gauss,
it seems, was unaware of the
method!
(1821-1896).
1874.
2
3
Before we consider Jacobi's method in the general case, we will look at a
modification of it that often converges faster to the solution. The Gauss-Seidel method
is the same as the Jacobi method except that we use each new value as soon as we can.
=*
So in our example, we begin by calculating x1 = +
as before, but
we now use this value of x 1 to get the next value of x 2 :
(5 0)/7 0.7 14
7 53. 1.829
=
Xz
=
2. 1 4.
+
2.
---7
=
We then use this value of x2 to recalculate x 1 , and so on. The iterates this time are
shown in Table
We observe that the Gauss-Seidel method has converged faster to the solu­
tion. The iterates this time are calculated according to the zigzag pattern shown in
Table
2. 1 5.
0
Ta b l e 2 . 1 4
n
1
2
3
4
5
00 0.1.872914 0.1.997685 0.1.999998 2.1.000000 2.1.000000
126
Chapter
2
Systems of Linear Equations
1
0
Ta b l e 2 . 1 5
n
2
3
Xz
The Gauss-Seidel method also has a nice geometric interpretation in the case of
two variables. We can think of x 1 and x2 as the coordinates of points in the plane. Our
starting point is the point corresponding to our initial approximation, (0, O). Our first
calculation gives x 1 = � , so we move to the point ( �, O) (0.714, 0). Then we compute
x2 = � 1 .829, which moves us to the point ( � , �) (0.714, 1 .829). Continuing in
this fashion, our calculations from the Gauss-Seidel method give rise to a sequence
of points, each one differing from the preceding point in exactly one coordinate. If
we plot the lines 7x 1 - x2 = 5 and 3x 1 - 5x2 = - 7 corresponding to the two given
equations, we find that the points calculated above fall alternately on the two lines,
as shown in Figure 2.27. Moreover, they approach the point of intersection of the
lines, which corresponds to the solution of the system of equations. This is what
convergence means!
=
=
=
2
05
.
0.2
-
0.4
05
.
-1
Figure 2 . 2 1
Converging iterates
The general cases of the two methods are analogous. Given a system of n linear
equations in n variables,
a i 1X1 + a i 2X2 + · · · +
a z1 X1 + a z 2X2 + · · · +
a i n Xn = b i
a z n Xn = b z
(2)
we solve the first equation for x 1 , the second for x2 , and so on. Then, beginning
with an initial approximation, we use these new equations to iteratively update each
Section
2. 5
Iterative Methods for Solving Linear Systems
121
variable. Jacobi's method uses all of the values at the kth iteration to compute the
(k + l )st iterate, whereas the Gauss-Seidel method always uses the most recent value
of each variable in every calculation. Example 2.39 later illustrates the Gauss-Seidel
method in a three-variable problem.
At this point, you should have some questions and concerns about these iterative
methods. (Do you?) Several come to mind: Must these methods converge? If not,
when do they converge? If they converge, must they converge to the solution? The
answer to the first question is no, as Example 2.38 illustrates.
Apply the Gauss-Seidel method to the system
Exa m p l e 2 . 3 8
X1 - Xz = 1
2x 1 + Xz = 5
with initial approximation
Solution
[ � ].
We rearrange the equations to get
X1 = 1 + X2
X2 = 5 - 2x 1
�
[:: ] [ � ].
The first few iterates are given in Table 2 . 1 6. (Check these.)
The actual solution to the given system is
=
Clearly, the iterates in
Table 2 . 1 6 are not approaching this point, as Figure 2.28 makes graphically clear in an
example of divergence.
0
Ta b l e 2 . 1 6
n
Xi
X2
0
0
1
3
2
4
-3
3
-2
9
4
10
- 15
5
- 14
33
-4
Fioure 2.28
Diverging iterates
128
Chapter
2
Systems of Linear Equations
S o when do these iterative methods converge? Unfortunately, the answer to this
question is rather tricky. We will answer it completely in Chapter 7, but for now we
will give a partial answer, without proof.
Let A be the n X n matrix
[ ""
a12
a
a
A = .2 1 22
a n) a n2
a,,
a2 n
a nn
l
We say that A is strictly diagonally dominant if
l I
a 11 >
l I
a22 >
l I l I
l l
a 1 2 + a 13 + · · · + a 1n
l I l I
l l
a2 1 + a2 3 + · · · + a2 n
That is, the absolute value of each diagonal entry a11 , aw . . . , ann is greater than the
sum of the absolute values of the remaining entries in that row.
Theorem 2 . 9
If a system of n linear equations in n variables has a strictly diagonally domi­
nant coefficient matrix, then it has a unique solution and both the Jacobi and the
Gauss-Seidel method converge to it.
Remark Be warned! This theorem is a one-way implication. The fact that a
system is not strictly diagonally dominant does not mean that the iterative meth­
ods diverge. They may or may not converge. (See Exercises 1 5 - 1 9.) Indeed, there are
examples in which one of the methods converges and the other diverges. However, if
either of these methods converges, then it must converge to the solution-it cannot
converge to some other point.
Theorem 2 . 1 0
If the Jacobi or the Gauss-Seidel method converges for a system of n linear
equations in n variables, then it must converge to the solution of the system.
Proof We will illustrate the idea behind the proof by sketching it out for the case of
Jacobi's method, using the system of equations in Example 2.37. The general proof
is similar.
Convergence means that "as iterations increase, the values of the iterates get
closer and closer to a limiting value:' This means that x1 and x2 converge to r and s,
respectively, as shown in Table 2 . 1 7.
We must prove that
[:: ] [:]
is the solution of the system of equations. In
other words, at the ( k + l ) st iteration, the values of x1 and x2 must stay the same as at
Section
2. 5
Iterative Methods for Solving Linear Systems
k+ I
k
Ta b l e 2 . 1 1
n
r
s
+
+
k+2
r
s
r
s
the kth iteration. But the calculations give x1 = (5
(7 3x1)/5 = (7 3r)/5. Therefore,
Rearranging, we see that
5
+s=r
--
7
and
7
129
+ x2 )/7 = (5 + s)/7 and x2 =
+ 3r = s
---
5
7r - s = 5
3r - 5s = - 7
Thus, x1 = r, x2 = s satisfy the original equations, as required.
By now you may be wondering: If iterative methods don't always converge to the
solution, what good are they? Why don't we just use Gaussian elimination? First, we
have seen that Gaussian elimination is sensitive to roundoff errors, and this sensitiv­
ity can lead to inaccurate or even wildly wrong answers. Also, even if Gaussian elimi­
nation does not go astray, we cannot improve on a solution once we have found it. For
example, if we use Gaussian elimination to calculate a solution to two decimal places,
there is no way to obtain the solution to four decimal places except to start over again
and work with increased accuracy.
In contrast, we can achieve additional accuracy with iterative methods simply by
doing more iterations. For large systems, particularly those with sparse coefficient
matrices, iterative methods are much faster than direct methods when implemented
on a computer. In many applications, the systems that arise are strictly diagonally
dominant, and thus iterative methods are guaranteed to converge. The next example
illustrates one such application.
Exa m p l e 2 . 3 9
Suppose we heat each edge of a metal plate to a constant temperature, as shown in
Figure 2 . 2 9 .
50°
Figure 2 . 2 9
A heated metal plate
oo
130
Chapter
2
Systems of Linear Equations
Eventually the temperature at the interior points will reach equilibrium, where the
following property can be shown to hold:
The temperature at each interior point P on a plate is the average of the tempera­
tures on the circumference of any circle centered at P inside the plate (Figure 2.30).
figure 2 . 3 0
To apply this property in an actual example requires techniques from calculus. As
an alternative, we can approximate the situation by overlaying the plate with a grid,
or mesh, that has a finite number of interior points, as shown in Figure 2.3 1 .
Figure 2 . 3 1
The discrete version of the heated
plate problem
The discrete analogue of the averaging property governing equilibrium tempera­
tures is stated as follows:
The temperature at each interior point P is the average of the temperatures at the
points adjacent to P.
For the example shown in Figure 2.3 1 , there are three interior points, and each is
adjacent to four other points. Let the equilibrium temperatures of the interior points
Section
2. 5
Iterative Methods for Solving Linear Systems
131
be t 1 , t2 , and t3 , as shown. Then, by the temperature-averaging property, we have
t! =
1 00 + 1 00 +
G
+ 50
------
4
t + t + 0 + 50
t1 = 1 3
4
1 00 + 1 00 + o +
t3 =
4
_
_
_
_
_
_
_
or
(3)
G
250
- t1 + 4t2 - t3 = 50
- t1 + 4t3 = 200
Notice that this system is strictly diagonally dominant. Notice also that Equa­
tions (3) are in the form required for Jacobi or Gauss-Seidel iteration. With an initial
approximation of t i = 0, t2 = 0, t3 = 0, the Gauss-Seidel method gives the following
iterates.
Iteration 1 :
t1 =
t1 =
t3 =
Iteration 2:
t1 =
t1 =
t3 =
�
1 00 + 1 00 + 0 + 50
= 62.5
4
62.5 + 0 + 0 + 50
= 28. 125
4
1 00 + 1 00 + 0 + 28. 125
= 57.03 1
4
100 + 1 00 + 28. 125 + 50
= 69.5 3 1
4
69.53 1 + 57.03 1 + 0 + 50
= 44. 141
4
100 + 1 00 + 0 + 44. 141
= 6 1 .035
4
Continuing, we find the iterates listed in Table 2 . 1 8 . We work with five-significant­
digit accuracy and stop when two successive iterates agree within 0.00 1 in all
variables.
Thus, the equilibrium temperatures at the interior points are (to an accuracy of
0.00 1 ) t i = 74. 1 08, t2 = 46.430, and t3 = 6 1 .607. (Check these calculations.)
By using a finer grid (with more interior points), we can get as precise informa­
tion as we like about the equilibrium temperatures at various points on the plate.
0
Ta b l e 2 . 1 8
n
ti
t1
t3
0
0
0
1
62.500
28. 125
57.03 1
2
69.53 1
44 . 1 4 1
6 1 .035
3
73.535
4 6. 1 4 3
6 1 .536
7
74. 107
46.429
6 1 .607
8
74. 107
46.429
6 1 .607
4
132
..
GAS
I
Chapter
2
Systems of Linear Equations
Exercises 2 . 5
In Exercises 1 -6, apply Jacobi's method to the given system.
Take the zero vector as the initial approximation and work
with four-significant-digit accuracy until two successive
iterates agree within 0.001 in each variable. In each case,
compare your answer with the exact solution found using
any direct method you like.
1 . 7X 1 - X2 = 6
2. 2X 1 + X2 = 5
X 1 - 5X2 = - 4
X 1 - X2 = 1
3. 4.5X 1 - 0.5X2 = 1
X 1 - 3.5X2 = - 1
4. 20X 1 + X2 - X3 = 1 7
X1 - 10x2 + x3 = 1 3
-x 1 + X2 + 10x3 = 1 8
5 . 3X 1 + X2
=1
X 1 + 4X2 + X3 = 1
X2 + 3X3 = 1
=1
6. 3X 1 - X2
=0
-X 1 + 3X2 - X3
-X2 + 3X3 - X4 =
-x3 + 3X4 = 1
In Exercises 7-12, repeat the given exercise using the Gauss­
Seidel method. Take the zero vector as the initial approxi­
mation and work with four-significant-digit accuracy until
two successive iterates agree within 0.001 in each variable.
Compare the number of iterations required by the Jacobi
and Gauss-Seidel methods to reach such an approximate
solution.
7. Exercise 1
8. Exercise 2
9. Exercise 3
10. Exercise 4
1 1 . Exercise 5
12. Exercise 6
In Exercises 13 and 14, draw diagrams to illustrate the con­
vergence of the Gauss-Seidel method with the given system.
13. The system in Exercise 1
14. The system in Exercise 2
In Exercises 15 and 1 6, compute the first four iterates,
using the zero vector as the initial approximation, to show
that the Gauss-Seidel method diverges. Then show that
the equations can be rearranged to give a strictly diagonally
dominant coefficient matrix, and apply the Gauss-Seidel
method to obtain an approximate solution that is accurate
to within 0.001.
16. X 1 - 4X2 + 2X3 = 2
15. X 1 - 2X2 = 3
3x 1 + 2x2 = 1
2x2 + 4x3 = 1
6X 1 - X2 - 2X3 = 1
17. Draw a diagram to illustrate the divergence of the
Gauss-Seidel method in Exercise 15.
In Exercises 18 and 1 9, the coefficient matrix is not strictly
diagonally dominant, nor can the equations be rearranged
to make it so. However, both the Jacobi and the Gauss-Seidel
method converge anyway. Demonstrate that this is true of
the Gauss-Seidel method, starting with the zero vector as
the initial approximation and obtaining a solution that is
accurate to within 0.01.
18. - 4x 1 + 5X2 = 14
X 1 - 3x2 = - 7
19. 5x 1 - 2x2 + 3x3 = - 8
X1 + 4X2 - 4X3 = 102
- 2x1 - 2x2 + 4x3 = - 90
20. Continue performing iterations in Exercise 18 to
obtain a solution that is accurate to within 0.00 1 .
2 1 . Continue performing iterations in Exercise 19 to
obtain a solution that is accurate to within 0.00 1 .
In Exercises 22-24, the metal plate has the constant tem­
peratures shown on its boundaries. Find the equilibrium
temperature at each of the indicated interior points by
setting up a system of linear equations and applying either
the Jacobi or the Gauss-Seidel method. Obtain a solution
that is accurate to within 0.001.
22.
50
50
Section
oo
oo
23.
t1
oo
100°
t2
t3
24.
oo
40°
100°
!00°
100°
oo
20°
f1
t2
f3
t4
20°
100°
100°
40°
In Exercises 25 and 26, we refine the grids used in Exer­
cises 22 and 24 to obtain more accurate information about
the equilibrium temperatures at interior points of the plates.
Obtain solutions that are accurate to within 0.001, using
either the Jacobi or the Gauss-Seidel method.
25.
oo
oo
t2
oo
t4
26.
oo
50
oo
50
20°
20°
oo
f1
t2
t3
t4
20°
oo
15
16
I]
lg
20°
40°
19
110
11 1
112
100°
40°
t1 3
t14
t15
t1 6
100°
40°
133
number line so that its ends are at 0 and 1 . The paper
is folded in half, right end over left, so that its ends
are now at 0 and i. Next, it is folded in half again, this
time left end over right, so that its ends are at ! and t .
Figure 2.32 shows this process. We continue folding
the paper in half, alternating right-over-left and left­
over-right. If we could continue indefinitely, it is clear
that the ends of the paper would converge to a point. It
is this point that we want to find.
(a) Let x 1 correspond to the left-hand end of the paper
and x2 to the right-hand end. Make a table with the
first six values of [x 1 , x2 ] and plot the correspond­
ing points on x 1 , x2 coordinate axes.
(b) Find two linear equations of the form x2 = ax1 + b
and x 1 = cx 2 + d that determine the new values
of the endpoints at each iteration. Draw the corre­
sponding lines on your coordinate axes and show
that this diagram would result from applying the
Gauss-Seidel method to the system of linear equa­
tions you have found. (Your diagram should resem ble Figure 2.27 on page 1 2 6.)
(c) Switching to decimal representation, continue
applying the Gauss-Seidel method to approximate
the point to which the ends of the paper are con­
verging to within 0.00 1 accuracy.
(d) Solve the system of equations exactly and compare
your answers.
28. An ant is standing on a number line at point A. It
walks halfway to point B and turns around. Then it
walks halfway back to point A, turns around again,
and walks halfway to point B. It continues to do this
indefinitely. Let point A be at 0 and point B be at 1 .
ts
50
Iterative Methods for Solving Linear Systems
Exercises 27 and 28 demonstrate that sometimes, if we are
lucky, the form of an iterative problem may allow us to use a
little insight to obtain an exact solution.
27. A narrow strip of paper 1 unit long is placed along a
oo
t4
2. 5
40° 100° 100°
The ant's walk is made up of a sequence of overlap­
ping line segments. Let x 1 record the positions of
the left-hand endpoints of these segments and x2
their right-hand endpoints. (Thus, we begin with
x 1 = 0 and x2 = t . Then we have x 1 = ! and x2 = t,
and so on. ) Figure 2.33 shows the start of the ant's
walk.
(a) Make a table with the first six values of [x1 , x2 ] and
plot the corresponding points on x 1 , x2 coordinate
axes.
(b) Find two linear equations of the form x2 = ax1 + b
and x 1 = cx2 + d that determine the newvalues ofthe
endpoints at each iteration. Draw the corresponding
Chapter
134
2
Systems of Linear Equations
�
0
I
........----..
!
I
0
D
0
l
4
l
2
0
!"tf
l
4
I
3
4
2
)
l
2
0
I
3
4
�
l
4
l
2
_5
0
l
4
�
l
2
Figure 2 . 3 2
Figure 2 . 3 3
Folding a strip of paper
The ant's walk
5
8
3
4
lines on your coordinate axes and show that this dia­
gram would result from applying the Gauss-Seidel
method to the system of linear equations you have
found. (Your diagram should resemble Figure 2.27
on page 126.)
(c) Switching to decimal representation, continue
applying the Gauss-Seidel method to approximate
the values to which x 1 and x2 are converging to
within 0.001 accuracy.
(d) Solve the system of equations exactly and compare
your answers. Interpret your results.
3
4
Chapter Review
Kev Defi nitions and concepts
augmented matrix, 61, 64
back substitution, 61
coefficient matrix, 64
consistent system, 60
convergence, 125- 126
divergence, 127
elementary row operations, 66
free variable, 71
Gauss-Jordan elimination, 73
Gauss-Seidel method, 124
Gaussian elimination, 68-69
homogeneous system, 76
inconsistent system, 60
iterate, 125
Jacobi's method, 124
leading variable (leading 1), 71 -73
linear equation, 58
linearly dependent vectors, 93
linearly independent
vectors, 93
1. Mark each of the following statements true or false:
Review Questions
(a) Every system of linear equations has a solution.
(b) Every homogeneous system of linear equations
has a solution.
(c) If a system of linear equations has more vari­
ables than equations, then it has infinitely many
solutions.
(d) If a system of linear equations has more equations
than variables, then it has no solution.
pivot, 66
rank of a matrix, 72
Rank Theorem, 72
reduced row echelon form, 73
row echelon form, 65
row equivalent matrices, 68
span of a set of vectors, 90
spanning set, 90
system of linear
equations, 59
(e) Determining whether b is in span(a 1 , . . . , an ) is
equivalent to determining whether the system
[A I b] is consistent, where A = [a 1 . . . an l ·
(f) In IR 3 , span( u, v) is always a plane through the origin.
( g) In IR 3 , if nonzero vectors u and v are not parallel,
then they are linearly independent.
(h) In IR 3 , if a set of vectors can be drawn head to tail,
one after the other so that a closed path (polygon)
is formed, then the vectors are linearly dependent.
Chapter Review
(i) If a set of vectors has the property that no
two vectors in the set are scalar multiples of
one another, then the set of vectors is linearly
independent.
(j) If there are more vectors in a set of vectors than
the number of entries in each vector, then the set
of vectors is linearly dependent.
2. Find the rnAA of the ma1'U
]
[ i �; �
_
3. Solve the linear system
3
3
-3
6
2
4
.
2
2
x + y - 2z = 4
x + 3y - z = 7
2x + y - Sz = 7
4. Solve the linear system
-
3w + 8x - 18y + z = 35
w + 2x 4y = 1 1
w + 3x - 7y + z = 10
5. Solve the linear system
2x + 3y = 4
x + 2y = 3
over 2 7 .
6. Solve the linear system
over 2 5 .
3x + 2y = 1
x + 4y = 2
[� I � ]
7. For what value(s) of k is the linear system with
augmented matrix
2
2k
inconsistent?
8. Find parametric equations for the line of intersection
of the planes x + 2y + 3z = 4 and Sx + 6y + 7z = 8.
9. Find the point of intersection of the following lines, if
it exists.
135
[} m
1 1 . Find the general equation of the plane spanned by
nd
[ � ] , [ -- � ] , [ ! ]
w)
u � [ } � [ Hw � [ : ]
u � [ -} � [ -n w � [ -:]
12. ? etermine whether
mdependent.
-3
13. Determine whether IR 3
( a)
(b)
=
2
are linearly
-2
span(u, v,
if:
14. Let a1, a2 , a3 be linearly independent vectors in IR 3 , and
let A = [ a1 a2 a3 ] . Which of the following statements
are true?
( a) The reduced row echelon form of A is 13 .
(b) The rank of A is 3.
(c) The system [A I b] has a unique solution for any
vector b in IR 3 .
(d) (a), (b), and (c) are all true.
(e) (a) and (b) are both true, but not (c) .
15. Let a1, a2 , a3 be linearly dependent vectors in IR 3 , not
all zero, and let A = [ a1 a2 a3 ] . What are the possible
values of the rank of A?
16. What is the maximum rank of a 5 X 3 matrix? What is
the minimum rank of a 5 X 3 matrix?
17. Show that if u and v are linearly independent vectors,
then so are u + v and u - v.
18. Show that span(u, v) = span(u, u + v) for any vectors
u and v.
19. In order for a linear system with augmented matrix
[A I b] to be consistent, what must be true about the
ranks of A and [A I b] ?
20. Are ilie matci=
O
U ! l [i -:i
-
row equivalent? Why or why not?
nd
M atri ces
We [Halmos and KaplanskyJ share
a philosophy about linear algebra:
we think basisjree, we write
basis-free, but when the chips are
down we close the office door and
compute with matrices like fury.
-Irving Kaplansky
In Paul Halmos: Celebrating
J. H. Ewing and Gehring,
eds. Springer-Verlag, 1991, p. 88
50
Years of Mathematics
F. W
3.0
I n t ro d u ctio n : M atrices i n Acti o n
In this chapter, we will study matrices in their own right. We have already used
matrices-in the form of augmented matrices-to record information about and to
help streamline calculations involving systems of linear equations. Now you will see
that matrices have algebraic properties of their own, which enable us to calculate
with them, subject to the rules of matrix algebra. Furthermore, you will observe that
matrices are not static objects, recording information and data; rather, they represent
certain types of functions that "act" on vectors, transforming them into other vectors.
These "matrix transformations" will begin to play a key role in our study of linear
algebra and will shed new light on what you have already learned about vectors and
systems of linear equations. Furthermore, matrices arise in many forms other than
augmented matrices; we will explore some of the many applications of matrices at the
end of this chapter.
In this section, we will consider a few simple examples to illustrate how matri­
ces can transform vectors. In the process, you will get your first glimpse of "matrix
arithmetic:'
Consider the equations
Y 1 = X 1 + 2Xz
3x2
Yz =
We can view these equations as describing a transformation of the vector x
[;:]
[ � � ],
into the vector y
by F, then F
=
=
( 1)
=
[:: ]
. If we denote the matrix of coefficients of the right-hand side
and we can rewrite the transformation as
or, more succinctly, y = Fx. [Think of this expression as analogous to the functional
notation y = f( x) you are used to: x is the independent "variable" here, y is the depen­
dent "variable;' and F is the name of the "function:']
136
Section
Thus, if x =
[ - � ],
3. 0
Introduction: Matrices in Action
then the Equations ( 1) give
3 3
[ � ] [ � � ] [ - � ].
y1 = - 2 + 2 · 1 = 0
or y =
·1 =
y2 =
We can write this expression as
Problem 1
131
[ o3 ]
=
Compute Fx for the following vectors x:
(a) x =
[�]
(b) x =
[ �]
_
(c) x =
[ = �J
Problem 2 The heads of the four vectors x in Problem 1 locate the four corners
of a square in the x 1 x2 plane. Draw this square and label its corners A, B, C, and D,
corresponding to parts (a), (b), (c), and (d) of Problem 1 .
O n separate coordinate axes (labeled y 1 and y2 ), draw the four points determined
by Fx in Problem 1 . Label these points A', B', C', and D'. Let's make the (reasonable)
assumption that the line segment AB is transformed into the line segment A ' B ', and
likewise for the other three sides of the square ABCD. What geometric figure is rep­
resented by A 'B' C'D'?
Problem 3 Th e center o f square ABCD i s the origin 0 =
A' B' C' D ' ? What algebraic calculation confirms this?
Now consider the equations
[ �]. What i s the center of
Z1
that transform a vector y =
[ �: ]
=
Y1 - Yz
=
2y
Zz
l
into the vector z =
transformation as z = Gy, where
G=
[ 1 -1]
-2
0
(2)
[:: J.
We can abbreviate this
Problem 4 We are going to find out how G transforms the figure A 'B' C'D'.
Compute Gy for each of the four vectors y that you computed in Problem 1 . [That
is, compute z = G ( Fx). You may recognize this expression as being analogous to
the composition of functions with which you are familiar.] Call the corresponding
points A " , B " , C " , and D " , and sketch the figure A "B " C "D " on z 1 z2 coordinate axes.
Problem 5 By substituting Equations ( 1 ) into Equations ( 2 ) , obtain equations for
z 1 and z2 in terms of x1 and x2 • If we denote the matrix of these equations by H, then
we have z = Hx. Since we also have z = GFx, it is reasonable to write
H = GF
Can you see how the entries of H are related to the entries of F and G?
Problem 6 Let's do the above process the other way around: First transform the
square ABCD, using G, to obtain figure A * B * C * D * . Then transform the resulting
figure, using F, to obtain A ** B ** C ** D ** . [Note: Don't worry about the "variables" x,
138
Chapter
3
Matrices
y,
(2)
and z here. Simply substitute the coordinates of A, B, C, and D into Equations
and then substitute the results into Equations ( l ).] Are A * *B* *C**D** and A"B"C"D"
the same? What does this tell you about the order in which we perform the transfor­
mations and G?
Problem 1 Repeat Problem 5 with general matrices
F
F = [ ] = [ggll21 gg2212 ] ,
(2) F
=
i
!ll
21
!
!1 2 ,
22
G
and H
= [hh2ll1 hh2212 ]
F
That is, if Equations ( 1) and Equations
have coefficients as specified by and G,
find the entries of H in terms of the entries of and G. The result will be a formula
for the "product" H GF.
Problem 8 Repeat Problems 1 -6 with the following matrices. (Your formula from
Problem 7 may help to speed up the algebraic calculations.) Note any similarities or
differences that you think are significant.
F = [� -�l = [� �] F = �l = [� � ]
(c) F= [ � � ],c = [ _� - � ] F = - 2 ] = [ 2 � ]
( a)
(b)
G
G
( d)
4
,G
1
M atrix O p e rati o n s
Although we have already encountered matrices, we begin by stating a formal
definition and recording some facts for future reference.
A matrix is a rectangular array of numbers called the entries, or
elements, of the matrix.
D e fi n it i o n
Although numbers will usually be
chosen from the set � of real numbers, they may also be taken from
the set C of complex numbers or
from "ll_P' where p is prime.
Technically, there is a distinction
between row/column matrices
and vectors, but we will not be­
labor this distinction. We will,
however, distinguish between
row matrices/vectors and column
matrices/vectors. This distinction
is important-at the very least­
for algebraic computations, as we
will demonstrate.
The following are all examples of matrices:
[� � l [ �
�l
-1
'TT
Ul
[1
1
1
1],
[ 1.2
51
6.9 0
- 7.3 9
-1
4.4 , [ 7 ]
8.5
l
The size of a matrix is a description of the numbers of rows and columns it has. A
matrix is called m X n (pronounced " m by n") if it has m rows and n columns. Thus,
the examples above are matrices of sizes X
X 3, 3 X 1, 1 X 4, 3 X 3, and 1 X 1,
respectively. A 1 X m matrix i s called a row matrix (or row vector), and an n X 1
matrix is called a column matrix (or column vector) .
We use double-subscript notation to refer to the entries of a matrix A. The entry of
A in row i and column j is denoted by a ij · Thus, if
=
=
2 2, 2
A = [�
9
5
then a 1 3 - 1 and a 22 5. (The notation A ij is sometimes used interchangeably with
a ij .) We can therefore compactly denote a matrix A by [a;j ] (or [a;j l m x n if it is impor­
tant to specify the size of A, although the size will usually be clear from the context) .
Section
[
3.1
Matrix Operations
139
With this notation, a general m X n matrix A has the form
A=
a ., a 12
a 21 a zz
a ml a m 2
If the columns of A are the vectors a1, a2 , . . . , an , then we may represent A as
A = [ a 1 a2 · · · a" ]
If the rows of A are A 1 , A2 , . . . , Am , then we may represent A as
A�
[]J
The diagonal entries of A are a11, aw a 33 , . . . , and if m = n (that is, if A has the same
number of rows as columns), then A is called a square matrix. A square matrix whose
nondiagonal entries are all zero is called a diagonal matrix. A diagonal matrix all
of whose diagonal entries are the same is called a scalar matrix. If the scalar on the
diagonal is 1, the scalar matrix is called an identity matrix.
For example, let
[
0
6
0
2 5
A=
-1 4
The diagonal entries of A are 2 and 4, but A is not square; B is a square matrix of size
2 X 2 with diagonal entries 3 and 5; C is a diagonal matrix; D is a 3 X 3 identity ma­
trix. The n X n identity matrix is denoted by In (or simply I if its size is understood).
Since we can view matrices as generalizations of vectors (and, indeed, matrices
can and should be thought of as being made up of both row and column vectors),
many of the conventions and operations for vectors carry through (in an obvious
way) to matrices.
Two matrices are equal if they have the same size and if their corresponding
entries are equal. Thus, if A = [a ij l m x n and B = [b ij lrxs' then A = B if and only if
m = r and n = s and a
ij = b;j for all i and j.
Exa m p l e 3 . 1
Consider the matrices
OJ
3
, and C =
[
2 0 x
5 3 y
]
Neither A nor B can be equal to C (no matter what the values of x and y), since A and
B are 2 X 2 matrices and C is 2 X 3. However, A = B if and only if a = 2, b = 0, c = 5,
and d = 3.
Exa m p l e 3 . 2
Consider the matrices
R � I 1 4 3 ] •nd c �
[r
4
140
Chapter
3
Matrices
Despite the fact that R and C have the same entries in the same order, R of- C since
R is 1 X 3 and C is 3 X 1 . (If we read R and C aloud, they both sound the same:
"one, four, three:') Thus, our distinction between row matrices/vectors and column
matrices/vectors is an important one.
Matrix Addilion a n d scalar M u ltiolicalion
Generalizing from vector addition, we define matrix addition componentwise. If A =
[a;) and B = [b;) are m X n matrices, their sum A + B is the m X n matrix obtained
by adding the corresponding entries. Thus,
[We could equally well have defined A + B in terms of vector addition by specifying
that each column (or row) of A + B is the sum of the corresponding columns (or
rows) of A and B.] If A and B are not the same size, then A + B is not defined.
Exa m p l e 3 . 3
Let
A =
Then
[
OJ [
]
- �]
[
1 4
-1
4 3
-3
, and C =
, B=
2 1
3 0 2
-2 6 5
but neither A + C nor B + C is defined.
]
The componentwise definition o f scalar multiplication will come a s n o surprise. If
A is an m X n matrix and c is a scalar, then the scalar multiple cA is the m X n matrix
obtained by multiplying each entry of A by c. More formally, we have
[In terms of vectors, we could equivalently stipulate that each column (or row) of
cA is c times the corresponding column (or row) of A.]
Exa m p l e 3 . 4
For matrix A in Example 3.3,
2A =
[
2 8
- 4 12
-4
-6
-�J
4
The matrix ( - l )A is written as -A and called the negative of A. As with vectors, we
can use this fact to define the difference of two matrices: If A and B are the same size,
then
A - B = A + ( - B)
Section
Exa m p l e 3 . 5
For matrices A and B in Example 3.3,
A-B=
[l
1
0
4
-2 6
3.1
-�]
Matrix Operations
3
6
141
�]
A matrix all of whose entries are zero is called a zero matrix and denoted by 0 (or
Om x n if it is important to specify its size) . It should be clear that if A is any matrix and
0 is the zero matrix of the same size, then
A+O=A=O+A
and
A - A = 0 = -A + A
M alrix M u ltiplication
Mathematicians are sometimes like
Lewis Carroll's Humpty Dumpty:
"When I use a word;' Humpty
Dumpty said, "it means just what
I choose it to mean-neither more
nor less" (from Through the Look­
ing Glass).
The Introduction in Section 3.0 suggested that there is a "product" of matrices that is
analogous to the composition of functions. We now make this notion more precise.
The definition we are about to give generalizes what you should have discovered in
Problems 5 and 7 in Section 3.0. Unlike the definitions of matrix addition and scalar
multiplication, the definition of the product of two matrices is not a componentwise
definition. Of course, there is nothing to stop us from defining a product of matrices
in a componentwise fashion; unfortunately such a definition has few applications and
is not as "natural" as the one we now give.
If A is an m x n matrix and B is an n x r matrix, then the product
C = AB is an m X r matrix. The (i, j) entry of the product is computed as
D e fi n it i o n
follows:
R e m a rks
Notice that A and B need not be the same size. However, the number of col­
umns of A must be the same as the number of rows of B. If we write the sizes of A, B,
•
and AB in order, we can see at a glance whether this requirement is satisfied. More­
over, we can predict the size of the product before doing any calculations, since the
number of rows of AB is the same as the number of rows of A, while the number of
columns of AB is the same as the number of columns of B, as shown below:
A
m X n
B
n X r
Size of AB
AB
m X r
142
Chapter
3
Matrices
The formula for the entries of the product looks like a dot product, and indeed
it is. It says that the (i, j) entry of the matrix AB is the dot product of the ith row of A
and the jth column of B:
•
a 11 a 1 2
a ln
a i1
a i2
a in
a ml a m 2
a mn
[
b"
b2 1
b l)
b2j
b,,
b2 r
b nl
b nj
b nr
:
Notice that, in the expression c iJ = a i 1 b11 + a i2 b 21 + · · · + a in b n) ' the "outer subscripts"
on each ab term in the sum are always i and j whereas the "inner subscripts" always
agree and increase from 1 to n. We see this pattern clearly if we write c iJ using sum­
mation notation:
n
Ci) = � a ikbkj
k� I
Exa m p l e 3 . 6
Compute AB if
[
1
3
A=
-2 - 1
-[ : �
-�]
3
- -1
and B =
-1 2 0
Since A is 2 X 3 and B is 3 X 4, the product AB is defined and will be a
2 X 4 matrix. The first row of the product C = AB is computed by taking the dot
product of the first row of A with each of the columns of B in turn. Thus,
C11 = 1 ( - 4) + 3(5) + ( - 1) ( - 1 ) = 12
C1 2 = 1 (0) + 3( - 2) + ( - 1 ) (2) = - 8
C1 3 = 1 ( 3) + 3( - 1) + ( - 1) (0) = 0
C1 = l ( - 1) + 3( 1) + ( - 1) (6) = - 4
The second row of C is computed by taking the dot product of the second row of A
with each of the columns of B in turn:
Solulion
4
Cz1 =
C22 =
Cz 3 =
c2 4 =
( - 2)( - 4) +
( - 2) ( 0) +
( - 2) ( 3) +
( - 2)( - l) +
( - 1)(5)
( - 1 ) ( - 2)
(- 1)(- 1)
( - l ) ( l)
Thus, the product matrix is given by
AB =
[
12 - 8 0
2 4 -5
+
+
+
+
( 1) ( - 1) = 2
( 1) ( 2 ) = 4
( 1) ( 0) = - 5
( 1 ) (6) = 7
- �]
(With a little practice, you should b e able t o d o these calculations mentally without
writing out all of the details as we have done here. For more complicated examples, a
calculator with matrix capabilities or a computer algebra system is preferable.)
4
Section
3.1
Matrix Operations
143
Before we go further, we will consider two examples that justify our chosen
definition of matrix multiplication.
Exa m p l e 3 . 1
Ann and Bert are planning to go shopping for fruit for the next week. They each want
to buy some apples, oranges, and grapefruit, but in differing amounts. Table 3.1 lists
what they intend to buy. There are two fruit markets nearby-Sam's and Theo's-and
their prices are given in Table 3.2. How much will it cost Ann and Bert to do their
shopping at each of the two markets?
Ta b l e 3 . 2
Ta b l e 3 . 1
Ann
Bert
Apples
6
4
Grapefruit
3
8
Oranges
10
5
Solution
Theo's
$0. 15
$0.30
$0.20
Sam's
$0. 10
$0.40
$0. 10
Apple
Grapefruit
Orange
If Ann shops at Sam's, she will spend
6(0.10) + 3(0.40) + 10(0. 10) = $2.80
If she shops at Theo's, she will spend
6(0. 1 5) + 3(0.30) + 10(0.20) = $3.80
Bert will spend
4(0. 10) + 8(0.40) + 5(0. 10) = $4. 10
at Sam's and
4(0. 15) + 8(0.30) + 5(0.20) = $4.00
at Theo's. (Presumably, Ann will shop at Sam's while Bert goes to Theo's.)
The "dot product form" of these calculations suggests that matrix multiplication
is at work here. If we organize the given information into a demand matrix D and a
price matrix P, we have
D=
[
[6
3
4 8
0. 10 0. 15
and P = 0.40 0.30
][
]
0. 10 0.20
]
The calculations above are equivalent to computing the product
Ta b l e 3 . 3
Ann
Bert
Sam's
$2.80
$4. 10
Theo's
$3.80
$4.00
[6
[
0. 10 0. 15
2.80 3.80
3 10
DP =
0.40 0.30 4. 10 4.00
4 8 5
0. 10 0.20
]
Thus, the product matrix DP tells us how much each person's purchases will cost at
each store (Table 3.3).
144
Chapter
3
Matrices
Exa m p l e 3 . 8
Consider the linear system
X 1 - 2X2 + 3X3
-X 1 + 3X2 + X3
2x 1 - x2 + 4x3
=
=
=
5
1
14
(1)
Observe that the left-hand side arises from the matrix product
so the system ( 1) can be written as
or A x = b, where A is the coefficient matrix, x is the (column) vector of variables, and
b is the (column) vector of constant terms.
You should have no difficulty seeing that every linear system can be written in the
form Ax = b. In fact, the notation [A I b] for the augmented matrix of a linear system
is just shorthand for the matrix equation A x = b. This form will prove to be a tre­
mendously useful way of expressing a system of linear equations, and we will exploit
it often from here on.
Combining this insight with Theorem 2.4, we see that Ax = b has a solution if
and only if b is a linear combination of the columns of A.
There is another fact about matrix operations that will also prove to be quite use­
ful: Multiplication of a matrix by a standard unit vector can be used to "pick out" or
]
[
4 2
l
and consider the
0 5 -1
products Ae 3 and e2A, with the unit vectors e 3 and e2 chosen so that the products
"reproduce" a column or row of a matrix. Let A
make sense. Thus,
2
5
[ �]
_
=
and e2A
=
[0 l ]
=
[O 5
[�
-
2
5
1]
-�J
Notice that Ae 3 gives us the third column of A and e2A gives us the second row of A.
We record the general result as a theorem.
Theorem 3 . 1
Let A be an m X n matrix, e; a 1 X
unit vector. Then
a. e; A is the ith row of A and
b. Aej is the jth column of A.
m
standard unit vector, and ej an n X 1 standard
Section
3. 1
Matrix Operations
145
We prove (b) and leave proving (a) as Exercise 4 1 . If a1, . . . , an are the columns
of A, then the product Aej can be written
Proof
Aej = Oa 1 + Oa2 +
[
·
·
·
+ laj +
·
·
·
We could also prove (b) by direct calculation:
Ae. =
}
+ Oan = aj
0
a1 1
a2 1
.
aml
0
since the 1 in ej is the jth entry.
Partitioned Matrices
It will often be convenient to regard a matrix as being composed of a number of
smaller submatrices. By introducing vertical and horizontal lines into a matrix, we
can partition it into blocks. There is a natural way to partition many matrices, par­
ticularly those arising in certain applications. For example, consider the matrix
1
0
A= 0
0
0
0
1
0
0
0
0
0
1
0
0
2 -1
3
4
1
7
0
7
2
It seems natural to partition A as
1 0 0 ! 2 -1
0 1 0i 1
3
� � � L� �
o 0 0 ! 1
7
0 0 0! 7 2
-
-----
-----
-
--------
-
[ � �]
where I is the 3 X 3 identity matrix, B is 3 X 2, 0 is the 2 X 3 zero matrix, and C is 2 X 2.
In this way, we can view A as a 2 X 2 matrix whose entries are themselves matrices.
When matrices are being multiplied, there is often an advantage to be gained by
viewing them as partitioned matrices. Not only does this frequently reveal underly­
ing structures, but it often speeds up computation, especially when the matrices are
large and have many blocks of zeros. It turns out that the multiplication of partitioned
matrices is just like ordinary matrix multiplication.
We begin by considering some special cases of partitioned matrices. Each gives
rise to a different way of viewing the product of two matrices.
Suppose A is m X n and B is n X r, so the product AB exists. If we partition B in
terms of its column vectors, as B = [b 1 : b 2 : . . . : b r ] , then
146
Chapter
3
Matrices
This result is an immediate consequence of the definition of matrix multiplication.
The form on the right is called the matrix-column representation of the product.
Exa m p l e 3 . 9
If
A
then
Ab I
=
=
[� - 1 �]
3
[ 1 - 1 :t ] � [ ' ]
�1 : 5
[3 ]
3
and Ab 2
Q
Therefore, AB = [Ab 1 : Ah2 ]
and B
=
.
2 : -2
=
=
4[ � - �1
i
[� - 1 :J [ -� ] � [ _�]
3
. (Check by ordinary matrix multiplication.)
�
Remark
Observe that the matrix-column representation of AB allows us to
write each column of AB as a linear combination of the columns of A with entries
from B as the coefficients. For example,
-1
3
(See Exercises 23 and 26.)
Suppose A is m X n and B is n X r, so the product AB exists. If we partition A in
terms of its row vectors, as
then
Once again, this result is a direct consequence of the definition of matrix multiplication.
The form on the right is called the row-matrix representation of the product.
Exa m p l e 3 . 1 0
Using the row-matrix representation, compute AB for the matrices in Example 3.9.
Section
Solution
A ,B
�
3. 1
Matrix Operations
141
We compute
[1 3 2]
Therefore ' AB
=
[ : - �]
�
[ 13 5
[-���-] [-�-�- - - - �-]
A2 B
=
and A2 B
=
[0 - 1 l ]
[� �]
4 -1
[2 -2]
as before.
2 -2 '
The definition of the matrix product AB uses the natural partition of A into rows
and B into columns; this form might well be called the row-column representation of
the product. We can also partition A into columns and B into rows; this form is called
the column-row representation of the product.
In this case, we have
so
(2)
Notice that the sum resembles a dot product expansion; the difference is that the in­
dividual terms are matrices, not scalars. Let's make sure that this makes sense. Each
term a;B; is the product of an m X 1 and a 1 X r matrix. Thus, each a;B; is an m X r
matrix-the same size as AB. The products a;B; are called outer products, and (2) is
called the outer product expansion of AB.
Exa m p l e 3 . 1 1
Compute the outer product expansion of AB for the matrices in Example 3.9.
Solution
We have
The outer products are
and
148
Chapter
3
Matrices
(Observe that computing each outer product is exactly like filling in a multiplication
table.) Therefore, the outer product expansion of AB is
]
] 4
[ -�J [: ] [
-1
3
+
0
-1
0
13 5
=
= AB
0
2 -2
+
We will make use of the outer product expansion in Chapters 5 and 7 when we
discuss the Spectral Theorem and the singular value decomposition, respectively.
Each of the foregoing partitions is a special case of partitioning in general. A ma­
trix A is said to be partitioned if horizontal and vertical lines have been introduced,
subdividing A into submatrices called blocks. Partitioning allows A to be written as a
matrix whose entries are its blocks.
For example,
1
0
A= 0
0
0
0 0 ! 2 -1
0i 1
3
0 l ..i 4 0
0 0 1
7
0 0 7 2
________________
_____________
4
and B =
-1
---------
1
0
3! 1 2
2!2 1
!
= � - _ _ } _ _ _ } _ _: _ }_
0!0 0i2
1 i0 0i3
are partitioned matrices. They have the block structures
and B =
[ BB2ll1
If two matrices are the same size and have been partitioned in the same way, it is clear
that they can be added and multiplied by scalars block by block. Less obvious is the
fact that, with suitable partitioning, matrices can be multiplied blockwise as well. The
next example illustrates this process.
Exa m p l e 3 . 1 2
Consider the matrices A and B above. If we ignore for the moment the fact that their
entries are matrices, then A appears to be a 2 X 2 matrix and B a 2 X 3 matrix. Their
product should thus be a 2 X 3 matrix given by
[
[
][
A 11 A 1 2 B 11
A 2 1 A 22 B2 1
A 11 B 11 + A , 2B2 1
A 2 1 B 1 1 + A 22B2 1
]
B 1 2 B 13
B22 B2 3
A 11 B 1 2 + A 1 2B 22 A ll B 13 + A , 2B2 3
A 2 1 B 1 2 + A 22B 22 A 2 1 B 13 + A 22B2 3
But all of the products in this calculation are actually matrix products, so we need to
AB =
]
make sure that they are all defined. A quick check reveals that this is indeed the case,
since the numbers of columns in the blocks of A ( 3 and 2) match the numbers of rows
in the blocks of B. The matrices A and B are said to be partitioned conformably for
block multiplication.
Carrying out the calculations indicated gives us the product AB in partitioned form:
A , , B , , + A ,, B,, � I, B ,, + A ,, I, � B ,, + A" �
6
]
�
[-: � i [ [
-1
3 = 0
+
4
0
5
-5
Section
3. 1
Matrix Operations
149
(When some of the blocks are zero matrices or identity matrices, as is the case here,
these calculations can be done quite quickly.) The calculations for the other five
blocks of AB are similar. Check that the result is
6
0
2 : 1 2 2
5 i 2 1 12
5 -5 i 3 3 ' 9
---------7-r-0----o-T-i3
'
�
'
2 i 0 0 i 20
7
(Observe that the block in the upper-left corner is the result of our calculations above.)
Check that you obtain the same answer by multiplying A by B in the usual way.
M atrix Powers
When A and B are two n X n matrices, their product AB will also be an n X n matrix.
A special case occurs when A = B. It makes sense to define A 2 = AA and, in general,
to define A k as
k
A = AA · · · A
k factors
�
if k is a positive integer. Thus, A 1 = A, and it is convenient to define A 0 = Iw
Before making too many assumptions, we should ask ourselves to what extent
matrix powers behave like powers of real numbers. The following properties follow
immediately from the definitions we have just given and are the matrix analogues of
the corresponding properties for powers of real numbers.
If A is a square matrix and r and s are nonnegative integers, then
1. A rAs = A r+s
2. (A r ) s = A'"
In Section 3.3, we will extend the definition and properties to include negative integer
powers.
Exa m p l e 3 . 1 3
(a) If A =
A2 =
[� �]
[ � � ] [ � � ] [� ]
, then
[ ] [ � ] [! :J
2 2 1
2 3
, A = A 2A =
2 2 1
2
and, in general,
2n - l
2n - 1
]
for all n 2: 1
The above statement can be proved by mathematical induction, since it is an
infinite collection of statements, one for each natural number n. (Appendix B gives a
150
Chapter
3
Matrices
1.
brief review of mathematical induction.) The basis step is to prove that the formula
holds for n = In this case,
A1 =
[ 22 11 -- 11
21 - J
21 - 1
J
as required.
The induction hypothesis is to assume that
1.
1.
[� �J
=A
for some integer k 2: The induction step is to prove that the formula holds for =
k + Using the definition of matrix powers and the induction hypothesis, we compute
n
�J
[ 22(k+(k+ JJ)) -- 1J
1
[ 01 - o1 J ' [ o1 - o1 J [ o1 - o1 J [ - o1 - 01 J.
[ - 01 - o1 J [ o1 - 01 J [ _ � � J
[ - 01 0l J [ O1 -l0 J [ 0l O1 J
Thus, the formula holds for all n 2: by the principle of mathematical induction.
(b) If B =
we find
then B 2 =
=
Continuing,
B 3 = B 2B =
and
B 4 = B 3B =
=
5
Thus, B = B, and the sequence of powers of B repeats in a cycle of four:
[ o1 -1J0 ' [ -10 - oJ1 ' [ - o1 0l J ' [ 01 o1 J ' [ o1 -1J0 '
The Tra nspose of a M atrix
Thus far, all of the matrix operations we have defined are analogous to operations on
real numbers, although they may not always behave in the same way. The next opera­
tion has no such analogue.
Section
3. 1
Matrix Operations
151
The transpose of an m x n matrix A is the n x m matrix A T
obtained by interchanging the rows and columns of A. That is, the ith column of
A T is the ith row of A for all i.
Definition
Exa m p l e 3 . 1 4
Let
A=
[ � � � l [ : �l
B=
and C = [ 5
-
1 2]
Then their transposes are
The transpose is sometimes used to give an alternative definition of the dot prod­
uct of two vectors in terms of matrix multiplication. If
then
[ �: j
u2 . . . u n l .
vn
A useful alternative definition of the transpose is given componentwise:
(A T) ij = Aji for all i and j
In words, the entry in row i and column j of A T is the same as the entry in row j and
column i of A.
The transpose is also used to define a very important type of square matrix: a
symmetric matrix.
Definition
A square matrix A is symmetric if A T = A-that is, if A is equal to
its own transpose.
Exa m p l e 3 . 1 5
Let
and B =
[
1 2
1
- 3
]
3
Chapter
152
Matrices
Then A is symmetric, since A T = A; but B is not symmetric, since B T =
[�
-
�]
* B.
4
A symmetric matrix has the property that it is its own "mirror image" across its
main diagonal. Figure 3.1 illustrates this property for a 3 X 3 matrix. The correspond­
ing shapes represent equal entries; the diagonal entries (those on the dashed line) are
arbitrary.
A componentwise definition of a symmetric matrix is also useful. It is simply the
algebraic description of the "reflection" property.
Figure 3 . 1
A symmetric matrix
A square matrix A is symmetric if and only if A ij = Aji for all i and j.
..
Let
I
Exercises 3 . 1
[ _ � �l = [�
0 -3
] E = (4
D = [_
2
-2
2
B
A=
l
,
c�
2], F =
[ � :J
[ �]
-
In Exercises 1 - 1 6, compute the indicated matrices (if
possible).
2. 3D - 2A
1 . A + 2D
4. c - B T
3. B - C
5. AB
6. BD
7. D + BC
8. BB T
10. F (DF)
9. E (AF)
1 1 . FE
12. EF
14. DA - AD
13. B TCT - ( CB ) T
15. A 3
16. U2 - D) 2
17. Give an example of a nonzero 2 X 2 matrix A such
that A 2
18. Let A
= 0.
=
[�
�] .
Find 2 X 2 matrices B and C such
that AB = AC but B
*-
C.
19. A factory manufactures three products (doohickies,
gizmos, and widgets) and ships them to two ware­
houses for storage. The number of units of each prod­
uct shipped to each warehouse is given by the matrix
[
200 75
A = 1 50 100
100 125
]
(where a ij is the number of units of product i sent to
warehouse j and the products are taken in alphabetical
order) . The cost of shipping one unit of each product
by truck is $ 1 .50 per doohickey, $1 .00 per gizmo, and
$2.00 per widget. The corresponding unit costs to ship
by train are $ 1 .75, $1 .50, and $1 .00. Organize these
costs into a matrix B and then use matrix multiplica­
tion to show how the factory can compare the cost of
shipping its products to each of the two warehouses by
truck and by train.
20. Referring to Exercise 19, suppose that the unit cost
of distributing the products to stores is the same for
each product but varies by warehouse because of the
distances involved. It costs $0.75 to distribute one unit
from warehouse 1 and $1 .00 to distribute one unit
from warehouse 2. Organize these costs into a matrix
C and then use matrix multiplication to compute the
total cost of distributing each product.
Section
3. 1
32. A =
[
In Exercises 23-28, let
34· A =
[! �L! J·
'
]
[ L:jH
[ ]
= -2
X 1 - Xz
Xz + x3 = - 1
A=
H -�i
u :J
0
35.
0 -1
and
B=
3
-1
6
23. Use the matrix-column representation of the product
24.
25.
26.
27.
28.
to write each column of AB as a linear combination of
the columns of A.
Use the row-matrix representation of the product to
write each row of AB as a linear combination of the
rows of B.
Compute the outer product expansion of AB.
Use the matrix-column representation of the product
to write each column of BA as a linear combination of
the columns of B.
Use the row-matrix representation of the product to
write each row of BA as a linear combination of the
rows of A.
Compute the outer product expansion of BA.
In Exercises 29 and 30, assume that the product AB makes
sense.
29. Prove that if the columns of B are linearly dependent,
then so are the columns of AB.
30. Prove that if the rows of A are linearly dependent, then
so are the rows of AB.
In Exercises 31 -34, compute AB by block multiplication,
using the indicated partitioning.
31. A �
[� H i �J
1 - 1 :: 0 0
n�
o
[ l
2 3:
33. A �
o of 1
:
-1 1 : 0
0 0: 1
153
o o
�
� ] . [ +!t -� l
In Exercises 21 -22, write the given system of linear equa­
tions as a matrix equation of the form Ax = b.
X 1 - 2X2 + 3X3 = 0
2x 1 + x2 - 5x3 = 4
- X1
+ 2X3 = 1
21.
22.
Matrix Operations
:1
2 3
4 5 0
B
-2 : 3 2
B=
0 0 Oi4
O 1
.
Let A =
-1 1
(a) Compute A 2 , A 3 , . • . , A 7 •
(b) What is A 2 0 1 5 ? Why?
36. Let B =
37. Let A =
0 0.
[[� ] ]
�
]
[
[
-
v2
Find, with justification, B 2 0 1 5 .
v2
. Find a formula for A n ( n 2 1) and
verify your formula using mathematical induction.
38. Let A =
cos tJ
sin tJ
- sin e
.
cos tJ
39.
]
cos W - sin W
.
. W
cos W
sm
(b) Prove, by mathematical induction, that
cos n tJ - sin n tJ c
An =
1or n 2 1
cos n tJ
sin n tJ
In each o f the following, find the 4 X 4 matrix A = [ a;j ]
that satisfies the given condition:
(b) a u = j - i
(a) a ij = ( - l) i +j
(i + j - 1 )7T
(c) a ij = (i - lY
(d) aiJ = sin
(a) Show that A 2 =
[
]
(
)
40. In each of the following, find the 6 X 6 matrix A = [aij]
41.
{
{
that satisfies the given condition:
i + j if i ::; j
(a) a ij =
(b) a ij =
0 if i > j
1 if 6 ::; i + j ::; 8
(c) a11 =
0 otherwise
Prove Theorem 3 . l (a) .
{ 01
4
if I i
if I i
l
- j l ::::: 1
-j > 1
154
Chapter
3
Matrices
M atrix A l g e b ra
In some ways, the arithmetic of matrices generalizes that of vectors. We do not expect
any surprises with respect to addition and scalar multiplication, and indeed there are
none. This will allow us to extend to matrices several concepts that we are already
familiar with from our work with vectors. In particular, linear combinations, span­
ning sets, and linear independence carry over to matrices with no difficulty.
However, matrices have other operations, such as matrix multiplication, that vec­
tors do not possess. We should not expect matrix multiplication to behave like multi­
plication of real numbers unless we can prove that it does; in fact, it does not. In this
section, we summarize and prove some of the main properties of matrix operations
and begin to develop an algebra of matrices.
Prooerlies of Addilion and scalar M u lliolicalion
All of the algebraic properties of addition and scalar multiplication for vectors
(Theorem 1 . 1 ) carry over to matrices. For completeness, we summarize these proper­
ties in the next theorem.
Theorem 3 . 2
Algebraic Properties of Matrix Addition and Scalar Multiplication
Let A, B, and C be matrices of the same size and let c and d be scalars. Then
a.
b.
c.
d.
e.
f.
g.
h.
A+B=B+A
(A + B) + C = A + (B + C )
A+0=A
A + ( -A) = 0
c (A + B) = cA + cB
( c + d )A = cA + dA
c (dA) = (cd )A
IA = A
Commutativity
Associativity
Distributivity
Distributivity
The proofs of these properties are direct analogues of the corresponding proofs
of the vector properties and are left as exercises. Likewise, the comments following
Theorem 1 . 1 are equally valid here, and you should have no difficulty using these
properties to perform algebraic manipulations with matrices. (Review Example 1.5
and see Exercises 17 and 18 at the end o f this section.)
The associativity property allows us to unambiguously combine scalar multiplica­
tion and addition without parentheses. IfA, B, and C are matrices of the same size, then
( 2A + 3B ) - C = 2A + ( 3B - C)
and so we can simply write 2A + 3B - C. Generally, then, if A 1 , A 2 , , Ak are matri­
ces of the same size and c 1 , c2 , . . . , ck are scalars, we may form the linear combination
•
.
.
c 1 A 1 + c 2A 2 + · · · + ckA k
We will refer to c 1 , c2 , . . . , ck as the coefficients of the linear combination. We can now
ask and answer questions about linear combinations of matrices.
Section
Exa m p l e 3 . 1 6
Let A 1 =
[ _ � �l [ � �l
[ � �]
[� !]
A2 =
and A 3 =
3. 2
Matrix Algebra
155
[ � � ].
(a) Is B =
a linear combination ofA 1 , A 2 , and A 3 ?
(b) Is C =
a linear combination of A 1 , A 2 , and A 3 ?
Solution
(a) We want to find scalars c 1 , c2 , and c 3 such that c 1A 1 + c2A 2 + c 3A 3 = B. Thus,
The left-hand side of this equation can be rewritten as
Comparing entries and using the definition of matrix equality, we have four linear
equations:
C1
- cl
C2 + c 3
+ c3
+ c3
C2 + C 3
1
=
=4
=2
=
1
[ � � � :j ----+ [ � 0 � � j
Gauss-Jordan elimination easily gives
�
1,
-
-1 0 1 2
0
1
0 0 1 3
0 0 0 0
(check this!), so c 1 = c 2 = - 2, and c3 = 3. Thus, A 1 - 2A 2 + 3A 3 = B, which can
be easily checked.
(b) This time we want to solve
Proceeding as in part (a), we obtain the linear system
C1
- c1
c2 + C3
+ c3
+ c3
C2 + C 3
1
=
=2
=3
=4
156
Chapter
3
Matrices
Row reduction gives
We need go no further: The last row implies that there is no solution. Therefore, in
this case, C is not a linear combination of Ai, A 2 , and A 3 •
Observe that the columns of the augmented matrix contain the entries
of the matrices we are given. If we read the entries of each matrix from left to right
and top to bottom, we get the order in which the entries appear in the columns of
the augmented matrix. For example, we read Ai as " O, 1, - 1 , O," which corresponds
to the first column of the augmented matrix. It is as if we simply "straightened out"
the given matrices into column vectors. Thus, we would have ended up with exactly
the same system of linear equations as in part (a) if we had asked
Remark
{]
a linemombillation of
[-l}[ � l [ '. }
and
We will encounter such parallels repeatedly from now on. In Chapter 6, we will
explore them in more detail.
We can define the span of a set of matrices to be the set of all linear combinations
of the matrices.
Exa m p l e 3 . 1 1
Describe the span of the matrices Ai, A 2 , and A 3 in Example 3 . 1 6.
One way to do this is simply to write out a general linear combination of
Ai, A 2 , and A 3 • Thus,
Solulion
�]
[;
:J
(which is analogous to the parametric representation of a plane) . But suppose we
want to know when the matrix
tion above, we know that it is when
is in span (Ai, A 2 , A 3 ) . From the representa­
for some choice of scalars Ci, c2 , c3 . This gives rise to a system of linear equations
whose left-hand side is exactly the same as in Example 3 . 1 6 but whose right-hand side
Section
3.2
Matrix Algebra
151
is general. The augmented matrix of this system is
and row reduction produces
__.
[ -l � : fl [�
0
�
0 0
0
0 1
0 0
--
! x tJ
!
- x tJ + w
!x + h
w-z
l
(Check this carefully.) The only restriction comes from the last row, where clearly we
must have w - z = in order to have a solution. Thus, the span of A 1 , A 2 , and A 3 con-
[ ; :]
sists of all matrices
for which w = z. That is, span (A 1 , A 2 , A 3 ) =
[ ]
[� �]
{ [ ; :] } .
4
If we had known this before attempting Example 3.16, we would have seen
Nole
is a linear combination of A 1 , A 2 , and A 3 , since it has
1 2
the necessary form (take w = 1, x = 4, and y = 2), but C =
cannot be a linear
3 4
combination of A 1 , A 2 , and A 3 , since it does not have the proper form ( 1 i= 4).
immediately that B =
Linear independence also makes sense for matrices. We say that matrices
A 1 , A2 ,
, Ak of the same size are linearly independent if the only solution of the
equation
(1)
•
•
•
0.
is the trivial one: c 1 = c2 = · · · = ck = If there are nontrivial coefficients that satisfy
( 1 ) , then A 1 , A 2 , , Ak are called linearly dependent.
•
Exa m p l e 3 . 1 8
.
.
Determine whether the matrices A 1 , A 2 , and A 3 in Example 3 . 1 6 are linearly
independent.
We want to solve the equation c 1A 1 + c2A 2 + c3A 3 = 0. Writing out the
matrices, we have
Solulion
This time we get a homogeneous linear system whose left-hand side is the same as
in Examples 3 . 1 6 and 3.17. (Are you starting to spot a pattern yet?) The augmented
matrix row reduces to give
158
Chapter
3
Matrices
Thus, c1 = c2 = c 3 = 0 , and we conclude that the matrices A1, A 2 , and A 3 are linearly
independent.
4
Prooenies of M alrix M ulliolicalion
Whenever we encounter a new operation, such as matrix multiplication, we must
be careful not to assume too much about it. It would be nice if matrix multiplication
behaved like multiplication of real numbers. Although in many respects it does, there
are some significant differences.
Exa m p l e 3 . 1 9
Consider the matrices
[ � �] [ � �]
[ � �] [ � �] [ : �]
[ � �] [ � �]
[� �]
A=
_
and B =
_
Multiplying gives
AB =
_
_
_
_
and BA =
_
_
Thus, AB -=fa BA. So, in contrast to multiplication of real numbers, matrix multiplica­
tion is not co m m uta tive the order of the factors in a product matters!
-
It is easy to check that A 2 =
[� �]
(do so!). So, for matrices, the equation
A 2 = 0 does not imply that A = 0 (unlike the situation for real numbers, where the
equation x2 = 0 has only x = 0 as a solution) .
However gloomy things might appear after the last example, the situation is not
really bad at all-you just need to get used to working with matrices and to constantly
remind yourself that they are not numbers. The next theorem summarizes the main
properties of matrix multiplication.
Theorem 3 . 3
Properties of Matrix Multiplication
Let A, B, and C be matrices (whose sizes are such that the indicated operations can
be performed) and let k be a scalar. Then
a.
b.
c.
d.
e.
A (B C ) = (AB) C
A ( B + C ) = AB + AC
(A + B)C = AC + BC
k (AB ) = ( kA ) B = A ( kB )
Im A = A = Aln if A is X n
m
Associativity
Left distributivity
Right distributivity
Multiplicative identity
We prove (b) and half of (e) . We defer the proof of property (a) until
Section 3.6. The remaining properties are considered in the exercises.
Proof
Section
3. 2
Matrix Algebra
159
(b) To prove A(B + C ) = AB + AC, we let the rows of A be denoted by A; and the
columns of B and C by bj and cf Then the jth column of B + C is bj + cj (since addi­
tion is defined componentwise), and thus
[A ( B + C) J u = A; · ( bj + c)
= A; . bj + A; . CJ
= (AB ) ;1 + (AC) ;1
= (AB + A C) ;j
Since this is true for all i and j, we must have A(B + C ) = AB + AC.
(e) To prove Ain = A, we note that the identity matrix In can be column-partitioned as
where e; is a standard unit vector. Therefore,
Ain = [Ae 1 : Ae2 : : Aen l
: an ]
= [ a 1 : az :
=A
•
·
·
•
·
·
by Theorem 3 . 1 (b) .
We can use these properties to further explore how closely matrix multiplication
resembles multiplication of real numbers.
Exa m p l e 3 . 2 0
If A and B are square matrices of the same size, is (A + B) 2 = A 2 + 2AB + B 2 ?
Solution
Using properties of matrix multiplication, we compute
( A + B ) 2 = ( A + B ) (A + B )
by left distributivity
= (A + B ) A + (A + B ) B
by right distributivity
= A 2 + BA + AB + B 2
Therefore, (A + B) 2 = A 2 + 2AB + B 2 if and only if A 2 + BA + AB + B 2 = A 2 +
2AB + B 2 . Subtracting A 2 and B 2 from both sides gives BA + AB = 2AB. Subtracting
AB from both sides gives BA = AB. Thus, (A + B) 2 = A 2 + 2AB + B2 if and only if A
and B commute. (Can you give an example of such a pair of matrices? Can you find
two matrices that do not satisfy this property?)
Properties of the Transpose
Theorem 3 . 4
Properties of the Transpose
Let A and B be matrices (whose sizes are such that the indicated operations can be
performed) and let k be a scalar. Then
a. (A T ) T = A
b. (A + B) T = A T + B T
T
d. (AB) T = B TA T
c. (kAl = k (A )
r
r
e. (A l = (A T ) for all nonnegative integers r
160
Chapter
3
Matrices
Proof
Properties (a) -(c) are intuitively clear and straightforward to prove (see Exercise
30). Proving property ( e) is a good exercise in mathematical induction (see Exercise 3 1).
We will prove (d), since it is not what you might have expected. [Would you have sus­
pected that (AB) T = A TB T might be true?]
First, ifA is m X n and B is n X r, then B T is r X n and A T is n X m. Thus, the product
B TA T is defined and is r X m. Since AB is m X r, (AB) T is r X m, and so (AB) T and B TA T
have the same size. We must now prove that their corresponding entries are equal.
We denote the ith row of a matrix X by row; (X) and its jth column by col/X) .
Using these conventions, we see that
[ (AB ) T ] ;j =
=
=
=
(AB )ji
row/A ) · col; ( B )
col/A T) · row; ( B T)
row; ( B T) · col/A T) = [ B TA T ] ij
(Note that we have used the definition of matrix multiplication, the definition of the
transpose, and the fact that the dot product is commutative.) Since i and j are arbi­
trary, this result implies that (AB) T = B TA T.
Remark
Properties (b) and (d) of Theorem 3.4 can be generalized to sums and
products of finitely many matrices:
(A 1 + A 2 + · · · + A k f = A[ + Af + · · · + A [ and (A 1 A 2 · · A k f
= A [ · · AfA [
assuming that the sizes o f the matrices are such that all o f the operations can be per­
formed. You are asked to prove these facts by mathematical induction in Exercises 32
and 33.
•
-
Exa m p l e 3 . 2 1
Let
Then A T =
We have
[� !]
�]
[
�
[� ]
[ :]
H �l
[� � i [ - r n � [ ': l: J
H : J [� � i � n � ]
A =
3
4
'
so A + AT =
and B =
2
5
-1
3
, a symmetric matrix.
BT =
so
BBT =
and
B TB =
-1
3
-1
3
2
10
3
Thus, both BB T and B TB are symmetric, even though B is not even square! (Check
that AA T and A TA are also symmetric.)
Section
3.2
Matrix Algebra
161
The next theorem says that the results of Example 3.2 1 are true in general.
a. If A is a square matrix, then A + A T is a symmetric matrix.
b. For any matrix A, AA T and A TA are symmetric matrices.
Theorem 3 . 5
Proof
We prove (a) and leave proving (b) as Exercise 34. We simply check that
(A + A Tf = A T + (A Tf = A T + A = A + AT
(using properties of the transpose and the commutativity of matrix addition). Thus,
A + A T is equal to its own transpose and so, by definition, is symmetric.
I
Exercises 3 . 2
[� !] [ - � �] .
- -
In Exercises 1 -4, solve the equation for X, given that
and B =
A=
1. X
2A + 3B = 0
In Exercises 9- 12, find the general form of the span of the
indicated matrices, as in Example 3. 1 7.
9. span(A 1 , A 2 ) in Exercise 5
10. span(A 1 , A 2 , A 3 ) in Exercise 6
1 1 . span(A 1 , A 2 , A 3 ) in Exercise 7
12. span(A 1 , A 2 , A 3 , A 4 ) in Exercise 8
2. 2X = A B
3. 2 (A + 2B ) = 3X
4. 2 (A - B + X) = 3 (X - A )
In Exercises 5-8, write B as a linear combination of the
other matrices, ifpossible.
5. B =
6. B =
A3 =
7. B
=
A2 =
8. B �
A, �
[� :l [ _ � �l
[ � l [ l �l
[� � ]
[ � �l [ l
[ �l [�
[: - n [ �
[: n n
A1 =
2
-4
AI =
Q
AI =
-1 2
0
Q
A3 =
-2
1
0
0
0
0
A, �
A3 =
Az =
Az =
[� � ]
[� ]
]
�]
n
i
0 -1
,
0
1
1
0
0
1
0
0 -1
0 ,
0 -1
-1
0
In Exercises 1 3- 1 6, determine whether the given matrices
are linearly independent.
,
13.
14.
15.
I •.
[� !l [� � ]
[� �l [ - � �l [ � � ]
,
[ � � ] , [ � � ] , [ � � i [ � - !]
[� -� m: ! m� : n
-[ � � i
-1 0
1
-1
0
0 -4
1
0
2
4
5
Chapter
162
3
Matrices
17. Prove Theorem 3.2(a) -(d).
18. Prove Theorem 3.2(e) -(h).
19. Prove Theorem 3.3(c).
20. Prove Theorem 3.3(d) .
21. Prove the half o f Theorem 3.3(e) that was not proved
in the text.
22. Prove that, for square matrices A2 and B,2 AB = BA if
and only if (A - B) (A + B) = A - B •
[:
[ l - ] 25. A
In Exercises 23-25, if B =
c, and d such that AB = BA.
23. A =
[�
1
1
] 24.
A=
�] ,find conditions on a, b,
1
1
-1
=
26. Find conditions on a, b, c, and d such that B =
commutes with both
[
2
3 4
1
[ ] [
1
0
o
o
and
o
o
27. Find conditions on a, b, c, and d such that B =
commutes with every 2 X 2 matrix.
]
[ ]
a b
c d
28. Prove that if AB and BA are both defined, then AB and
BA are both square matrices.
A square matrix is called upper triangular if all of the en­
tries below the main diagonal are zero. Thus, the form of an
upper triangular matrix is
*
0
0
0
*
*
*
*
*
*
*
*
0
0
0
*
where the entries marked are arbitrary. A more formal
definition of such a matrix A = [ a ij ] is that a ij = 0 if i > j.
Prove that the product of two upper triangular n X n
*
29.
matrices is upper triangular.
30. Prove Theorem 3.4(a) - (c) .
31. Prove Theorem 3.4(e).
32. Using induction, prove that for all n 2: 1,
(A 1 + A 2 + · · · + A n ) r = A f + Af + · · · + A�.
33. Using induction,
prove that for all n 2: 1 ,
T
(A 1 A 1 · · · A n ) = A�· · · AfA f.
34. Prove Theorem 3.S(b).
35. (a) Prove that ifA and B are symmetric n X n matrices,
then so is A + B.
36.
37.
(b) Prove that if A is a symmetric n X n matrix, then
so is kA for any scalar k.
(a) Give an example to show that if A and B are
symmetric n X n matrices, then AB need not be
symmetric.
(b) Prove that if A and B are symmetric n X n matrices,
then AB is symmetric if and only if AB = BA.
A square matrix is called skew-symmetric if A r = -A.
Which of the following matrices are skew-symmetric?
(a)
(<)
[� -�]
[ _� �]
-[ : -� - �] [ - : � �]
(b)
(d)
38. Give a componentwise definition of a skew-symmetric
matrix.
39. Prove that the main diagonal of a skew-symmetric ma­
trix must consist entirely of zeros.
40. Prove that if A and B are skew-symmetric n X n
matrices, then so is A + B.
41. If A and B are skew-symmetric 2 X 2 matrices, under
what conditions is AB skew-symmetric?
42. Prove that if A is an n X n matrix, then A - A r is
skew-symmetric.
43. (a) Prove that any square matrix A can be written
as the sum of a symmetric matrix and a skew­
symmetric matrix. [Hint: Consider Theorem 3.5
and Exercise 42.]
2
5
(b) Illu,tmte pact (o) fo, the nrnt'ix A �
[�
8
The trace of an n X n matrix A = [ a ij ] is the sum of the en­
tries on its main diagonal and is denoted by tr(A). That is,
tr (A ) = a 11 + a 22 + · · · + a ""
If A and B are n X n matrices, prove the following
44.
45.
46.
47.
properties of the trace:
(a) tr (A + B) = tr (A) + tr (B)
(b) tr (kA) = ktr (A), where k is a scalar
Prove that if A and B are n X n matrices, then
tr (AB) = tr (BA).
If A is any matrix, to what is tr (AA T ) equal?
Show that there are no 2 X 2 matrices A and B such
that AB - BA = 12 •
Section
3. 3
The Inverse of a Matrix
163
T h e Inverse of a M atrix
In this section, we return to the matrix description Ax = b of a system of linear equa­
tions and look for ways to use matrix algebra to solve the system. By way of analogy,
consider the equation ax = b, where a, b, and x represent real numbers and we want
to solve for x. We can quickly figure out that we want x = b/a as the solution, but we
must remind ourselves that this is true only if a * 0. Proceeding more slowly, assum­
ing that a * 0, we will reach the solution by the following sequence of steps:
( )
1
1
1
b
b
b
= -( =:> -( a) x = - =:> l · x = - =:> x = ax = b =:> -(
a ax) a b)
a
a
a
a
(This example shows how much we do in our head and how many properties of arith­
metic and algebra we take for granted! )
To imitate this procedure for the matrix equation Ax = b, what do we need? We
need to find a matrix A ' (analogous to 1 /a) such that A ' A = I, an identity matrix
(analogous to 1 ) . If such a matrix exists (analogous to the requirement that a * O),
then we can do the following sequence of calculations:
Ax = b =:> A ' (Ax) = A 'b =:> (A 'A)x = A 'b =:> Ix = A 'b =:> x = A 'b
�
�
(Why would each of these steps be justified?)
Our goal in this section is to determine precisely when we can find such a matrix
A ' . In fact, we are going to insist on a bit more: We want not only A ' A = I but also
AA' = I. This requirement forces A and A ' to be square matrices. (Why?)
If A is an n
the property that
Definition
where I = In is the
invertible.
Exa m p l e 3 . 2 2
If A =
AA , =
Exa m p l e 3 . 2 3
[ � 53 ]
[ 2 35 ] [ 3
n
X
n
-1
n
matrix, an inverse of A is an n
-1
[� �]
n
matrix A ' with
o
1
is an inverse of A, since
[3
and A 'A = - 1
Show that the following matrices are not invertible:
(a) 0 =
x
AA ' = I and A ' A = I
identity matrix. If such an A ' exists, then A is called
, [ 3 -52 ]
-52 ] [ � ]
, then A =
1
x
(b) B =
- 25 ] [ 2 35 ]
1
[� �]
Solution
(a) It is easy to see that the zero matrix 0 does not have an inverse. If it did, then there
would be a matrix O ' such that 00' = I = O' O. But the product of the zero matrix
with any other matrix is the zero matrix, and so 00' could never equal the identity
164
Chapter
3
Matrices
matrix I. (Notice that this proof makes no reference to the size of the matrices and so
is true for n X n matrices in general.)
[; :]
[ � ! ][; :J [ � � ]
(b) Suppose B has an inverse B'
. The equation BB'
=
=
I gives
=
from which we get the equations
+ 2y
w
x
2w
+ 4y
2x
=
+ 2z
=
=
+ 4z
=
1
0
O
1
Subtracting twice the first equation from the third yields 0 = - 2, which is clearly
absurd. Thus, there is no solution. (Row reduction gives the same result but is not
really needed here.) We deduce that no such matrix B' exists; that is, B is not invert­
ible. (In fact, it does not even have an inverse that works on one side, let alone two!)
4
R e m a rks
Even though we have seen that matrix multiplication is not, in general, commutative, A' (if it exists) must satisfy A' A = AA' .
•
The examples above raise two questions, which we will answer in this section:
•
( 1) How can we know when a matrix has an inverse?
(2) If a matrix does have an inverse, how can we find it?
We have not ruled out the possibility that a matrix A might have more than
•
one inverse. The next theorem assures us that this cannot happen.
Theorem 3 . 6
If A is an invertible matrix, then its inverse is unique.
Proof
In mathematics, a standard way to show that there is just one of something is
to show that there cannot be more than one. So, suppose that A has two inverses-say,
A ' and A " . Then
Thus,
AA ' = I = A 'A and AA " = I = A " A
A ' = A 'I = A ' (AA " ) = (A 'A )A " = IA " = A "
Hence, A ' = A", and the inverse is unique.
Thanks to this theorem, we can now refer to the inverse of an invertible matrix.
From now on, when A is invertible, we will denote its (unique) inverse by A - 1 (pro­
nounced "A inverse").
Warning
Do not be tempted to write A - l
=
1
A
! There is no such operation as
"division by a matrix." Even if there were, how on earth could we divide the scalar 1 by
Section
3. 3
The Inverse of a Matrix
165
the matrix A? If you ever feel tempted to "divide" by a matrix, what you really want to do
is multiply by its inverse.
We can now complete the analogy that we set up at the beginning of this section.
Theorem 3 . 1
If A is an invertible n X n matrix, then the system of linear equations given by
Ax = b has the unique solution x = A - l b for any b in !R n .
Theorem 3.7 essentially formalizes the observation we made at the beginning
of this section. We will go through it again, a little more carefully this time. We are
asked to prove two things: that Ax = b has a solution and that it has only one solution.
(In mathematics, such a proof is called an "existence and uniqueness" proof.)
To show that a solution exists, we need only verify that x = A - l b works. We check
that
Proof
So A - l b satisfies the equation Ax = b, and hence there is at least this solution.
To show that this solution is unique, suppose y is another solution. Then Ay = b,
and multiplying both sides of the equation by A - l on the left, we obtain the chain of
implications
A - 1 (Ay)
=
A - 1 b => (A- 1A) y = A - 1 b => Iy
=
A - 1 b => y = A - 1 b
Thus, y is the same solution as before, and therefore the solution is unique.
So, returning to the questions we raised in the Remarks before Theorem 3.6, how
can we tell if a matrix is invertible and how can we find its inverse when it is invert­
ible? We will give a general procedure shortly, but the situation for 2 X 2 matrices is
sufficiently simple to warrant being singled out.
Theorem 3 . 8
If A
=
[ : �],
If ad - be
=
then A is invertible if ad - be * 0, in which case
A-1 -
[ ]
1
d
ad - be - e
0 , then A is not invertible.
[ ]
-
b
a
The expression ad - be is called the determinant of A, denoted det A. The formula
a b
1
for the inverse of
(when it exists) is thus
times the matrix obtained by
e d
dct A
interchanging the entries on the main diagonal and changing the signs on the other
two entries. In addition to giving this formula, Theorem 3.8 says that a 2 X 2 matrix
A is invertible if and only if det A * 0. We will see in Chapter 4 that the determinant
can be defined for all square matrices and that this result remains true, although there
is no simple formula for the inverse of larger square matrices.
[ ae
Proof
[
Suppose that det A
][ ]
b d -b
d -e a
=
--
]
ad - be * 0. Then
0
ad - be
a d - be - ab + ba
=
0
ad - be
ed - de -eb + da
=
[
]
=
det A
[ OJ
l
0 1
166
Chapter
3
Matrices
Similarly,
[ ] [ ] [ OJ
[ : �] ( � [ � - � ] ) [ � � ]
( � [ � - � ] ) [ : �] [ � � ]
[ ]
d -b a b
-e a e d
det A
=
l
0 1
Since det A * 0, we can multiply both sides of each equation by 1 / det A to obtain
=
de A -
and
=
de A -
[Note that we have used property (d) of Theorem 3.3.] Thus, the matrix
1
d -b
det A - e
a
satisfies the definition of an inverse, so A is invertible. Since the inverse of A is unique,
by Theorem 3.6, we must have
A-1 -
[ ]
d -b
det A - e
a
1
--
Conversely, assume that ad - be = 0. We will consider separately the cases where
a * 0 and where a = 0. If a * 0, then d = be/a, so the matrix can be written as
where k = e/a. In other words, the second row of A is a multiple of the first. Referring
to Example 3.23(b), we see that if A has an inverse
[; ;] ,
then
and the corresponding system of linear equations
+ by
aw
ax
kaw
�
+ kby
kax
has no solution. (Why?)
If a = 0, then ad - be
Thus, A is of the form
=
1
+ bz = 0
=
=
+ kbz
0 implies that be
=
=
O
1
0, and therefore either b or e is 0.
[ � �] [ � �]
[ � �] [; ;] [ � � ] [ � � ] .
or
In the first case,
�
*
have an inverse. (Verify this.)
Consequently, if ad - be = 0, then A is not invertible.
Similarly,
[ � �]
cannot
Section
Exa m p l e 3 . 2 4
Find the inverses o f A =
Solution
Exa m p l e 3 . 2 5
The Inverse of a Matrix
[ l 2 ] [ 12 --515 ]
1(4)1- 2 --2 2 - : �
_ _-_2 [ 1 ] [ - --]
12 ( - 5) - ( - 15)
3 4
We have <let A =
A-1
�
3. 3
and B =
(3) =
4
, if they exist.
4
* 0, so A is invertible, with
=
-3
(Check this.)
On the other hand, <let B =
2
2
(4) = 0, so B is not invertible.
2yy - 2
Use the inverse o f the coefficient matrix to solve the linear system
x+ =
3x + 4 =
Solution
-+-
3
The coefficient matrix is the matrix A =
[ _ � ];
161
[� ! l
whose inverse we com­
1
puted in Example 3.24. By Theorem 3.7, Ax = b has the unique solution x = A - b.
Here we have b =
thus, the solution to the given system is
x=
[ - � -tJ [ _ � ] [ -�]
22
1
1
Solving a linear system Ax = b via x = A - b would appear to be a good
method. Unfortunately, except for X coefficient matrices and matrices with cer­
tain special forms, it is almost always faster to use Gaussian or Gauss-Jordan elimi­
nation to find the solution directly. (See Exercise 3 .) Furthermore, the technique of
Example
works only when the coefficient matrix is square and invertible, while
elimination methods can always be applied.
Remark
3.25
Properties of Invertible Malrices
The following theorem records some of the most important properties of invertible
matrices.
Theorem 3 . 9
a. If A is an invertible matrix, then A - I is invertible and
b. If A is an invertible matrix and c is a nonzero scalar, then cA is an invertible
matrix and
_l_
(cA) - 1 = c A - 1
c. If A and B are invertible matrices of the same size, then AB is invertible and
168
Chapter
3
Matrices
d. If A is an invertible matrix, then A T is invertible and
e. If A is an invertible matrix, then A n is invertible for all nonnegative inte­
gers n and
(A " ) - 1 = (A - 1 ) "
Proof We will prove properties (a), (c), and (e), leaving properties (b) and (d) to be
proven in Exercises 1 4 and 1 5 .
(a) To show that A - i i s invertible, we must argue that there i s a matrix X such that
A - 1x = I = XA - 1
But A certainly satisfies these equations in place of X, so A - l is invertible and A is an
inverse of A - 1 . Since inverses are unique, this means that (A - 1 ) - 1 = A.
( c) Here we must show that there is a matrix X such that
(AB ) X = I = X (AB )
The claim is that substituting B - 1A - l for X works. We check that
�
where we have used associativity to shift the parentheses. Similarly, (B - 1A - l ) (AB ) = I
(check!), so AB is invertible and its inverse is B - 1A - 1 .
(e) The basic idea here is easy enough. For example, when n = 2, we have
A 2 (A - 1 ) 2 = AAA - IA - I = A L4. - I = AA - I = I
Similarly, (A - I ) 2A 2 = I. Thus, (A - I ) 2 is the inverse of A 2 . It is not difficult to see that
a similar argument works for any higher integer value of n. However, mathematical
induction is the way to carry out the proof.
The basis step is when n = 0, in which case we are being asked to prove that A 0 is
invertible and that
�
This is the same as showing that I is invertible and that r 1 = I, which is clearly true.
(Why? See Exercise 1 6.)
Now we assume that the result is true when n = k, where k is a specific nonnega­
tive integer. That is, the induction hypothesis is to assume that A k is invertible and that
The induction step requires that we prove that Ak+I is invertible and that
k
k1
k
1 1
1k1
(A + ) - = (A - ) + . Now we know from (c) that A + = A A is invertible, since A
and (by hypothesis) A k are both invertible. Moreover,
(A - 1 ) k+ 1 = (A - 1 ) kA - 1
by the induction hypothesis
= (A k ) - 1A - 1
= (AA k ) - 1
by property (c)
= (A k+ 1 ) - 1
Section
3. 3
The Inverse of a Matrix
Therefore, A n is invertible for all nonnegative integers n , and (A n ) - 1
principle of mathematical induction.
=
169
(A - l ) n by the
R e m a rks
While all of the properties of Theorem 3.9 are useful, ( c) is the one you should
highlight. It is perhaps the most important algebraic property of matrix inverses. It
is also the one that is easiest to get wrong. In Exercise 17, you are asked to give a
counterexample to show that, contrary to what we might like, (AB) - 1 * A - l B - 1 in
general. The correct property, (AB) - 1 = B - 1A - 1 , is sometimes called the socks-and­
shoes rule, because, although we put our socks on before our shoes, we take them off
in the reverse order.
•
Property (c) generalizes to products of finitely many invertible matrices: If A 1 ,
A 2 , . . . , A n are invertible matrices of the same size, then A 1 A 2 • • • A n is invertible and
•
(A 1A 2 · · · A n ) - 1
=
A ;;- 1 · · · A2 1A� 1
(See Exercise 1 8.) Thus, we can state:
The inverse of a product of invertible matrices is the product of their inverses in
the reverse order.
1
1 1
a
b
a
+
b
square matrices, (A + B) - 1 = A - 1 + B - 1 (and, indeed, this is not true in general; see
Exercise 1 9 ) . In fact, except for special matrices, there is no formula for (A + B) - 1 .
•
Since, for real numbers, -- * - + -, we should not expect that, for
Property (e) allows us to define negative integer powers of an invertible
matrix:
•
Definition
defined by
If A is an invertible matrix and n is a positive integer, then A - n is
With this definition, it can be shown that the rules for exponentiation, A rA s
and (AT = A'5, hold for all integers r and s, provided A is invertible.
=
A'· + s
One use of the algebraic properties of matrices is to help solve equations involving
matrices. The next example illustrates the process. Note that we must pay particular
attention to the order of the matrices in the product.
Exa m p l e 3 . 2 6
Solve the following matrix equation for X (assuming that the matrices involved are
such that all of the indicated operations are defined):
A - 1 (BX) - 1
=
(A - 1 B 3 ) 2
110
Chapter
3
Matrices
Solulion
There are many ways to proceed here. One solution is
A - 1 (BX)- 1
�
=
(A - 1 B 3 ) 2 ==? ((BX)A) - 1 = (A - 1 B 3 ) 2
=} [ (( BX)A ) - 1 ] - 1 = [ (A - 1 B 3 ) 2 ] - 1
==? (BX)A = [ (A - 1 B 3 )(A- 1 B 3 ) J - 1
==? ( BX)A = B - 3 (A- 1 ) - 1 B - 3 (A- 1 ) - 1
==? BXA = B - 3AB - 3A
==? B - 1 BXAA - 1 = B - 1 B - 3AB - 3AA - 1
==? IXI = B - 4AB - 3I
==? X = B - 4AB - 3
(Can you justify each step?) Note the careful use of Theorem 3.9(c) and the expansion
of (A - 1 B 3 ) 2 • We have also made liberal use of the associativity of matrix multiplica­
tion to simplify the placement (or elimination) of parentheses.
Elemen1arv Malrices
We are going to use matrix multiplication to take a different perspective on the row
reduction of matrices. In the process, you will discover many new and important
insights into the nature of invertible matrices.
If
we find that
In other words, multiplying A by E (on the left) has the same effect as interchanging
rows 2 and 3 of A. What is significant about E? It is simply the matrix we obtain by
applying the same elementary row operation, R 2 � R 3 , to the identity matrix J3 . It
turns out that this always works.
An elementary matrix is any matrix that can be obtained by per­
forming an elementary row operation on an identity matrix.
DefiniliOD
Since there are three types of elementary row operations, there are three cor­
responding types of elementary matrices. Here are some more elementary matrices.
Exa m p l e 3 . 2 1
Let
Section
3. 3
The Inverse of a Matrix
111
Each of these matrices has been obtained from the identity matrix I4 by applying a
single elementary row operation. The matrix E 1 corresponds to 3 R2 , E2 to R 1 � R 3 ,
and E3 to R4 - 2R 2 . Observe that when we left-multiply a 4 X n matrix by one of these
elementary matrices, the corresponding elementary row operation is performed on
the matrix. For example, if
A=
then
and
[
"[ "
a12
a 2 1 a 22 a 2 3
a 31 a 3 2 a 33
a41 a42 a43
""
]
"
"
[
au a12
3a 2 1 3a 22 3a 2 3 ' E A =
E1A =
2
a 31 a 3 2 a 33
a41 a42 a43
a12
au
a 22
a ,,
E 3A =
a32
a 31
a
a41 - 2a 2 1 42 - 2a 22
] ""
[ :
a 3 2 a ,,
a z1 a 22 a z3
a 11 a 1 2 a 13
a 41 a42 a 43
a 13
a,,
a 33
a43 - 2a 2 3
'
l
Example 3.27 and Exercises 24-30 should convince you that any elementary
row operation on any matrix can be accomplished by left-multiplying by a suitable
elementary matrix. We record this fact as a theorem, the proof of which is omitted.
Theor em 3 . 1 0
Let E be the elementary matrix obtaine d by performing an elementary row opera­
A,
tion on In - If the same elementary row operation is performed on an n X r matrix
the result is the same as the matrix EA.
Remark
From a computational point of view, it is not a good idea to use el­
ementary matrices to perform elementary row operations-just do them directly.
However, elementary matrices can provide some valuable insights into invertible
matrices and the solution of systems of linear equations.
We have already observed that every elementary row operation can be "undone:'
or "reversed:' This same observation applied to elementary matrices shows us that
they are invertible.
Exa m p l e 3 . 2 8
Let
0
0
�
0
4
0
�]
1
, and E3 =
[ � � �]
-2 0 1
Then E 1 corresponds to R 2 � R3 , which is undone by doing R2 � R3 again. Thus,
E 1 - l = E 1 . (Check by showing that E l = E 1 E 1 = I.) The matrix E2 comes from 4R2 ,
112
Chapter
3
Matrices
which is undone by performing iR 2 • Thus,
which can be easily checked. Finally, E3 corresponds to the elementary row opera­
tion R 3 - 2R 1 , which can be undone by the elementary row operation R 3 + 2R 1 . So,
in this case,
(Again, it is easy to check this by confirming that the product of this matrix and E3 ,
in both orders, is I.)
Notice that not only is each elementary matrix invertible, but its inverse is another
elementary matrix of the same type. We record this finding as the next theorem.
Theorem 3 . 1 1
Each elementary matrix is invertible, and its inverse is an elementary matrix of the
same type.
The Fundam en1a1 Theorem ol lnvenible Malrices
We are now in a position to prove one of the main results in this book-a set of
equivalent characterizations of what it means for a matrix to be invertible. In a sense,
much of linear algebra is connected to this theorem, either in the development of
these characterizations or in their application. As you might expect, given this intro­
duction, we will use this theorem a great deal. Make it your friend!
We refer to Theorem 3 . 1 2 as the first version of the Fundamental Theorem, since
we will add to it in subsequent chapters. You are reminded that, when we say that a set
of statements about a matrix A are equivalent, we mean that, for a given A, the state­
ments are either all true or all false.
Theorem 3 . 1 2
The Fundamental Theorem of Invertible Matrices: Version 1
Let A be an n X
a.
b.
c.
d.
e.
n
matrix. The following statements are equivalent:
A is invertible.
Ax = b has a unique solution for every b in !R n .
Ax = 0 has only the trivial solution.
The reduced row echelon form of A is Iw
A is a product of elementary matrices.
Section
3. 3
The Inverse of a Matrix
113
We will establish the theorem by proving the circular chain of implications
Proof
( a ) =} ( b ) =} ( c ) =} ( d ) =} ( e ) =} ( a )
(a) =} (b) We have already shown that if A is invertible, then Ax = b has the
unique solution x = A - 1 b for any b in !R n (Theorem 3.7).
(b) =} (c) Assume that Ax = b has a unique solution for any b in !R n . This implies,
in particular, that A x = has a unique solution. But a homogeneous system A x =
always has x = as one solution. So in this case, x = must be the solution.
(c) =} (d) Suppose that Ax = has only the trivial solution. The corresponding
system of equations is
0
0
0
0
0
a 11 x 1 + a 1 2x2 + · · · + a 1nxn
a 11 X 1 + a 2 1X2 + . . . + a 1nXn
=
=
0
0
and we are assuming that its solution is
=
=
O
O
In other words, Gauss-Jordan elimination applied to the augmented matrix of the
system gives
[A l o J
=
[ ""
�
a12
a , a21
a ln
a2n
a n ] a n2
a nn
f l [!
0
�
0
0
0
1
!]
[In l o J
Thus, the reduced row echelon form of A is In ( d) =} (e) If we assume that the reduced row echelon form of A is In , then A can be
reduced to In using a finite sequence of elementary row operations. By Theorem 3. 10,
each one of these elementary row operations can be achieved by left-multiplying by an
appropriate elementary matrix. If the appropriate sequence of elementary matrices is
E 1 , E2 , , Ek (in that order), then we have
•
•
.
Ek · · · E2 E 1 A
=
In
According to Theorem 3. 1 1 , these elementary matrices are all invertible. Therefore,
so is their product, and we have
Again, each E; - 1 is another elementary matrix, by Theorem 3. 1 1 , so we have written
A as a product of elementary matrices, as required.
(e) =} (a) If A is a product of elementary matrices, then A is invertible, since
elementary matrices are invertible and products of invertible matrices are invertible.
114
Chapter
3
Matrices
Exa m p l e 3 . 2 9
If possible, express A =
Solution
[� �J
[2
as a product of elementary matrices.
[ 2l J [ l
[ l J � [ l OJ
We row reduce A as follows:
A=
1
J
3 �
3
�
0
3 �·
3
0
0
-3
0
3
-3
J
= I2
1
Thus, the reduced row echelon form of A is the identity matrix, so the Fundamental
Theorem assures us that A is invertible and can be written as a product of elementary
matrices. We have E4E3 E2 E 1A = I, where
are the elementary matrices corresponding to the four elementary row operations
used to reduce A to I. As in the proof of the theorem, we have
[ O l J [ 2l O J [ l - l J [ l O J
1 0
as required.
1
0
1
0
-3
Remark
Because the sequence of elementary row operations that transforms A
into I is not unique, neither is the representation of A as a product of elementary
matrices. (Find a different way to express A as a product of elementary matrices.)
"Never bring a cannon o n stage in
Act I unless you intend to fire it by
the last act:' - Anton Chekhov
Theorem 3 . 1 3
The Fundamental Theorem is surprisingly powerful. To illustrate its power, we
consider two of its consequences. The first is that, although the definition of an in vertible matrix states that a matrix A is invertible if there is a matrix B such that both
AB = I and BA = I are satisfied, we need only check one of these equations. Thus, we
can cut our work in half!
Let A be a square matrix. If B is a square matrix such that either AB = I or BA = I,
then A is invertible and B = A - 1 •
0.
0.
0
Proof Suppose BA = I. Consider the equation Ax =
Left-multiplying by B, we have
BAx = BO. This implies that x = Ix = Thus, the system represented by Ax = has the
unique solution x = From the equivalence of (c) and (a) in the Fundamental Theo­
rem, we know that A is invertible. (That is, A - 1 exists and satisfies AA i = I = A iA.)
If we now right-multiply both sides of BA = I by A - i , we obtain
BAA - i = IA - i =} BI = A - i =} B = A i
0.
-
-
-
(The proof in the case of AB = I is left as Exercise 4 1 .)
The next consequence of the Fundamental Theorem is the basis for an efficient
method of computing the inverse of a matrix.
Section
Theorem 3 . 1 4
3. 3
The Inverse of a Matrix
115
Let A be a square matrix. If a sequence of elementary row operations reduces A
to I, then the same sequence of elementary row operations transforms I into A - i .
Proof If A is row equivalent to I, then we can achieve the reduction by left­
multiplying by a sequence E 1 , E2 ,
, Ek of elementary matrices. Therefore, we have
Ek · · · E2 E 1A = I. Setting B = Ek · · · E2 E 1 gives BA = I. By Theorem 3 . 1 3, A is invert­
ible and A - l = B. Now applying the same sequence of elementary row operations to
I is equivalent to left-multiplying I by Ek · · · E2 E 1 = B. The result is
Ek · · · E2E 1 I = BI = B = A - 1
•
•
•
Thus, I is transformed into A - i by the same sequence of elementary row operations.
The Gauss-Jordan Melhod for Compuling lhe Inverse
We can perform row operations on A and I simultaneously by constructing a "super­
augmented matrix" [A I I ] . Theorem 3 . 1 4 shows that if A is row equivalent to I [which,
by the Fundamental Theorem (d) � (a), means that A is invertible] , then elementary
row operations will yield
If A cannot be reduced to I, then the Fundamental Theorem guarantees us that A is
not invertible.
The procedure just described is simply Gauss-Jordan elimination performed on an
n X 2n, instead of an n X (n + 1), augmented matrix. Another way to view this pro­
cedure is to look at the problem of finding A -I as solving the matrix equation AX = In
for an n X n matrix X. (This is sufficient, by the Fundamental Theorem, since a right
inverse of A must be a two-sided inverse.) If we denote the columns of X by x1 , . . . , Xn'
then this matrix equation is equivalent to solving for the columns of X, one at a time.
Since the columns of In are the standard unit vectors e 1 , . . . , en , we thus have n systems
of linear equations, all with coefficient matrix A:
Since the same sequence of row operations is needed to bring A to reduced row
echelon form in each case, the augmented matrices for these systems, [A I e 1 ] , . . . ,
[A I e n ] , can be combined as
We now apply row operations to try to reduce A to In , which, if successful, will simul­
taneously solve for the columns of A - i , transforming In into A - i .
We illustrate this use of Gauss-Jordan elimination with three examples.
Exa m p l e 3 . 3 0
Find the inverse of
A=
if it exists.
[� �
1 3
�
-3
-
i
116
Chapter
3
Matrices
Solulion
Gauss-Jordan elimination produces
[A I I J
R2 - 2R,
R3 - R1
�
H:J R,
R3-R2
�
R, + R3
R2 + 3R3
�
R, - 2R2
�
Therefore,
[:
[:
[:
[:
[:
[:
A-1 =
[
2
2
3
2
-2
1
2
1
2
1
0
�:
0
-1
4 0
-3 0 0
0
6 -2 1
-2 - 1 0
-1
-1
-3
-2 - 1
-1
-3
0
I
-2
0
0
1
-2
2 0 -1
1 0 -5
-2
0
0 0 9
1 0 -5
0 1 -2
9 -t - 5
3
-5
-2
t 1
I
2
I
-2
I
2
1
2
!
3
-2
1
]
2
!
1]
�]
�]
�:
- :i
(You should always check that AA - I = I by direct multiplication. By Theorem 3 . 1 3,
we do not need to check that A - iA = I too.)
Remark
Notice that we have used the variant of Gauss-Jordan elimination that
first introduces all of the zeros below the leading ls, from left to right and top to
bottom, and then creates zeros above the leading l s, from right to left and bottom to
top. This approach saves on calculations, as we noted in Chapter 2, but you may find
it easier, when working by hand, to create all of the zeros in each column as you go.
The answer, of course, will be the same.
Exa m p l e 3 . 3 1
Find the inverse of
A=
if it exists.
[ ! - :i
-2
-1
2
-2
3. 3
Section
The Inverse of a Matrix
111
Solution
We proceed as in Example 3.30, adjoining the identity matrix to A and
then trying to manipulate [A I I] into [I I A - 1 ] .
[ -:
[:
,-,., [ I
[A I IJ
1
-1
2
-2
1
1
3
R2 + 2R1
R3 + R1
�
0
-4
-2 2 1
-6 1 0
2 -1
0
-3 2
0 0
0 -5
�
�]
-4 1 0
6 0 1
-2 0 0
�]
0
1
-3
�]
At this point, we see that it is not possible to reduce A to I, since there is a row of zeros
on the left-hand side of the augmented matrix. Consequently, A is not invertible.
4
As the next example illustrates, everything works the same way over "11.P. , where
p is prime.
Exa m p l e 3 . 3 2
Find the inverse of
A
=
if it exists, over "11.. 3 .
Solution 1
in "11.. 3 .
We use the Gauss-Jordan method, remembering that all calculations are
[A I I]
=
�
2R1
�
R1 + R1
�
R1 + 2R2
Thus , A - 1
[� �]
=
[
0
2
�] ,
[� 1 �]
[� 1 �]
[� 1 �]
[� 1 �]
2
0
1
0
1
1
0
1
1
0
2
0
2
2
0
2
and it is easy to check that, over "11.. 3 , AA - l
=
I.
Since A is a 2 X 2 matrix, we can also compute A - l using the formula
given in Theorem 3.8. The determinant of A is
Solution 2
det A
=
2 (0) - 2 (2)
=
-1
=
2
in "11.. 3 (since 2 + 1 = O ) . Thus, A - l exists and is given by the formula in Theorem 3.8.
We must be careful here, though, since the formula introduces the "fraction" l /det A
Chapter
118
3
Matrices
and there are no fractions in Z 3 . We must use multiplicative inverses rather than
division.
Instead of l /det A = 1 /2, we use T 1 ; that is, we find the number x that satisfies
the equation 2x = 1 in Z 3 • It is easy to see that x = 2 is the solution we want: In Z 3 ,
T 1 = 2, since 2(2) = 1 . The formula for A - 1 now becomes
A-1 = T1
[_�
-� ] =
2
which agrees with our previous solution.
..
I
[� �]
[! !]
[[i n ]
[IO. � -�]
[ ]
1.
2.
3.
4.
7.
=
Exercises 3 . 3
In Exercises 1 - 1 0, find the inverse of the given matrix (if it
exists) using Theorem 3.8.
5.
[� �] [� �]
- 1 .5
0.5
[ �_� - �� ]
[ ]
[[ ] ]
l / V2 l / V2
- 1 ; V2 1 / \/2
3.55 0.25
8.
8.52 0.60
6.
- 4.2
2.4
9.
•
l /a l /b
, where neither a, b, c, nor d is O
1 /c l /d
•
In Exercises 1 1 and 12, solve the given system using the
method of Example 3.25.
12. X 1 - X2 = 1
2x 1 + x2 = 2
1 1 . 2X + y = - 1
5x + 3y = 2
13. Let A =
For larger systems, the difference is even more
pronounced, and this explains why computer
systems do not use one of these methods to solve
linear systems.
14. Prove Theorem 3.9(b).
15. Prove Theorem 3.9(d).
16. Prove that the n X n identity matrix In is invertible and
that ln- l = Iw
17. (a) Give a counterexample to show that (AB) - 1 *
A - 1 B - 1 in general.
(h) Under what conditions on A and B is (AB) - 1 =
A - 1 B - 1 ? Prove your assertion.
18. By induction, prove that if A 1 , A 2 , , A n are invertible
matrices of the same size, then the product A 1A 2 · · · A n
A n ) - 1 = A ;;- 1 · · · A2 1 A � 1 •
is invertible and (A 1A 2
19. Give a counterexample to show that (A + B) - 1 *
A - 1 + B - 1 in general.
[ � !l [ ;] [ �J
h1 =
, h2 =
-
, and b 3 =
(a) Find A - 1 and use it to solve the three systems
Ax = h 1 , Ax = h2 , and Ax = h 3 .
[ �] .
(h) Solve all three systems at the same time by row re­
ducing the augmented matrix [A I h 1 h 2 h 3 ] using
Gauss-Jordan elimination.
(c) Carefully count the total number of individual
multiplications that you performed in (a) and in
(b) . You should discover that, even for this 2 X 2
example, one method uses fewer operations.
•
•
•
•
In Exercises 20 -23, solve the given matrix equation for X.
Simplify your answers as much as possible. (In the words of
Albert Einstein, "Everything should be made as simple as pos­
sible, but not simpler.") Assume that all matrices are invertible.
21. AXB = ( BA ) 2
20. XA 2 = A - 1
1
1
1
22. (A - X) - = A ( B - 2A ) - 23. ABXA - 1 B - 1 = I + A
In Exercises 24-30, let
A=
c
�
[:
[:
2
1
-1
2
] [ � �l
] H -:]
�
-1
,
B=
-1
1 ' D=
-1
1
-1
1
2
2
-1
-1
-1
Section
In each case, find an elementary matrix E that satisfies the
given equation.
24. EA = B
25. EB = A
26. EA = C
28. EC = D
27. EC = A
29. ED = C
30. Is there an elementary matrix E such that EA = D?
Why or why not?
31.
50.
33.
35.
37.
0
1
0
0
c
0
34.
36.
38.
[� �]
[ _� � ]
[: :J
[i l H
0
1
0
0
.
39. A =
•
.
[ _ � -�J
52.
54.
0
In Exercises 39 and 40, find a sequence of elementary
matrices E1, E2 , , Ek such that Ek · · · E2 E1A = I. Use this
sequence to write both A and A i as products of elementary
matrices.
56.
-
40. A =
[ � �]
41. Prove Theorem 3 . 1 3 fo r the case o f AB = I.
42. (a) Prove that if A is invertible and AB = 0, then
[ � :J
[� - �]
[: -� -� ]
:[ � : J
[: � �]
0
49.
Thus, something that is idempotent has the "same
power" when squared.)
( a) Find three idempotent X matrices.
(b) Prove that the only invertible idempotent n X n
matrix is the identity matrix.
45. Show that if A is a square matrix that1 satisfies the
equation A 2 - 2A + I = 0, then A - = 2I - A.
46. Prove that if a symmetric matrix is invertible, then its
inverse is symmetric also.
22
[ -2 4]
3
-1
-1
[
0 2 V2
V2
0
- 4 V2 V2
58.
1
0
0
3
0
0
B = 0.
(b) Give a counterexample to show that the result in
part (a) may fail if A is not invertible.
43. (a) Prove that if A is invertible and BA = CA, then
B = C.
(b) Give a counterexample to show that the result in
part (a) may fail if A is not invertible.
44. A square matrix A is called idempotent if A 2 = A.
(The word idempotent comes from the Latin idem,
meaning "same;' and potere, meaning "to have power:'
119
In Exercises 48-63, use the Gauss-Jordan method to find the
inverse of the given matrix (if it exists).
48.
32.
The Inverse of a Matrix
47. Prove that if A and B are square matrices and AB is
invertible, then both A and B are invertible.
In Exercises 3 1 -38, find the inverse of the given elementary
matrix.
[� �]
[� �]
[ i - �]
[ i n' H
3. 3
60.
61.
[! !]
[: 2 OJ
over Z 5
5
63.
62.
[
0
1
[:
4 over Z 7
6 1
Partitioning large square matrices can sometimes make their
inverses easier to compute, particularly if the blocks have
a nice form. In Exercises 64-68, verify by block multip lica­
tion that the inverse of a matrix, ifpartitioned as shown, is
as claimed. (Assume that all inverses exist as needed.)
Chapter
180
65.
66.
67.
3
[ � �r [
[� �r [
[� �r
l
l
l
Matrices
]
(BC) - 1 B
- (BC) - 1
1
C(BC) - I - C(BC) - 1 B
(I - BC) - 1
- (I - BC) - 1 B
- C(I - BC) - 1 I + C(I - BC) - 1 B
[
]
(BD- 1 C) - 1 BD - 1
- (BD- 1 C) - 1
- D - 1 c(BD- 1 c) - l D - 1 - D - 1 c(BD- 1 c) - 1 BD - l
A B -1 - p Q
1 1
68.
- R S , where P = (A - BD - C) - ,
C D
- 1 , R = -D - 1 CP, and S = D - 1
Q = -PBD
1
+ D - CPBD - 1
[ ] [ ]
•
In Exercises 69-72, partition the given matrix so that you
can apply one of the formulas from Exercises 64-68, and
then calculate the inverse using that formula.
]
�]
0 0
1 0
3 1
2 0
70. The matrix in Exercise 58
71.
[� �:
0
0 1
-1 1
1 0
72.
T h e LU Factorizatio n
Just as it is natural (and illuminating) to factor a natural number into a product of
other natural numbers-for example, 30 = 2 3 5-it is also frequently helpful to fac­
tor matrices as products of other matrices. Any representation of a matrix as a product
of two or more other matrices is called a matrix factorization. For example,
·
·
[ ] [ ][ ]
3
9
-1
-5
=
l 0 3
3 1 0
-1
-2
is a matrix factorization.
Needless to say, some factorizations are more useful than others. In this section,
we introduce a matrix factorization that arises in the solution of systems of linear
equations by Gaussian elimination and is particularly well suited to computer imple­
mentation. In subsequent chapters, we will encounter other equally useful matrix
factorizations. Indeed, the topic is a rich one, and entire books and courses have been
devoted to it.
Consider a system of linear equations of the form Ax = b, where A is an n X n
matrix. Our goal is to show that Gaussian elimination implicitly factors A into a prod­
uct of matrices that then enable us to solve the given system (and any other system
with the same coefficient matrix) easily.
The following example illustrates the basic idea.
Exa m p l e 3 . 3 3
Let
1
-1
5
Section
3.4
The LU Factorization
Row reduction of A proceeds as follows:
181
u
-1
5
(1)
The three elementary matrices £ 1 , E2 , E3 that accomplish this reduction of A to
echelon form U are (in order) :
0
1
0
0
1
0
Hence,
Solving for A, we get
A = E 1- 1 E2- 1 E3- 1 U =
[� �J U �m �J u
LU
u
�
J
�
u
0
1
0
0
1
0
0
1
-2
0
-2
Thus, A can be factored as
A = LU
where U is an upper triangular matrix (see the exercises for Section 3.2), and L is unit
lower triangular. That is, L has the form
1948
(1912-1954)
The LU factorization was introduced
in
by the great English
mathematician Alan M. Turing
in a paper entitled
"Rounding-off Errors in Matrix
Processes" ( Quarterly Journal of
(1948), 287-308).
Mechanics and Applied Mathematics,
1
pp.
During
World War II, Turing was
instrumental in cracking the
German "Enigma'' code. However,
he is best known for his work in
mathematical logic that laid the
theoretical groundwork for the
development of the digital computer
and the modern field of artificial
intelligence. The "Turing test"
that he proposed in
is still
used as one of the benchmarks in
addressing the question of whether
a computer can be considered
"intelligent:'
1950
r � [:
: !]
with zeros above and l s on the main diagonal.
The preceding example motivates the following definition.
D e f i n i t i o n Let A be a square matrix. A factorization of A as A = LU, where
L is unit lower triangular and U is upper triangular, is called an LU factorization
of A.
R e m a rks
Observe that the matrix A in Example 3.33 had an LU factorization because no
row interchanges were needed in the row reduction of A. Hence, all of the elementary
•
matrices that arose were unit lower triangular. Thus, L was guaranteed to be unit
182
Chapter
3
Matrices
�
lower triangular because inverses and products of unit lower triangular matrices are
also unit lower triangular. (See Exercises 29 and 30.)
If a zero had appeared in a pivot position at any step, we would have had to swap
rows to get a nonzero pivot. This would have resulted in L no longer being unit lower
triangular. We will comment further on this observation below. (Can you find a ma­
trix for which row interchanges will be necessary?)
•
The notion of an LU factorization can be generalized to nonsquare matrices
by simply requiring U to be a matrix in row echelon form. (See Exercises 1 3 and 1 4 . )
•
Some books define an LU factorization of a square matrix A to be any factor­
ization A = LU, where L is lower triangular and U is upper triangular.
The first remark above is essentially a proof of the following theorem.
Theorem 3 . 1 5
If A is a square matrix that can be reduced to row echelon form without using any
row interchanges, then A has an LU factorization.
To see why the LU factorization is useful, consider a linear system Ax = b, where
the coefficient matrix has an LU factorization A = LU. We can rewrite the system
Ax = b as LUx = b or L( Ux) = b. If we now define y = Ux, then we can solve for x in
two stages:
1. Solve Ly = b for y byforward substitution (see Exercises 25 and 26 in Section 2 . 1 ) .
2. Solve Ux = y fo r x by back substitution.
Each of these linear systems is straightforward to solve because the coefficient matri­
ces L and U are both triangular. The next example illustrates the method.
Exa m p l e 3 . 3 4
Use an LU factorization of A =
[ ! !]
to ,olve Ax � b, whece b �
-2
0
1
-1
5
-2
Solution
In Example 3.33, we found that
A =
[f:J
[ � � i [� - � -� ]
2
-1
1
0
2
[-: l
= LU
As outlined above, to solve Ax = b (which is the same as L( Ux) = b ), we first solve
Ly � b fm y �
Thi' ' ' ju,t the Hnem- 'l"tem
Y1
-4
2y 1 + Y2
-y 1 - 2y2 + y3 = 9
-
Forward substitution (that is, working from top to bottom) yields
Y i = l , y2 = - 4
2y i = - 6 , y = 9 + Y i + 2y2 = - 2
3
Section
2x 1 +
3.4
The LU Factorization
183
x2 + 3x3 =
- 3x2 - 3x3 = - 6
2x3 = - 2
and back substitution quickly produces
X3 = - 1 ,
- 3x2 = - 6 + 3x3 = - 9 so that x2 = 3, and
2x 1 = 1 - x2 - 3x3 = 1 so that x1 = t
Thmfoce, the rnlution to the given 'Y'tem Ax � b ; , x
�
[ _!l
·
A n Easv wav to Find L U Facto rizations
ii>-"'--
In Example 3.33, we computed the matrix L as a product of elementary matrices.
Fortunately, L can be computed directly from the row reduction process without our
needing to compute elementary matrices at all. Remember that we are assuming that
A can be reduced to row echelon form without using any row interchanges. If this is
the case, then the entire row reduction process can be done using only elementary
row operations of the form R; - kR1 . (Why do we not need to use the remaining
elementary row operation, multiplying a row by a nonzero scalar?) In the operation
R; - kRj, we will refer to the scalar k as the multiplier.
In Example 3.33, the elementary row operations that were used were, in order,
2)
-1)
-2)
R2 - 2R 1
(multiplier
R3 + R 1 = R 3 - ( - l)R 1
(multiplier
R3 + 2R2 = R 3 - ( - 2)R2
(multiplier
The multipliers are precisely the entries of L that are below its diagonal! Indeed,
=
=
=
and L 2 1 = 2, L 3 1 = - 1 , and L 3 2 = - 2. Notice that the elementary row operation
R; - kR1 has its multiplier k placed in the (i, j) entry of L.
Exa m p l e 3 . 3 5
Find an LU factorization of
A=
r ! 4 ! --�4�1
3 2
-9 5
5
-2
-1
184
Chapter
3
Matrices
Solulion
] ['
[;
. -HJ•, [ :
Reducing A to row echelon form, we have
A
=
[j
4
2
5
3
8
5
-2
-4
- 10
-1
-4
R2 - 2R1
R3-R1
R4- ( - 3)R1
-----+
R, - !R2
R4-4R1
-----+
-----+
]
!
]
]
1 3 -4
Q 2 2 -2
2
0
0 8 7 -1
2
0
0
3 -4
2 -2
4
- 1 -8
3 -4
2 2 -2
4
0 0 1
0 0 0 -4
=
U
The first three multipliers are 2, 1 , and - 3, and these go into the subdiagonal entries
of the first column of L. So, thus far,
The next two multipliers are t and 4, so we continue to fill out L:
The final multiplier, - 1 , replaces the last * in L to give
Thus, an LU factorization of A is
A
=
![
4
3 2
-9 5
as is easily checked.
!
5
-2
-��i
-1
-4
� � �i �
[ [
0
1 t
-3 4
-1
0
1
�
2
0 0 1
0 0 0
=�i
4
-4
=
LU
Section
3.4
The LU Factorization
185
R e m a rks
In applying this method, it is important to note that the elementary row opera­
tions R; - kRj must be performed from top to bottom within each column (using the
diagonal entry as the pivot), and column by column from left to right. To illustrate
what can go wrong if we do not obey these rules, consider the following row reduction:
•
U
�
[
�Li L32-=� L2-�1 ]=
L � [i : �]
LUL
A.)
A = [� � �i [� �
2 2
1
�'
0 0
-1
�
0
0
This time the multipliers would be placed in as follows:
get
.,._..
but
A LU.
-1
2,
1 . We would
(Check this! Find a correct factorization of
An alternative way to construct is to observe that the multipliers can be
obtained directly from the matrices obtained at the intermediate steps of the row
reduction process. In Example 3.33, examine the pivots and the corresponding col­
umns of the matrices that arise in the row reduction
•
*
1
-1
5
u
A.
The first pivot is 2, which occurs in the first column of Dividing the entries of
this column vector that are on or below the diagonal by the pivot produces
[± J [ _ � -
A1
The next pivot is - 3, which occurs in the second column of . Dividing the entries
of this column vector that are on or below the diagonal by the pivot, we obtain
U.
The final pivot (which we did not need to use) is 2, in the third column of Divid­
ing the entries of this column vector that are on or below the diagonal by the pivot,
we obtain
±[J [J
If we place the resulting three column vectors side by side in a matrix, we have
L
u -� J
which is exactly once the above-diagonal entries are filled with zeros.
186
Chapter
3
Matrices
In Chapter 2, we remarked that the row echelon form of a matrix is not unique.
However, if an invertible matrix A has an LU factorization A = LU, then this factoriza­
tion is unique.
Theorem 3 . 1 6
�
If A is an invertible matrix that has an LU factorization, then L and U are unique.
Proof Suppose A = LU and A = L 1 U1 are two LU factorizations of A. Then LU =
L 1 U1 , where L and L 1 are unit lower triangular and U and U1 are upper triangular. In
fact, U and U1 are two (possibly different) row echelon forms of A.
By Exercise 30, L 1 is invertible. Because A is invertible, its reduced row echelon
form is an identity matrix I by the Fundamental Theorem of Invertible Matrices.
Hence U also row reduces to I (why?) and so U is invertible also. Therefore,
Lj 1 ( L U) u - 1 = Lj 1 ( L 1 U1 ) u - 1 so ( Ll 1 L ) ( uu - 1 ) = ( Ll 1 L 1 ) ( UI u - 1 )
Hence,
�
( Ll 1 L ) I = I( UI u - 1 ) so Li 1 L = U1 u - I
But Lj 1 L is unit lower triangular by Exercise 29, and U1 u - 1 is upper triangular.
(Why?) It follows that Lj 1 L = U1 u- 1 is both unit lower triangular and upper tri­
angular. The only such matrix is the identity matrix, so Lj 1 L = I and U1 u - 1 = I. It
follows that L = L 1 and U = U1 , so the LU factorization of A is unique.
The pr LU Facto rization
We now explore the problem of adapting the LU factorization to handle cases where
row interchanges are necessary during Gaussian elimination. Consider the matrix
A straightforward row reduction produces
which is not an upper triangular matrix. However, we can easily convert this into
upper triangular form by swapping rows 2 and 3 of B to get
Alternatively, we can swap rows 2 and 3 of A first. To this end, let P be the elementary
matrix
Section
3.4
The LU Factorization
181
corresponding to interchanging rows 2 and 3, and let E be the product of the
elementary matrices that then reduce PA to U (so that E - 1 = L is unit lower triangu­
lar). Thus EPA = U, so A = (EP) - 1 U = P - 1 E - 1 U = P - 1 LU.
Now this handles only the case of a single row interchange. In general, P will be
the product P = Pk · · ·P2 P 1 of all the row interchange matrices P 1 ,P2 , , Pk (where
P 1 is performed first, and so on) . Such a matrix P is called a permutation matrix. Ob­
serve that a permutation matrix arises from permuting the rows of an identity matrix
in some order. For example, the following are all permutation matrices:
•
[� �] . [ :
•
•
0
0
Fortunately, the inverse of a permutation matrix is easy to compute; in fact, no calcu­
lations are needed at all!
Theorem 3 . 1 1
If P is a permutation matrix, then p - l = P T.
We must show that p Tp = I. But the ith row of p T is the same as the ith
column of P, and these are both equal to the same standard unit vector e, because P
is a permutation matrix. So
Proof
( prp) ;; = ( ith row of p r ) ( ith column of P) = e re = e · e = 1
This shows that diagonal entries of pTp are all ls. On the other hand, if j i= i, then
the jth column of P is a different standard unit vector from e-say e ' . Thus, a typical
off-diagonal entry of prp is given by
( p rp) iJ = ( ith row of p r ) (jth column of P) = e re ' = e · e ' = 0
Hence pTP is an identity matrix, as we wished to show.
Thus, in general, we can factor a square matrix A as A = P - 1 LU = P rLU.
Let A be a square matrix. A factorization of A as A = PTLU, where
P is a permutation matrix, L is unit lower triangular, and U is upper triangular, is
called a PrLUfactorization of A.
Definition
Exa m p l e 3 . 3 6
Find o pTLU fadmizotion of A
�
[: � n
First we reduce A to row echelon form. Clearly, we need at least one row
interchange.
2
2
0
R1 +-* Rz
R, - 2R1
�
�
0
0
2
A =
1
-3
Solution
[ : ! ] [ i ! ] [ i - �l
[: -!]
�
Rz +-* R3
2
-3
0
188
Chapter
3
Matrices
We have used two row interchanges (R 1
permutation matrix is
p = P2 P 1 =
�
R 2 and then R 2
R 3 ), so the required
�
[i �m :J [: �]
1
0
0
1
0
0
0
0
We now find an LU factorization of PA.
PA =
[: �m !] [� !] [: -!]
[ � : rn � � : J [ i -� - ! :
1
0
0
2
1
0
0
2
1
2
-3
0
R2 - 2R1
u
Hence L 2 1 = 2, and so
A � P'L U �
The discussion above justifies the following theorem.
Theorem 3 . 1 8
Every square matrix has a PTLU factorization.
Even for an invertible matrix, the P TLU factorization is not unique. In
Example 3.36, a single row interchange R 1 � R 3 also would have worked, leading to
a different P. However, once P has been determined, L and U are unique.
Remark
Comp utational Considerations
IfA is n X n, then the total number of operations (multiplications and divisions) required
to solve a linear system Ax = b using an LU factorization of A) is T(n) n 3 /3, the same
as is required for Gaussian elimination. (See the Exploration "Counting Operations;'
in Chapter 2.) This is hardly surprising since the forward elimination phase produces
the LU factorization in n 3 /3 steps, whereas both forward and backward substitution
require n 2 /2 steps. Therefore, for large values of n, the n 3 /3 term is dominant. From
this point of view, then, Gaussian elimination and the LU factorization are equivalent.
However, the LU factorization has other advantages:
=
=
=
From a storage point of view, the L U factorization is very compact because
we can overwrite the entries of A with the entries of L and U as they are computed. In
Example 3.33, we found that
•
A=
�[ - � � i [ � � � i [ � - � � ]
[ � =� �]
-
-2
5 5
-1
-2
1
This can be stored as
-1
-2
-
2
0
0
2
= LU
Section
3.4
The LU Factorization
189
with the entries placed in the order ( 1 , 1 ) , ( 1 ,2), ( 1 ,3), (2, 1 ) , (3, 1 ) , (2,2), (2,3), (3,2),
(3,3). In other words, the subdiagonal entries of A are replaced by the corresponding
multipliers. (Check that this works! )
•
Once an LU factorization of A has been computed, it can be used to solve as
many linear systems of the form Ax
as we like. We just need to apply the method
of Example 3.34, varying the vector each time.
•
For matrices with certain special forms, especially those with a large number
of zeros (so-called "sparse" matrices) concentrated off the diagonal, there are methods
that will simplify the computation of an LU factorization. In these cases, this method
is faster than Gaussian elimination in solving Ax
•
For an invertible matrix A, an LU factorization of A can be used to find A - 1 ,
if necessary. Moreover, this can be done in such a way that it simultaneously yields a
factorization of A - 1 . (See Exercises 1 5 - 1 8.)
b= b
= b.
Remark If you have a CAS (such as MATLAB) that has the LU factorization
built in, you may notice some differences between your hand calculations and the
computer output. This is because most CAS's will automatically try to perform partial
pivoting to reduce roundoff errors. (See the Exploration "Partial Pivoting;' in Chapter
2.) Turing's paper is an extended discussion of such errors in the context of matrix
factorizations.
I
This section has served to introduce one of the most useful matrix factorizations.
In subsequent chapters, we will encounter other equally useful factorizations.
Exercises 3 . 4
=b
1. A = [ l ] [ J [ l ] ' b = [ S J
= [� -� ] = [t � ][� - ! l b = [ � ]
= [ - ! : =�J H ; � l
�[ = !], b [ - � 1
= [ � = � � 1 [ -�; � 1
[� -� H b � U_
In Exercises 1 -6, solve the system Ax
LU factorization of A.
1 O
-1 1
-2
2 5
5. A
=
6. A
=
using the given
-2
0 6
1
2. A
3. A
_
_
x
4. A
4
0 0
-1
x
-�
2 2
=
0
0
0 1
]
[ ]
In Exercises 7-12,find an LUfactorization of the given matrix.
7·
[
1
-3
2
-1
s.
2
3
-4
1
Chapter
190
9.
11.
Matrices
[: ; :l
[ � : _; -:i
[ -� ! -� � i
-1
12.
3
-2
4 4
6 9
-9
In Exercises 1 9-22, write the given permutation matrix as a
product of elementary (row interchange) matrices.
19. [ : � ]
0
0
0
21.
7 3
5 8
Generalize the definition of LUfactorization to nonsquare
matrices by simply requiring U to be a matrix in row ech­
elon form. With this modification, find an LUfactorization
of the matrices in Exercises 13 and 14.
[
1 c
13. 0 3 3
0 0 0
14.
[�
1
-2
2
-7
1
3
-�
0
3
3
-3
]
-1
8
5
-6
-�]
Fo r a n invertible matrix with a n L Ufactorization1 A1 = LU,
both L and U will be invertible
and1 A - l = u1 - L - . In
1
Exercises 15 and 1 6, find L , u - , and A - for the given
matrix.
15. A in Exercise 1
16. A in Exercise 4
The inverse of a matrix can also be computed by solving sev­
eral systems of equations using the method ofExample 3.34.
For an n X n matrix A, to find its inverse we need to solve
AX = In for the n X n matrix X. Writing this equation as
A [ x1 x 2 · · x n ] = [ e 1 e2 · · · e n ] , using the matrix-column
form ofAX, we see that we need to solve n systems of linear
equations: Ax1 = e 1 , Ax2 = e2 , . . . , AXn = en · Moreover, we
can use the factorization A = LU to solve each one of these
systems.
0
1
22. 0
0
0
0
0
0
1
0
0
0
0
1
0
0
0
0
0
0
1
0
0
0
0
0
1
0
[ -; �]
In Exercises 23-25, find a PTLUfactorization of the given
matrix A.
23. A =
25. A =
H !]
2
3
[ -;
-1
1
1
0
-1
1
0
24. A =
il
2
1
3
-1
26. Prove that there are exactly n! n X n permutation
matrices.
In Exercises 27-28, solve the system Ax = b using the given
factorization A = PTLU. Because ppT = I, PTLUx = b can
be rewritten as LUx = Pb. This system can then be solved
using the method of Example 3.34.
27. A =
•
In Exercises 1 7 and 18, use the approach just outlined to
find A - l for the given matrix. Compare with the method of
Exercises 15 and 1 6.
18. A in Exercise 4
17. A in Exercise 1
20.
[ ;]
1
0
0
0
[� � ]
0 0
0 1
1 0
0 0
[� � �l [� � �l [� C :l
m
[ : �l l
[ ! �- [ : : J [ : : J
[ �:i
[� r
-
1
-1
3
1
0
x
0 0 1
�
P'L U, b
2
�
3
28. A =
0
0
0
x
1
-1
0
= P TL U, b =
-t 1
0
1
-1
Section
29. Prove that a product of unit lower triangular matrices
is unit lower triangular.
30. Prove that every unit lower triangular matrix is
invertible and that its inverse is also unit lower
triangular.
An LDUfactorization of a square matrix A is a factoriza­
tion A = LDU, where L is a unit lower triangular matrix,
D is a diagonal matrix, and U is a unit upper triangular matrix (upper triangular with ls on its diagonal). In
Exercises 31 and 32, find an LDUfactorization of A.
3. 5
Subspaces, Basis, Dimension, and Rank
191
31. A in Exercise 1
32. A in Exercise 4
33. If A is symmetric and invertible and has an LDU
factorization, show that U = L r.
34. If A is symmetric and invertible and A = LDL T (with L
unit lower triangular and D diagonal), prove that this
factorization is unique. That is, prove that if we also
have A = L1D1L[ (with L1 unit lower triangular and D1
diagonal), then L = L1 and D = D1•
S u b s p aces , Basis , D i m e n s io n , and R a n k
z
2u + v
x
y
Figure 3 . 2
This section introduces perhaps the most important ideas in the entire book. We have
already seen that there is an interplay between geometry and algebra: We can often
use geometric intuition and reasoning to obtain algebraic results, and the power of
algebra will often allow us to extend our findings well beyond the geometric settings
in which they first arose.
In our study of vectors, we have already encountered all of the concepts in this
section informally. Here, we will start to become more formal by giving definitions
for the key ideas. As you'll see, the notion of a subspace is simply an algebraic
generalization of the geometric examples of lines and planes through the origin. The
fundamental concept of a basis for a subspace is then derived from the idea of direc­
tion vectors for such lines and planes. The concept of a basis will allow us to give a
precise definition of dimension that agrees with an intuitive, geometric idea of the
term, yet is flexible enough to allow generalization to other settings.
You will also begin to see that these ideas shed more light on what you already
know about matrices and the solution of systems of linear equations. In Chapter 6,
we will encounter all of these fundamental ideas again, in more detail. Consider this
section a "getting to know you" session.
A plane through the origin in IR 3 "looks like" a copy of IR 2 • Intuitively, we would
agree that they are both "two-dimensional:' Pressed further, we might also say that
any calculation that can be done with vectors in IR 2 can also be done in a plane through
the origin. In particular, we can add and take scalar multiples (and, more generally,
form linear combinations) of vectors in such a plane, and the results are other vec­
tors in the same plane. We say that, like IR 2 , a plane through the origin is closed with
respect to the operations of addition and scalar multiplication. (See Figure 3.2.)
But are the vectors in this plane two- or three-dimensional objects? We might
argue that they are three-dimensional because they live in IR 3 and therefore have three
components. On the other hand, they can be described as a linear combination of just
two vectors-direction vectors for the plane-and so are two-dimensional objects liv­
ing in a two-dimensional plane. The notion of a subspace is the key to resolving this
conundrum.
192
Chapter
3
Matrices
Definition
A subspace of !R n is any collection S of vectors in !R n such that:
1 . The zero vector 0 is in S.
2. If u and v are in S, then u + v is in S. (S is closed under addition.)
3. If u is in S and c is a scalar, then cu is in S. (S is closed under scalar
multiplication.)
We could have combined properties (2) and (3) and required, equivalently, that S be
closed under linear combinations:
If u 1 , u2 , • . • , uk are in S and c 1 , c2 , . • . , ck are scalars,
then c 1 u 1 + c2 u2 + · · · + ckuk is in S.
Exa m p l e 3 . 3 1
Every line and plane through the origin in IR 3 is a subspace of IR 3 . It should be clear
geometrically that properties ( 1 ) through (3) are satisfied. Here is an algebraic proof
in the case of a plane through the origin. You are asked to give the corresponding
proof for a line in Exercise 9.
Let <JP be a plane through the origin with direction vectors v1 and v2 . Hence, <JP =
span (v1 , v2 ) . The zero vector 0 is in <JP, since 0 = Ov1 + Ov2 • Now let
be two vectors in <JP. Then
Thus, u + v is a linear combination of v1 and v2 and so is in <JP.
Now let c be a scalar. Then
4
which shows that cu is also a linear combination of v1 and v2 and is therefore in <JP. We
have shown that <JP satisfies properties ( 1 ) through ( 3) and hence is a subspace of IR 3
If you look carefully at the details of Example 3.3 7, you will notice that the fact
that v1 and v2 were vectors in IR 3 played no role at all in the verification of the prop­
erties. Thus, the algebraic method we used should generalize beyond IR 3 and apply
in situations where we can no longer visualize the geometry. It does. Moreover, the
method of Example 3.37 can serve as a "template" in more general settings. When we
generalize Example 3.3 7 to the span of an arbitrary set of vectors in any !R n , the result
is important enough to be called a theorem.
Theorem 3 . 1 9
Let v1 , v2 , . . . , vk be vectors in !R n . Then span(v1 , v2 , . . . , vk) is a subspace of !R n .
Let S = span (v1 , v2 , . . . , vk) · To check property ( 1 ) of the definition, we simply
observe that the zero vector 0 is in S, since 0 = Ov1 + Ov2 + · · · + Ovk.
Proof
3. 5
Section
Subspaces, Basis, Dimension, and Rank
193
Now let
be two vectors in S. Then
u + v = ( c 1 v1 + c2v2 + · · · + ckvk ) + ( d 1 v1 + d2v2 + · · · + dkvk )
= ( c 1 + d 1 ) v1 + ( cz + dz ) V2 + · · · + ( ck + dk ) vk
Thus, u + v is a linear combination of v1 , v2 , , vk and so is in S. This verifies prop­
erty ( 2).
To show property (3), let c be a scalar. Then
•
.
.
cu = c ( c 1 v1 + c2v2 + · · · + c kvk )
= ( cc 1 ) v1 + ( cc2 ) v2 + · · · + ( cck ) vk
which shows that cu is also a linear combination of v1 , v2 , . . . , vk and is therefore
in S. We have shown that S satisfies properties ( 1 ) through (3) and hence is a subspace
of ll�r.
We will refer to span (v1 , v2 ,
, vk) as the subspace spanned by v1 , v2 ,
, vk.
We will often be able to save a lot of work by recognizing when Theorem 3 . 1 9 can be
applied.
.
Exa m p l e 3 . 3 8
Show that the set of all vectors
forms a subspace of IR\ 3 .
Solullon
•
.
[;]
.
that satisfy the conditions x
=
3y and z
.
=
•
- 2y
z
Substitufog the two conditions into
[�]
Since y is arbitrary, the given set of vectors is span
of lR\ 3 , by Theorem 3 . 1 9.
[ _n
yields
( [ �] )
-2
and is thus a subspace
4
Geometrically, the set of vectors in Example 3.38 represents the line through the
migin in GI' with dimtion vectm
194
Chapter
3
Matrices
- 2y
Exa m p l e 3 . 3 9
[; ]
Determine whether the set of all vectors
and z =
is a subspace of IR 3 .
Solulion
....,..
3y
that satisfy the conditions x =
z
3[ y
-:2y l l
+ 1
This time, we have all vectors of the form
The zero vector is not of this form. (Why not? Try solving
3[ y [ )
-:2y l H H'""'
1
property ( 1 ) does not hold, so this set cannot be a subspace of IR 3 .
Exa m p l e 3 . 4 0
[;], y
[ :2 ]
where
Determine whether the set o f all vectors
S o lulion
These are the vectors of the form
= :x2-, i s a
- call this set S. This time 0
belongs to S (take x = O), so property ( 1 ) holds. Let u
Then
u
2.
+
v =
subspace o f IR 2 .
=
[ J
x� + x�
[ :� ]
and v
=
[:� ]
=
[�]
be in S.
X 1 + Xz
which, in general, is not in S, since it does not have the correct form; that is,
x � + x? * (x1 + x2 ) To be specific, we look for a counterexample. If
u =
(2)
[�]
[!]
2IR . [ ! ]
and
then both u and v are in S, but their sum u +
property fails and S is not a subspace of
(3)
v =
v =
is not in S since 5 *
32
. Thus,
.+
In order for a set S to be a subspace of some !R n , we must prove that
properties ( 1 ) through hold in general. However, for S to fail to be a subspace of !R n ,
it is enough to show that one of the three properties fails to hold. The easiest course is
usually to find a single, specific counterexample to illustrate the failure of the property.
Once you have done so, there is no need to consider the other properties.
Remark
Subspaces Associaled wilh Malrices
2;
A great many examples of subspaces arise in the context of matrices. We have already
encountered the most important of these in Chapter we now revisit them with the
notion of a subspace in mind.
Section
Definition
Let A be an m
3. 5
x
Subspaces, Basis, Dimension, and Rank
195
n matrix.
1 . The row space of A is the subspace row(A) of !R n spanned by the rows of A.
2. The column space of A is the subspace col (A) of !R m spanned by the columns
of A.
Remark
Observe that, by Example 3.9 and the Remark that follows it, col(A)
consists precisely of all vectors of the form Ax where x is in !R n .
Exa m p l e 3 . 4 1
Consider the matrix
( ' ) Dctmnine wheth" b
�
[�]
[
1
A= 0
3
;,
-1
1
-3
i
in the column 'P"' of A.
(b) Determine whether w = [ 4 5 ] is in the row space of A.
(c) Describe row(A) and col (A) .
Solution
(a) By Theorem 2.4 and the discussion preceding it, b is a linear combination of the
columns of A if and only if the linear system Ax = b is consistent. We row reduce
the augmented matrix as follows:
[ � �: � ] [ � � � ]
�
Thus, the system is consistent (and, in fact, has a unique solution) . Therefore, b is
in col (A) . (This example is just Example 2 . 1 8, phrased in the terminology of this
section.)
(b) As we also saw in Section 2.3, elementary row operations simply create linear
combinations of the rows of a matrix. That is, they produce vectors only in the row
space of the matrix. If the vector w is in row(A), then w is a linear combination of the
rows of A, so if we augment A by w as
___...
[�],
[�' ]
it will be possible to apply elementary row
operations to this augmented matrix to reduce it to form
using only elementary
row operations of the form R; + kRj , where i > j -in other words, workingfrom top
to bottom in each column. (Why?)
In this example, we have
[�]
[� - � i [� - � i [� -� i
3 -3
4 5
��
�:=
!�:
�
0
0
0
9
��
�
R 4 - 9R2
0
0
0
0
��
196
Chapter
3
Matrices
�
Therefore, w is a linear combination of the rows of A (in fact, these calculations show
that w = 4 [ l - 1 ] + 9 [ O l ] -how?), and thus w is in row(A) .
( c) It is easy to check that, fo r any vector w = [ x y ] , the augmented matrix
reduces to
[�]
in a similar fashion. Therefore, every vector in IR 2 is in row(A), and so row(A) = IR 2 .
Finding col (A) is identical to solving Example 2.2 1 , wherein we determined that
it coincides with the plane (through the origin) in IR 3 with equation 3x - z = 0. (We
will discover other ways to answer this type of question shortly.)
Remark We could also have answered part (b) and the first part of part (c) by
observing that any question about the rows of A is the corresponding question about
the columns of A T. So, for example, w is in row(A) if and only if wT is in col (A T ) . This
is true if and only if the system A Tx = wT is consistent. We can now proceed as in
part (a) . (See Exercises 2 1 -24.)
The observations we have made about the relationship between elementary row
operations and the row space are summarized in the following theorem.
Theorem 3 . 2 0
Let B be any matrix that is row equivalent to a matrix A. Then row(B) = row(A).
The matrix A can be transformed into B by a sequence of row operations.
Consequently, the rows of B are linear combinations of the rows of A; hence, linear
combinations of the rows of B are linear combinations of the rows of A. (See Exer­
cise 2 1 in Section 2.3.) It follows that row(B) � row(A) .
O n the other hand, reversing these row operations transforms B into A. There­
fore, the above argument shows that row(A) � row(B) . Combining these results, we
have row(A) = row(B) .
Proof
There is another important subspace that we have already encountered: the set
of solutions of a homogeneous system of linear equations. It is easy to prove that this
subspace satisfies the three subspace properties.
Theorem 3 . 2 1
Let A be an m X n matrix and let N be the set of solutions of the homogeneous
linear system Ax = 0. Then N is a subspace of !R n .
[Note that x must be a (column) vector in !R n in order for Ax to be defined and
that 0 = O m is the zero vector in !R m .] Since AOn = Om , O n is in N. Now let u and v be
in N . Therefore, Au = 0 and Av = 0. It follows that
A ( u + v) = Au + Av = 0 + 0 = 0
Proof
Section
3. 5
Subspaces, Basis, Dimension, and Rank
Hence, u + v is in N . Finally, for any scalar c,
A ( c u)
=
c (Au )
=
cO
191
= 0
and therefore cu is also in N. It follows that N is a subspace of !R n .
Definition
Let A be an
m
X
n
matrix. The null space of A is the subspace of
0. It is denoted
!R n consisting of solutions of the homogeneous linear system Ax =
by null(A).
The fact that the null space of a matrix is a subspace allows us to prove what intuition and examples have led us to understand about the solutions of linear systems:
They have either no solution, a unique solution, or infinitely many solutions.
Theorem 3 . 2 2
Let A be a matrix whose entries are real numbers. For any system of linear
equations Ax = b, exactly one of the following is true:
a. There is no solution.
b. There is a unique solution.
c. There are infinitely many solutions.
At first glance, it is not entirely clear how we should proceed to prove this theo­
rem. A little reflection should persuade you that what we are really being asked to
prove is that if (a) and (b) are not true, then ( c) is the only other possibility. That is, if
there is more than one solution, then there cannot be just two or even finitely many,
but there must be infinitely many.
If the system Ax = b has either no solutions or exactly one solution, we are
done. Assume, then, that there are at least two distinct solutions of Ax = b-say, x 1
and x2 . Thus,
Ax , = b and Ax2 = b
Proof
with x1 * x2 . It follows that
A ( x, - x2 ) = Ax 1 - Ax2 = b - b = 0
Set Xo = x 1 - x2 . Then Xo * 0 and AXo = 0. Hence, the null space of A is nontrivial,
and since null (A) is closed under scalar multiplication, CXo is in null (A) for every
scalar c. Consequently, the null space of A contains infinitely many vectors (since it
contains at least every vector of the form CXo and there are infinitely many of these) .
Now, consider the (infinitely many) vectors of the form x1 + CXo , as c varies through
the set of real numbers. We have
A ( x, + c x0 ) = Ax, + cAXo = b + c O = b
Therefore, there are infinitely many solutions of the equation Ax = b.
Basis
We can extract a bit more from the intuitive idea that subspaces are generalizations
of planes through the origin in IR 3 . A plane is spanned by any two vectors that are
198
Chapter
3
Matrices
parallel to the plane but are not parallel to each other. In algebraic parlance, two
such vectors span the plane and are linearly independent. Fewer than two vectors will
not work; more than two vectors is not necessary. This is the essence of a basis for a
subspace.
Definition
A basis for a subspace S of ll�r is a set of vectors in S that
1 . spans S and
2. is linearly independent.
Exa m p l e 3 . 4 2
Exa m p l e 3 . 4 3
In Section 2.3, we saw that the standard unit vectors e1, e2 , . . . e n in !R n are linearly
independent and span !R n . Therefore, they form a basis for !R n , called the standard
basis.
In Example 2 . 1 9, we showed that IR 2 = span
([
�
4
] [� ] ) [ � ] [� ]
. Since _ and
_ ,
also linearly independent (as they are not multiples), they form a basis for IR 2 .
{ [�l [�] }
{ [ �l [�] }
are
4
A subspace can (and will) have more than one basis. For example, we have just
seen that IR 2 has the standard basis
. How_
ever, we will prove shortly that the number of vectors in a basis for a given subspace will
always be the same.
Exa m p l e 3 . 4 4
and the basis
Find a basis for S = span (u, v, w) , where
The vectors u, v, and w already span S, so they will be a basis for S if they
are also linearly independent. It is easy to determine that they are not; indeed, w =
2u - 3v. Therefore, we can ignore w, since any linear combinations involving u, v,
and w can be rewritten to involve u and v alone. (Also see Exercise 47 in Section 2.3.)
This implies that S = span (u, v , w) = span ( u, v) , and since u and v are certainly
linearly independent (why?), they form a basis for S. (Geometrically, this means that
u, v, and w all lie in the same plane and u and v can serve as a set of direction vectors
for this plane.)
Solution
�
4
Section
Exa m p l e 3 . 4 5
3. 5
Subspaces, Basis, Dimension, and Rank
199
Find a basis for the row space of
A=
Solution
[ -!
3
-1 0
2 1 -2
1 6
The reduced row echelon form of A is
0
1
0
0
1
2
0
0
0
0
1
0
By Theorem 3.20, row(A) = row(R), so it is enough to find a basis for the row space
of R. But row(R) is clearly spanned by its nonzero rows, and it is easy to check that
the staircase pattern forces the first three rows of R to be linearly independent. (This
is a general fact, one that you will need to establish to prove Exercise 33.) Therefore,
a basis for the row space of A is
{[1 0
0
- 1 ] , [0
1 2 0 3 ] , [0 0 0
4]}
We can use the method of Example 3.45 to find a basis for the subspace spanned
by a given set of vectors.
Exa m p l e 3 . 4 6
Rework Example 3.44 using the method from Example 3.45.
We transpose u, v, and w to get row vectors and then form a matrix with
these vectors as its rows:
Solution
[
!> ]
� � -!
Proceeding as in Example 3.45, we reduce B to its reduced row echelon form
1 0
and use the nonzero row vectors as a basis for the row space. Since we started with
column vectors, we must transpose again. Thus, a basis for span (u, v, w) is
{ [ H Ul l
R e m arks
In fact, we do not need to go all the way to reduced row echelon form-row ech­
elon form is far enough. If U is a row echelon form of A, then the nonzero row vectors
•
200
Chapter
3
Matrices
of U will form a basis for row(A) (see Exercise 33). This approach has the advantage of
(often) allowing us to avoid fractions. In Example 3.46, B can be reduced to
which gives us the basis
\HH - rn
for span (u, v, w) .
•
Observe that the methods used in Example 3.44, Example 3.46, and the Remark
above will generally produce different bases.
We now turn to the problem of finding a basis for the column space of a matrix A.
One method is simply to transpose the matrix. The column vectors of A become the
row vectors of A T' and we can apply the method of Example 3.45 to find a basis for
row(A T ) . Transposing these vectors then gives us a basis for col (A) . (You are asked to
do this in Exercises 2 1 -24.) This approach, however, requires performing a new set
of row operations on A T.
Instead, we prefer to take an approach that allows us to use the row reduced form
of A that we have already computed. Recall that a product Ax of a matrix and a vec­
tor corresponds to a linear combination of the columns of A with the entries of x as
coefficients. Thus, a nontrivial solution to Ax = 0 represents a dependence relation
among the columns of A. Since elementary row operations do not affect the solution
set, if A is row equivalent to R, the columns of A have the same dependence relation­
ships as the columns of R. This important observation is the basis (no pun intended!)
for the technique we now use to find a basis for col (A).
Exa m p l e 3 . 4 1
Find a basis for the column space of the matrix from Example 3.45,
A=
[-�
[�
1 3
1
1
-1 0
2
-2
6
1
-:i
Let a; denote a column vector of A and let r; denote a column vector of the
reduced echelon form
Solution
R
�
0 1 0
1 2 0
0 0 1
0 0 0
-il
We can quickly see by inspection that r3 = r 1 + 2r2 and rs = - r 1 + 3r2 + 4r4 • (Check
that, as predicted, the corresponding column vectors of A satisfy the same depen­
dence relations.) Thus, r3 and rs contribute nothing to col (R) . The remaining column
Section
3. 5
Subspaces, Basis, Dimension, and Rank
201
vectors, r 1 , r2 , and r4 , are linearly independent, since they are just standard unit vec­
tors. The corresponding statements are therefore true of the column vectors of A.
Thus, among the column vectors ofA, we eliminate the dependent ones ( a3 and a5 ) ,
and the remaining ones will be linearly independent and hence form a basis for col(A).
What is the fastest way to find this basis? Use the columns of A that correspond to the
columns of R containing the leading ls. A basis for col(A) is
r . . . ., . .,
} � { [ - ;H - �H - m
Warning
Elementary row operations change the column space! In our example,
col (A) * col (R), since every vector in col (R) has its fourth component equal to 0 but
this is certainly not true of col (A) . So we must go back to the original matrix A to get
the column vectors for a basis of col (A) . To be specific, in Example 3.47, r 1 , r2 , and r4
do not form a basis for the column space of A.
Exa m p l e 3 . 4 8
Find a basis for the null space of matrix A from Example 3.47.
There is really nothing new here except the terminology. We simply have
to find and describe the solutions of the homogeneous system Ax = 0. We have al­
ready computed the reduced row echelon form R of A, so all that remains to be done
in Gauss-Jordan elimination is to solve for the leading variables in terms of the free
variables. The final augmented matrix is
0 1 0
2 0
0 0 1
0 0 0
Solution
If
then the leading l s are in columns 1 , 2, and 4, so we solve for x 1 , x2 , and x4 in terms of
the free variables x3 and x5 • We get x 1 = - x3 + x5 , x2 = - 2x3 - 3x5, and x4 = - 4x5.
Setting x3 = s and x5 = t, we obtain
-s + t
1
-1
X1
-3
- 2s - 3t
-2
X2
= s 1 + t 0 = su + tv
s
x = X3
0
-4t
-4
X4
0
X5
Thus, u and v span null(A ), and since they are linearly independent, they form a basis
for null(A).
4
202
Chapter
3
Matrices
Following is a summary of the most effective procedure to use to find bases for
the row space, the column space, and the null space of a matrix A.
1. Find the reduced row echelon form R of A.
2. Use the nonzero row vectors of R (containing the leading l s) to form a basis for
row(A) .
3. Use the column vectors o fA that correspond t o the columns o f R containing the
leading ls (the pivot columns) to form a basis for col (A) .
4. Solve fo r the leading variables o f Rx = 0 i n terms o f the free variables, set the
free variables equal to parameters, substitute back into x, and write the result as
a linear combination off vectors (where f is the number of free variables) . These
f vectors form a basis for null (A) .
I f we d o not need t o find the null space, then it i s faster t o simply reduce A t o row
echelon form to find bases for the row and column spaces. Steps 2 and 3 above remain
valid (with the substitution of the word "pivots" for "leading l s").
Dimension and Rank
We have observed that although a subspace will have different bases, each basis has
the same number of vectors. This fundamental fact will be of vital importance from
here on in this book.
Theorem 3 . 2 3
The Basis Theorem
Let S be a subspace of !R n . Then any two bases for S have the same number of
vectors.
Proof
Let B = {u 1 , u2 , . . . , u,.} and C = {v1 , v2 ,
, v, } be bases for S. We need to
prove that r = s. We do so by showing that neither of the other two possibilities, r < s
or r > s, can occur.
Suppose that r < s. We will show that this forces C to be a linearly dependent set
of vectors. To this end, let
.
Sherlock Holmes noted, "When
you have eliminated the impos­
sible, whatever remains, however
improbable, must be the truth"
(from The Sign of Four by Sir
Arthur Conan Doyle).
•
•
(1)
Since B i s a basis for S , we can write each
elements "f
V;
a s a linear combination o f the
= a 1 1 u, + a 12 u2 + · · · + a , rur
Vz = a 1 1 u , + a z 2U2 + · · · + a zrUr
V1
Substituting the Equations (2) into Equation ( 1 ) , we obtain
(2)
Section
3. 5
Subspaces, Basis, Dimension, and Rank
203
Regrouping, we have
( c 1 a ll + C2 a2 1 + . . . + c, as1 ) U1 + ( c 1 a 1 2 + C2 a22 + . . . + cs asz ) Uz
+ + ( c 1 a 1r + c2a2 r + + c, a ) u, = 0
Now, since B is a basis, the u/s are linearly independent. So each of the expressions in
·
·
·
·
·
·
parentheses must be zero:
c 1 a 11 + c2a2 1 + + c, a, 1 = 0
C 1 a 1 2 + C2 a22 + . . . + csas 2 = 0
·
·
·
This is a homogeneous system of r linear equations in the s variables c 1 c 2
, c5 (The
fact that the variables appear to the left of the coefficients makes no difference.) Since
r < s, we know from Theorem 2.3 that there are infinitely many solutions. In particu­
lar, there is a nontrivial solution, giving a nontrivial dependence relation in Equa­
tion ( 1 ) . Thus, C is a linearly dependent set of vectors. But this finding contradicts the
fact that C was given to be a basis and hence linearly independent. We conclude that
r < s is not possible. Similarly (interchanging the roles of B and C), we find that r > s
leads to a contradiction. Hence, we must have r = s, as desired.
,
, •
.
.
•
Since all bases for a given subspace must have the same number of vectors, we can
attach a name to this number.
If S is a subspace of !R n , then the number of vectors in a basis for S
is called the dimension of S, denoted dim S.
Definition
The zero vector 0 by itself is always a subspace of !R n . (Why?) Yet any set
containing the zero vector (and, in particular, { 0}) is linearly dependent, so { 0} cannot
have a basis. We define dim {O} to be 0.
Remark
Exa m p l e 3 . 4 9
Since the standard basis for !R n has n vectors, dim !R n = n . (Note that this result agrees
with our intuitive understanding of dimension for n :s 3.)
Exa m p l e 3 . 5 0
In Examples 3.45 through 3.48, we found that row(A) has a basis with three vectors,
col (A) has a basis with three vectors, and null (A) has a basis with two vectors. Hence,
dim(row(A)) = 3, dim(col (A)) = 3, and dim(null (A)) = 2.
A single example is not enough on which to speculate, but the fact that the row
and column spaces in Example 3.50 have the same dimension is no accident. Nor is
the fact that the sum of dim(col (A)) and dim(null (A)) is 5, the number of columns of
A. We now prove that these relationships are true in general.
204
Chapter
3
Matrices
Theorem 3 . 2 4
The row and column spaces of a matrix A have the same dimension.
Proof
Let R be the reduced row echelon form of A. By Theorem 3.20, row(A) =
row(R), so
dim(row (A)) = dim(row(R))
= number of nonzero rows of R
= number of leading l s of R
Let this number be called r.
Now col (A) * col (R), but the columns of A and R have the same dependence
relationships. Therefore, dim(col (A)) = dim(col (R)). Since there are r leading ls, R
has r columns that are standard unit vectors, e1, e2 , . . . , er. (These will be vectors in
!R m if A and R are m X n matrices.) These r vectors are linearly independent, and the
remaining columns of R are linear combinations of them. Thus, dim(col (R)) = r. It
follows that dim(row(A)) = r = dim(col (A)), as we wished to prove.
1878
(1849-1917),
The rank o f a matrix was first de­
fined in
by Georg Frobenius
although he defined
it using determinants and not as we
have done here. (See Chapter
Frobenius was a German
mathematician who received his
doctorate from and later taught
at the University of Berlin. Best
known for his contributions to
group theory, Frobenius used
matrices in his work on group
representations.
4.)
The rank of a matrix A is the dimension of its row and column
spaces and is denoted by rank(A) .
Definition
For Example 3.50, we can thus write rank(A) = 3.
R e m a rks
The preceding definition agrees with the more informal definition of rank that
was introduced in Chapter 2. The advantage of our new definition is that it is much
more flexible.
•
The rank of a matrix simultaneously gives us information about linear
dependence among the row vectors of the matrix and among its column vectors. In
particular, it tells us the number of rows and columns that are linearly independent
(and this number is the same in each case!).
•
Since the row vectors of A are the column vectors of A r, Theorem 3.24 has the
following immediate corollary.
Theorem 3 . 2 5
For any matrix A,
rank (A T) = rank (A )
Proof
We have
rank (A T) = dim ( col (A T))
= dim ( row (A ))
= rank (A )
The nullity of a matrix A is the dimension of its null space and is
denoted by nullity(A) .
Definition
Section
3. 5
Subspaces, Basis, Dimension, and Rank
205
In other words, nullity(A) is the dimension of the solution space of Ax = 0, which
is the same as the number of free variables in the solution. We can now revisit the
Rank Theorem (Theorem 2.2), rephrasing it in terms of our new definitions.
Theorem 3 . 2 6
The Rank Theorem
If A is an m X n matrix, then
rank (A ) + nullity (A ) = n
-
Proof
Let R be the reduced row echelon form of A, and suppose that rank(A) = r.
Then R has r leading ls, so there are r leading variables and n r free variables in the
solution to Ax = 0. Since dim(null (A)) = n r, we have
-
rank (A ) + nullity (A ) = r + (n
= n
-
r)
Often, when we need to know the nullity of a matrix, we do not need to know the
actual solution of Ax = 0. The Rank Theorem is extremely useful in such situations,
as the following example illustrates.
Exa m p l e 3 . 5 1
Find the nullity of each of the following matrices:
M=
N
[I �:
� [�
1
4
7
and
-2
-3
-1
:
Since the two columns of M are clearly linearly independent, rank(M ) = 2.
Thus, by the Rank Theorem, nullity(M ) = 2 - rank(M ) = 2 - 2 = 0.
There is no obvious dependence among the rows or columns of N, so we apply
row operations to reduce it to
Solution
[:
2
0
-2
1
0
-� ]
We have reduced the matrix far enough (we do not need reduced row echelon form
here, since we are not looking for a basis for the null space). We see that there are only
two nonzero rows, so rank(N) = 2. Hence, nullity(N) = 4 rank(N) = 4 2 = 2.
-
-
4
The results of this section allow us to extend the Fundamental Theorem of
Invertible Matrices (Theorem 3 . 1 2 ) .
206
Chapter 3 Matrices
Theorem 3 . 2 1
The Fundamental Theorem of Invertible Matrices: Version 2
Let A be an n X
a.
b.
c.
d.
e.
f.
g.
h.
i.
j.
k.
The
of a matrix
wasSyldefined
in(1814-1887),
1884nullbyityJames
Joseph
vesterin
who
was
i
n
terested
-properties
of certain
matrices
that
do
not
change
under
types
of transformations.
Bornthe
insecond
England,
Syl
v
ester
became
presidentSociety.
of theInLondon
Mathematical
1878,
whi
l
e
teaching
at
Johns
Hopkins
University
imore, he
founded thein Balthetfirst
mathematical
journal in the United States.
invariants
American Journal of
Mathematics,
n
matrix. The following statements are equivalent:
A is invertible.
Ax = b has a unique solution for every b in ll�r.
Ax = 0 has only the trivial solution.
The reduced row echelon form of A is In A is a product of elementary matrices.
rank(A) = n
nullity(A) = 0
The column vectors of A are linearly independent.
The column vectors of A span ll�r.
The column vectors of A form a basis for ll�r.
The row vectors of A are linearly independent.
1. The row vectors of A span !R n .
m. The row vectors of A form a basis for !R n .
We have already established the equivalence of (a) through (e). It remains to
be shown that statements (f) to (m) are equivalent to the first five statements.
(f) � (g) Since rank(A) + nullity(A) = n when A is an n X n matrix, it follows from
the Rank Theorem that rank(A) = n if and only if nullity(A) = 0.
(f) ==> (d) ==> ( c) ==> (h) If rank(A) = n , then the reduced row echelon form of A has
n leading ls and so is I - From (d) ==> (c) we know that Ax = 0 has only the trivial
n
solution, which implies that the column vectors of A are linearly independent, since
Ax is just a linear combination of the column vectors of A.
(h) ==> (i) If the column vectors of A are linearly independent, then Ax = 0 has only
the trivial solution. Thus, by (c) ==> (b), Ax = b has a unique solution for every b in
!R n . This means that every vector b in !R n can be written as a linear combination of the
column vectors of A, establishing (i).
(i) ==> (j) If the column vectors of A span W, then col (A) = !R n by definition,
so rank(A) = dim(col (A)) = n. This is (f), and we have already established that
(f) ==> (h) . We conclude that the column vectors of A are linearly independent and so
form a basis for !R n , since, by assumption, they also span !R n .
(j) ==> (f) If the column vectors of A form a basis for !R n , then, in particular, they are
linearly independent. It follows that the reduced row echelon form of A contains n
leading l s, and thus rank(A) = n .
The above discussion shows that (f) ==> (d) ==> (c) ==> (h) ==> (i) ==> (j) ==>
(f) � (g) . Now recall that, by Theorem 3.25, rank(A T ) = rank(A), so what we have
just proved gives us the corresponding results about the column vectors of A r_ These
are then results about the row vectors of A, bringing (k), (1), and (m) into the network of
equivalences and completing the proof.
Proof
Theorems such as the Fundamental Theorem are not merely of theoretical inter­
est. They are tremendous labor-saving devices as well. The Fundamental Theorem
has already allowed us to cut in half the work needed to check that two square matri­
ces are inverses. It also simplifies the task of showing that certain sets of vectors are
bases for !R n . Indeed, when we have a set of n vectors in !R n , that set will be a basis for
!R n if either of the necessary properties of linear independence or spanning set is true.
The next example shows how easy the calculations can be.
Section 3. 5 Subspaces, Basis, Dimension, and Rank
Exa m p l e 3 . 5 2
201
m
n
mr
Show that the vectors
and
form a basis for IR 3 .
Solution
According to the Fundamental Theorem, the vectors will form a basis for
IR 3 if and only if a matrix with these vectors as its columns (or rows) has rank 3. We
[� � � f l --7 [ � - � J
perform just enough row operations to determine this:
A
-
We see that A has rank 3, so the given vectors are a basis for IR 3 by the equivalence of
(f) and (j) .
4
The next theorem is an application of both the Rank Theorem and the Funda­
mental Theorem. We will require this result in Chapters 5 and 7.
Theorem 3 . 2 8
Let A be an m X n matrix. Then:
a. rank(A TA) = rank(A)
b. The n X n matrix A TA is invertible if and only if rank(A) =
n.
Proof
(a) Since A TA is n X
then tells us that
n,
it has the same number of columns as A. The Rank Theorem
rank (A ) + nullity (A ) =
n
= rank (A TA ) + nullity (A TA )
Hence, to show that rank(A) = rank(A TA), it is enough to show that nullity(A) =
nullity(A TA). We will do so by establishing that the null spaces of A and A TA are the
same.
To this end, let x be in null (A) so that Ax = 0. Then A TAx = A To = 0, and thus
x is in null (A TA). Conversely, let x be in null (A TA). Then A TAx = 0, so xTA TAx =
xTO = 0. But then
and hence Ax = 0, by Theorem 1 .2 (d) . Therefore, x is in null (A), so null (A)
null (A TA), as required.
(b) By the Fundamental Theorem, the n X n matrix A TA is invertible if and only if
rank(A TA) = n. But, by (a) this is so if and only if rank(A) = n .
Coordin ates
We now return to one of the questions posed at the very beginning of this section:
How should we view vectors in IR 3 that live in a plane through the origin? Are they
two-dimensional or three-dimensional? The notions of basis and dimension will help
clarify things.
208
Chapter 3 Matrices
A plane through the origin is a two-dimensional subspace o f IR 3 , with any set of
two direction vectors serving as a basis. Basis vectors locate coordinate axes in the
plane/subspace, in turn allowing us to view the plane as a "copy" of IR 2 • Before we
illustrate this approach, we prove a theorem guaranteeing that "coordinates" that arise
in this way are unique.
Theorem 3 . 2 9
Let S be a subspace of !R n and let B = {v1 , v2 , . . . , vk } be a basis for S. For every
vector v in S, there is exactly one way to write v as a linear combination of the basis
vectors in B:
Proof
Since B is a basis, it spans S, so v can be written in at least one way as a linear
combination of v1, v2 , . . . , vk . Let one of these linear combinations be
Our task is to show that this is the only way to write v as a linear combination of
v1, v2 , . . . , vk . To this end, suppose that we also have
v = d1 v1 + d2v2 + · · · + dkvk
Then
Rearranging (using properties of vector algebra), we obtain
(c 1 - d 1 )v1 + (c2 - d2 )v2 + · · · + (ck - dk ) vk = 0
Since B is a basis, v1, v2 , . . . , vk are linearly independent. Therefore,
(c 1 - d1 ) = (c2 - d2 ) = · · · = (ck - dk ) = 0
In other words, c1 = d1, c 2 = d2 ,
, ck = dk, and the two linear combinations are
actually the same. Thus, there is exactly one way to write v as a linear combination of
•
•
•
the basis vectors in B.
Let S be a subspace of !R n and let B = {v1 , v2 , . . . , vk } be a basis for
S. Let v be a vector in S, and write v = c1v1 + c2v2 + · · · + ckvk . Then c1, c2 , . . . , ck
are called the coordinates of v with respect to B, and the column vector
Definition
is called the coordinate vector of v with respect to B.
Exa m p l e 3 . 5 3
Let E = { e 1 , e2 , e 3 } be the standard basis for IR 3 . Find the coordinate vector of
with respect to E.
Section 3. 5 Subspaces, Basis, Dimension, and Rank
Solution
Since v =
209
2e1 7e 4e 2
[ 47 ]
+ 2 + 3,
[v] E =
It should be clear that the coordinate vector of every (column) vector in !R n with
respect to the standard basis is just the vector itself.
Exa m p l e 3 . 5 4
ID Exompk 3.44, we <aw iliat u �
2u
[ -} m
[ _�]
�
and w �
[ ;l
-
ace th"e v"
tors in the same subspace (plane through the origin) S of IR 3 and that B = { u, v} is a
basis for S. Since w =
- 3v, we have
[w] B =
See Figure 3.3.
z
- 3v
,w
=
2u
-
3v
x ---.
The
respectcoordinates
to a basisof a vector with
Figure 3 . 3
I
y
Exercises 3 . 5
In Exercises 1 -4, let S be the collection of vectors
[; ]
in IR 2
that satisfy the given property. In each case, either prove that
S forms a subspace of IR 2 or give a counterexample to show
that it does not.
1. x =
2. x O, y
3. y = 2x
4. xy
0
2 20
20
In Exm;,,, 5-8, let S be the collea;on of vato"
[;]
;n D;l'
7. x - y + z = 1
9. Prove that every line through the origin in IR 3 is a sub­
space of IR 3 .
10. Suppose S consists of all points in IR 2 that are on the
x-axis or the y-axis (or both). (S is called the union of
the two axes.) Is S a subspace of IR 2 ? Why or why not?
In Exercises 1 1 and 12, determine whether b is in col(A)
and whether w is in row(A), as in Example 3.4 1 .
11. A =
b=
,w = [-1
1]
that satisfy the given property. In each case, either prove that
S forms a subspace of IR 3 or give a counterexample to show
that it does not.
5. x = y = z
6. z = 2x, y =
0
8. I x - y I = I Y - z I
12. A �
[ � � � l [�]
[ � -� � } [i} 2 4
�
w� [
-5
Chapter 3 Matrices
210
13. In Exercise 1 1 , determine whether w is in row(A),
using the method described in the Remark following
Example 3.4 1 .
14. In Exercise 1 2 , determine whether w is in row(A),
using the method described in the Remark following
Example 3.4 1 .
1 5 . IfA ;, tho mateix in F.xmi" 1 1 , b
�
1 6 . I fA i' 1he mateix in Exmi" 12, i< v
�
[ �: }
[ -:}
n null (A) ?
n nu!l(A) ?
In Exercises 1 7-20, give bases for row(A), col(A), and null(A).
17. A =
19. A =
20. A =
[� -�]
[
[� J
!]
H
1
18. A = 0
1
0
1
1
1
1
2
-1
-3
1
-4
l
0
-1
-1
-4 0 2
2
2
4
-2
In Exercises 21-24, find bases for row(A) and col(A) in the
given exercises using A T.
30. [ 0 1
-2
1
1 ], [3
- 1 0], [2
1 5
1]
For Exercises 31 and 32, find bases for the spans of the
vectors in the given exercises from among the vectors
themselves.
3 1 . Exercise 29
32. Exercise 30
33. Prove that if R is a matrix in echelon form, then a basis
for row(R) consists of the nonzero rows of R.
34. Prove that if the columns of A are linearly indepen­
dent, then they must form a basis for col (A) .
For Exercises 35-38, give the rank and the nullity of the
matrices in the given exercises.
35. Exercise 1 7
36. Exercise 1 8
37. Exercise 1 9
38. Exercise 2 0
39. I f A i s a 3 X 5 matrix, explain why the columns o f A
must be linearly dependent.
40. If A is a 4 X 2 matrix, explain why the rows of A must
be linearly dependent.
41. If A is a 3 X 5 matrix, what are the possible values of
nullity(A) ?
42. I f A i s a 4 X 2 matrix, what are the possible values of
nullity(A) ?
l
In Exercises 43 and 44, find all possible values of rank(A) as
a varies.
2 a
[ � � =�l
2 1 . Exercise 1 7
22. Exercise 1 8
23. Exercise 1 9
24. Exercise 20
25. Explain carefully why your answers to Exercises 1 7
and 2 1 are both correct even though there appear to be
differences.
26. Explain carefully why your answers to Exercises 1 8
and 2 2 are both correct even though there appear to
be differences.
Answer Exercises 45-48 by considering the matrix with the
given vectors as its columns.
In Exercises 27-30, find a basis for the span of the given
vectors.
46. Do
27.
[ - il [ - � J [ J [ - :J m. [ : J m
29. [ 2
28.
-3
1], [1
- 1 O ] , [4 -4 1 ]
4a 2
-2 1
45. Do
47. Do
44. A =
[J [ H [ : J
[ -J [ - n [ -1 ]
-2
-1
form a b"i' foe � ' ?
rn
ril m m
focm a bM1' fm � ' ?
focm a bMi' foe � ' '
a
Section 3. 6 Introduction to Linear Transformations
49. Do
50. Do
m [:J m
m [:J [�l
form a b.,i, fo, Zl?
fmm a bn<i• fo, Zj?
In Exercises 51 and 52, show that w is in span(B) andfind
the coordinate vector [w] 8.
In Exercises 53-56, compute the rank and nullity of the
given matrices over the indicated "ll..P "
53.
55.
56.
[ � :J
[� �]
1
1
0
ov°' z,
3 1
3 0
0 4
[!
4 0 0
3 5 1
0 2 2
1
54.
om Z;
�]
[ � �]
1
1
0
om Z,
211
57. If A is m X n, prove that every vector in null (A) is
orthogonal to every vector in row(A) .
58. I f A and B are n X n matrices o f rank n, prove that AB
has rank n.
59. (a) Prove that rank(AB) :::::: rank(B) . [Hint: Review
Exercise 29 in Section 3 . 1 . ]
(b) Give an example in which rank(AB) < rank(B).
60. ( a) Prove that rank(AB) :::::: rank(A). [Hint: Review
Exercise 30 in Section 3 . 1 or use transposes and
Exercise 59(a).]
(b) Give an example in which rank(AB) < rank(A) .
61. ( a) Prove that i f U i s invertible, then rank( UA) =
rank(A) . [Hint: A = U- 1 ( UA).]
(b) Prove that if V is invertible, then rank(A V) =
rank(A) .
62. Prove that an m X n matrix A has rank 1 i f and only if
A can be written as the outer product uvT of a vector u
in !Rm and v in !R n .
63. If an m X n matrix A has rank r, prove that A can be
written as the sum of r matrices, each of which has
rank 1. [Hint: Find a way to use Exercise 62.]
64. Prove that, for m X n matrices A and B, rank (A + B) ::::::
rank(A) + rank(B).
65. Let A be an n X n matrix such that A 2 = 0. Prove that
rank(A) :::::: n/2. [Hint: Show that col(A) � null(A) and
use the Rank Theorem.]
66. Let A be a skew-symmetric n X n matrix.
(See page 1 6 2 ) .
( a) Prove that xT Ax = 0 for all x in !R n .
(b) Prove that I + A is invertible. [Hint: Show that
null(I + A) = {O}.]
om Z,
0
I n t ro d u ct i o n to l i n e a r Tra n s f o r m a t i o n s
In this section, we begin to explore one o f the themes from the introduction t o this
chapter. There we saw that matrices can be used to transform vectors, acting as a type
of "function'' of the form w = T (v) , where the independent variable v and the de­
pendent variable w are vectors. We will make this notion more precise now and look
at several examples of such matrix transformations, leading to the concept of a linear
transformation- a powerful idea that we will encounter repeatedly from here on.
212
Chapter 3 Matrices
We begin by recalling some of the basic concepts associated with functions. You will
be familiar with most of these ideas from other courses in which you encountered func­
tions of the form f: IR ---+ IR [such as f(x) = x 2 ] that transform real numbers into real num­
bers. What is new here is that vectors are involved and we are interested only in functions
that are "compatible" with the vector operations of addition and scalar multiplication.
Consider an example. Let
Then
Thi, ,hows that A tmmfmmn into w
�
[ _ :J
We can describe this transformation more generally. The matrix equation
[ � - � l [; ] [ � l
=
3
2x y
3x + 4y
4
gives a formula that shows how A transforms an arbitrary vector
vectm
[ "_ ]
2x y
3x + 4y
in
D;l'.
[; ]
in IR 2 into the
We denote this tmmfmmotion by T, and w'ite
r,( [x] ) [ 3x2x � l
y
�
+ 4y
Y
(Although technically sloppy, omitting the parentheses in definitions such as this one
is a common convention that saves some writing. The description of TA becomes
with this convention.)
With this example in mind, we now consider some terminology. A transformation
(or mapping or function) T from !R n to !R m is a rule that assigns to each vector v in !R n
a unique vector T (v) in !R m . The domain of T is !R n , and the codomain of T is !R m . We
indicate this by writing T : !R n ---+ !R m . For a vector v in the domain of T, the vector T (v)
in the co domain is called the image of v under (the action of) T. The set of all possible
images T (v) (as v varies throughout the domain of T) is called the range of T.
In our example, the domain of TA is IR 2 and its codomain is IR 3 , so we write
T,,
D;l' --. D;l'.
The image of v
�
[_:J
is w
�
T, ( v)
�
[ _iJ
What is the cange of
Section 3. 6 Introduction to Linear Transformations
213
TA? It consists of all vectors in the codomain IR 3 that are of the form
TA
�
[ -�i
[ ] [ � ] [�] [ �]
x
y
2x y = x
+y 3x + 4y
3
4
=
whi<h d mib" the "t of •ll linm wmbimtiom of the wlumn ve<to"
and
of A. In othe; wo'd', the 'ange of T
IB
[�]
the wlumn 'P"'e of A ! (We
will have more to say about this later-for now we'll simply note it as an interesting
observation.) Geometrically, this shows that the range of TA is the plane through the
origin in IR 3 with direction vectors given by the column vectors of A. Notice that the
range of TA is strictly smaller than the co domain of TA.
linear Tra nsformations
The example TA above is a special case of a more general type of transformation called
a linear transformation. We will consider the general definition in Chapter 6, but the
essence of it is that these are the transformations that "preserve" the vector operations
of addition and scalar multiplication.
D e fi n it i o n
A transformation T : !R n ---+ !R m is called a linear transformation if
1 . T (u + v) = T (u) + T (v) for all u and v in !R n and
2. T(cv) = cT (v) for all v in !R n and all scalars c.
Exa m p l e 3 . 5 5
Consider once again the transformation T : IR 2 ---+ IR 3 defined by
r[;] � [ :: ; � l
[;: J [;: J
Let's check that T is a linear transformation. To verify ( 1 ) , we let
u=
Then
and v =
214
Chapter 3 Matrices
To show (2 ), we let v =
[; ]
T( cv)
=
and let c b e a scalar. Then
r(c [;] ) r( [ :;] )
:x__
c�
[ 32((cx)cx) 4(cy)(cy) ] [ cc((3x2 4y)y) ]
=
=
+
+
=
cT(v)
Thus, T is a linear transformation.
Remark
The definition of a linear transformation can be streamlined by com­
bining ( 1 ) and (2 ) as shown below.
T : !R n � !R m is a linear transformation if
In Exercise 53 , you will be asked to show that the statement above is equivalent
to the original definition. In practice, this equivalent formulation can save some
writing-try it!
Although the linear transformation T in Example 3.55 originally arose as a matrix
transformation TA, it is a simple matter to recover the matrix A from the definition of
T given in the example. We observe that
rn
T � TA, whm A
�
[: � J
-
(N otke that when the vmiahles x ond y "' J;ned
up, the matrix A is just their coefficient matrix.)
Recognizing that a transformation is a matrix transformation is important, since,
as the next theorem shows, all matrix transformations are linear transformations.
Theorem 3 . 3 0
Let A be an m X
n
matrix. Then the matrix transformation TA : !R n � !R m defined by
is a linear transformation.
Section 3. 6 Introduction to Linear Transformations
Proof
215
Let u and v be vectors in !R n and let c be a scalar. Then
and
TA (cv)
=
A(cv)
=
c (Av)
=
cTA (v)
Hence, TA is a linear transformation.
Exa m p l e 3 . 5 6
Let F : IR 2 ---+ IR 2 be the transformation that sends each point to its reflection in the
x-axis. Show that F is a linear transformation.
From Figure 3.4, it is clear that F sends the point (x, y) to the point (x, -y).
Thus, we may write
Solution
y
( 1 , 2)
T
(x, y)
T
I
I
I
�+----+1�--+-��
:I 1--• x
I
I
I
I
We could proceed to check that F is linear, as in Example 3.55 (this one is even easier
to check! ), but it is faster to observe that
I
I
•
•
(x, - y )
( 1 , - 2)
Therefore, F
Reflection in the x-axis
Figure 3 . 4
[; ] [; ]
=
A
, where A =
[ � � ],
_
so F is a matrix transformation. It
now follows, by Theorem 3.30, that F is a linear transformation.
Exa m p l e 3 . 5 1
Let R : IR 2 ---+ IR 2 be the transformation that rotates each point 90° counterclockwise
about the origin. Show that R is a linear transformation.
Solution
we have
As Figure 3.5 shows, R sends the point (x, y) to the point ( - y, x). Thus,
Hence, R is a matrix transformation and is therefore linear.
-y
90° rotation
Figure 3 . 5
A
x
Observe that if we multiply a matrix by standard basis vectors, we obtain the col­
umns of the matrix. For example,
We can use this observation to show that every linear transformation from !R n to
!R m arises as a matrix transformation.
216
Chapter 3 Matrices
Theorem 3 . 3 1
Let T : !R n � !R m b e a linear transformation. Then T is a matrix transformation.
More specifically, T = TA > where A is the m X n matrix
Let e1, e2 ,
, en be the standard basis vectors in !R n and let x be a vector
n
in !R . We can write x = x1e1 + x 2 e2 +· · · + xn e n (where the x/s are the components
of x) . We also know that T(e1), T(e2 ),
, T(e n ) are (column) vectors in !R m . Let A =
[ T(e1) : T(e2 ) : · : T(en )l be the m X n matrix with these vectors as its columns.
Then
Proof
•
•
.
•
•
.
•
•
T ( x) = T(x1 e 1 + X 2 e 2 + . . . + xn en )
= x1 T(e 1 ) + x2 T(e2 ) + · · · + Xn T(en )
[ T(e , ) T(e, ) : · · · T(e,) ]
�
as required.
[}]
�
Ax
The matrix A in Theorem 3.31 is called the standard matrix of the linear trans­
formation T.
Exa m p l e 3 . 5 8
,......
Show that a rotation about the origin through an angle e defines a linear transforma­
tion from IR 2 to IR 2 and find its standard matrix.
Solulion
Let Re be the rotation. We will give a geometric argument to establish
the fact that R0 is linear. Let u and v be vectors in IR 2 • If they are not parallel, then
Figure 3.6(a) shows the parallelogram rule that determines u + v. If we now apply R0 ,
the entire parallelogram is rotated through the angle e, as shown in Figure 3.6(b ). But the
diagonal of this parallelogram must be R0(u) + R0(v), again by the parallelogram rule.
Hence, R0(u + v) = R0(u) + R0(v) . (What happens if u and v are parallel?)
y
y
u+v
r
(a)
-
I
u
I
I
I
Re ( v)'
x
x
(b)
Figure 3 . 6
,......
Similarly, if we apply Re to v and cv, we obtain R0(v) and Re(cv) , as shown
in Figure 3.7. But since the rotation does not affect lengths, we must then have
R8 ( cv) = cR8 ( v) , as required. (Draw diagrams for the cases 0 < c < 1 , - 1 < c < 0,
and c < - 1 .)
Section 3. 6 Introduction to Linear Transformations
211
y
y
'--_,-'
cos 8
CV
Re (cv)
Figure 3 . 1
( 1 , 0)
R 0(e1)
Figure 3 . 8
Therefore, Re is a linear transformation. According to Theorem 3.3 1 , we can find
its matrix by determining its effect on the standard basis vectors e1 and e2 of IFR 2 • Now,
1
c s0
as Figure 3.8 shows, R 0
= � .
0
sm ()
We can find Re
[] [ ]
[�]
[]
similarly, but it is faster to observe that Re
[�]
[] [ ]
must be per-
1
0
- sin ()
=
pendicular (counterclockwise) to R0
and so, by Example 3.57, R8
.
1
cos ()
0
(F 1gure 3 . 9) .
Therefore, the standard matrix of Re is
[
cos ()
. ()
sm
]
- sin ()
.
cos ()
y
R e(ez)
Figure 3 . 9
The result of Example 3.58 can now be used to compute the effect of any rota­
tion. For example, suppose we wish to rotate the point (2, - 1 ) through 60° about the
origin. (The convention is that a positive angle corresponds to a counterclockwise
Chapter 3 Matrices
218
rotation, while a negative angle is clockwise.) Since cos 60° = 1 /2 and sin 60° =
'\/3/2, we compute
y
R60
[] [
cos 60°
2
=
-1
sin 60°
- sin 60°
cos 60°
(2, I )
-
60° rotation
1 /2
'\/3/2
- '\/3/2
1 /2
( 2 + '\/3) /2
( 2 '\/3 - 1 ) /2
]
][ ]
2
-1
=
Exa m p l e 3 . 5 9
(a) Show that the transformation P : IR 2 ---+ IR 2 that projects a point onto the x-axis is
a linear transformation and find its standard matrix.
(b) More generally, if e is a line through the origin in IR 2 , show that the transforma­
tion Pe : IR 2 ---+ IR 2 that projects a point onto e is a linear transformation and find its
standard matrix.
Solution
y
(a) As Figure 3. 1 1 shows, P sends the point (x, y) to the point (x, O). Thus,
It follows that P is a matrix transformation (and hence a linear transformation) with
(x, y )
T
-+------+---+ x
(x,
0)
projection
Figure 3 . 1 1
A
2
-1
Thus, the image of the point (2, - 1 ) under this rotation is the point ((2 + '\/3)/2,
(2'\/3 - 1 )/2) ( 1 .87, 1 .23), as shown in Figure 3.10.
Figure 3 . 1 0
A
][ ] [
[
standard matrix
[� �] .
(b) Let the line e have direction vector d and let v be an arbitrary vector. Then Pe is
given by projd(v), the projection of v onto d, which you'll recall from Section 1 .2 has
the formula
Thus, to show that Pe is linear, we proceed as follows:
Pe( u + v) =
=
=
=
( �
(
(
( )
d · (u +
d·d
v)
)
d
)
)
( )
d·v
d·u
d
d d
d·u
-
d·d
+
d·v
-
d·d
d·u
d +
d·d
d
d•v
d
d·d
= Pe(u)
+
Pe(v)
Similarly, Pe(cv) = cPe(v) for any scalar c (Exercise 52). Hence, Pe is a linear
transformation.
Section 3. 6 Introduction to Linear Transformations
To find the standard matrix of Pe, we apply Theorem 3 .3 1 . If we let d
=
219
[�: ] ,
then
and
Thus, the standard matrix of the projection is
d 1 d2
d�
]
=
[
d U (d f + d D d 1 d2/(d i + d i )
d 1 d2/(d? + d i ) d U (dl + d i )
As a check, note that in part (a) we could take d
x-axis. Therefore, d1
=
1 and d2
=
=
0, and we obtain A
e1
=
]
as a direction vector for the
[� �] ,
4
as before.
New Linear Tra nsformations from Old
If T : !R m � !R n and S : !R n � [RP are linear transformations, then we may follow T by
S to form the composition of the two transformations, denoted S T. Notice that, in
order for S T to make sense, the codomain of T and the domain of S must match
(in this case, they are both !R n ) and the resulting composite transformation S T goes
from the domain of T to the codomain of S (in this case, S T : !R m � [RP ). Figure 3 . 1 2
shows schematically how this composition works. The formal definition o f composi­
tion of transformations is taken directly from this figure and is the same as the cor­
responding definition of composition of ordinary functions:
0
0
0
0
( S o T ) (v)
=
S ( T ( v))
Of course, we would like S T to be a linear transformation too, and happily we
find that it is. We can demonstrate this by showing that S T satisfies the definition of
a linear transformation (which we will do in Chapter 6), but, since for the time being
we are assuming that linear transformations and matrix transformations are the same
thing, it is enough to show that S T is a matrix transformation. We will use the nota­
tion [ T ] for the standard matrix of a linear transformation T.
0
0
0
S
T
v •�•�• S(T(v)) (S T)(v)
T(v)
�m
�n
�P
=
The composition of transformations
Figure 3 . 1 2
0
220
Chapter 3 Matrices
Theorem 3 . 3 2
Let T : !R m ---+ !R n and S : !R n ---+ !RP b e linear transformations. Then S T : !R m ---+ !RP
is a linear transformation. Moreover, their standard matrices are related by
0
[S TJ = [SJ [ TJ
0
Let [S J = A and [ T J = B. (Notice that A is p X n and B is n X m.) If v is a vector
in !R m , then we simply compute
Proof
( S T ) (v) = S ( T ( v)) = S ( Bv) = A ( Bv) = (AB ) v
0
.......
(Notice here that the dimensions of A and B guarantee that the product AB makes
sense.) Thus, we see that the effect of S T is to multiply vectors by AB, from which
it follows immediately that S T is a matrix (hence, linear) transformation with
[S T J = [S J [ T J .
0
0
0
Isn't this a great result? Say it in words: "The matrix of the composite is the prod­
uct of the matrices:' What a lovely formula!
Exa m p l e 3 . 6 0
Consider the linear transformation T : IR 2 ---+ IR 3 from Example 3.55, defined by
r [:: J [ 3x1 + ]
=
x
2x 1 � x2
4X2
and the linear transformation S : IR 3 ---+ IR 4 defined by
Solulion
We see that the standard matrices are
so Theorem 3.32 gives
[S o TJ
[SJ [ TJ
It follows that
( S a T)
[ � -: l
[! � -�l
: -�
i
[� m� � [ ]
+
[:; ] +
[ :J
[ -1 -l l [ '" + �,l
_
[SJ
and
0
3
-1
ITI
-1
6
3x 1 - 7x2
- xi X2
6x 1 3x2
1
3
Section 3. 6 Introduction to Linear Transformations
221
(In Exercise 29, you will be asked to check this result by setting
[;:]
=
y3
r[:: J
=
[
x
2x 1 � x2
3x, + 4x2
]
and substituting these values into the definition of S, thereby calculating ( S T)
dice<tly.)
0
Exa m p l e 3 . 6 1
x,
.:+
Find the standard matrix of the transformation that first rotates a point 90° counter­
clockwise about the origin and then reflects the result in the x-axis.
[ 1 -1O ] .
3.56, respectively, where we found their standard matrices to be [ R ] =
[F] =
[ � �]
The rotation R and the reflection F were discussed in Examples 3.57 and
Solution
0
-
and
It follows that the composition F R has for its matrix
0
[F a R ] = [ F ] [R ] =
ii--""--
[]
[ 1 -1o ] [ o1 1 ] [ -1o 1 ]
-
0
0
-
=
0
(Check that this result is correct by considering the effect of F R on the standard
basis vectors e1 and e 2 . Note the importance of the order of the transformations:
R is performed before F, but we write F R. In this case, R F also makes sense. Is
R o F = F o R?)
0
0
°
4
Inverses of linear Transformations
Consider the effect of a 90° counterclockwise rotation about the origin followed by
a 90° clockwise rotation about the origin. Clearly this leaves every point in IR 2 un­
changed. If we denote these transformations by R90 and R_90 (remember that a nega­
tive angle measure corresponds to clockwise direction), then we may express this
as ( R90 R_90) (v) = v for every v in IR 2 . Note that, in this case, if we perform the
transformations in the other order, we get the same end result: (R_90 R9 0) (v) = v
for every v in IR 2 •
Thus, R9 0 R_9 0 (and R_9 0 R9 0 too) is a linear transformation that leaves every
vector in IR 2 unchanged. Such a transformation is called an identity transformation.
Generally, we have one such transformation for every !R n -namely, I : !R n � !R n such
that I(v) = v for every v in !R n . (If it is important to keep track of the dimension of the
space, we might write In for clarity.)
So, with this notation, we have R90 R_90 = I = R_90 R90. A pair of transforma­
tions that are related to each other in this way are called inverse transformations.
°
°
°
°
°
°
Let S and T be linear transformations from !R n to !R n . Then S and T
are inverse transformations if S T = In and T S = In -
D e fi n i t i o n
0
0
222
Chapter 3 Matrices
Remark Since this definition is symmetric with respect to S and T, we will say
that, when this situation occurs, S is the inverse of T and T is the inverse of S. Further­
more, we will say that S and T are invertible.
�
In terms of matrices, we see immediately that if S and T are inverse transformations, then [S J [ T J = [S T J = [ JJ = I, where the last I is the identity matrix. (Why
is the standard matrix of the identity transformation the identity matrix?) We
must also have [ T ] [S J = [ T S J = [ JJ = I. This shows that [S J and [ T J are inverse
matrices. It shows something more: If a linear transformation T is invertible, then its
standard matrix [ T J must be invertible, and since matrix inverses are unique, this
means that the inverse of T is also unique. Therefore, we can unambiguously use the
notation r - 1 to refer to the inverse of T. Thus, we can rewrite the above equations as
1
1
1
[ T J [ T - J = I = [ T - J [ T J , showing that the matrix of r - is the inverse matrix of [ T J .
We have just proved the following theorem.
0
0
Theorem 3 . 3 3
Let T : !R n ---+ !R n b e an invertible linear transformation. Then its standard matrix
[ T J is an invertible matrix, and
Remark Say this one in words too: "The matrix of the inverse is the inverse of
the matrix:' Fabulous!
Exa m p l e 3 . 6 2
Find the standard matrix of a 60° clockwise rotation about the origin in IR 2 •
Earlier we computed the matrix of a 60° counterclockwise rotation about
the origin to be
- V3/2
1 /2
Solulion
]
Since a 60° clockwise rotation is the inverse of a 60° counterclockwise rotation, we can
apply Theorem 3.33 to obtain
)-1 1 /2 V3/2
1 /2 - V3/2 - 1
[ R - 6o J - [ ( R6o J 1 /2
1 /2
- V3/2
V3/2
[
lt--l'--
Exa m p l e 3 . 6 3
]
] [
(Check the calculation of the matrix inverse. The fastest way is to use the 2 X 2 short­
cut from Theorem 3.8. Also, check that the resulting matrix has the right effect on the
standard basis in IR 2 by drawing a diagram.)
4
Determine whether projection onto the x-axis is an invertible transformation, and if
it is, find its inverse.
Solulion
The standard matrix of this projection P is
[�
since its determinant is 0. Hence, P is not invertible either.
�] ,
which is not invertible
4
Section 3. 6 Introduction to Linear Transformations
T (a,
I
223
Remark Figure 3 . 1 3 gives some idea why P in Example 3.63 is not invertible. The
projection "collapses" IR 2 onto the x-axis. For P to be invertible, we would have to have
a way of "undoing" it, to recover the point ( a, b) we started with. However, there are
infinitely many candidates for the image of ( a, O) under such a hypothetical "inverse:'
Which one should we use? We cannot simply say that P - 1 must send (a, 0) to (a, b),
since this cannot be a definition when we have no way of knowing what b should be.
(See Exercise 42.)
b)
� (a, b ')
I
(a, 0)
I
• (a, b ' ')
Projections are not invertible
Figure 3 . 1 3
Associalivilv
�
Theorem 3.3 (a) in Section 3.2 stated the associativity property for matrix multiplication: A (BC) = (AB) C. (If you didn't try to prove it then, do so now. Even with all
matrices restricted 2 X 2, you will get some feeling for the notational complexity
involved in an "elementwise" proof, which should make you appreciate the proof we
are about to give.)
Our approach to the proof is via linear transformations. We have seen that every
m
n
m X n matrix A gives rise to a linear transformation TA : !R ---+ !R ; conversely, every
m
n
linear transformation T : !R ---+ !R has a corresponding m X n matrix [ T ] . The two
correspondences are inversely related; that is, given A, [ TA] = A, and given T, Tr T J = T.
Let R = TA, S = Tn, and T = Tc. Then, by Theorem 3.32,
A ( BC) = (AB ) C if and only if R ( S T ) = ( R S ) a T
We now prove the latter identity. Let x be in the domain of T [and hence in the do­
main of both R (S T) and (R S) T-why?] . To prove that R (S T) = (R S) T,
it is enough to prove that they have the same effect on x. By repeated application of the
definition of composition, we have
0
�
0
0
0
0
0
0
0
0
0
0
( R a ( S a T ) ) ( x) = R ( ( S T ) ( x))
= R ( S ( T ( x)))
= ( R S ) ( T ( x)) = ( ( R S ) T ) ( x)
0
0
�
I
0
0
as required. (Carefully check how the definition of composition has been used four
times.)
This section has served as an introduction to linear transformations. In Chap­
ter 6, we will take a more detailed and more general look at these transformations.
The exercises that follow also contain some additional explorations of this important
concept.
Exercises 3 . 6
1 . Let TA : IR 2
IR 2 be the matrix transformation corre­
[� -:] .
[ � ] [ - �l
---+
sponding to A =
where u =
and v =
Find TA (u) and TA (v),
2. Let TA : IR 2
---+
[ ]
� !
[�]
IR 3 be the matrix transformation corre-
sponding to A =
TA (v), where u =
3
-1
. Find TA ( u) and
and v =
[ _�J .
224
Chapter 3 Matrices
In Exercises 3-6, prove th at th e given transformation is a
linear transformation, using th e definition (or th e Remark
following Example 3 . 55) .
In Exercises 7- 1 0, give a counterexample to sh ow th at th e
given transformation is not a linear transformation.
I I
7. r
8. r
:2
Y
+
9. r
10. r
x7
y- 1
[; J
[; J
[; J [ J
[; J [ J
[ lxl ]
[x l ]
In Exercises 1 1 - 1 4, find th e standard matrix of th e linear
transformation in th e given exercise.
12. Exercise 4
14. Exercise 6
1 1 . Exercise 3
13. Exercise 5
In Exercises 215-18,2 sh ow th at th e given transformation from IR to IR is linear by sh owing th at it is a matrix
transformation.
15. F reflects a vector in the y-axis.
16. R rotates a vector 45° counterclockwise about the
origin.
17. D stretches a vector by a factor of 2 in the x-component
and a factor of 3 in the y-component.
18. P projects a vector onto the line y = x.
19. The three types of elementary matrices give rise to five
types of 2 X 2 matrices with one of the following forms:
[� �] [� �]
[� �]
[� �] [� �]
In Exercises 20-25, find th e standard
matrix of th e given
2
2
linear transformation from IR to IR .
20. Counterclockwise rotation through 1 20° about the
origin
2 1 . Clockwise rotation through 30° about the origin
22. Projection onto the line y = 2x
23. Projection onto the line y = - x
24. Reflection in the line y = x
25. Reflection in the line y = - x
26. Let e be a line through the origin in IR 2 , Pe the linear
transformation that projects a vector onto e, and Fe the
transformation that reflects a vector in e.
(a) Draw diagrams to show that Fe is linear.
(b) Figure 3.14 suggests a way to find the matrix of Fe,
using the fact that the diagonals of a parallelogram
bisect each other. Prove that Fe (x) = 2Pe (x) - x,
and use this result to show that the standard matrix
of Fe is
(where the direction vector of e is d =
[ �: ] ).
(c) If the angle between e and the positive x-axis is e ,
show that the matrix of Fe is
[
cos W
sin W
sin W
- cos W
]
y
or
or
Each of these elementary matrices corresponds to a linear
transformation from IR 2 to IR 2 • Draw pictures to illustrate
the effect of each one on the unit square with vertices at
(0, O), ( 1 , O), (0, 1), and ( 1 , 1 ) .
Figure 3 . 1 4
In Exercises 27 and 28, apply part (b) or (c) of Exercise 26
to find th e standard matrix of th e transformation.
27. Reflection in the line y = 2x
Section 3. 6 Introduction to Linear Transformations
\/3
28. Reflection in the line y =
x
29. Check the formula for S T in Example 3.60, by
performing the suggested direct substitution.
0
In Exercises 30-35, verify Th eorem 3.32 byfinding th e
matrix of S T (a) by direct substitution and (b) by matrix
multiplication of [S] [T] .
0
225
41. If () is the angle between lines e and m (through the
origin), then Fm Fe = R + w · (See Exercise 26.)
42. (a) If P is a projection, then P P = P.
°
0
(b) The matrix of a projection can never be invertible.
43. If e, m , and n are three lines through the origin, then
Fn Fm Fe is also a reflection in a line through the
origin.
44. Let T be a linear transformation from IR 2 to IR 2 (or
from IR 3 to IR 3 ). Prove that T maps a straight line to a
straight line or a point. [Hint: Use the vector form of
the equation of a line.]
45. Let T be a linear transformation from IR 2 to IR 2 (or
from IR 3 to IR 3 ). Prove that T maps parallel lines to
parallel lines, a single line, a pair of points, or a single
point.
°
°
In Exercises 46-51, let ABCD be th e square with vertices
( - 1 , 1 ) , ( 1 , 1 ) , ( 1 , - 1 ), and ( l, - 1 ) . Use th e results in
Exercises 44 and 45 to find and draw th e image ofABCD
under th e given transformation.
46. T in Exercise 3
-
47. D in Exercise 1 7
48. P in Exercise 1 8
49. The projection in Exercise 2 2
In Exercises 36-39, find th e standard matrix of th e compos­
ite transformation from IR 2 to IR 2 •
36. Counterclockwise rotation through 60°, followed by
reflection in the line y = x
37. Reflection in the y-axis, followed by clockwise rotation
through 30°
38. Clockwise rotation through 45°, followed by projec­
tion onto the y-axis, followed by clockwise rotation
through 45°
39. Reflection in the line y = x, followed by counterclock­
wise rotation through 30°, followed by reflection in the
line y = -x
In Exercises 40-43, use matrices to prove th e given state­
ments about transformations from IR 2 to IR 2 •
40. If Re denotes a rotation (about the origin) through the
angle (), then R" R13 = Ra + f3 ·
0
50. T in Exercise 3 1
5 1 . The transformation in Exercise 37
52. Prove that Pe ( cv) = cPe ( v) for any scalar c
[Example 3.59(b) ] .
53. Prove that T : !R n ---+ !R m is a linear transformation if and
only if
for all v1 , v2 in !R n and scalars c1 , c2 .
54. Prove that (as noted at the beginning of this section)
the range of a linear transformation T : !R n ---+ !R m is the
column space of its matrix [ T ] .
55. If A is an invertible 2 X 2 matrix, what does the
Fundamental Theorem of Invertible Matrices assert
about the corresponding linear transformation TA in
light of Exercise 19?
Vi gnette
Rob o t i c s
I n 1 98 1 , the U.S. Space Shuttle Columbia blasted o ff equipped with a device called the
Shuttle Remote Manipulator System (SRMS). This robotic arm, known as Canadarm,
has proved to be a vital tool in all subsequent space shuttle missions, providing strong,
yet precise and delicate handling of its payloads (see Figure 3 . 1 5).
Canadarm has been used to place satellites into their proper orbit and to retrieve
malfunctioning ones for repair, and it has also performed critical repairs to the shut­
tle itself. Notably, the robotic arm was instrumental in the successful repair of the
Hubble Space Telescope. Since 1 998, Canadarm has played an important role in the
assembly and operation of the International Space Station.
;;;
Canadarm
<(
z
35
<(
z
Figure 3 . 1 5
226
A robotic arm consists of a series of links of fixed length connected at joints where
they can rotate. Each link can therefore rotate in space, or (through the effect of the
other links) be translated parallel to itself, or move by a combination (composition) of
rotations and translations. Before we can design a mathematical model for a robotic
arm, we need to understand how rotations and translations work in composition. To
simplify matters, we will assume that our arm is in IR 2 .
]
[
In Section 3.6, we saw that the matrix of a rotation R about the origin through an
cos e - sin e
angle e is a linear transformation with matrix .
(Figure 3 . 1 6(a)). If
sm e
cos e
, then a translation along v is the transformation
v
=
[:]
T ( x)
x + v or, equivalently, T
=
(Figure 3 . 1 6(b)) .
y
[; ] [; : : ]
=
y
T(x)
=
x+v
x
R(x)
(b) Translation
(a) Rotation
Figure 3 . 1 6
[�]
Unfortunately, translation is not a linear transformation, because T(O) * 0. How­
ever, there is a trick that will get us around this problem. We can represent the vector
x
�
[; ]
"'
th, mtoc
in D;l'. This i& <oll,d "P""'nting x in homogrn,om coor­
dinates. Then the matrix multiplication
[
o ] [x]
[
l
represents the translated vector T (x) in homogeneous coordinates.
We can treat rotations in homogeneous coordinates too. The matrix multiplication
[i
cos e
sin e
0
l[
- sin e
cos e
0
� ;
=
x cos e - y sin e
x sin e y cos e
l� [ �
:
represents the rotated vector R(x) in homogeneous coordinates. The composition T R
that gives the rotation R followed by the translation T is now represented by the product
0 a
1 b
0 1
cos e
sin e
0
- sin e o
cos e
0
=
cos e
si e
0
- sin e
cos e
0
�]
[Note that R o T * T o R. ]
To model a robotic arm, we give each link its own coordinate system (called a
frame) and examine how one link moves in relation to those to which it is directly
connected. To be specific, we let the coordinate axes for the link A; be X; and y;, with
the X;-axis aligned with the link. The length of A; is denoted by a;, and the angle
221
between X; and X; - 1 is denoted by e;. The joint between A; and A; - 1 is at the point
(0, O) relative to A; and (a;_1, O) relative to A;_ 1 . Hence, relative to A;_ 1 , the coordinate
system for A; has been rotated through e i and then translated along
[ a� I ]
(Figure 3 . 1 7) . This transformation is represented in homogeneous coordinates by
the matrix
Figure 3 . 1 1
To give a specific example, consider Figure 3 . 1 8(a). It shows an arm with three
links in which A 1 is in its initial position and each of the other two links has been
rotated 45° from the previous link. We will take the length of each link to be 2 units.
Figure 3 . 1 8(b) shows A 3 in its initial frame. The transformation
T3
=
[
cos 45
sin 45
0
- sin 45
cos 45
0
- 1 / \/2
l / v2
0
causes a rotation of 45° and then a translation by 2 units. As shown in 3 . 1 8(c), this
places A 3 in its appropriate position relative to A 2 's frame. Next, the transformation
T2
228
=
[
cos 45
sin 45
0
]� [
- sin 45 2
cos 45
0
=
:2
l / v2
1/
- 1 / v2
l / v2
0
is applied to the previous result. This places both A 3 and A 2 in their correct posi­
tion relative to A 1 , as shown in Figure 3 . 1 8 (d) . Normally, a third transformation Ti
(a rotation) would be applied to the previous result, but in our case, Ti is the identity
transformation because A 1 stays in its initial position.
Typically, we want to know the coordinates of the end (the "hand") of the robotic
arm, given the length and angle parameters-this is known as forward kinematics.
Following the above sequence of calculations and referring to Figure 3 . 1 8, we see that
Y3
Y1
(b) A 3 in its initial frame
(a) A three-link chain
YI
Y2
(c) T3 puts A3 in Az's initial frame
Figure 3 . 1 8
(d)
T2 T3 puts A3 in A 1's initial frame
we need to determine where the point (2, O) ends up after T3 and T2 are applied. Thus,
the arm's hand is at
- 1 / Vl
1 / V2
0
which represents the point ( 2 + V2, 2 + Vl) in homogeneous coordinates. It is easily
checked from Figure 3 . l S(a) that this is correct.
The methods used in this example generalize to robotic arms in three dimen­
sions, although in IR 3 there are more degrees of freedom and hence more variables.
The method of homogeneous coordinates is also useful in other applications, notably
computer graphics.
229
230
Chapter 3 Matrices
A p p l icati o n s
Markov Chains
A market research team is conducting a controlled survey to determine people's pref­
erences in toothpaste. The sample consists of 200 people, each of whom is asked to
try two brands of toothpaste over a period of several months. Based on the responses
to the survey, the research team compiles the following statistics about toothpaste
preferences.
Of those using Brand A in any month, 70% continue to use it the following month,
while 30% switch to Brand B; of those using Brand B in any month, 80% continue to
use it the following month, while 20% switch to Brand A. These findings are summa­
rized in Figure 3 . 1 9, in which the percentages have been converted into decimals; we
will think of them as probabilities.
�
0.30
Andreia Russian
A. Markovmathematician who
was
studied andoflaterSt. Petersburg.
taught at theHe
University
was
interested
in theory
numberoftheory,
analysis,
and
the
con­
tinued
fractions,
a
recentl
y
devel­
oped
field
that
Markov
applied
towasprobabi
lterested
ity theory.in Markov
also
i
n
poetry,
and
one
of
the
uses
to
whi
c
h
he
put
Markov
chains
was theandanalotherysis
ofliterary
patternstexts.
in poems
0.70
( 1 856- 1 922)
Exa m p l e 3 . 6 4
0.80
0.20
Figure 3 . 1 9
Figure 3.19 is a simple example of a (finite) Markov chain. It represents an evolv­
ing process consisting of a finite number of states. At each step or point in time, the
process may be in any one of the states; at the next step, the process can remain in its
present state or switch to one of the other states. The state to which the process moves
at the next step and the probability of its doing so depend only on the present state
and not on the past history of the process. These probabilities are called transition
probabilities and are assumed to be constants (that is, the probability of moving from
state i to state j is always the same).
In the toothpaste survey described above, there are just two states-using Brand
A and using Brand B-and the transition probabilities are those indicated in
Figure 3.19. Suppose that, when the survey begins, 1 20 people are using Brand A and
80 people are using Brand B. How many people will be using each brand 1 month
later? 2 months later?
The number of Brand A users after 1 month will be 70% of those initially
using Brand A (those who remain loyal to Brand A) plus 20% of the Brand B users
(those who switch from B to A):
Solution
0.70 ( 120 ) + 0.20 ( 80)
=
1 00
Similarly, the number of Brand B users after 1 month will be a combination of those
who switch to Brand B and those who continue to use it:
0.30 ( 1 20 ) + 0.80 ( 80)
=
1 00
Section 3. 7 Applications
231
We can summarize these two equations in a single matrix equation:
[ �:�� �:!� ] [l !� ] [ � �� ]
[ ]
=
[ ]
1 00
120
. (Note
and x 1 =
1 00
80
that the components of each vector are the numbers of Brand A and Brand B users,
Let,s call the matrix P and label the vectors Xo =
in that order, after the number of months indicated by the subscript.) Thus, we have
x1 = Px 0 .
Extending the notation, let xk be the vector whose components record the distri­
bution of toothpaste users after k months. To determine the number of users of each
brand after 2 months have elapsed, we simply apply the same reasoning, starting with
x 1 instead of x 0 . We obtain
0.70 0.20 1 00
90
X2 = Px1 =
0.30 0.80 1 00
1 10
][ ] [ ]
[
from which we see that there are now 90 Brand A users and 1 1 0 Brand B users.
The vectors xk in Example 3.64 are called the state vectors of the Markov chain,
and the matrix P is called its transition matrix. We have just seen that a Markov chain
satisfies the relation
xk + 1 = Pxk for k = 0, 1 , 2, . . .
From this result it follows that we can compute an arbitrary state vector iteratively
once we know x 0 and P. In other words, a Markov chain is completely determined by
its transition probabilities and its initial state.
R e m arks
Suppose, in Example 3.64, we wanted to keep track of not the actual numbers
of toothpaste users but, rather, the relative numbers using each brand. We could con­
vert the data into percentages or fractions by dividing by 200, the total number of
users. Thus, we would start with
0.60
Xo = �s�o� =
0.40
2 00
to reflect the fact that, initially, the Brand A-Brand B split is 60%-40%. Check by
0.50
, which can then be taken as x1 (in agreement
direct calculation that PXo =
0.50
with the 50-50 split we computed above) . Vectors such as these, with nonnegative
components that add up to 1 , are called probability vectors.
•
Observe how the transition probabilities are arranged within the transition
matrix P. We can think of the columns as being labeled with the present states and the
rows as being labeled with the next states:
•
[ ]
Next
[] [ ]
[
Present
B
A
A 0.70 0.20
B 0.30 0.80
]
Chapter 3 Matrices
232
The
fromword
the Greekmeaning
adjectiis "capabl
vderie vede of
aiming"
(or guessi
ng). nItghasthatcome
togoverned
be applied
to
anythi
is
by
the
laws
of
probability
in the sense about
that probabi
lelitiyhood
makesof
predictions
the
l
i
k
things
happening.
Inprocesses"
probabiliform
ty
theory,
"stochastic
a generalization of Markov chains.
stochastic
stokhastikos,
Note also that the columns of P are probability vectors; any square matrix with this
property is called a stochastic matrix.
We can realize the deterministic nature of Markov chains in another way. Note
that we can write
and, in general,
xk = P k x0 for k =
0, 1, 2, ...
This leads us to examine the powers of a transition matrix. In Example
have
[ 0.0.3700 0.0.8200 ] [ 0.0.3700 0.0.8200 ] [ 0.0.4555 0.0.3700 ]
1.
14.)
0.45.
3.20
22 2
0.7(0.3) 0.2 1)0., 3(0.8) 0.24). 1
0.45.
0.45
2
reverse
p2 =
7
�
Y
� B 0.2 1 '
A 0.49
A
A
Y
� B 0.24*
A 0.06
B
figure 3 . 2 0
3.64,
we
=
What are we to make of the entries of this matrix? The first thing to observe is that P 2
is another stochastic matrix, since its columns sum to (You are asked to prove this
in Exercise
Could it be that P 2 is also a transition matrix of some kind? Consider
one of its entries-say, (P 2 ) 2 1 =
The tree diagram in Figure
clarifies where
this entry came from.
There are four possible state changes that can occur over months, and these
correspond to the four branches (or paths) of length in the tree. Someone who
initially is using Brand A can end up using Brand B months later in two different
ways (marked * in the figure) : The person can continue to use A after month and
then switch to B (with probability
=
or the person can switch to B after
month and then stay with B (with probability
=
The sum of these
probabilities gives an overall probability of
Observe that these calculations are
exactly what we do when we compute (P 2 ) n.
It follows that (P 2 ) n =
represents the probability of moving from state
(Brand A) to state (Brand B) in two transitions. (Note that the order of the sub­
scripts is the
of what you might have guessed.) The argument can be general­
ized to show that
1
1
k
(P ) ij is the probability of moving from state j to state i in k transitions.
3.64,
In Example
what will happen to the distribution of toothpaste users in the
long run? Let's work with probability vectors as state vectors. Continuing our calcula­
tions (rounding to three decimal places), we find
x0 X3 =
X6 =
[ 0.0.3700 0.0.8200 ] [ 0.0.5500 ]
[ 0.0.4600 ] [ 0.0.5500 ]
[ 0.0.3700 0.0.8200 ] [ 0.0.4555 ] [ 0.0.457525 ] [ 0.0.458812 ]
[ 0.0.459703 ] [ 0.0.450298 ] [ 0.0.459901 ] [ 0.0.460000 ]
'
x1 -
'
x2 - Px 1 -
Px2 =
=
' X?
=
' Xg
=
'
x4 =
' X9
=
=
[0.0.4555 ],
[0.0.459406 ] '
[ 0.0.460000 ]
' x5 =
' X io
=
Section 3. 7 Applications
233
40%
60%
[ 0.0.3700 0.0.8200 ] [ 0.0.46] [0.0.46 ]
and so on. It appears that the state vectors approach (or converge to ) the vector
[ 0.0.46 ]
,
implying that eventually
of the toothpaste users in the survey will be using
Brand A and
will be using Brand B. Indeed, it is easy to check that, once this
distribution is reached, it will never change. We simply compute
=
4,
A state vector x with the property that Px = x is called a steady state vector. In
Chapter we will prove that every Markov chain has a unique steady state vector. For
now, let's accept this as a fact and see how we can find such a vector without doing any
iterations at all.
We begin by rewriting the matrix equation Px = x as Px = Ix, which can in turn be
rewritten as (I - P)x = 0. Now this is just a homogeneous system of linear equations
with coefficient matrix I - P, so the augmented matrix is [I - P I OJ . In Example
we have
[I
-
I
p OJ =
which reduces to
3.64,
[ 1 --0.0.3070 1 --0.0.2080 I 0 J - [ - 0.0.3300 - 0.0.2200 l 0o ]
O
So, if our steady state vector is x =
solution is
1
[ :: ]
, then x2 is a free variable and the parametric
If we require x to be a probability vector, then we must have
Therefore, x2 =
t = 53 =
0.6
= X 1 + Xz = t t + t = � t
and x 1 =
0.4, [ 0.0.46]actual
200,
2=
5
so x =
iterative calculations above. (If we require x to contain the
in this example we must have x1 +
Exa m p l e 3 . 6 5
x2 =
, in agreement with our
[ 12080 ] .)
3.2 1.
distribution, then
from which it follows that x =
A psychologist places a rat in a cage with three compartments, as shown in Figure
The rat has been trained to select a door at random whenever a bell is rung and to
move through it into the next compartment.
2
1,
(a) If the rat is initially in compartment what is the probability that it will be in
compartment after the bell has rung twice? three times?
(b) In the long run, what proportion of its time will the rat spend in each compartment?
[p;j J be the transition matrix for this Markov chain. Then
P21 = P31 = 2I , P12 = Pn = 3I , P32 = p23 = 32 , and P11 = P22 = p 33 = O
Solution
Let P =
234
Chapter 3 Matrices
Figure 3 . 2 1
�
(Why? Remember that pij is the probability of moving from j to i.) Therefore,
and the initial state vector is
(a) After one ring of the bell, we have
Continuing (rounding to three decimal places), we find
x2 = Px1 =
and
x3 = Px2 =
[
[ l m11 [!]
l
�l
[ l ! ] [!] [ [
l
3
0
=
2
3
I
3
l
2
3
18
0
=
0333 0.333
0.333
0 222
0.389
0.389
Therefore, after two rings, the probability that the rat is in compartment 2 is !
0.333, and after three rings, the probability that the rat is in compartment 2 is
fs 0.389. [Note that these questions could also be answered by computing ( P 2 )n
and ( P 3 ) 2 1 .]
=
=
Section 3. 7 Applications
-
235
(b) This question is asking for the steady state vector x as a probability vector. As we
saw above, x must be in the null space of I P, so we proceed to solve the system
[
J
x, x,
-i 0
_l O
Hence, if x
�
[ :: l x,
x1 x2 x3
then
�
ability vector, we need 1 =
+
l
1 0
idm and
+
-----*
�
l t,
1 0
0
0 0
�
l.
-� O
-1 0
0 0
J
Since x mu•t be ' prnb·
= � t. Thus, t = i and
which tells us that, in the long run, the rat spends � of its time in compartment 1 and
i of its time in each of the other two compartments.
linear Economic Models
We now revisit the economic models that we first encountered in Section 2.4 and
recast these models in terms of matrices. Example 2.33 illustrated the Leontief closed
model. The system of equations we needed to solve was
] [xx21 ]
In matrix form, this is the equation Ex = x, where
E=
[ � �: � � !
1 /2
1 /3
1 /2
1 /4 and x =
1 /4
X3
The matrix E is called an exchange matrix and the vector x is called a price vector.
In general, if E [ e;j ] , then e;j represents the fraction (or percentage) of industry j's
output that is consumed by industry i and X; is the price charged by industry i for its
output.
In a closed economy, the sum of each column of E is 1 . Since the entries of E are
also nonnegative, E is a stochastic matrix and the problem of finding a solution to the
equation
=
Ex = x
(1)
is precisely the same as the problem of finding the steady state vector of a Markov
chain! Thus, to find a price vector x that satisfies Ex = x, we solve the equivalent
homogeneous equation ( I E ) x = 0. There will always be infinitely many solu­
tions; we seek a solution where the prices are all nonnegative and at least one price
is positive.
-
236
Chapter 3 Matrices
2.34,
1010
30
The Leontief open model is more interesting. In Example
system
xXz=1 = 0.0.42xx1 0.0.25Xx2 O0..2lX3x3
1
2
x3= O. lx1 0.3x2 0.3x3
x = Cx
(2)
- C) x =
0.
1
x
10
0.
2
0.
5
1
= [ 0.0.41 0.0.23 0.0.23 ] = [ X3Xz ] = [ 3100 ]
C
C= x
X;
d;
xx
+
+
+
+
+
+
+
+
+
In matrix form, we have
+d
where
we needed to solve the
or ( I
, X
c
d
, d
The matrix is called the consumption matrix, x is the production vector, and d is
the demand vector. In general, if
[c;j ] ,
[x;J , and d [d;J , then c;j represents
the dollar value of industry i's output that is needed to produce one dollar's worth of
industryj's output, is the dollar value (price) of industry i's output, and is the dol­
lar value of the external demand for industry i 's output. Once again, we are interested
in finding a production vector with nonnegative entries such that at least one entry
is positive. We call such a vector a feasible solution.
=
Exa m p l e 3 . 6 6
=
Determine whether there is a solution to the Leontief open model determined by the
following consumption matrices:
C = [ 1/1/24 1/1/33 ]
C = [�
- C) x
C= [ � �� 2/1/23 ]
J1 [ 1/1/24 1/1/33 ] [ - 31/2/4 -1/2/33 ]
[ - 31/2/4 - 2/31/3 ] [Xx21 ] = [ddz1 ]
3.7.
-C
[XX21 ] = [ - 31/2/4 - 2/1/33 ] - 1 [dd21 ] - [ 3/22 9/l 4 ] [dd:]
d 1 , d2 ,
x1 x2•
- C) - 1
(a)
Solulion
(b)
(a) We have
I-
so the equation ( I
O
_
= d becomes
In practice, we would row reduce the corresponding augmented matrix to determine
a solution. However, in this case, it is instructive to notice that the coefficient matrix
I
is invertible and then to apply Theorem
We compute
Since
and all entries of ( I
are nonnegative, so are
find a feasible solution for any nonzero demand vector.
and
Thus, we can
Section 3.7 Applications
(b) In this case,
I-
C
=
so that
[
1 /2
- 1 /2
x
- 1 /2
2/3
= (I -
]
(I - c) - l =
and
C) - 1d
[ -- 46
=
-- 66 ]d
[ -- 46
231
-- 66 ]
Since all entries of (I - C) - 1 are negative, this will not produce a feasible solution for
any nonzero demand vector d.
4
Motivated by Example 3.66, we have the following definition. (For two m X n
matrices A = [a ij ] and B [b;j ] , we will write A 2 B if a iJ 2 b;Jor all i and j. Similarly,
we may define A > B, A :::::: B, and so on. A matrix A is called nonnegative if A 2 0
and positive if A > 0.)
=
Definition
and (I - c) - l
A consumption matrix C is called productive if I - C is invertible
2 0.
We now give three results that give criteria for a consumption matrix to be
productive.
Theorem 3 . 3 4
Let C be a consumption matrix. Then C is productive if and only if there exists a
production vector x 2 0 such that x > Cx.
Assume that C is productive. Then I - C is invertible and (I - C) - 1
Proof
2 0. Let
Then x = (I - c) - 1j 2 0 and (I - C)x = j > 0. Thus, x - Cx > 0 or, equiva­
lently, x > Cx.
Conversely, assume that there exists a vector x 2 0 such that x > Cx. Since
C 2 0 and C * 0, we have x > 0 by Exercise 35. Furthermore, there must exist a
real number ,\ with 0 < ,\ < 1 such that Cx < ,\x. But then
C 2x
�
=
C( Cx) :::::: C(,\x)
=
,\( Cx) < ,\ (,\x)
=
,\ 2 x
By induction, it can be shown that 0 :::::: C "x < ,\ " x for all n 2 0. (Write out the de­
tails of this induction proof.) Since 0 < ,\ < 1, ,\ " approaches 0 as n gets large. There­
fore, as n ---+ oo , ,\ " x ---+ 0 and hence C"x ---+ 0. Since x > 0, we must have C " ---+ 0 as
n ---+
oo .
Now consider the matrix equation
(I -
C) ( I
+ C + C2 +
·
·
·
+ C" - 1) = I - C "
238
Chapter 3 Matrices
As n �
oo ,
en
�
0 , s o we have
(I - C) (I + C + C2 + . . . )
=
I- 0
=
I
Therefore, I - C is invertible, with its inverse given by the infinite matrix series
I + C + C2 + . . . . Since all the terms in this series are nonnegative, we also have
(I - c) - 1
=
r
+ c + c2 +
. . .
:::: o
Hence, C is productive.
R e m a rks
The infinite series I + C + C2 + . . . is the matrix analogue of the geomet­
ric series 1 + x + x2 + . . . . You may be familiar with the fact that, for l x l < 1 ,
1 + x + x2 + . . . = 1 / ( 1 - x) .
•
Since the vector Cx represents the amounts consumed by each industry, the in­
equality x > Cx means that there is some level of production for which each industry
is producing more than it consumes.
•
For an alternative approach to the first part of the proof of Theorem 3.34, see
Exercise 42 in Section 4.6.
•
corollarv 3 . 3 5
TheLatin
wordword comes from
therefers
which
to
a
garland
given
as
a
re­tle
ward.
Thus,
a
corollary
is
a
l
i
t
extra reward that follows from a
theorem.
corollary
corollarium,
Corollarv 3 . 3 6
Let C be a consumption matrix. If the sum of each row of C is less than 1, then
C is productive.
Proof
If
then Cx is a vector consisting of the row sums of C. If each row sum of C is less than
1 , then the condition x > Cx is satisfied. Hence, C is productive.
Let C be a consumption matrix. If the sum of each column of C is less than 1 , then
C is productive.
Proof
If each column sum of C is less than 1, then each row sum of cT is less than 1 .
Hence, CT is productive, by Corollary 3.35. Therefore, by Theorems 3.9(d) and 3.4,
It follows that (I - c) - l 2: 0 too and, thus, c is productive.
You are asked to give alternative proofs of Corollaries 3.35 and 3.36 in Exercise 52 of
Section 7.2.
It follows from the definition of a consumption matrix that the sum of column
j is the total dollar value of all the inputs needed to produce one dollar's worth of
industry j 's output-that is, industry j 's income exceeds its expenditures. We say that
such an industry is profitable. Corollary 3.36 can therefore be rephrased to state that
a consumption matrix is productive if all industries are profitable.
Section 3. 7 Applications
P.Matrices
H. Leslie,in Certain
"On thePopulation
Use of
Mathematics;'
(1945), pp. 183-212.
Biometrika
33
Exa m p l e 3 . 6 1
Population G rowlh
239
1945.
One of the most popular models of population growth is a matrix-based model, first
introduced by P. H. Leslie in
The Leslie model describes the growth of the fe­
male portion of a population, which is assumed to have a maximum lifespan. The
females are divided into age classes, all of which span an equal number of years. Using
data about the average birthrates and survival probabilities of each class, the model is
then able to determine the growth of the population over time.
1
3 (0-1
( 1 -2
(2-3
A certain species of German beetle, the Vollmar-Wasserman beetle (or VW beetle,
for short), lives for at most years. We divide the female VW beetles into three age
classes of year each: youths
year), juveniles
years), and adults
years).
The youths do not lay eggs; each juvenile produces an average of four female beetles;
and each adult produces an average of three females.
The survival rate for youths is
(that is, the probability of a youth's surviving to
become a juvenile is
and the survival rate for juveniles is
Suppose we begin
with a population of
female VW beetles: youths, juveniles, and adults.
Predict the beetle population for each of the next years.
Solution
that year:
0.1005), 50% 40 40 25%. 20
5
1
40 4 20 3 220
40 0.5 20
40 0.25 10
4
[ � s 00.25
2
;n,
[
[ :�
1
0, 1 , 2,
0, 1 , 2,
110
220
3
4
[ � s 00.25 00 ] [ 2010 ] [ 1105 ]
4s 0 03 110110 45555 l
[ � 0.25 0 ][ 5 ] [ 27.5
After year, the number of youths will be the number produced during
+
x
=
The number of juveniles will simply be the number of youths that have survived:
x
=
Likewise, the number of adults will be the number of juveniles that have survived:
x
=
We can combine these into a single matrix equation
x
° ' Lx o
�
x,, whm x,
the initilli populotion di'tcibution vcrtor nnd x,
�
�
-
is the distribution after year. We see that the structure of the equation is exactly the
same as for Markov chains: xk + i = Lxk for k =
. . . (although the interpretation
is quite different) . It follows that we can iteratively compute successive population
distribution vectors. (It also follows that xk = L k x0 for k =
. . . , as for Markov
chains, but we will not use this fact here.)
We compute
x,
�
L x,
x,
�
Lx,
�
240
Chapter 3 Matrices
40 3 l [ 455 l [ 227.302.55 l
0.4 25 �3 302.��.55 13.951.752
00.25 00 ][ 227.13.755 l [ 151.56.828 l
Ther
e
for
e
,
t
h
e
model
pr
e
di
c
t
s
t
h
at
af
t
e
r
year
s
t
h
er
e
wi
l
be
appr
o
xi
m
at
e
l
y
5
You exampl
could are,g28ue
e VWrobeetundedles,to151thjeuveninearleesst, iandnteger57 adulat eachts. (Note:
t951adulhat young
wets aftsehroulfestmedpalhave3-whi
s
t
e
p-for
ch would have affected the subsequent iterations. We elected
not to do this, since the calculations are only approxim ations anyway and it is much
easier to use a calculator or if you do not round as you go.)
Theatiomatn wirixth ninageExampl
es 3.of6equal
7 is caldurledataion, wil be anInn genern matal, irfixwewihaveth thae
popul
cl
a
s
e
following structure: b b b b b
, 2 3
n-1 n
s, 0 0
0
0
00 S02 S03 00 00
0 0 0 Sn - 1 0
Here, b1by, beach2, fearmale teheinbirtclahs parameters
(b; the average numbers of females pro­
duced
i ) and s1 , s2 , are the survival probabilities (s; t h e
probabiWhatlityartehweat atofemalmakee inofclouras cali sucrulviavtesionsin?toOverclasali , th1)e beet. le population appears
t225o befrionmcreyearasing,1 talotyearhough2. tFihgerure ear3.e 2s2omeshowsfluctthueatchange
ions, suichn tasheapopul
decreaastieonfroinmeach250 tofo
the three age clas es and clearly shows the growth, with fluctuations.
=
CAS
L
Leslie matrix.
L
X
L=
•
•
=
•
•
•
=
•
+
4000
3000
·�
:;
§' 2000
i::
0
Cl.,
1 000
I
Adults
o l..:::::+:::::
::
:i:::
:: ___,_�::;..._.
:::
"""'*���=::::::i:
: ::::
::: ::::=:::+:+
_.
2
4
8
6
10
0
Figure 3 . 2 2
Time (in years)
Section 3. 7 Applications
241
0.9
0.8
0.7
-�
c
0
:;
0..
0
0..
4-<
0
...
c
0)
(.)
....
0)
0...
0.6
0.5
0.4
0.3
0.2
I
Juveniles
I
Adults
0. 1
Time (in years)
5
10
Figure 3 . 2 3
20
15
Ifcl, iansst, eaaddiffoferplentotpatingtetrhneemergespopul
attihoin,s, wewe needplot ttohecompute tpopul
acttiioonn iofn
eachthe popul
.
To
do
h
e
fr
a
as in each year;s. Forthexampl
at is, wee, afneedter 1toyeardiv,idwee eachhavedistribu­
tion vectoartbyionthine eachsum ageof itcls component
-2501-x1 250 [ 2202010 ] [ 0.0.0.800488 ]
sgetts ofayoutgraphhsl,ik8%e tihsejuoneveniinleFis, gandure43%.23,is
adulwhiwhiccthhs.stIehflowswes usplcltohetaratthl88%yisthtyatpeofthtofheeprdatpopuloaporovertaitoiontinmofconse,thweeipopul
each cleasis is approaching
a steady state. turns out that the steady0.st7a2te vectoraitniotnhiisnexampl
[ 0.0.0244 ]
Thatadultsis. ,(iInn totheherlongworrudn,s, 7th2%e popul
of theatpopul
aditiostnriwibutl ebed among
youths,t2h4e%thjruevenie agelescl, aands es4i%n
i
o
n
i
s
the ratio 18: 6 : 1 .) We wil see how to determine this ratio exactly in Chapter 4.
Therrelatieonsarheipmanys amongsituata ifionnsiteinsetwhiofcobjh itectiss.imForporexampl
tant toe,beweablmiegthto model
thdese icnrtiebre­
wi
s
h
t
o
varies,iocommuni
us types cofatineton wlinorksksconnect
(roads iconnect
iintegs,toetwnsc.) ,orairrleilnateiornsoutheipssconnect
igrngoupscit­
n
g
s
a
t
e
l
among
or individuals (friendship relationships in a society, predator-prey relationships in
actual
=
It
Graphs and Digraphs
relative
l
__
=
242
Chapter 3 Matrices
B
A
an ecosinygstseum,chdominetwnoranceks andrelarteiloanstiohnsiphsipins,aandsporittt,uetrncs.).outGraphsthat matareriidceales arlyesuaiustedefulto
model
tool in their stconsudy.ists of a finite set of points (called vertices) and a finite set of
connectif theys twaroe (tnhote endpoi
necessnaritslofy dianstedge.
inct) verFiguretices3.. 2We4 shsayowsthanat
tedges,
wexampl
o verteacheiofcestofharewhiesameadcjhacentgraph
ent ways.relatTheionsgrhiapphss tharateidthente "isfayme"the
iedges.
n the sense that all we caredrawn
about iarn etwthoediadjffearcency
We
can
r
e
cor
d
t
h
e
es
s
e
nt
i
a
l
i
n
for
m
at
i
o
n
about
a
gr
a
ph
i
n
a
mat
r
i
x
and
us
e
mat
r
i
x
questioensrs canabouthandlthe egrthaph.e calThiculsaitsioparns tveryiculaquirlycuskley­.
fulalgebrif thaetogrhelaphsp usareanslawrgere,cersintceaincomput
If is a grdefiaphnedwitbyh n vertices, then its
i s the
n n matrix [or
if thhererweiissean edge between vertices i and j
a { 1 ot
Figure 3.25 shows a graph and its associated adjacency matrix.
A graph
D
c
A B c D
Two representations of the same
graph
Figure 3 . 2 4
The termcomes from the Latin
is theverb
plural)
whichof graphs
means "to(andturn:'
In
theetry),context
geom­
aanvertex
is"turns"
a corner-a
poidif­nt
where
edge
i
n
to
a
ferent edge.
vertex (vertices
vertere,
G
A(G)]
Definition
X
A
..
'1
=
adjacency matrix
0
Vl
V2
A=
V4
V3
graph with adjacency matrix
Figure 3 . 2 5
A
A
1
1
[ tl
0
0 0
sleysmmettherericismataObser
rloopix. atv(Weverhy?thtate)x Noti.thIenicsadjeomealascency
ositthuatatimatoansdir,aiaxgonalgrofaphaentmaygrryapha;;haveofis mornecesise tzerhsaanroilyoneun­a
of vertmaticesr.ixInsosutchhatcasaijeequal
s, it mays themakenumbersensofe edges
to modibetfyweenthe
defiveredgetnicibetestioinwandeenof jth.ae paiadjracency
ofofedgesa paththisatthale lnumber
ows us tofo tedgesravel
frit ocontmWeoneaidefinvers, nandteexa weto anotwilinhrereafegrcontr atophiantuouspato behlwiya. Thetshequence
as a is a 5-pForath.exampl
et,hiatn tthhee
Not
i
c
e
grfirastphofofthFiesgeuriseclosed
3.25, (it beginiss aand3-pendsath, andat thedges
e sametwiverce; taepatx); hsutchhatadoespathnotis calinclleuddea
The
s
e
cond
us
e
s
t
h
e
edge
bet
w
een
and
the same edge more than once is called a path.
Remark
�
A
path
length
k
v 1 v3 v2 v 1
circuit.
v1
k-path.
v4 v 1 v2 v2 v 1 v3
simple
v2
Section 3. 7 Applications
243
x to gievofe ustheinadjforamcencyationmataboutrix
tinheFipatWegurhcanes of3.2usevar5: itohuse power
lengthssofinathgreagrph'aph.s adjConsacencyidermattherisquar
represtehntat? Look at the (2, 3) entry. From the definition of
matWhatrixdomulthteipentlicatrieiosn,ofwe know
The onlys way thtihsatexprmakees iupon tcanhe sruemsulits inonzer
n a nonzero. Buto numberis inonzer
s if at loeasiftandoneonlof tyhief
product
welbot3 (vhliaasveranandteedgex k) .betIarnweournonzer
eenexamplando, whie, thcThushismeans
,happensthertheatforkwitlherbeeais2-andanpatedgeforkh betbetweenw2,eensvero ticandes 2 andas
2
whithe crhemaitel sniusngthentat rtiheers ofe are tcorrwo 2-ectplatyhgis vbete 2-wpeenathvers inticthese 2grandaph.3.) (TheCheckargtument
o see thweat
haveas Exerjusctisgie v72.en can be generalized to yield the following result, whose proof we leave
If is thofe adjk-pacency
number
aths betmatweenrix verof atigrcesaphand then the entry of is equal to the
A2
a2k
a 2 ka k 3
a k3
a 2 k a k3
vk
v3 .
=1
v2
vk
=
(A 2 ) 23 = a 2 , a l3 + a 22 a 23 + a 23 a 33 + a 24a43
= l·l + l·l + l·O + O·O
=
�
A2
A
Exa m p l e 3 . 6 8
i
G,
Ak
(i, j)
j.
How manyWe3-pneedaths arthee ther2)e betentwryeenof andwhichinis Fithgeurdote 3.pr25?oduct of row of and
column 2 of The calculation gives
3 2
6
so there are six 3-paths between vertices and 2, which can be easily checked.
beposmodel
verticesForareexampl
orderede,
bydirectsIonmeemany
td edgesypeapplofmiicgrathteliaotbensionusthteathdatcantoimrepres
eesnta one-diedrectbywiayaongraph,
onroutetshthieneedges.
graphmodel
that model
s
aecostransystpeorm.taAtiographn netwiworthkdiorrectpredated edgesor-priesycalrelleadtioanships in aFiagraph
i
n
g
an
examplAn eeas. y modification to the definition of adjacency matricgesureall3.ows26 sushowsto usane
them with digraphs.
Solution
A.
v1
A3,
(1,
(A 3 ) 1 2
=
v2
1
·1 + ·1 + 1·1 + 0·0
=
1
digraph.
A digraph
Figure 3 . 2 6
A2
Chapter 3 Matrices
244
i s the
matrix If[or is a didefigraphnedwibyth vertices, then its
otif thhererweiissean edge from vertex to vertex j
Thus, the adjacency matrix for the digraph in Figure is
Definition
n
X
n
A
G
A(G)]
adjacency matrix
n
i
3.26
Not(Whensurwoulprisindglity,be?th)e Youadjascency
mathaverixnoofdiaffidiculgrtayphseeiisngnotthatsymmetnowriccontin agenerins thale.
h
oul
d
sist thate algilvedges
pattnumber
hion offlotwhs iofisnidthea.e samek-dipratecthsiobetn. (wSeenee Exervertciicsese , wherThee wenextinexampl
es an applalongica­a
Firovund-e tennirobisnpltoayerurnsament(Djokoviin whic, Federch eacher, Nadal
, plRoddiays everck, yandothSafier pln)ayercompetonce.e iThen a
pl
a
yer
ditexgrjameans
ph in Fithgurat eplayersudefmmareateizdespltahyere rej.su(lAts.diAgdiraphrectiendwhiedgechfrthomereveris texactex tloy verone­
direThectedadjedgeacencybetwmateen reverix fory paither diofgverraphticiesn Fiis gcalurleed a is
Ak
directed
Exa m p l e 3 . 6 9
3.27
D
F
tournament
Figure 3 . 2 1
A
i
i
tournament.)
3.27
s
R
72.)
N
0
0
A= 1
0
0
1 0
0
0 0 1 0
0 0 0 1
0 1 0 0
wher
e thiecalorlyd.erThusof th, eFedervertiecrescor(arndesphence
toherorwowsandandcolcoluumnmns offor exampl
is deteer.mined
alphabetSuppos
onds
t
ayers, basofewid nons fortheeachresulpltsaofyerth. eiObsr matervcehesthat. Onethe
waynumberto dooftwiehiwesnmis wieachgshth tplbeoaryertaonkcounthadtheifistvhjueespltnumber
ng roew;
equivalently, the vector containing all tthheersouwmsuofmstheisentgivreniesbyin tthhee prcororductespondiwher
2
2,
A)
Aj,
j=
1
1
1
1
1
Section 3. 7 Applications
245
In our case, we have 0 1 0
3
01 00 01 1 01 23
00 00 01 00 01 11
which produces the followinFigrrsat:nkinDjg:okovic, Federer (tie)
Second:
Nadal
Thi
r
d
:
Roddi
c
k,
Safi
n
(
t
i
e
)
e thdefe pleatayered sFederwho etire, dheindesthiesrrvaesnkifinrsgtequalplace.lyRoddi
strong?ckDjwoulokovid usc mie tghhte same
arguettyhpeat
sofinarceArgheument
the tiee hewitbeath SafiNadal
n. However
,defeat
Safin ecould d arotghueertsh; atfurhetherhasmtorweo,
"ihendimiregctht" notvictteoorbrtiheeatsakbecaus
,
who
Roddick has only indirect victory (over Safin, who then
defeatSineced Nadal
)
.
grcororup,esptondsheinnotatogrioaonup2-ofpofatinhtdiieinrsettcthherewidie nmaygsrasph,eemsnotsobemorwea canple ausyerusefulewhot.hMore defesquareoverateedof, alanthl teihnadjediotraehcency
ctervis icntma­othrye
tofrixth. Toe matcomput
rix e bothwhiwincshandare giinvdienrebyct wins for each player, we need the row sums
00 01 01 1 01 00 21 1 21
001 000 00 001 00 001 001 00 01 020
0 01 22 22 23 11 87
01 001 011 02 201 11 23
Thus
Unfor, tweunatwoulely,dthriasnkapprtheoachplayeris nots as guarfolloaws:nteDjedotkovio brecak, Federall tieers,.Nadal, Safin, Roddick.
Aj
=
two
one
A + A2,
(A + A 2 )j
+
=
6
I
Exercises 3 . 1
M a rkov C h a i n s
[ 0.0.55 0.0.73 ]
[ 0.0.55 ]
In Exercises 1 -4, let P
be the transition mabe
trix for a Markov chain with two states. Let x0
the initial state vector for the population.
=
=
1.
2.
3.
4.
Computpreoporandtion of the state 1 population wil be in
What
state 2prafoteporr twtioonstofeps?the state 2 population wil be in
What
sFitantde 2thafetsetreadytwossttaetps?e vector.
x1
x2 •
Chapter 3 Matrices
246
P [! ! I]
: � [ :m
Comput
e
and
What
ation wil be in
sWhattate 1prprafooteporporr twttiiooonnstofofeps?tthhee ssttaattee 21 popul
sFitantde t3hafetsetreadytwossttaetps?e vector. population wil be in
Supposeaccor
that tdhiengweatto ahMarer inkaovparchaiticnul. aSpecir regifiocnally,
behaves
sauwetpposdaye thisat0.t6h62e prifotbabiodaylitiys wetthatandtomor0.2r50owifwitoldaybe
idays driys. 0.The750prifotbabiodaylitiys drthaty andtomor0.3r38owifwitoldaybe aisdrwety .
studyR.ofGabrrainfaliell
iand[nThiTels exerAviNeumann,
vcioverse is abas27-eMardyearonkovperian Chaiactod.ualnSeeModel
Rainfall Occurrence at Tel Aviv;' ( 1 962),for Daily
pp. 90-95.
]
Wrchaiitne. down the transition matrix for this Markov
If Mondaydayiswialdrbey day,wet?what is the probability that
Wednes
In thdre lyodaysng rube?n, what wil the distribution of wet
and
ated onetthheatheithgehtprs oofbabichillidtireesn
rtDathelatataiavhaveteatlopartbeenheientr paraccumul
ewilnthaves. Suppos
sabihorlittiechis tlhdataraemedi0.6, 0.u2m-, andheiag0.thta2l,par,rmediesepntectuwim-ivlelhhaveyei; tghhtea,prtoraol b­,
medi
shavepectuaivm-telaylh;, eiandmedightt,huorem-prshhooreibabigthtchil,iortlidessarhtheorat0.t 1chia, s0.hld7or,artandpare 0.0.e2nt,20., wire­4,l
andWr0.4,itreesdownpectivtehley.transition matrix for this Markov
chain.is the probability that a short person wil
What
havef 20%a tofal thgreacurndchirentldpopul
? ation is tal , 50% is of
Imedi
is shaortiotns, what? wil the
diWhatstriubpropor
utmiheiongbethtio,innandofthrt30%eheegener
popul
a
t
i
o
n
wi
l
be
t
a
l
,
of
medium height, and short in the long run?
In Emdm 5-8, let
�
11.
be the tmn;;t;on ma-
i,;xfor n Ma,kov cha;n wUh th " al" Let x,,
�
H.
be
6.
7.
8.
9.
x1
J.
x2 .
''A
(a)
(b)
K.
Quarterly Journal of
the Royal Meteorological Society, 88
''A
Models in Archaeology
the initial state vector for the population.
5.
Asoutstuhdywesoft pifrofmwn1940(pine)tonut1947cropshypotinhtesheizAmer
ican
ed
t
h
at
nutD. prThomas
oduction, folComput
lowed a eMarr Sikmovulachaitionn.Model
[See of
etaternBass;' iinn ShosD. L.hCloneanarke,Subsed., istence and Settlement
PatGr(London:
.] tThehendattheaprsuoggesbabitleidtieths atthat
itfhonee folyearlowiMet'ns gcryearhouen,p was's cr1972)good,
wercrope was0.08,fa0.ir0, 7,thanden th0.eo8prp5,woul
robabiespdectlitbeiievsgood,
eltyh;atiftonefaheirfol,yearorlopoorw-'s
0.itnh1gen1,yeartandhe'spr0.cro8obabi0,p woulrelsiptiecteds betihvatelgood,yt;hief onefolfailro,yearwiornpoor'gs yearcropwer'swascreo0.ppoor09, ,
woul
respWrectd beiivteegood,
lydown. faithre, ortrapoornsitiowern mate 0.ri1x1,fo0.r0t5,hisandMar0.k8ov4,
chaiIf thne .pifion nut crop was good in 1940, find the
probabi
crop in the years 1941
tIhnrtohughe lolngi1945.tiesruofn,awhatgoodpropor
tion of the crops wil
be
good,
fai
r
,
and
poor?
Robot
e mazey
schoos
hownesiwhihaven Ficghbeenurwaye 3.pr2to8ogo.andgrammed
at eachtojutrnctaversionertahndoml
(c)
12.
(a)
(b)
( c)
10.
Figure 3 . 2 8
(b)
(b)
( c)
(d)
Cons
he trasnsthitiisosnitmatuatiroixn.for the Markov
chaiSupposntrthueatctwetmodel
t(iAosn.uFimendththatsetiasttrettaadywikesthseacht1a5terobotdirosbottrisbatuttheacheiosnameofjurnc-amount
obots.
of
t
i
m
e
t
o
t
r
a
vel
bet
w
een
t
w
o
adj
a
cent
j
u
nct
i
o
ns.
)
e a roiwvevectmatorrixconsPisisatisntogchasentitrieclymatof rlis.x Prif ove
tandLethatjonladenotnonnegat
y if j P = j .
(a)
(a)
13.
Section 3. 7 Applications
02
0.
1
0.
4
04
]
Show
t
h
at
t
h
e
product
of
t
w
o
2
2
s
t
o
chas
t
i
c
[
0.
2
[
0
35
5
0.0.41 55 0.0.3505 �0.6350 : 0.0 3 0.0.42 0.0.25 0.0.31
Prmatmatoverriiccteseshatiisstalhalsesoproaaosducttsotochaschasoftittcwicmatomatrirxi.x.stochastic
0.5 0 0.2 0.2
IPf-a1 2is als2osatoschastochastictmatic matrixrPixi. s invertible, prove that
C
x
1/21/2 1/41/2 ] [ 3l ]
[
P
0.31 0.0.42 ] , [ 21 ]
[
0.
In Exerecdisnumber
e if Monda
y iunts a dril ya wetday,day?what is the
5 0.0.42 010.2 ] '2 ]
expect
of
days
[
[
�
ItinonsExeruntciislea10,shorwhatt perissothnehasexpecta tael ddesnumber
of
gener
a
­
0
0.
5
4
c
endant
?
IisntExerhe expectcise 11,ed inumber
fthe pifiofonyearnutscruntopiilsafagoodir onecryearop occur, whats? C [ 010 0.0.42 010.2 ] [ 3.u5 ]
ItinonsExer, whatcise i12,s thsetaexpect
rting fredomnumber
each ofofthmoves
e otheruntjunc­il a Let A be0.3an 0.2 mat0.3 rix, A 2 2.0 Suppose that
robot reaches junction 4?
Ax<
x
fo
r
s
o
me
x
i
n
x
2 Prove that x
and ixtiandes:
vectLet A,IofrAs in2C, and2ProvebeandtheCfol2lomatwin2rgicinesequal
ACIf A 2 and2 x 2 x thentAxhen
[ 11/22 3/41/4 ]
[ 1/21/3 2/31 2 ]
[ 0.0.46 0.0.47]
[ 0.0. 1 0.0.46]
A
popul
a
t
i
o
n
wi
t
h
t
w
o
age
cl
a
s
e
s
has
a
Les
l
i
e
mat
r
i
x
2
[ 1/31/3 3/20
1/2
1
�] . If the initial population vector is
[
�
0.
6
[
0
0
1
3
L
]
�
1/2
1/3
[ 030.3 0.0 5 0.023 ] 0500.1/225 00.0.37002/30350.l 25 ] x0 [ 1 � ], compute x1, x2, and x3.
0.4 0.5 0.5 [ 0.25 0 0.40 A population with three age clas es has a Leslie matrix
L � H 0.05 1 � l If the m;t;ru populahon
00.2300 0.0. 1105 0100.45 vector is x0 [ : . compute x,. x,, ond x,.
[ 0.0.25 0.0.63 ]
[ 0. 1 5 0.30 0.50 ]
]
X
14. (a)
X
(c)
In Exercises 31 -34, a consumption matrix and a demand
vector d are given. In each case, find a feasible production
vector that satisfies Equation (2).
Suppose we want to know the average (or expected) number
of steps it will take to go from state i to state j in a Markov
chain. It can be shown that the following computation
answers this question: Delete the jth row and the jth column
of the transition matrix to get a new matrix Q. (Keep
the rows and columns of Q labeled as they were in P.) The
expected number of steps from state i to state j is given by
the sum of the entries in the column of (I Q) - 1 labeled i.
15.
9,
17.
18.
30.
29.
nXn
(b)
16.
CAS
31. c =
,d =
32. c =
d=
33. c =
,d =
34.
,d =
=
nXn
35.
36.
Linear E c o n o m i c M o d e l s
In Exercises 1 9-26, determine which of the matrices are
exchange matrices. For those that are exchange matrices,
find a nonnegative price vector that satisfies Equation (1).
19.
21.
23.
I
20.
22.
I
9
B,
D
!R n .
ll�r,
B 0
D 0,
BD 0.
>B
0, * 0,
(a)
(b)
P o p u l at i o n G r owth
37.
24.
26.
38.
1
·
In Exercises 27-30, determine whether the given consumption matrix is productive.
27.
28.
0.
0.
=
> 0.
y
nXn
=
25.
241
> Bx.
Chapter 3 Matrices
248
39.
A popul1 ation1 with3three age clas es has a Leslie matrix
[ 00. 00.5 00 l . If the initial population vector is
[ 100100 ] , compute and
100
A population with four age clas es has a Leslie matrix
-_[ �.5 0.0� 7 0.0 3 00 . Ifthe initial population
� � �i
voctodno rn l rnmpute Md
nasspuecirvievsalwiprthobabitwolageity ofcla80%s es frofom1 yearclas's dur1 toa­
tAclioacerns has2.taiEmpi
eachtwo posfemalsibleegiLesrivcesalliebievimatrtdhencertiocesfivshareowsfee maltheats ,peronyearaver.aThusge, ,
[ 0.0 8 05] and [ 40. 8 01 ]
Steacharticasng ewi. th [ � � l compute , in
of eachdoageyour
clgrForaasphseachoversucasggestime,etpl?(aost itnheFirgeurlateiv3.e2s3)ize. What
Suppose the2Leslie matrix for the VW beetle is
[ �.! L � l sta,ting with acb;tmy Xo, detec­
miSupposne theetbehavi
o
r
of
t
h
i
s
popul
a
t
i
o
n.
he Leslie matrix for the VW beetle is
[ 0� �0.5 20� l Investigate the effect of varying
tWoodl
he suravndivalcarpriobbabiou arelityfoundof thpre young
beet
l
e
s
.
imaricianly inorn ththeweswestt.ern
prTheoviaverncesageoflCanada
and
t
h
e
Amer
ThegivenbiirnthTablandeifs3.esu4rpv,anwhiivalofcrhaatfseehsmalowsforeeachtihsataboutcarageibbr1ou4ayearcketcowssar. doe
not give birth at all during their first 2 years and give
L
=
x0
40.
7
=
x1, x2 ,
biyearrths.tTheo aboutmoronetalitcaly rfatpere foryearyoungduricalngvtesheiisr vermidydlhiegh.
x3 .
L
x,. x,.
�
41.
LI
L2
=
(a)
x0
x,
=
x1,
=
.
•
.
x1 0
(b)
42.
L
·
43.
L=
cAs 44.
AA
·
s
=
Ta b l e 3 . 4
Age
(years)
Birth
Rate
Survival
Rate
2-44-60-2 0.0.1.408 0.0.0.93
6-88-10 1.1.88 0.0.99
10-12
12-14 0.1.66 0.0.60
JTablaspTheere 3.Nat5number
i. UsonalingParsaofCAS,kwoodl
in Alprbeadierndcttacartihnei1990bcarouibrareoupore popul
shteownd inatiinon
ctyoutheconclpopulude?atio(nWforhatthe
yearforassu1992mpts 2010iandonsanddoes1994.2020.thThenisWhatmodelprodojemake,
and how could it
be improved?)
7
Ta b l e 3 . 5
Age
(years)
0-22-4
4-66-8
8-10
10-12
12-14
Woodland Caribou
Population in Jasper
National Park. 1 9 9 0
102
85
120
Number
Source: World Wildlife Fund Canada
Section 3. 7 Applications
Graphs and Digraphs
In Exercises 45-48, determine the adjacency matrix of the
given graph.
45
VI
In Exercises 53-56, determine the adjacency matrix of the
given digraph.
V2
.--------.
V4
V3
V4
46.
249
V1
54.
V3
v,
V2
V2
V4
47.
55.
V4
V5
56. V J
V2
V4
V3
In Exercises 49-52, draw a graph that has the given adjacency matrix.
49.
51.
001 001
[ 0� 00 01 �]1 0
01 00 00 01 11
10 0 00 00
50.
52.
0
[ 0; 0 00 ll1 1
00 00 00 1 1
11 00 00
In Exercises 57-60, draw a digraph that has the given adjacency matrix.
57.
10 00
�[ 01 0 ; ]
58.
10 00
0
0
[: 0 1 �:
Chapter 3 Matrices
250
0 1 0 1
0 0 1 0
0 0 0 1
0 1 0 0
0 1 0
0
1
59. 0
1
0
60.
sadjweracency
the folmatlowirinxg quesfor tthioisns.digraph and use it to an­
A
0 1 0 0 1
0 0 0
0
0 0
0 1 0 0
1 0 0 0
Rodent
In Exercises 61-68, use powers of adjacency matrices to
determine the number ofpaths of the specified length
between the given vertices.
61.
50,
2, v 1
v2
62.
2, v 1
v2
63.
50,
3, v 1
v3
64.
52,
4, v2
v2
65.
57,
2, v 1 v3
66.
57,
3, v4 v 1
67.
60,
3, v4 v 1
68.
60,
4, v 1 v4
69. A
(a)
i A
ExerExercciissee 52, lleengtngthh andand
ExerExercciissee lleengtngthh andand
Fish
Bird
ExerExercciissee lleengtngthh ttoo
WhiHowcdoesh speciesshhasow tthhies?most direct sources of food?
ExerExercciissee lleengtngthh ttoo
Whimostchotshperecispeseciiseas?diHowrect sdoesource ofshfoodow thforis? the
Let Ifberowtheofadjaicency
mat
r
i
x
of
a
gr
a
ph
G.
s all zeros, what does this imply
Iinf direatects sbandourcebeatof food.s weHowsaycanthatwehasuse astoande­
about
G?
IfaboutcoluG?mn of is all zeros, what does this imply
tseormiurcesne?whiWhichchspspeciecieseshashasththe emosmost tindidirreectctandfood
i
n
di
r
e
ct
food
s
o
ur
c
es
combi
n
ed?
Let Ifberowtheofadjacency
mat
r
i
x
of
a
di
g
r
a
ph
D.
Suppos
e
t
h
at
pol
l
u
t
a
nt
s
ki
l
t
h
e
pl
a
nt
s
i
n
t
h
i
s
food
i
s
al
l
zer
o
s
,
what
does
t
h
i
s
i
m
pl
y
web,
and
we
want
t
o
det
e
r
m
i
n
e
t
h
e
eff
e
ct
t
h
i
s
aboutcoluD?mn ofA2 is all zeros, what does this imply change
wi
l
have
on
t
h
e
ecos
y
s
t
e
m.
Cons
t
r
u
ct
a
Ifabout
new
adj
a
cency
mat
r
i
x
fr
o
m
by
del
e
t
i
n
g
t
h
e
D?
r
o
w
and
col
u
mn
cor
r
e
s
p
ondi
n
g
t
o
pl
a
nt
s
.
Repeat
a tourmatnamentrices,wiratnkh stixhe
parthetmoss (a)t tando (c)leandast affdetecteremdibyne twhihe change.
ch species are
plplFiaagyeryeruress,firsttoibys thdeteUsdierignmrgaiphnadjinofgacency
What wimatl rtihxecallong-culatetiromnseffwielctsofhowthethpolis?lution be?
usas iinngExampl
the noteion of combinedwiwinsnonls andy andindithreenctbywins, FiveWhat
peopl
e
ar
e
al
l
connect
e
d
by
em
ai
l
.
Whenever
tithalemonghearbyse-amjuaiiclyinpigeitcetoofsogosmeonesip, heelsore isnhtehe
pasgroneoupseofs accor
d
i
n
g
t
o
Tabl
e
netDrawworthke" diandgrafiphndtihtats adjmodelacencys thmatis "grosixsip
(b)
70.
(a)
A
(b)
(c)
a
A
c,
a
j A
A
(h)
j
3.29
P1
c
A
(d)
i A2
(a)
71.
Figure 3 . 3 0
A*
A
P6.
(e)
3.69.
73.
3.6.
(a)
A.
Ta b l e 3 . 6
amfoodtowebinidin ­
aFicatsgmeursaltehl atecoshasiyssatedim.asgrAaaphdisourrreectprceeedofsedgeentfood.infrg oCons
truct the
Figure 3 . 2 9
72.
3.30
a
BerAnnCartla
Dana
Ehaz
Sender
p6
b
a
b
CarCarllaa,, Ehaz
Dana
Ehaz
Ann,
Car
l
a
Bert
Recipients
Chapter Review
evergosDefisyinponeegeta ons frhioasms torhAnnehertimtoleisbotitt. t(ahTkeshusCara,lpersainandoneonEhaz.sttoep,e-)mIfail
steps wiWhatl it tmatakerix
forcalBercteverulhearsatyioonenarreeluvealmor,se tsothhowhearis? many
the rumor?
mat
r
i
x
i
n
Exer
c
i
s
e
49
forcalIf Annceverulathearyioonen rseaelvealrsue mortsothheari,show? themanyrumor?stepsWhatwil matit tarkeix TheThe adjadjaacency
cency
mat
r
i
x
i
n
Exer
c
i
s
e
52
The
adj
a
cency
mat
r
i
x
i
n
Exer
c
i
s
e
51
Idinggeneral
,
i
f
i
s
t
h
e
adj
a
cency
mat
r
i
x
of
a
ed to 00 00 00 11
ver[Thetreaxgosph,byshowipanetpatcanwhor(wekofinstoetlmehiisf verexerlengttecxhis)e?iiss connect
r
e
mi
n
i
s
c
ent
ofplatyheandnotfiiolmn ofby"tshixatdegrname)ees, ofwhisecparh sautggesion"ts(fothundat anyin the 0 0 01 01 01 01
0
0
0
1
1
twhoswo people lengte arh eisconnect
e
d
by
a
pat
h
of
acquai
n
t
a
nces
at mose frivt ol6.oTheuslygameassert"sStihxatDegrall acteesoofrs are 1 Prove0that1 a 0graph0 is bipartite if and only if its
Keviconnectn Bacon"
mor
e
d
t
o
t
h
e
act
o
r
Kevi
n
Bacon
i
n
s
u
ch
a
wa
y
.
]
vercantibecesparcantitbeionedlabelased so that its adjacency matrix
Let Bybeintducthe adjion,acencyprovematthatrixforofalalgraph1, the
of n-paths
[�--H�-]
betentHowwryeendoof tverheitssitcequal
esatementandto tandhe number
tIf ios tbehemodiadjacency
fied ifmatisriaxdiofgraproof
adiph?graphin partwhat(a) havedoes grUsaiphng hasthe rnoesuciltrcinuiparts oft (odda), prleongtvehth. at a bipartite
the entry of represent if ?
(b)
A graph is called bipartite if its vertices can be subdi­
vided into two sets U and V such that every edge has one
endpoint in U and the other endpoint in V For example,
the graph in Exercise 48 is bipartite with U = { v 1 , v2 , v3 }
and V = {v4 , v5 }. In Exercises 76-79, determine whether a
graph with the given adjacency matrix is bipartite.
step
(c)
(d)
76.
77.
78.
A
i
j
79.
80. ( a)
74.
251
(a)
G.
A
An
(b)
75. A
n 2::
i
(i, j)
(b)
G
(i, j )
A=
j.
AA T
i *j
G,
Chapter Review
basBasiiss,Theor198em, 202
colcoluumnmn matspacerixof(vaector), 138
matrixi,tion195oflinear
compos
coortbasransdisi,nfoatrm208e vectationsor,wit219h respect to a
didiamgonalensiomatn, r203ix, 139
elementary matrix, 170
Kev Defi nitions and concepts
Fundament
al172,Theor206em of Invertible
Mat
r
i
c
es
,
iindverentisteyofmata srqixuar, e139
intmatverranssreifoxof,rma163atlinioearn, 221
combinationiofndependence
matrices, 154
lliinnofearearmatdependence/
linearfacttrorarinsiczesatfo,iromn,157atio181n, 213
LU
matmatrriixx,addi138tion, 140
matmatrriixx mulfactotirpizliatcatioin,on, 180141
matnegatriixvpowers
, r149ix, 140
e
of
a
mat
nulnulllitsypaceof aofmata matrix,rix204, 197
outparteirtipronedoductmat, ri147ces (block
permulmuttipaltiicoatniomatn),rix145,, 187148
Chapter 3 Matrices
252
sttarndaransfodrmatmatrioixn,of a216linear
ssyummet
bspace,ric mat192rix, 151
tzerranso matposeriofx, a mat141 rix, 151
MarFork eachanyofmattherifolx lowibotngh statementand s truearore defifalsne:ed. uctIf posofselibelment
e, exprareys mattherimatces.rix [41 ! ] as a prodIf andthBenareBmatrices such that AB and If is a square matrix such that show that
If A, B,B,andthen are invertible matrices such that
Thetary matinverrisxe. of an elementary matrix is an elemen­ Find an faotmization of [: - l :J
Find bases for the row space, column space, and null
Theelementtransarpyosmate ofrixan. elementary matrix is an
Theelementproductary matof trwixo. elementary matrices is an sp�<ofA [� =� � � n
aEverIfsubsiyspplanaceaneofin matis aritxw, o-thendimthense nulionall spsaceubsofpaceis haveSupposthee smatamericroesw spandace?BarWhye roorwwhyequinotvale?ntDo. Do andthey
BIfhaveis antheinsverametibcolle umatmnrisxp, ace?explaWhyin whyor whyandnot? must
ofThe transformation T: defined by
nulmatl srpixace.? ExplIs tahinis. true if is a nonin­
T(If xT:) - xis aislianlearineartratnsrafonsrfomratmiaton.ion, then verhavetibtlheessame
q
uar
e
whose berowsinveraddtibuple.to the zero
talhlerxeinistahe4 domai5 matn ofrix such that T(x) Ax for vectIf oisr,aexplsquarainewhymatrixcannot
Let
be
an
mat
r
i
x
wi
t
h
l
i
n
ear
l
y
i
n
dependent
col
u
mns
.
Expl
a
i
n
why
mus
t
be
an
i
n
ver
t
i
b
l
e
]
-�
�].
[� � [�
mat
r
i
x
.
Mus
t
al
s
o
be
i
n
ver
t
i
b
l
e
?
Expl
a
i
n
.
Find a linear transformation T: such that
A(B2BBT) - 1 A(B2TBB2) - 1 B -1B
�
]
].
]
J
and
r
r
�
[
[
[
[
_�
�
The outer product expansion of
Fi
n
d
t
h
e
s
t
a
ndar
d
mat
r
i
x
of
t
h
e
l
i
n
ear
t
r
a
ns
fo
r
m
at
i
o
n
2
/
-l
l
If is a matrix such that [ - 3 /2 4 ], find rT:otation of 45°thataboutcorrtehspeondsorigintofola count
ebyrcloackwiprojseec­
l
o
wed
ion onteoththate lTine is a linear transformation
If [ � : = : ] ond X is a matdx such thot tSuppos
andT2 (vs)uppos(ewtherhatevTis2 a vectT orT)su. PrchotvehatthTat(vv) and T(butv)
-[ � - �], findX.
are linearly independent.
3 -2
pr158,oper159,ties of167matrix algebra, 154,
rank ofTheora matem,rix, 205204
Rank
reprproeductsentast, ions146-148
of matrix
rrooww matspacerixof(vaectmatorr)i,x, 138195
ssspccalalanaarrofmulmata setrtiipxofl,e ofmat139a rmaticesr,ix, 156140
square matrix, 139
Review Questions
1.
(a)
(b) A
A * 0,
(c)
XA =
A,
AA T
= 0.
X
A TA
10.
=0
12.
(e)
LU
A
�
13.
(f)
�
mXn
(g) A
(h)
14.
IR 3
=
IR 4 ---+ IR 5
(j)
A
ll�r .
IR 3 .
(i)
X
A
8. A
�
AX =
A
A
A
16. A
=
17.
In Exercises 2-7, let A =
and B =
Compute the indicated matrices, ifpossible.
4. TA
3.
2.
5.
6.
7.
AA T
A-1 =
A
15. A
IR 2 ---+ IR 2
T.
9. A
A 3 = 0,
11. A
(I - A) - 1 = I + A + A 2 •
X = A - 1 B.
(d)
A=
mXn
A
AA T
A
A TA
IR 2 ---+ IR 2
18.
A.
19.
AT
IR 2 ---+ IR 2
y = - 2x.
: !R n ---+ !R n
20.
=O
=
0
*0
E i genva l ues and
E i genvectors
latent,secular,
characteristic,proper,
eiroot,
gen number
value,
Almost every combination of
the adjectives
and
with the nouns
and
has been used in
the literature for what we call a
proper value.
-Paul R. Halmos
Van Nostrand,(2nd1958,edition)
p. 102
Finite Dimensional Vector Spaces
4.0
Introduction : A ovnamical svstem o n Graphs
Weestinsgawreisnultths.eBotlasthchapt
eovr tchaihat intesrandatintghmate Lesrilxiemulmodeltipliofcatpopul
ion ofatteionnprgrooduces
intebri­t
Mar
k
wt
h
exhi
steadystsatndatessuichn cerbehavitain osri.tuFiatrsiot ns.we wiOnel loofokthate anotgoalhserofittehraistivchapte proceser iss,toro help you
under
oblemslititehsattofolfacilolw,itayoute thwie comput
l find iattihelons.pful)
to usOure a CASexamplthorateausicalnevolscmatulvesatorgrircwiaesphs.th(Inmat(seethreSectixprcapabi
iphs anyhasgraverph tiin­
whices,ciht ievers denoty veredtebyx iKws adjForacentexamplto evere, Fiy goturhieoernverstheowsx.AIfaarcompl
e
t
e
gr
a
e
pr
e
sent
a
t
i
o
n
of
K4
•
4
Pick any vects ofx,or x ison IRthatwithisnonnegat
ivethentriandes andso on.labelComput
the veretitchese
ofadjK4acencywith matthe rcomponent
l
a
bel
e
d
wi
iAx.x A TrofyK4thandis forrelsaebelvertahlevectvertoircsesxofandtheexplgraphainwi, inthtethrme cors ofrethspeondigraph,ng
component
s
of
how the new laNowbels can be dettheerprmoinceseds frinomProthbleeolmd laThatbels. is, for a given choice of x,
ces asges.descSirnibceedcomponent
above and tshofentapplhe vecty Aoagairs thnems(andelvagaies win, landgetagaiquinte)
untrlaerlagibele,l awetpathewitverelrntiemer
hem byatdiiovnidriensguleachts in tvecthe vector byor its
after each
iteration. Thus, if a tcomput
CAS
dynami­
cal system,
3.7).
complete graph
n
4. 1
Problem 1
v1
x1,
1.
iterate
Problem 2
scale
VJ
V2
------
we wil replace it by
largest component
.!. [ � ] [ � ]
4 1
Figure 4 . 1
K4
1
·5
0.25
0.25
253
Chapter 4 Eigenvalues and Eigenvectors
254
thisisforprocestshenguarantandees thatUsteheatlalerasgest ttecomponent
ofandeachtwo-vectdeciormwiall-pnowlace
beNotaccure Dotahcy.atthWhat
n
i
t
e
r
a
t
i
o
ns
appearshouls todbehavehappeninotincedg? that, in each case, the labeling vector is
You
grappr(waiphsthooutachiwistnchgaltiahnicerg)s s.ttWhat
eaadyin vectsitsatotehrevect(aresltoaetradyioandnsshtapplaiptebetlaybeltwheene!)adj. Labelthaecencynewthelmatverabeltrisicxandes ofonetthheemorolcompl
ed onestimet?ee
Make aifconjwe elactbelure aboutwithththeegenersteadyal casstatee vectWhator andis tapplhe sytetadyhe adjstaatelcency
abel? What
happens
matrix Thewithout scaling? is shown in Figure Repeat the process in
ProblWeemswil tnowhroughexplowiretthhtehpris ogrcesaph.s with some other clas es of graphs to see if they
behave
me way.e, Theis the graphisshthowne grainphFiwigurthe 4.3ver. tices arranged in a cyclic
fashion.thFore saexampl
Repeat
oblemsthe tgener
hroughal caswie.th cycles for various
values of Repeat
and makePrthoeblaprconjeomcesesctwiofutrhPre about
val
u
es
of
What
happens
?
A
bi
p
ar
t
i
t
e
gr
a
ph
i
s
a
(
s
e
e
Exer
c
i
s
e
s
i
n
Sec­
tisioadjn a.cent) ifittos verevertiycesvercantexbeinparV, tanditionedviceinvertossae.tIsf andandVVeach
such thhaveat eververy vertictesex, tihnen
the graph is denotRepeated bythe prForocesexampl
e,blemsis thtehrgroughaph in wiFigthurcompl
e ete bipartite
s
of
Pr
o
graphsBy the endforofvarthioisuschaptvalueesr, youof wiWhatl be ihappens
?
have made in this Introduction. n a position to explain the observations you
1.
K4 ,
K3
Ks .
Problem 3
A
Kw
Problem 4
Kn
A
Figure 4 . 2
1
3
cycle
Cs
Problem 6
odd
Figure 4 . 3
Problem 1
Cn
n
1
n
6
3
even
Cn
n.
74-78
complete bipartite graph
37
Kn , n -
Problem 8
K3 , 3
Kn , n
Figure 4 . 4
4.2.
Petersen graph
Problem 5
n.
U
U
1
n
4.4.
3
I n t ro d u ct i o n to E i g e n v a l u e s a n d E i g envectors
eady stofatepopulvectoartioi nn tgrheowtconth. eForxt ofa
tInMarwoChapt
applkov chaiiecratinons:wewithMarencount
ktransovitiechaiornedmatntsheandrinotx tihoaensLestofeadylaiestmodel
or atihadon vecttheoprr opersatitsyfythinatg
for, whera Lese lireematpresreintx edathsteeadysteadystatsetavectte groorwtswastahteravectapopul
te. For example, we saw that
[ . ] [ ] [ ] and [�5
phenomenon
morrs seugener
al yAx. Thatis juisst, aforscaalsaqruarmuletma­iple
tIofrnixthiThis wechapts asis ktehrwhet,ewe ihnervesthtiergateeexithsitsnonzer
o
vect
o
ch
t
h
at
and
i
t
i
s
one
of
t
h
e
mos
t
cent
r
a
l
pr
o
bl
e
ms
i
n
l
i
n
ear
algebra. has applications throughout mathematics and in many other fields as well.
therofe is acornonzerrespondioLetvectngobertoansuch thatmatrix. A sSuchcalar a vectis calolrexid ans called an of if
3,
Px = x;
Lx = rx
The
adjective of' means
"own'German
' orand"characteristic
aresensecharac­
teristic
of
a
matrix
in
the
that
they
contai
n
i
m
portant
informa­
tion
aboutThetheletter,.\
nature(lambda),
of the the
matrix.
Greek equiis used
valentforofeigenvalues
the English
letter
becauseasat one time theyThewereprefialsox
known
is pronounced "EYE-gun:'
eigen
Eigen­
values
eigenvectors
L,
latent values.
eigen
U
P,
L,
r
It
D e fi n it i o n
tor
A
x
eigenvalue problem,
A
x
nXn
A.
x
4
0
0.25
0.4
0.7 0.2 0.4
=
0.6
0 3 0.8 0.6
A,
x.
x
A x = Ax.
A
eigenvalue A
eigenvec­
Section 4.1 Introduction to Eigenvalues and Eigenvectors
Exa m p l e 4 . 1
255
Show that x [ � ] is an eigenvector of [ � � ] and find the corresponding
eigenvalue.
We compute
[ � � ] [ � ] [:] [ � ]
from which it follows that xis an eigenvector of corresponding to the eigenvalue
A=
=
Solution
= 4x
=4
Ax =
A
Exa m p l e 4 . 2
4.
4
genvalue.ue of [ � ] and determine all eigenvectors corresShowponditnhgatto tihs ians eigeienval
Weequimusvaletntshtowo ththeatequattherieonis a vectosroxwesuchneedthatto comput
x Sx.e Butthe nulthisl
equat
i
o
n
i
s
space of the matrix We find that
[ I 23 ] [ J [ ]
Since ethme ofcolIunmnsvertibofletMathis matricesriximarpleiecls etarhatly iltisnnulearllyspdependent
, theo.Fundament
al
Theor
ace
i
s
nonzer
Thus
,
x
x has ainnontg therinulviall sspoace:lution, so is an eigenvalue of We find its eigenvectors by
comput
[ 1 ] [I t 1 J
Thus, if x [:J is an eigenvector corresponding to the eigenvalue it satisfies
t or t so these eigenvectors are of the form
5
A=
nonzero
(A - SI)x = 0,
Solution
A - 51.
A
_
SJ =
S
1
4
_
4
5 O
0 5
=
5
[A - SI I O J =
A =
-4 2
4 -2
A =
A.
-4 2 0
- 0
�
4 -2 0
0
0 0
=
x 1 - x2 = 0,
5,
x 1 = x2 ,
That is, they are the nonzero multiples of [ t ] (or, equivalently, the nonzero multiples
of [ :Ji
.+
Thest thseetsetofofall eigenvectvectorsocorrs inretshpeondinulnl gsptaceo anofeigenval,\Iu. eIt folofloanws that tmathis setrix of
ieisgjuenvect
ors, together with the zero vector in the null space of
1
,\
nonzero
!R n , is
A-
nXn
A - AI.
A
256
Chapter 4 Eigenvalues and Eigenvectors
Let
be
an
mat
r
i
x
and
l
e
t
be
an
ei
g
enval
u
e
of
The
colcallleedcttihone of all eigenvectof oandrs coris rdenot
espondied nbyg to together with the zero vector, is
Therefore, in Example { t[ � ] } .
D e fi n it i o n
A
n
n
4.2, E5
A,
A.
=
eiShowgenspthace.at is an eigenvalue of
produces As in Example we compute the null space of Row reduction
[ - � � - !] [ � � - � i
ee thatotrhsecornulrelssppondiace ofng to this ieis gnonzer
oe. satHence,isfy x1 is an ei-g2x3envalue
oforfroxm1 andwhicthheweeig2x3senvect
enval
u
. It follows that
A
=
6
A
=
4.2,
Solulion
A
-
6I
A
=
-
2
2
-4
A
A,
=
-
-
-----+
0
0
-
6I.
0
6I
6
x2 +
+ x2
=
0
,
Inion we can gisavyse athgeomet
ric ionrtserpreandtationarofe tparhe anotl eli.oThusn of ,an ieis ganenvecteigenvect
or. Theor
equat
at
t
h
e
vect
ofif if andis paronlalyeilfto trwher
ansfoerms is itnhteomata parrixaltrealnsvectforomrat[oior,nequicorvrealspentondily, nifgandto only
Find the eigenvectors and eigenvalues of [ � O ] geometrically.
Example We rTheecognionlzyevectthators thisatthemapsmatparrix aofl elatoretfhleemselctionves arein tvecthe x-orasxipars a(sleeel
to the y-axis (i.e., multiples of [ �] ), w�ich are reversed (eigenvalue and vectors
paral el to the x-axis (i.e., multiples of [ ] ), which are sent to themselves (eigenvalue
cor(seereFispgondiure n.g ei. gAccor
enspacesdinglarye, and are the eigenvalues of and the
span([ � ] ) and span([ � ] )
IR 2 ,
Ax Ax
Ax
=
A
TA (x)
Exa m p l e 4 . 4
A
EA-
A
eigenspace
Exa m p l e 4 . 3
x
A
x,
TA
x
A
3.56 ) .
x
A] .
A
Solulion
x
=
-1
F
F
- 1 ),
1)
0
4 5)
A
E_ 1
=
=
-1
A
=
E1
1
A,
=
Section 4.1 Introduction to Eigenvalues and Eigenvectors
251
y
3
-3
The eigenvectors of a reflection
Figure 4 . 5
y
4
3
2
3
4
Figure 4 . 6
The discussion
is basedPictur­
on the
article
"Eigenpictures:
ingSteventhe Schonefeld
EigenvectorinProblem" by
(1996),
pp. 316-319.
The College
Mathematics Journal 26
of eigoenvect
oirfsandgeometonlyriifcalx landy is Axto drareawalxigandnedAxin ahead-straigthto­
tlianile..AnotThenIn Fihgxerurwiwayel4.be6to, xianthseiinangkenvect
r
of
A
ei
g
envect
o
r
of
A
but
i
s
not
.
eigenvalorsuegeometthenricsoalliys,anywe need
non­
zeronloyIfconsmulxistiiandperleeitofghenvect
x.e effSo,ectoiroffofweAAonwantcorretsopvectsearondiocrnhsg. forFitogurteihgeeenvect
4.7(a) shows what happens when
we transform unit vectors with the matrix A [ � �] of Example 4.[\�ny,]isplay
the results head-to-tail, as in Figure 4.6. We can see that the vector x 1 is an
eigenvector, but we also notice that there appears to be an eigenvector in the second
quadrant. Indeed, this is the case, and it turns out to be the vector [ � j �] .
y
A,
unit
=
=
-
/ v'2
258
Chapter 4 Eigenvalues and Eigenvectors
y
y
(b)
(a)
Figure 4 . 1
1
(b), weorsseate what
happens
when
we
us
e
t
h
e
mat
r
i
x
A
ITherne arFiegurnoeei4.g7envect
[
1
al
l
!
to firincdineitegrenvect
oiorns onceofthem-but
we have tonehe corrquesestpioondin renmaig eingsenval
­
uesdo we,Weandfinowrwest fihavenknowd thaehowgeomet
p
r
e
t
a
t
:
How
an eigenvalue ofA ieif gandenvalonluyesifofthaeginulvenl smatpaceriofx?AThe- keyisisnontthe robsiviaelr.vation that is
Recall from Section that the determinant of a matrix A [ ; �] is
tFurhe texprhermesorioe,nthdete Fundament
A -al TheorandeAmiofls innververtitbibleleifMatandriconles yguarif detantAeesisthnonzer
o
.
atonla ma­y if
titrsixdethaserma nontinantriivsizeral nulo. Putl sptaceing itfhandese factonlys tiofgetit ishernoni, wenverseettibhlate-hence,
i
f
and
(fosr fact charmatactriecresizesat
leiegasenvalt) iusesan, andeigenval
u
e
of
A
i
f
and
onl
y
i
f
det(
A
0.
Thi
moment, though,weletwi's sleesohowon generto usaeliiztewiitttho squarematmatrirciceses. ofarbitrary size. For the
Find all of the eigenvalues and corresponding eigenvectors of the matrix A
[ � �] from Example 4. 1 .
tion det(A The- precedi0. nSignrceemarks show that we must find all solutions of the equa­
det (A det [ 1 l ]
-1
wetionneedare easto isloylvfound
e the quadr
to be atic equat
4 andion These are th0.erTheeforesotlhuetieionsgenvalto thuiess equa­ofA.
=
A
AI
3.3
=
2 X 2
=
be,
ad
A
-
2X2
AI ) =
2X2
Exa m p l e 4 . 5
A
Solulion
AI )
- AI)
=
=
3 -A
A
=
=
3-A
(3 - A ) ( 3 - A )
A 2 - 6A + 8 =
A 2.
=
=
A 2 - 6A + 8
Section 4.1 Introduction to Eigenvalues and Eigenvectors
259
To find the eigenvectors corresponding to the eigenvalue A = 4, we compute the
null space of A - 4I. We find
l
[-1 1 J [1 1 J
[:J
{ [:: ] } { [ � ] }
[A - 4I oJ =
from which it follows that x =
l
1 O
-+
-1 0
0
-1 O
0 0
is an eigenvector corresponding to A = 4 if and
only if x 1 - x2 = 0 or x 1 = x2 • Hence, the eigenspace E4 =
span
([ � ] )
.
Similarly, for A = 2, we have
so y =
[A - 2I I 0 l =
[;J
= x2
[ � � I �J [� � I �J
{ [ �2 ] } { [ � ] } ([ � ] )
=
---+
is an eigenvector corresponding to A = 2 if and only if y 1 + y2 = 0 or
y 1 = -Yi- Thus , the eigenspace E2 =
= y2
= span
.
Figure 4.8 shows graphically how the eigenvectors of A are transformed when
multiplied by A: an eigenvector x in the eigenspace E4 is transformed into 4x, and an
eigenvector y in the eigenspace E2 is transformed into 2y. As Figure 4.7(a) shows, the
eigenvectors of A are the only vectors in IR 2 that are transformed into scalar multiples
of themselves when multiplied by A.
y
Ay = 2y
-4
-3
-2
2
-1
-1
-2
-3
-4
How transforms eigenvectors
Figure 4 . 8
A
3
4
Chapter 4 Eigenvalues and Eigenvectors
260
Remark
You will recall that a polynomial equation with real coefficients (such as
the quadratic equation in Example 4.5) need not have real roots; it may have complex
roots. (See Appendix C.) It is also possible to compute eigenvalues and eigenvectors
when the entries of a matrix come from "ll_P ' where p is prime. Thus, it is important to
specify the setting we intend to work in before we set out to compute the eigenvalues
of a matrix. However, unless otherwise specified, the eigenvalues of a matrix whose
entries are real numbers will be assumed to be real as well.
Exa m p l e 4 . 6
Interpret the matrix in Example 4.5 as a matrix over "11_ 3 and find its eigenvalues in
that field.
A2 = -2 = 1, A2 - A=6A 1 = A= -1 = 2A2 2 =
The solution proceeds exactly as above, except we work modulo 3. Hence,
the quadratic equation
+ 8 0 becomes +
0. This equation is the
same as
giving
and
as the eigenvalues in Z 3 • (Check
that the same answer would be obtained by first reducing A modulo 3 to obtain
S o lulion
[� �]
Exa m p l e 4 . 1
and then working with this matrix.)
= [� -�]
= - AI) = [ -1A --1A ] = ,\2 1
A= A=
Find the eigenvalues of A
S o lulion
(a) over IR and (b) over the complex numbers C.
We must solve the equation
O
det (A
det
+
(a) Over IR, there are no solutions, so A has no real eigenvalues.
(b) Over C, the solutions are
i and
- i. (See Appendix C.)
In the next section, we will extend the notion of determinant from
2 2
X
to
n X n matrices, which in turn will allow us to find the eigenvalues of arbitrary square
..
I
matrices. (In fact, this isn't quite true-but we will at least be able to find a polynomial
equation that the eigenvalues of a given matrix must satisfy.)
Exercises 4 . 1
In Exercises 1 -6, show that v is an eigenvector ofA and find
the corresponding eigenvalue.
[� �l = [ � ]
= [� �l v = [ -�]
= [ - ! �l = [ _� ]
1. A =
2. A
3. A
v
v
4. A
= [! = �l = [�]
= [ 1 1 -; ] .F [ -: ]
-1
-2
= [ : 2 1 ] = [ 11 ]
v
0
5. A
6. A
0
0
,v
•
Section 4.1 Introduction to Eigenvalues and Eigenvectors
A
In Exercises 7- 12, show that is an eigenvalue of A and
find one eigenvector corresponding to this eigenvalue.
7. A =
8. A =
9. A =
IO. A =
11. A =
12. A �
[� 2] A
[� 2] ,
[ _� 4] A
[! -2]
0 : J A -I
H 0 -I ] �
[: 2 � , A = 2
, =3
-1
3
IR2
In Exercises 1 9-22, the unit vectors x in and their
images Ax under the action of a 2 X 2 matrix A are drawn
head-to-tail, as in Figure 4.7. Estimate the eigenvectors and
eigenvalues ofA from each ''eigenpicture."
19.
y
,\ = - 1
,
5
-7 ,
=1
A = -6
1
In Exercises 1 3 - 1 8, find the eigenvalues and eigenvectors of
A geometrically.
13. A =
14. A =
15. A =
[ -� �]
[� �]
[� �]
[il� �d]
[� �]
[� -�]
(reflection in the y-axis)
20.
(reflection in the line y = x)
(projection onto the x-axis)
(projection onto the line through the
2
origin ith irection vector
16. A =
17. A =
[i])
and a factor of 3 vertically)
18. A =
2
(stretching by a factor of horizontally
(counterclockwise rotation of
about the origin)
90°
261
y
262
Chapter 4 Eigenvalues and Eigenvectors
21.
y
29. A =
[� �]
30. A =
[1
0
1 + i
1
i
]
In Exercises 3 1 -34, find all of the eigenvalues of the ma­
trix A over the indicated Zp .
31. A =
33. A =
[ � �]
[ ! �]
over Z 3
32. A =
over Zs
34. A =
[�
[: �]
over Zs
35. (a) Show that the eigenvalues of the 2 X 2 matrix
A=
-2
22.
[ : �]
are the solutions of the quadratic equation
A 2 - tr(A)A + det A = 0, where tr (A) is the trace
of A. (See page 1 62.)
(b) Show that the eigenvalues of the matrix A in
part (a) are
y
A = H a + d ± V(a
-
d) 2 + 4bc )
( c) Show that the trace and determinant of the matrix A
in part (a) are given by
tr(A) = A 1 + A 2 and det A = A 1 A 2
35.
In Exercises 23-26, use the method of Example 4.5 to find
all of the eigenvalues of the matrix A. Give bases for each of
the corresponding eigenspaces. Illustrate the eigenspaces and
the effect of multiplying eigenvectors by A as in Figure 4. 8.
23. A =
[� - � ]
[� �]
24. A =
[! �]
[ � �]
where A 1 and A 2 are the eigenvalues of A.
36. Consider again the matrix A in Exercise Give
conditions on a, b, c, and d such that A has
(a) two distinct real eigenvalues,
(b) one real eigenvalue, and
(c) no real eigenvalues.
37. Show that the eigenvalues of the upper triangular
matrix
A=
[ � �]
are A = a and A = d, and find the corresponding
eigenspaces.
25. A =
26. A =
,E:S738. Let a and b be real numbers. Find the eigenvalues and
,E:S7
corresponding eigenspaces of
In Exercises 27-30, find all of the eigenvalues of the matrix
A over the complex numbers C. Give bases for each of the
_
corresponding eigenspaces.
27. A =
[ � �]
_
28. A =
[ -3]
2
l
O
over the complex numbers.
11 f
Section 4.2 Determinants
263
Determ i n a nts
Historically, determinants preceded matrices-a curious fact in light of the way linear
algebra is taught today, with matrices before determinants. Nevertheless, determi­
nants arose independently of matrices in the solution of many practical problems,
and the theory of determinants was well developed almost two centuries before
matrices were deemed worthy of study in and of themselves. A snapshot of the his­
tory of determinants is presented at the end of this section.
Recall that the determinant of the 2 X 2 matrix A =
[ aa2l 1l
det A = a l l a 22 - a 1 2 a 2 1
We first encountered this expression when we determined ways to compute the
inverse of a matrix. In particular, we found that
[ aa2111 aa2212 ]
I
The determinant of a matrix A is sometimes also denoted by I A , so for the 2 X 2
matrix A =
Warning
we may also write
. I aa ll1 aa 12 I
2 22
tation. It 1s. easy to mistake
·
[ aa2ll1 aa2212 ]
This notation for the determinant is reminiscent of absolute value no-
. for determmant,
.
, the notat10n
for
,
the notation for the matrix itself. Do not confuse these. Fortunately, it will usually
be clear from the context which is intended.
11
We define the determinant of a X matrix A = [a] to be
det A = l a l = a
I I
(Note that we really have to be careful with notation here: a does not denote the
absolute value of a in this case.) How then should we define the determinant of a
3 X 3 matrix? If you ask your CAS for the inverse of
A=
[
the answer will be equivalent to
�
[� � ;]
g h
l
l
ei - fh ch - bi bf - ce
Jg - di ai - cg cd - af
dh - eg bg - ah ae - bd
where Li = aei - afh - bdi + bfg + cdh - ceg. Observe that
Li = aei - afh - bdi + bfg + cdh - ceg
= a ( ei - fh) - b (di - Jg) + c(dh - eg)
A-1 =
=a
I � �I - b I � �I c I � �I
+
264
Chapter 4 Eigenvalues and Eigenvectors
and that each of the entries in the matrix portion of A - i appears to b e the determi­
nant of a 2 X 2 submatrix of A. In fact, this is true, and it is the basis of the definition
of the determinant of a 3 X 3 matrix. The definition is recursive in the sense that the
determinant of a 3 X 3 matrix is defined in terms of determinants of 2 X 2 matrices.
Defi n ition
[
a ll
Let A = a 21
a 31
]
a 13
a 23 . Then the determinant of A is the scalar
a 33
(1)
Notice that each of the 2 X 2 determinants is obtained by deleting the row and col­
umn of A that contain the entry the determinant is being multiplied by. For example,
the first summand is a 11 multiplied by the determinant of the submatrix obtained by
deleting row 1 and column 1 . Notice also that the plus and minus signs alternate in
Equation ( 1 ) . If we denote by A;j the submatrix of a matrix A obtained by deleting row
i and column j, then we may abbreviate Equation ( 1 ) as
det A = a ll det A ll - a 12 det A 12 + a 13 det A 13
3
2: ( - l ) 1 +ja 1j det A 1j
j= l
For any square matrix A, det A;j is called the (i, j)-minor of A.
Exa m p l e 4 . 8
Compute the determinant of
Solution
We compute
det A = 5
1 � �I
-
- ( - 3)
I� �I I� �I
+2
_
= 5 ( 0 - ( - 2 )) + 3 ( 3 - 4 ) + 2 ( - 1 - 0 )
= 5 (2 ) + 3 ( - 1 ) + 2 ( - 1 ) = 5
:+
With a little practice, you should find that you can easily work out 2 X 2 determinants
in yom head. Wciting out the secood line in the above solution is then unnecess
Another method for calculating the determinant of a 3 X 3 matrix is analogous
to the method for calculating the determinant of a 2 X 2 matrix. Copy the first two
columns of A to the right of the matrix and take the products of the elements on the six
Section 4.2 Determinants
265
diagonals shown below. Attach plus signs to the products from the downward-sloping
diagonals and attach minus signs to the products from the upward-sloping diagonals.
(2)
+
This method gives
In Exercise 19, you are asked to check that this result agrees with that from Equa­
tion ( 1) for a 3 X 3 determinant.
Exa m p l e 4 . 9
Calculate the determinant of the matrix in Example 4.8 using the method shown in (2).
Solution
products:
We adjoin to A its first two columns and compute the six indicated
0
- 10
-9
-2
Adding the three products at the bottom and subtracting the three products at the
top gives
det A
=
0 + ( - 1 2) + ( - 2) - 0 - ( - 1 0) - ( - 9)
=
5
as before.
Warning
We are about to define determinants for arbitrary square matrices.
However, there is no analogue of the method in Example 4.9 for larger matrices. It is
valid only for 3 X 3 matrices.
Delerminanls of n x n Malrices
The definition of the determinant of a 3 X 3 matrix extends naturally to arbitrary
square matrices.
D e fi D i l i O D
Let A
minant of A is the scalar
=
n
[a;) be an n
_L ( - l ) 1 +ia 11 det A 11
j� I
x
n matrix, where n 2: 2. Then the deter-
(3)
266
Chapter 4 Eigenvalues and Eigenvectors
It is convenient to combine a minor with its plus o r minus sign. To this end, we define
the (i, j)-cofactor ofA to be
With this notation, definition (3) becomes
<let A
2: a 1j C ij
j= l
n
=
(4)
Exercise 20 asks you to check that this definition correctly gives the formula for the
determinant of a 2 X 2 matrix when n 2.
Definition ( 4) is often referred to as cofactor expansion along the first row. It is
an amazing fact that we get exactly the same result by expanding along any row (or
even any column) ! We summarize this fact as a theorem but defer the proof until the
end of this section (since it is somewhat lengthy and would interrupt our discussion
if we were to present it here).
=
Theorem 4 . 1
The Laplace Expansion Theorem
The determinant of an n X n matrix A
=
[a;) , where n 2: 2, can be computed as
2: a ;j Cij
j= l
n
(5)
(which is the cofactor expansion along the ith row) and also as
2: a ij cij
i= l
n
=
(6)
(the cofactor expansion along the jth column).
A
Since Cij = ( - l ) i +j <let A;j , each cofactor is plus or minus the corresponding minor,
with the correct sign given by the term ( - 1 y +j . quick way to determine whether
the sign is + or - is to remember that the signs form a "checkerboard" pattern:
+
+
+
+
+
+
+
+
Section 4.2 Determinants
Exa m p l e 4 . 1 0
261
Compute the determinant of the matrix
by (a) cofactor expansion along the third row and (b) cofactor expansion along the
second column.
Solution
(a) We compute
PierrebornSimonin Normandy,
Laplace (17France,
49-1827)
was
and
was expected
tos mathematical
become a
clergyman
unti
l
hi
talents
at school.
Hecontributions
madeweremanynoticed
important
to astronomy.
calculus, He was
probabi
l
i
t
y,
and
anBonaparte
examinerat ofthetheRoyal
youngArtillery
Napoleon
andserved
later, when
Napoleon
was
inCorps
power,
bri
e
fl
y
as
Minister
InteriorLaplace
and thenwasChancellor
oftheof thethetitlSenate.
granted
e
of
Count
of
the
Empire
in 1806 andde Laplace
receivedinthe1817.title of
Marquis
Exa m p l e 4 . 1 1
= 2 1 -� � 1 - (- 1) 1 � � 1 + 3 1 � -� 1
2 ( - 6) + + 3 ( 3 )
=
=5
8
(b) In this case, we have
= - ( - 3) 1 � � 1 + 0 1 � � 1 - ( - 1 ) 1 � � I
= 3( - 1) + 0 + 8
=5
4. 1 0
we needed to do fewer calculations than
Notice that in part (b) of Example
in part (a) because we were expanding along a column that contained a zero entry­
namely, a 22 ; therefore, we did not need to compute C22 • It follows that the Laplace
Expansion Theorem is most useful when the matrix contains a row or column with
lots of zeros, since, by choosing to expand along that row or column, we minimize the
number of cofactors we need to compute.
Compute the determinant of
3
+
First, notice that column has only one nonzero entry; we should there­
fore expand along this column. Next, note that the / - pattern assigns a minus sign
Solution
268
Chapter 4 Eigenvalues and Eigenvectors
to the entry a 2 3 = 2 . Thus, we have
det A = a 13 C 13 + a 23 C 23 + a 33 C 33 + a 43 C43
= O(C 13 ) + 2C 23 + O(C 33 ) + O(C43 )
2 -3 1
-2 1 - 1 3
-2
0
We now continue by expanding along the third row of the determinant above
(the third column would also be a good choice) to get
( I =�
det A = - 2 - 2
� I - I � � I)
- 2 ( - 2 ( - 8) 5)
- 2 ( 1 1 ) = - 22
-
(Note that the + I pattern for the 3 X 3 minor is not that of the original matrix but
that of a 3 X 3 matrix in general.)
-
4
The Laplace expansion is particularly useful when the matrix is (upper or lower)
triangular.
Exa m p l e 4 . 1 2
Compute the determinant of
2
0
A= 0
0
0
Solution
0
-3
3 2 5
0 1 6
0 0 5
0 0 0
4
7
0
2
-1
We expand along the first column to get
3
0
det A = 2
0
0
2
1
0
0
7
5
0
6
5
2
0 -1
(We have omitted all cofactors corresponding to zero entries.) Now we expand along
the first column again:
0
1 6
2
det A = 2 · 3 0 5
0 0 -1
Continuing to expand along the first column, we complete the calculation:
det A = 2 · 3 · 1
I�
1
2
= 2 . 3 . 1 . ( 5 ( - 1 ) - 2 . 0 ) = 2 . 3 . 1 . 5 . ( - 1 ) = - 30
-1
4
Section 4.2 Determinants
269
Example 4. 1 2 should convince you that the determinant of a triangular matrix is the
product of its diagonal entries. You are asked to give a proof of this fact in Exercise 2 1 .
We record the result as a theorem.
Theorem 4 . 2
The determinant of a triangular matrix is the product of the entries on its main
diagonal. Specifically, if A = [a;) is an n X n triangular matrix, then
<let A = a 1 1 a 22 · • · a ""
Nole
In general (that is, unless the matrix is triangular or has some other special
form), computing a determinant by cofactor expansion is not efficient. For example,
the determinant of a 3 X 3 matrix has 6 = 3 ! summands, each requiring two multipli­
cations, and then five additions and subtractions are needed to finish off the calcula­
tions. For an n X n matrix, there will be n! summands, each with n - 1 multiplications,
and then n ! - 1 additions and subtractions. The total number of operations is thus
T(n) = (n - l ) n ! + n ! - 1 > n!
Even the fastest o f supercomputers cannot calculate the determinant o f a mod­
erately large matrix using cofactor expansion. To illustrate: Suppose we needed to
calculate a 50 X 50 determinant. (Matrices much larger than 50 X 50 are used to store
the data from digital images such as those transmitted over the Internet or taken by a
digital camera.) To calculate the determinant directly would require, in general, more
6
than 50! operations, and 50! 3 X 1 0 4 . If we had a computer that could perform
1
2
a trillion ( 1 0 ) operations per second, it would take approximately 3 X 1 0 52 sec­
onds, or almost 1 0 4 5 years, to finish the calculations. To put this in perspective, con­
sider that astronomers estimate the age of the universe to be at least 10 billion ( 1 0 1 0 )
years. Thus, on even a very fast supercomputer, calculating a 50 X 50 determinant by
cofactor expansion would take more than 1 0 30 times the age of the universe!
Fortunately, there are better methods-and we now turn to developing more
computationally effective means of finding determinants. First, we need to look at
some of the properties of determinants.
=
Properties of Determinants
The most efficient way to compute determinants is to use row reduction. However,
not every elementary row operation leaves the determinant of a matrix unchanged.
The next theorem summarizes the main properties you need to understand in order
to use row reduction effectively.
Theorem 4 . 3
Let A = [a ij ] be a square matrix.
a. If A has a zero row (column), then <let A = 0.
b. IfB is obtained by interchanging two rows (columns) of A, then <let B = - <let A.
c. If A has two identical rows (columns), then <let A = 0.
d. If B is obtained by multiplying a row (column) of A by k, then <let B = k <let A.
e. If A, B, and C are identical except that the ith row (column) of C is the sum of
the ith rows (columns) of A and B, then <let C = <let A + <let B.
f. If B is obtained by adding a multiple of one row (column) of A to another row
(column), then <let B = <let A.
210
Chapter 4 Eigenvalues and Eigenvectors
Proof We will prove (b) as Lemma 4. 14 at the end of this section. The proofs of
properties (a) and (f) are left as exercises. We will prove the remaining properties in
terms of rows; the corresponding proofs for columns are analogous.
(c) If A has two identical rows, swap them to obtain the matrix B. Clearly, B = A, so
<let B = det A. On the other hand, by (b), det B = - det A. Therefore, det A = - det A,
so det A = 0.
( d) Suppose row i ofA is multiplied by k to produce B; that is, b ij = ka;j for j = 1 , . . . , n.
Since the cofactors Cij of the elements in the ith rows of A and B are identical (why?),
expanding along the ith row of B gives
n
n
n
det B = � b ;j Cij = � ka ij cij = k � a ij cij = k det A
j= l
j= l
j= l
(e) As in (d), the cofactors Cij of the elements in the ith rows of A, B, and C are
identical. Moreover, c ij = a;j + b;j for j = 1 , . . . , n. We expand along the ith row of C
to obtain
n
n
n
n
det C = � c ;j Cij = � ( au + b ;)Cu = � a ;j CiJ + � bu Cu = det A + det B
j= l
j= l
j= l
j= l
Notice that properties (b), ( d), and ( f) are related to elementary row operations.
Since the echelon form of a square matrix is necessarily upper triangular, we can
combine these properties with Theorem 2 to calculate determinants efficiently. (See
Exploration: Counting Operations in Chapter 2, which shows that row reduction of
an n X n matrix uses on the order of n 3 operations, far fewer than the n ! needed for
cofactor expansion.) The next examples illustrate the computation of determinants
using row reduction.
Exa m p l e 4 . 1 3
[ 20
Compute <let A if
(a)
A=
(b) A �
Solution
-4
-
3 -1
5 3
2
6
]
[ i -� j il
(a) Using property (f) and then property (a), we have
2
3
5
<let A = 0
-4 -6
-1
3
2
R, + 2R1
2 3
0 5
0 0
-1
3 =O
0
Section 4.2 Determinants
211
(b) We reduce A to echelon form as follows (there are other possible ways to do this):
2 -4 5
3
0 -3
2 -4
0 - 3 6 R ,<->R, 0
2
4
5
4
5 7
5 -1 -3
- 1 -3
1
1
2
0 -1
R3 - 2R 1
1tt
4
R4 - SR 1
R
R
2 -4
5
0
Q
= -3
= - ( - 3)
0
4
0
7
3
2 -9
0 -1
0
0
3
<let A =
2
5
6
5 Ri/ 3
= -3
7
1
0 -1
2
-1
4
7
2 -4
1
0
2
5
0
2
4
-1
-1
-4
5
-3
2
5
7
1
2
-9
3
5
1
0 -1
2
Q -1
2 -9
=3
0 1 5 - 33
0
0
0
0 - 13
= 3 . 1 . ( - 1 ) . 1 5 . ( - 1 3 ) = 585
R, + 4R,
R4 + 2 R2
Remark By Theorem 4.3, we can also use elementary column operations in the
process of computing determinants, and we can "mix and match'' elementary row and
column operations. For example, in Example 4. l 3(a), we could have started by adding
column 3 to column 1 to create a leading 1 in the upper left-hand corner. In fact, the
method we used was faster, but in other examples column operations may speed up
the calculations. Keep this in mind when you work determinants by hand.
oe1erminan1s of Eleme n1arv Malrices
Recall from Section 3.3 that an elementary matrix results from performing an ele­
mentary row operation on an identity matrix. Setting A = In in Theorem 4.3 yields
the following theorem.
Theorem 4 . 4
Let E be an n X n elementary matrix.
a. If E results from interchanging two rows of In , then <let E = - 1 .
b. If E results from multiplying one row of In by k, then <let E = k.
c. If E results from adding a multiple of one row of In to another row, then
det E = 1 .
TheGreek
word verb is derivedwhifromch
themeans
"toisgrasp:
'pIner mathematics,
athatlemma
a
"hel
theorem"
we
"grasp
hol
d
of"
andmoreuse
toimprove
another,
usually
portant, theorem.
lemma
lambanein,
Since <let In = 1 , applying (b), (d), and (f) of Theorem 4.3 immediately gives
(a), (b), and (c), respectively, of Theorem 4.4.
Proof
Next, recall that multiplying a matrix B by an elementary matrix on the left per­
forms the corresponding elementary row operation on B. We can therefore rephrase
(b), (d), and (f) of Theorem 4.3 succinctly as the following lemma, the proof of which
is straightforward and is left as Exercise 43.
212
Chapter 4 Eigenvalues and Eigenvectors
Lem m a 4 . 5
Let B b e an n X n matrix and let E b e an n X n elementary matrix. Then
det ( EB )
=
( det E ) ( det B )
We can use Lemma 4.5 to prove the main theorem of this section: a characteriza­
tion of invertibility in terms of determinants.
Theorem 4 . 6
A
square matrix A is invertible if and only if det A * 0.
Let A be an n X n matrix and let R be the reduced row echelon form of A.
We will show first that det A * 0 if and only if det R * 0. Let E 1 , E2 ,
, Er be the
elementary matrices corresponding to the elementary row operations that reduce
A to R. Then
Proof
•
•
•
Taking determinants of both sides and repeatedly applying Lemma 4.5, we obtain
( det E) · · · ( det E 2 ) ( det E , ) ( det A )
det R
By Theorem 4.4, the determinants of all the elementary matrices are nonzero. We
conclude that det A * 0 if and only if det R * 0.
Now suppose that A is invertible. Then, by the Fundamental Theorem oflnvertible
Matrices, R In , so det R 1 * 0. Hence, det A * 0 also. Conversely, if det A * 0,
then det R * 0, so R cannot contain a zero row, by Theorem 4.3 (a). It follows that
R must be In (why?), so A is invertible, by the Fundamental Theorem again.
=
...--
=
=
Determinants and Matrix Operations
....,...
Theorem 4 . 1
Let's now try to determine what relationship, if any, exists between determinants and
some of the basic matrix operations. Specifically, we would like to find formulas for
det (kA), det(A + B), det(AB), det(A - i ), and det(A T ) in terms of det A and det B .
Theorem 4.3(d) does not say that det(kA) = k det A. The correct relationship
between scalar multiplication and determinants is given by the following theorem.
If A is an n
x
n matrix, then
det ( kA )
....,...
=
kn det A
You are asked to give a proof of this theorem in Exercise 44 .
Unfortunately, there is no simple formula for det (A + B), and in general,
det (A + B) * det A + det B. (Find two 2 X 2 matrices that verify this.) It therefore
comes as a pleasant surprise to find out that determinants are quite compatible with
matrix multiplication. Indeed, we have the following nice formula due to Cauchy.
Section 4.2 Determinants
wasth. bornbriliniaParis
andprolstudied
engineering
buthe published
switched
toovermathemati
c
s
because
of
poor
heal
nt
and
i
fi
c
mathemati
c
i
a
n,
700 papers,inmany
on quite
difficult problems.
His name
can betheory,
found alongebra,
manyandtheorems
and
definitions
di
ff
erenti
a
l
equations,
infinite
series,
probability
physics.
He is noted
for asintroducing
rigor intoconservati
calculus, laying
the foundation
for theandbranch
ofhe mathematics
known
analysis.
Politically
v
e,
Cauchy
was
a
royalist,
in post
1830
followed
Charles
i
n
to
exi
l
e.
He
returned
to
France
in
1838
but
did
not
return
to
his
atloyalthetySorbonne
to the newuntiking.l the university dropped its requirement that faculty swear an oath of
213
Augustin Louis Cauchy ( 1 789- 1 857)
A
X
Theorem 4 . 8
If A and B are n X n matrices, then
det (AB ) = ( det A ) ( det B )
We consider two cases: A invertible and A not invertible.
If A is invertible, then, by the Fundamental Theorem of lnvertible Matrices, it can
be written as a product of elementary matrices-say,
Proof
A = E , E2 • · · Ek
Then AB = E 1 E2
· · ·
EkB, so k applications of Lemma 4.5 give
Continuing to apply Lemma 4.5, we obtain
det (AB ) = det ( E 1 E 2 • • • E k ) det B = ( det A ) ( det B )
On the other hand, if A is not invertible, then neither is AB, by Exercise 47
in Section 3.3. Thus, by Theorem 4.6, det A = 0 and det(AB) = 0. Consequently,
det (AB) = (det A)(det B), since both sides are zero.
Exa m p l e 4 . 1 4
Applying Theorem 4.8 to A =
[ 22 �]
[ � � ],
[ 1216 � ]
12
and B =
AB =
and that det A = 4, det B = 3, and det(AB) =
(Check these assertions! )
we find that
= 4 · 3 = (det A) (det B), as claimed.
4
The next theorem gives a nice relationship between the determinant of an invertible
matrix and the determinant of its inverse.
214
Chapter 4 Eigenvalues and Eigenvectors
Theorem 4 . 9
If A is invertible, then
1
det (A - 1 ) = -det A
Since A is invertible, AA - I = I, so <let (AA - I ) = <let I = 1 . Hence,
(<let A)(det A - 1 ) = 1 , by Theorem 4.8, and since <let A * 0 (why?), dividing by
<let A yields the result.
Proof
.......
Exa m p l e 4 . 1 5
Verify Theorem 4.9 fo r the matrix A of Example 4.14.
Solulion
We compute
so
det A - 1 =
Remark
(�) (�) - ( - �) ( - �) = % - t = � =
�
de A
The beauty of Theorem 4.9 is that sometimes we do not need to know
what the inverse of a matrix is, but only that it exists, or to know what its determinant
is. For the matrix A in the last two examples, once we know that <let A = 4 * 0, we
immediately can deduce that A is invertible and that <let A - l = � without actually
computing A - l .
We now relate the determinant of a matrix A to that of its transpose A T_ Since the
rows of A T are just the columns of A, evaluating <let A T by expanding along the first
row is identical to evaluating <let A by expanding along its first column, which the
Laplace Expansion Theorem allows us to do. Thus, we have the following result.
Theorem 4 . 1 0
For any square matrix A,
det A = det AT
Gabriel
Cramer (1704-1752)
was
athatSwissbearsmathematician.
The
rule
in 1750, inhishisname
treatisewas published
Ascasesearlyof theas 1730,
however,
special
formula
were
known
other
mathematicians,
including
theto(1698-1746),
Scotsman
Colin
Maclaurin
perhaps the greatest
ofwerethetheBritish
mathematicians
"successors of Newton:who'
Introduction
to the Analysis ofAlgebraic Curves.
Cra m e r 's Rule and lhe Adioinl
In this section, we derive two useful formulas relating determinants to the solution
of linear systems and the inverse of a matrix. The first of these, Cramer's Rule, gives
a formula for describing the solution of certain systems of n linear equations in n
variables entirely in terms of determinants. While this result is of little practical use
beyond 2 X 2 systems, it is of great theoretical importance.
We will need some new notation for this result and its proof. For an n X n ma­
trix A and a vector b in !R n , let A;(b) denote the matrix obtained by replacing the ith
column of A by b. That is,
Column i
-1.-
A ; ( b ) = [ a 1 • · · h · · · an l
Section 4.2 Determinants
Theorem 1 . 1 1
215
Cramer's Rule
Let A be an invertible n X n matrix and let b be a vector in !R n . Then the unique
solution x of the system Ax = b is given by
X·l =
Proof
e2 ,
.
•
•
det (A; ( b ))
for i = 1 , . . . , n
det A
The columns of the identity matrix I = In are the standard unit vectors e 1 ,
, en - If Ax = b, then
AI; ( x) = A [ e 1
x · · · en ] = [Ae 1
= [ a 1 · · · b · · · an ] = A; ( b )
•
•
•
Ax · · · Ae n ]
Therefore, by Theorem 4.8,
( detA ) ( det I; ( x)) = det (AI; ( x)) = det (A ; ( h ))
Now
1
0
0
1
X1
Xz
0
0
0
0
det I; ( x) = 0
0
X;
0
0 = X;
0
0
0
0
Xn - 1
Xn
1
0
0
as can be seen by expanding along the ith row. Thus, (<let A) x; = det(A; (h)), and the
result follows by dividing by <let A (which is nonzero, since A is invertible) .
Exa m p l e 4 . 1 6
Use Cramer's Rule t o solve the system
X 1 + 2X 2 = 2
- x 1 + 4x 2 = 1
Solution
det A
We compute
= 1-� !I =
6, det (A 1 ( b ))
= I� !I =
= 1-� �I
6, and det (A 2 ( b ))
=3
By Cramer's Rule,
det (A 2 ( b ))
det (A 1 ( b )) 6
X 1 = ---- = - = 1 and x2 = ---­
det A
6
det A
3
6
2
216
Chapter 4 Eigenvalues and Eigenvectors
As noted previously, Cramer's Rule is computationally inefficient for
all but small systems of linear equations because it involves the calculation of many
determinants. The effort expended to compute just one of these determinants, using
even the most efficient method, would be better spent using Gaussian elimination to
solve the system directly.
Remark
The final result of this section is a formula for the inverse of a matrix in terms of
determinants. This formula was hinted at by the formula for the inverse of a 3 X 3
matrix, which was given without proof at the beginning of this section. Thus, we have
come full circle.
Let's discover the formula for ourselves. If A is an invertible n X n matrix, its
inverse is the (unique) matrix X that satisfies the equation AX = I. Solving for X one
column at a time, let "i be the jth column of X. That is,
X·1
=
Therefore, Axj = ej , and by Cramer's Rule,
det (A/e) )
det A
However,
i th column
a 11 a i 2
a z, a 22
<let (A; ( e) ) =
-J,
0
0
ai , a1 2
a n , an 2
a 1n
a2n
aJn
0
= ( - l )j + i det A = C
JI
JI
a nn
which is the (j, i ) -cofactor of A.
It follows that xij = ( 1 /det A)Cji • so A - i = X = ( 1 / <let A) [Cji ] = ( 1 /det A) [Cij f.
In words, the inverse of A is the transpose of the matrix of cofactors of A, divided by
the determinant of A.
The matrix
is called the adjoint (or adjugate) of A and is denoted by adj A. The result we have
just proved can be stated as follows.
Section 4.2 Determinants
Theorem 4 . 1 2
Let A be an invertible n X n matrix. Then
A-1 =
Exa m p l e 4 . 1 1
1
--
det A
211
adj A
Use the adjoint method to compute the inverse of
A=
Solution
C1 1 =
C2 1 =
C31 =
We compute det A =
+ I�
- I�
+ I�
41
1
1-1
41
-3
-
-3
=
-18
-
[ �1 � -�i
3
-3
2 and the nine cofactors
C1 2 =
=3
C22 =
= 10
C3 2 =
41
1
1-1
41
- I�
+I�
- I�
-3
-
-3
C1 3 =
= -2
C2 3 =
= -6
C33 =
The adjoint is the transpose of the matrix of cofactors-namely,
adj A =
Then
A-1 =
-1
[ : ��
10
1
�J
-6 -2
=
[ �4� -1- �
-
=
-
=
= -2
-2
[ l� -1 �� 1 [ - � - :1 1
1 -18
adj A = - 2
det A
r
+ I� �I 4
I � � I -1
+I� �I
�� 1
= 10
3
-2
-2
=
-t
-2
t
which is the same answer we obtained (with less work) in Example 3.30.
Proof of lhe Laplace Expansion Theorem
Unfortunately, there is no short, easy proof of the Laplace Expansion Theorem. The
proof we give has the merit of being relatively straightforward. We break it down into
several steps, the first of which is to prove that cofactor expansion along the first row
of a matrix is the same as cofactor expansion along the first column.
Le m m a 4 . 1 3
Let A be an n X n matrix. Then
218
Chapter 4 Eigenvalues and Eigenvectors
We prove this lemma by induction on n . For n = 1, the result is trivial. Now
assume that the result is true for (n - 1) X ( n - 1) matrices; this is our induction
hypothesis. Note that, by the definition of cofactor (or minor), all of the terms con­
taining au are accounted for by the summand au Cu. We can therefore ignore terms
containing au.
The ith summand on the right-hand side of Equation (7) is a; 1 C; 1 = a; 1 ( - 1 ) i + l
det A; 1 . Now we expand det A; 1 along the first row:
Proof
a12
a 13
a ; - 1, 2 a ; - 1,3
a ; +1, 2 a ; +1,3
an2
a n3
a 11
a ln
a ; - 1,J
a ; +1,J
a ; - 1,n
a ; +J,n
a n)
a n,n
1
The jth term in this expansion of det A; 1 is a 11( - 1) +J - l det A 1 ;, 11, where the nota­
tion A kl, rs denotes the submatrix of A obtained by deleting rows k and l and columns
r and s. Combining these, we see that the term containing a; 1 a 11 on the right-hand
side of Equation ( 7) is
a ; 1 ( - l ) ; +1 a 1/ - l ) 1 +J - I det A 1 u1 = ( - l ) i +J+ 1 a ; 1 a 11 det A 1 ;,11
What is the term containing a i l a 11 on the left-hand side of Equation (7)? The
factor a 11 occurs in the jth summand, a 11 C 1i = a 1/ - 1) 1 +J det A 11 . By the induction
hypothesis, we can expand det A 1i along its first column:
a z1
a 31
a z, J - 1 a z, J+ 1
a 3, j - I a 3, j+ I
a2n
a 3n
a;1
a ;,j - 1 a ;, J+ l
a ;n
an !
a n, j - 1 a n, )+ !
a nn
The ith term in this expansion of det A 11 is a i l ( - l ) (i - l ) + l det A 1 ;, 11 , so the term con­
taining a; 1 a 1i on the left-hand side of Equation (7) is
i
a 1J ( - 1 ) 1 +Ja ; 1 ( - 1 ) ( ; - i ) + i det A 1 ;, 11 - ( - l) +J+ i a il a 1J det A 1 ;, 11
which establishes that the left- and right-hand sides of Equation (7) are equivalent.
Next, we prove property (b) of Theorem 4.3.
Le m m a 4 . 1 4
Let A be an n X n matrix and let B be obtained by interchanging any two rows
(columns) of A. Then
det B = - det A
Section 4.2 Determinants
219
Proof
Once again, the proof is by induction on n. The result can be easily checked
when n = 2, so assume that it is true for ( n - 1 ) X (n - 1 ) matrices. We will prove
that the result is true for n X n matrices. First, we prove that it holds when two adja­
cent rows of A are interchanged-say, rows r and r + 1 .
By Lemma 4. 13, we can evaluate <let B by cofactor expansion along its first col­
umn. The ith term in this expansion is ( - 1 ) 1 + i b; 1 <let B; 1 . If i * r and i * r + 1, then
b; 1 = a; 1 and B; 1 is an (n - 1) X (n - 1) submatrix that is identical to A; 1 except that
two adjacent rows have been interchanged.
al l
a12
aln
a; 1
a;2
a; n
a r + l, l a r + l, 2
a rl
a r2
a n2
an l
a r + l ,n
a rn
a nn
Thus, by the induction hypothesis, <let B; 1 = - <let A; 1 if i * r and i * r + 1 .
If i = r, then b; 1 = a r + l , l and B i l = A r + l , l ·
al l
Row
a12
i -+ a r + l, l a r + l, 2
aln
a rl
a r2
a r + l ,n
a rn
an ]
an2
a nn
Therefore, the rth summand in <let B is
( - l ) r + I b r l <let B r ! = ( - l ) r + I a r + I,I <let A r + I, I = - ( - l ) ( r + I) + I a r + I,I <let A r + I, I
Similarly, if i = r + 1 , then b; 1 = a r 1 , B; 1 = A r 1 , and the (r + l )st summand in <let B is
( - l ) ( r + I) + I b r I,I det B r I,I = ( - l ) 'a r l det A r l = - ( - l ) r + l a r l det A r l
+
+
In other words, the rth and (r + l)st terms in the first column cofactor expansion of
<let B are the negatives of the ( r + 1 )st and rth terms, respectively, in the first column
cofactor expansion of <let A.
Substituting all of these results into <let B and using Lemma 4. 1 3 again, we obtain
n
� ( - 1 ) i + l b; 1 <let B; 1
<let B = ,L,.;
i= l
n
� ( - l ) i + l b; 1 <let B; 1 + ( - l ) r+ l b rl <let B r1 + ( - l ) ( r + J ) + l b r + l, l <let B r + l,l
i= l
i ;" r, r + l
n
� ( - l ) i + 1 a; 1 ( - det A; 1 ) - ( - l ) ( r + l) + l a r + l , l det A r + l, l - ( - l ) r + l a r1 det A r1
i= l
i t= r,r + l
n
- ,L,.;
� ( - 1 ) i + l a; 1 <let A; 1
i= l
- det A
280
Chapter 4 Eigenvalues and Eigenvectors
This proves the result for n X n matrices if adjacent rows are interchanged. To
see that it holds for arbitrary row interchanges, we need only note that, for example,
rows r and s, where r < s, can be swapped by performing 2( s - r) - 1 interchanges of
adjacent rows (see Exercise 67) . Since the number of interchanges is odd and each one
changes the sign of the determinant, the net effect is a change of sign, as desired.
The proof for column interchanges is analogous, except that we expand along
row 1 instead of along column 1 .
We can now prove the Laplace Expansion Theorem.
Proof of Theorem 4 . 1
Let B be the matrix obtained by moving row i of A to the top,
using i - 1 interchanges of adjacent rows. By Lemma 4. 14, <let B = ( - 1 ) i - l det A. But
b 1j = a;j and B 1j = A;j for j = 1 , . . . , n.
a;1
a in
a ;j
a ll
a 1j
a ln
<let B = a ; - 1,1
a; + 1,1
a nl
Thus,
a ; - 1,j
a; + 1,j
a i 1, n
ai+ l, n
a nj
a nn
-
det A = ( - l ) i - 1 det B = ( - l ) i - 1 2: ( - 1 ) 1 +j b 1j det B 1j
j� I
n
n
n
= ( - 1 ) ; - i 2: ( - 1 ) 1 +ja ij <let A ij = 2: ( - l ) i +ja ij <let A ij
j� I
j� I
which gives the formula for cofactor expansion along row i.
The proof for column expansion is similar, invoking Lemma 4. 1 3 so that we can
use column expansion instead of row expansion (see Exercise 68).
A Brief Hislorv of oe1erm ina n1s
ATakakazu
self-taughtSekichiKowa
ld prodi(1642-1708)
gy,
was
descended
fromIn aaddition
family of
samurai
warriors.
discovering
determinants,
hetoequations,
wrote about
diophantine
magic
squares,
and
Bernoulli
numbers
(before
Bernoulli) andin calculus.
quite likely made
discoveries
As noted at the beginning of this section, the history of determinants predates that of
matrices. Indeed, determinants were first introduced, independently, by Seki in 1 683
and Leibniz in 1 693. In 1 748, determinants appeared in Maclaurin's Treatise on A lge ­
bra, which included a treatment of Cramer's Rule up to the 4 X 4 case. In 1 7 50, Cramer
himself proved the general case of his rule, applying it to curve fitting, and in 1 772,
Laplace gave a proof of his expansion theorem.
The term determinant was not coined until 1 80 1 , when it was used by Gauss.
Cauchy made the first use of determinants in the modern sense in 1 8 1 2 . Cauchy, in
fact, was responsible for developing much of the early theory of determinants, in­
cluding several important results that we have mentioned: the product rule for de­
terminants, the characteristic polynomial, and the notion of a diagonalizable matrix.
Determinants did not become widely known until 1 84 1 , when Jacobi popularized
them, albeit in the context of functions of several variables, such as are encountered in
a multivariable calculus course. (These types of determinants were called "Jacobians"
by Sylvester around 1 850, a term that is still used today.)
Section 4.2 Determinants
Gottfried phiWilhelm
vonandLeibniz
(1646-1716)
wasprobabl
bornyinbestLeipzig
andforstudied
law, (with
theology,
l
osophy,
mathematics.
He
is
known
developing
Newton,
independentl
ybranches
) the mainof imathematics
deas of differenti
alsand
integralve.calculus.
However,the notion
his
contributions
to
other
are
al
o
impressi
He
developed
ofothers
a determinant,
knew
versions
ofandCramer'
s Rulefoundation
and the Laplace
Expansion
Theoremworkbeforehe
were
given
credi
t
for
them,
laid
the
for
matrix
theory
through
didHe believed
on quadratic
forms.
Leibnizofalsogoodwasnotation
the firstand,to develop
the binary
systemnotation
of arithmetic.
i
n
the
importance
along
with
the
familiar
fora
deri
v
ati
v
es
and
integrals,
i
n
troduced
a
form
of
subscri
p
t
notation
for
the
coefficients
of
linear system that is essential y the notation we use today.
281
By the late 1 9th century, the theory of determinants had developed to the stage
that entire books were devoted to it, including Dodgson's An Elementary Treatise on
Determinants in 1 867 and Thomas Muir's monumental five-volume work, which
appeared in the early 20th century. While their history is fascinating, today deter­
minants are of theoretical more than practical interest. Cramer's Rule is a hopelessly
inefficient method for solving a system of linear equations, and numerical methods
have replaced any use of determinants in the computation of eigenvalues. Determi­
nants are used, however, to give students an initial understanding of the characteristic
polynomial (as in Sections 4. 1 and 4.3).
..
I
Exercises 4 . 2
•
Compute the determinants in Exercises 1 -6 using cofactor
expansion along the first row and along the first column.
1 0 3
1. 5
0
2
1
3. - 1
0
-1
0
2 3
5. 2 3 1
2
3
0 1
2. 2 3
-1 3
0
-1
-1
-2
0
1 1 0
4. 1 0
0
2 3
6. 4 5 6
7 8 9
8. 2
3
1
0
-2
-1
1
2
0
tan e
- sin e
cos e
12. b c d
0 e 0
-1 0 3
5 2 6
0 0
4 2 1
0 0 0 a
0 0 b c
15.
0 d e f
g h
cos e sin e
10.
0 cos e
0 sin e
0 a 0
a b 0
11. 0 a b
a 0 b
13.
Compute the determinants in Exercises 7-15 using cofactor
expansion along any row or column that seems convenient.
5 2 2
7. - 1
2
3 0 0
1 3
-4
9. 2 - 2 4
-1 0
j
2
1
14.
0
2
0 3
0 2
-1
0
-1
2
4
-3
Chapter 4 Eigenvalues and Eigenvectors
282
In Exercises 1 6- 1 8, compute the indicated 3 X 3 determi­
nants using the method of Example 4. 9.
16. The determinant in Exercise 6
17. The determinant in Exercise 8
18. The determinant in Exercise 1 1
19. Verify that the method indicated in (2) agrees with
Equation ( 1 ) for a 3 X 3 determinant.
20. Verify that definition ( 4) agrees with the definition of a
2 X 2 determinant when n = 2.
2 1 . Prove Theorem 4.2. [Hint: proof by induction would
be appropriate here.]
A
In Exercises 22-25, evaluate the given determinant using
elementary row and/or column operations and Theorem 4.3
to reduce the matrix to row echelon form.
22. The determinant in Exercise 1
23. The determinant in Exercise 9
24. The determinant in Exercise 1 3
25. Th e determinant i n Exercise 14
1
-2
2
1 0
3
27. 0 - 2 5
0 4
0
0 1
0
28. 0
5 2
3 -1 4
29.
2 3
30. 0 4 1
6 4
4 1
31. -2 0
5 4
1
0
32.
0
0
0
-3
33.
0
0
0 0 0
0 1 0
1 0 0
0 0 1
1 0 1 0
0
0 1
34.
1 1 0 0
1
0 0
2a 2b 2c
36.
e f
g h
d e f
37. a b c
38.
g h
2c b a
39. 2f e d
2i h g
a + 2g b + 2h c + 2i
40. 3d + 2g 3e + 2h 3f + 2i
h
g
35. d
41. Prove Theorem 4.3(a).
43. Prove Lemma 4.5.
In Exercises 26-34, use properties of determinants to
evaluate the given determinant by inspection. Explain
your reasoning.
1 1
26. 3 0
2 2
Find the determinants in Exercises 35-40, assuming that
a b c
d e f =4
g h
2
-1
3
-3
5
-4
-2
2
3
-2
1
2 0 0
0 0 0
0 0 4
0
0
2a b/ 3
2 d e/ 3
2g h/ 3
a-c b
d -f e
g- i h
-c
-f
-i
c
f
42. Prove Theorem 4.3(f) .
44. Prove Theorem 4.7.
In Exercises 45 and 46, use Theorem 4. 6 to find all values of
k for which A is invertible.
-k
k+1
45. A =
-8 k 1
k
46. A
�
[1 � :
[ � �]
2
k
In Exercises 47-52, assume that A and B are n X n matrices
with det A = 3 and det B = - 2. Find the indicated
determinants.
47. det (AB )
50. det ( 2A )
48. det (A 2 )
5 1 . det ( 3B r )
49. det ( B- 1 A )
52. det (AA r)
In Exercises 53-56, A and B are n X n matrices.
53. Prove that det (AB) = det (BA).
54. If B is invertible, prove that det (B - 1AB) = det (A) .
5 5 . I f A i s idempotent (that is, A 2 = A), find all possible
A
values of det (A) .
56. square matrix A i s called nilpotent i f A m = 0 for
some m > 1. (The word nilp otent comes from the
Latin nil, meaning "nothing;' and potere, meaning
"to have power:' nilpotent matrix is thus one that
A
Section 4.2 Determinants
becomes "nothing" -that is, the zero matrix-when
raised to some power.) Find all possible values of
det (A) if A is nilpotent.
In Exercises 57-60, use Cramer's Rule to solve the given
linear system.
58. 2x - y = 5
57. x + y = 1
x + 3y = - 1
x-y=2
59. 2x + y + 3z = 1
60. x + y - z = 1
y+ z=1
x+y+z=2
=3
z=1
x-y
In Exercises 61-64, use Theorem 4. 12 to compute the in­
verse of the coefficient matrix for the given exercise.
62. Exercise 58
61. Exercise 57
63. Exercise 59
64. Exercise 60
65. If A is an invertible n X n matrix, show that adj A is
where P and S are square matrices. Such a matrix is
said to be in block (upper) triangularform. Prove that
det A = ( det P) ( det S)
[Hint: Try a proof by induction on the number of rows
of P.]
70. (a) Give an example to show that if A can be
partitioned as
A =
(b) Assume that A is partitioned as in part (a) and
that P is invertible. Let
B=
[-�-f-�-]
Writi n g Project
[��-�� -d-?-]
Compute det (BA) using Exercise 69 and use the
result to show that
det A = det P det ( S - RP - 1 Q )
[The matrix S - RP - 1 Q is called the Schur com­
1
( adj A ) - 1 = --A = adj (A - 1 )
det A
66. If A is an n X n matrix, prove that
det ( adj A ) = ( det A ) n - I
A =
[�- f{j
where P, Q, R, and S are all square, then it is not
necessarily true that
det A = ( det P) ( det S) - ( det Q ) ( det R )
also invertible and that
67. Verify that if r < s, then rows r and s of a matrix can
be interchanged by performing 2 (s - r) - 1 inter­
changes of adjacent rows.
68. Prove that the Laplace Expansion Theorem holds for
column expansion along the jth column.
69. Let A be a square matrix that can be partitioned as
283
plement of P in A, after Issai Schur ( 1 875- 1 94 1 ),
who was born in Belarus but spent most of his
life in Germany. He is known mainly for his fun­
damental work on the representation theory of
groups, but he also worked in number theory,
analysis, and other areas.]
(c) Assume that A is partitioned as in part (a), that
P is invertible, and that PR = RP. Prove that
detA = det ( PS - RQ )
Which Came First: The Matrix or the Determinant?
The way in which matrices and determinants are taught today-matrices before
determinants-bears little resemblance to the way these topics developed histori­
cally. There is a brief history of determinants at the end of Section 4.2.
Write a report on the history of matrices and determinants. How did the nota­
tions used for each evolve over time? Who were some of the key mathematicians
involved and what were their contributions?
1. Florian Cajori, A History of Mathematical Notations (New York: Dover, 1 993).
2. Howard Eves, An Introduction to the History of Mathematics (Sixth Edition)
(Philadelphia: Saunders College Publishing, 1 990).
3. Victor J. Katz, A History of Mathematics: An Introduction (Third Edition)
(Reading, MA: Addison Wesley Longman, 2008).
4. Eberhard Knobloch, Determinants, in Ivor Grattan-Guinness, ed., Compan­
ion Encyclopedia of the History and Philosophy of the Mathematical Sciences
(London: Routledge, 2013).
Vi g n ette
L e wis C arro ll's C o n d e n s at i o n M e th o d
In 1 866, Charles Dodgson-better known by his pseudonym Lewis Carroll­
published his only mathematical research paper. In it, he described a "new and brief
method" for computing determinants, which he called "condensation:' Although not
well known today and rendered obsolete by numerical methods for evaluating determi­
nants, the condensation method is very useful for hand calculation. When calculators or
computer algebra systems are not available, many students find condensation to be their
method of choice. It requires only the ability to compute 2 X 2 determinants.
We require the following terminology.
If A is an n x n matrix with n 2: 3, the interior of A, denoted
int(A), is the (n - 2 ) X (n - 2 ) matrix obtained by deleting the first row, last
row, first column, and last column of A.
D e fi n it i o n
is muchLewisbetterCarroll,
knownunderby his
pen
name,
which he wroteand
He aland
so wrote
several
mathematics
books
collections
oflogic puzzles.
Charles Lutwidge Dodgson ( 1 8321 898)
in Wonderland
Looking Glass.
Alice's Adventures
Through the
This
visgCarrol
nette ils'sbased
on theonarticle
"Lewi
Condensati
atiEveng Determi
nants"
byMethod
Adriaforn RiEvalce anduNovember
Torrence
in
2006,
12-15. For ofurther
detaiseels of
thepp.Davicondensati
n
method,
d M. Bressoud,
We will illustrate the condensation method for the 5 X 5 matrix
0
2
3 -1 2
2
3 1 -4
A =
1
2 -1
2 1
3
1 - 1 2 -2
-4
2
0
Begin by setting A 0 equal to the 6 X 6 matrix all of whose entries are 1 . Then, we set
A 1 = A. It is useful to imagine A 0 as the base of a pyramid with A 1 centered on top of
A 0 . We are going to add successively smaller and smaller layers to the pyramid until
we reach a 1 X 1 matrix at the top-this will contain <let A. (Figure 4.9)
Math Horizons,
Proofs and
Confirmations: The Story of the
Alternating Sign Matrix Conjecture,
MAA
UniversiSpectrum
ty Press, Series
1999).(Cambridge
284
Ao
Figure 4 . 9
alNextl , we s"ucondens
bmatricees" Aof1 iAnto a matrix A; whose entries are the determinants of
2X2
1:
I �1 1 2
I 2 -� 1 1 -�
2
I 3 - 11 1 1 - 11
1
1 -! � 1 1 1
2
A; =
3
4X4
-� 1 1 -� � I I �
3
1� 1 2 � 1 1 �
2 2 1 1
-1 1 1 -1 2 1 12
-� 1 1 -� � I I�
-� I
-�1
[ -j
-� I
-� 1
11
7
-1
1
-7
1
5
-1
-
:
-4
6
1
of A;Aby2 =theA;corr. esponding entry of int(A0) to get matrix
ANow2. WeSiwenceredipeatA0viidstehaleachelprso, cedur
tenthisrymeans
dividing each entry ofA� bye,thconse cortrruectspinondig A�ngfrentomrtyhofe int(Asubmatand sroicon.es ofWeA2obtandaitnh: en
1 - � �l I � - ; 1 1 - ; - : 1
]
A� = 1 : _; 1 1 _; � I I � ! I
,
1
2X2
1
1),
1
-
A'
4
1
1
3 1 20
30 1 8
30
12
18
4
11
11
60/3
36/2
- 4/ - 1
20
18
18
4
1
- 29
13 1
- 27
- 29
= ���� l = [ �� ��
26/2
=[
]=[
- 42
- 96
12
4
- 27
- 29
13
],
- 27
- 29
26
]
- 94
,
350
]
[
A � = [ 1 : -�� I ] =
A5 = =
be checked
by hotodherwimetl prhoodsduce, <leat A = matIrnixgener
al, aforinianng <let A. matrix A,
tAshecancondens
a
t
i
o
n
met
A
cont
n
y, thedmetthenhodbe brtryeaksing down
idfethbye inzerteoriotor ofconsanytruofctthA;e +A; sHowever
contain,s cara zerefulo,
susineceClofweeelarelwoulment
t
o
di
v
i
that we can proaceed.ry row and column operations can be used to eliminate the zeros so
- 42/7
- 96/ ( - 1 )
6
[ 8604/ 1 8 ]
- 94/ 1
350/5
-6
96
- 94
70
[ 8604 ] ,
[ 478 ]
478.
1X1
nXn
1.
285
Exp loration
G e o m etric App l i c at i o n s of D eterminants
This exploration will reveal some of the amazing applications o f determinants to
geometry. In particular, we will see that determinants are closely related to area and
volume formulas and can be used to produce the equations of lines, planes, and
certain other curves. Most of these ideas arose when the theory of determinants was
being developed as a subject in its own right.
The Cross Product
[ :: ]
Recall from Exploration: The Cross Product in Chapter 1 that the cross product
of u
�
[ ::J
ond v
�
;,
thrndor n X v defined by
[
]
U 2 V 3 - U 3 V2
U 3 V 1 - U 1 V3
X
U 1 V2 - U 2 V 1
If we write this cross product as ( u 2 v3 - u 3 v2 ) e1 - ( u1v3 - u 3 v1 ) e2 + ( u1v2 - u 2 v1 ) e3 ,
where e1, e2 , and e3 are the standard basis vectors, then we see that the form of this
U
V
=
formula is
if we expand along the first column. (This is not a proper determinant, of course,
since e1, e2 , and e 3 are vectors, not scalars; however, it gives a useful way of remem­
bering the somewhat awkward cross product formula. It also lets us use properties of
determinants to verify some of the properties of the cross product.)
Now let's revisit some of the exercises from Chapter 1 .
286
1. Use the determinant version of the cross product to compute u X v.
[} HJ [ l m
[
[
i
l
n =:
lm
[:l [ :: l [ :l
[ U3UU21
( a) u �
u�
(o)
2. If u �
�
(b) u �
v�
( d) u �
and w �
�
U •
-
�
�
how that
(V X W) = det
3. Use properties of determinants (and Problem 2 above, if necessary) to prove
the given property of the cross product.
( a)
( c)
(e)
( g)
(b) u x 0 = 0
v X u = - ( u X v)
( d ) u X kv = k ( u X v)
uXu=0
u X ( v + w) = u X v + u X w ( f ) u · ( u X v) = 0 and v · ( u X v) = 0
u · ( v X w) = ( u X v) · w ( the triple scalar product identity)
Area a n d Vol u m e
We can now give a geometric interpretation of the determinants of 2 X 2 and 3 X 3
matrices. Recall that if u and v are vectors in IFR 3 , then the area A of the parallelogram
determined by these vectors is given by A = II u X v 11 . (See Exploration: The Cross
Product in Chapter 1 . )
4. Let u =
y
and v =
[ ::] .
Show that the area A o f the parallelogram
determined by u and v is given by
[H;nt
Fioure 4 . 1 0
[ �:]
w,;,,
u and v "'
[ �:] [�:Ji
and
5 . Derive the area formula in Problem 4 geometrically, using Figure 4. 1 0 as a
guide. [Hint: Subtract areas from the large rectangle until the parallelogram remains.]
Where does the absolute value sign come from in this case?
281
h{
vxw
v
Figure 4 . 1 1
6. Find the area of the parallelogram determined by u and v.
( a) u =
Figure 4 . 1 2
[ � ], v [ - � ]
=
(b) u =
[!l [�]
v=
Generalizing from Problems 4-6, consider a parallelepiped, a three-dimensional
solid resembling a "slanted" brick, whose six faces are all parallelograms with oppo­
site faces parallel and congruent (Figure 4. 1 1 ). Its volume is given by the area of its
base times its height.
7. Prove that the volume V of the parallelepiped determined by u, v, and w is
given by the absolute value of the determinant of the 3 X 3 matrix [ u v w] with u,
v, and w as its columns. [Hint: From Figure 4. 1 1 you can see that the height h can be
expressed as h = ll u ll cos e, where e is the angle between u and v X w. Use this fact to
show that V = l u · (v X w) I and apply the result of Froblem 2.]
8. Show that the volume V of the tetrahedron determined by u, v, and w
(Figure 4. 12) is given by
V
= i l u · (v X w) I
[Hint: From geometry, we know that the volume of such a solid is V = t (area of the
base) (height).]
Now let's view these geometric interpretations from a transformational point of
view. Let A be a 2 X 2 matrix and let P be the parallelogram determined by the vectors
u and v. We will consider the effect of the matrix transformation TA on the area of P.
Let TA ( P) denote the parallelogram determined by TA (u) = Au and TA ( v) = Av.
9. Prove that the area of TA ( P ) is given by I det Al (area of P ).
10. Let A be a 3 X 3 matrix and let P be the parallelepiped determined by the
vectors u, v, and w. Let TA ( P ) denote the parallelepiped determined by TA (u) = Au,
TA (v) = Av, and TA (w) = Aw. Prove that the volume of TA (P) is given by l det A I
(volume of P ).
The preceding problems illustrate that the determinant of a matrix captures what
the corresponding matrix transformation does to the area or volume of figures upon
which the transformation acts. (Although we have considered only certain types of fig­
ures, the result is perfectly general and can be made rigorous. We will not do so here.)
288
lines and Planes
Suppose we are given two distinct points (x1, y 1) and (x2 , y2 ) in the plane. There is a
unique line passing through these points, and its equation is of the form
ax + by + c = 0
Since the two given points are on this line, their coordinates satisfy this equation.
Thus,
ax1 + by 1 + c = 0
ax2 + by2 + c = 0
The three equations together can be viewed as a system of linear equations in the vari­
ables a, b, and c. Since there is a nontrivial solution (i.e., the line exists), the coefficient
matrix
cannot be invertible, by the Fundamental Theorem of Invertible Matrices. Conse­
quently, its determinant must be zero, by Theorem 4.6. Expanding this determinant
gives the equation of the line.
The equation of the line through the points (x1, y 1) and (x2 , y2 ) is given by
1 1 . Use the method described above to find the equation of the line through the
given points.
( a ) ( 2, 3 ) and ( - 1 , 0 )
( b ) ( 1 , 2 ) and ( 4, 3 )
12. Prove that the three points (xi, Y i) , ( x2 , y2 ), and (x3 , y3 ) are collinear (lie on
the same line) if and only if
X i Yi 1
Xz Y2 1 = 0
X3 Y3
1 3 . Show that the equation of the plane through the three noncollinear points
(x1, Y 1 , z 1 ), (x2 , y2 , z2 ), and (x3 , y3 , z3 ) is given by
x
Xi
Xz
X3
y z
Y1 Z i
=O
Y2 Zz
Y3 Z3
What happens if the three points are collinear? [Hint: Explain what happens when
row reduction is used to evaluate the determinant.]
289
14. Prove that the four points (x1, y 1 , z 1), (x2 , y2 , z2 ), (x3 , y3 , z3 ), and ( x4 , y4 , z4 ) are
coplanar (lie in the same plane) if and only if
X1
Xz
X3
X4
Y1
Y2
Y3
Y4
Z1
Zz
Z3
Z4
=O
Curve Filling
y
�+--+--+---+---. x
-2
2
4
6
Figure 4 . 1 3
When data arising from experimentation take the form of points (x, y) that can be
plotted in the plane, it is often of interest to find a relationship between the variables x
and y. Ideally, we would like to find a function whose graph passes through all of the
points. Sometimes all we want is an approximation (see Section 7. 3 ), but exact results
are also possible in certain situations.
1 5 . From Figure 4. 1 3 it appears as though we may be able to find a parabola
passing through the points A( - 1 , 10), B(O, 5), and C(3, 2). The equation of such a
parabola is of the form y = a + bx + cx 2 . By substituting the given points into this
equation, set up a system of three linear equations in the variables a, b, and c. Without
solving the system, use Theorem 4.6 to argue that it must have a unique solution. Then
solve the system to find the equation of the parabola in Figure 4. 13.
1 6. Use the method of Problem 1 5 to find the polynomials of degree at most 2
that pass through the following sets of points.
(a) A ( l , - 1 ), B ( 2 , 4), C(3, 3)
l
(b) A ( - 1 , - 3 ), B ( , - 1 ), C(3, 1 )
1 7. Generalizing from Problems 1 5 and 1 6, suppose a 1 , a 2 , and a 3 are distinct
real numbers. For any real numbers b 1 , b 2 , and b 3 , we want to show that there is a
unique quadratic with equation of the form y = a + bx + cx 2 passing through the
points (a1, b 1) , (a 2 , b 2 ), and (a 3 , b 3 ). Do this by demonstrating that the coefficient
matrix of the associated linear system has the determinant
a 1 ai
a 2 a� = ( a 2 - a 1 ) ( a 3 - a 1 ) ( a 3 - a z )
a 3 a�
which is necessarily nonzero. (Why?)
1 8. Let a 1 , a 2 , a 3 , and a 4 be distinct real numbers. Show that
a1
az
a3
a4
ai
a�
a�
a4z
ai
a 23
= (a 2 - a 1 ) ( a 3 - a 1 ) ( a4 - a 1 ) ( a 3 - a 2 ) ( a4 - a 2 ) ( a4 - a 3 ) * 0
a�
a 43
For any real numbers b 1 , b 2 , b 3 , and b4 , use this result to prove that there is a unique
cubic with equation y = a + bx + cx 2 + dx 3 passing through the four points (a1 , b 1) ,
(a 2 , b2 ), (a 3 , b 3 ) , and (a4 , b4 ) . (Do not actually solve for a, b, c, and d. )
290
19. Let a1, a 2 ,
•
•
•
, a n be n real numbers. Prove that
a 1 ai
a 1 a�
a 3 a�
a n1 - 1
a zn - 1
a 3n - 1
l
II (aj - a ;)
5' i < j -:5 n
a n a zn
a nn - 1
where TI1 :s i <J:s n (a1 - a;) means the product of all terms of the form (a1 - a;), where
i < j and both i and j are between 1 and n. [ The determinant of a matrix of this form
(or its transpose) is called a Vandermonde determinant, named after the French
mathematician A. T. Vandermonde ( 1 735- 1 796).]
Deduce that for any n points in the plane whose x-coordinates are all distinct,
there is a unique polynomial of degree n - 1 whose graph passes through the given
points.
291
292
Chapter 4 Eigenvalues and Eigenvectors
'
Eigenvalues a n d E i g envectors o t n x n M atrices
Now that we have defined the determinant o f an n X n matrix, we can continue our
discussion of eigenvalues and eigenvectors in a general context. Recall from Section 4. 1
that A is an eigenvalue of A if and only if A - AI is noninvertible. By Theorem 4.6, this
is true if and only if det(A - AI) = 0. To summarize:
The eigenvalues of a square matrix A are precisely the solutions A of the equation
det (A - AI) = 0
When we expand det(A - AI ), we get a polynomial in A, called the characteristic
- AI ) = 0 is called the characteristic equation
polynomial of A. The equation det(A
of A. For example, if A =
det (A - AI) =
l a -e A
[ : �],
its characteristic polynomial is
I
b
= (a - i\)(d - A) - be = A 2 - (a + d)i\ + (ad - be)
d - i\
If A is n X n, its characteristic polynomial will be of degree n. According to the Fun­
damental Theorem of Algebra (see Appendix D), a polynomial of degree n with real
or complex coefficients has at most n distinct roots. Applying this fact to the charac­
teristic polynomial, we see that an n X n matrix with real or complex entries has at
most n distinct eigenvalues.
Let's summarize the procedure we will follow (for now) to find the eigenvalues
and eigenvectors ( eigenspaces) of a matrix.
Let A be an n X n matrix.
1 . Compute the characteristic polynomial det(A
-
U ) of A.
2. Find the eigenvalues ofA by solving the characteristic equation det (A - AI ) = 0
for i\.
3. For each eigenvalue A, find the null space of the matrix A - i\I. This is
the eigenspace EA , the nonzero vectors of which are the eigenvectors of A
corresponding to A.
4. Find a basis for each eigenspace.
Exa m p l e 4 . 1 8
Find the eigenvalues and the corresponding eigenspaces of
Section 4.3 Eigenvalues and Eigenvectors of
Solution
n
X
n
Matrices
293
We follow the procedure outlined previously. The characteristic polynomial is
det (A - AI) =
0
1
0 - ;\
- ;\
1
2 - 5 4 - ;\
1
0
1
= - ;\
- 5 4 - ;\
2 4 - ;\
2
= - ;\(;\ - 4;\ + 5 ) - ( - 2 )
= - +4 2
+2
-I A
I 1
I
A3 A - SA
To find the eigenvalues, we need to solve the characteristic equation det(A - AI) = 0
for A. The characteristic polynomial factors as (A - 1 ) 2 (;\ - 2). (The Factor
- the characteristic equation is
Theorem is helpful here; see Appendix D.) Thus,
- (;\ - 1 ) 2 (;\ - 2) = 0, which clearly has solutions ;\ = 1 and ;\ = 2. Since ;\ = 1 is
a multiple root and A = 2 is a simple root, let us label them ;\1 = ;\ 2 = 1 and A3 = 2.
To find the eigenvectors corresponding to ;\1 = ;\ = 1 , we find the null space of
A - II =
Row reduction produces
� J n !]
[
]
n � : �]
n
[A - I I O J =
9--!>-
2
1
-1
-1
-5 4
-5
1 0
-1 1
0
-----+
-5 3
-1
1 -1
0
0
(We knew ;n odvan<e thot we mu" get ot lmt one mo rnw. Why?) Thu', x
=
�
[:: l'
in the eigenspace E1 if and only if x1 - x3 = 0 and x2 - x3 = 0. Setting the free variable
x3 = t, we see that x1 = t and x2 t, from which it follows that
-2
[A - 2I I O J =
So x
�
[:: l
A3 = 2, we find the null space of A - 21
[ � � � ----+ [ � =i �
i � i
To find the eigenvectors corresponding to
by row reduction:
1
-2
-5 2 0
; , ;n the e;gmpoce E, ;f ond only ;f x,
free variable x3 = t, we have
0 0
�
)x, ond x,
0 0
�
!x,. Sethng the
294
Chapter 4 Eigenvalues and Eigenvectors
where we have cleared denominators in the basis by multiplying through by the least
common denominator 4. (Why is this permissible?)
4
Remark
Notice that in Example 4. 1 8, A is a 3 X 3 matrix but has only two distinct
eigenvalues. However, if we count multiplicities, A has exactly three eigenvalues (A = 1
twice and A = 2 once). This is what the Fundamental Theorem of Algebra guarantees.
Let us define the algebraic multiplicity of an eigenvalue to be its multiplicity as a root
of the characteristic equation. Thus, A = 1 has algebraic multiplicity 2 and A = 2 has
algebraic multiplicity 1 .
Next notice that each eigenspace has a basis consisting of just one vector. In other
words, dim E 1 = dim E2 = 1 . Let us define the geometric multiplicity of an eigenvalue
A to be dim E;"' the dimension of its corresponding eigenspace. As you will see in
Section 4.4, a comparison of these two notions of multiplicity is important.
Exa m p l e 4 . 1 9
Find the eigenvalues and the corresponding eigenspaces of
A=
Solulion
[ -� � -�]
1 0
-1
The characteristic equation is
-1 - A
0 = det (A - AI) =
3
1
1
1
-1 - A
-3 = -A
1
-1 - A
0
-A
0
Hence, the eigenvalues are A 1 = A 2 = 0 and A 3 = -2. Thus, the eigenvalue 0 has alge­
braic multiplicity 2 and the eigenvalue - 2 has algebraic multiplicity 1 .
For A 1 = A 2 = 0 , we compute
[A - OJ I O J
[A I OJ
0
0
0
&om which it follow' th"t '"' eigenwdm x
-� �i [� � J
[ :: l
-1 0
�
�
in E0 "ti'fi" x,
both x2 and x3 are free. Setting x2 = s and x3 = t, we have
For A 3 = - 2,
[A - ( - 2)I I O J
[A + 2I I O J
0 0
-1 O
0 0
0 0
�
x,. Thecefoce,
Section 4.3 Eigenvalues and Eigenvectors of
n
X
n
Matrices
295
so x3 = t is free and x 1 = -x3 = - t and x2 = 3x3 = 3t. Consequently,
It follows that ,\ 1 = ,\ 2 = 0 has geometric multiplicity 2 and ,\ 3 = - 2 has geometric
multiplicity 1 . (Note that the algebraic multiplicity equals the geometric multiplicity
for each eigenvalue.)
4
In some situations, the eigenvalues of a matrix are very easy to find. If A is a
triangular matrix, then so is A - AI, and Theorem 4.2 says that det (A - AI ) is just
the product of the diagonal entries. This implies that the characteristic equation of
a triangular matrix is
(a ll - .\ )(a 22 - A) · · · (a nn - A) = 0
,
from which it follows immediately that the eigenvalues are A 1 = a 11 , A 2 = a 22 ,
A n = a n,,. We summarize this result as a theorem and illustrate it with an example.
•
Theorem 4 . 1 5
Exa m p l e 4 . 2 0
.
•
The eigenvalues of a triangular matrix are the entries on its main diagonal.
The eigenvalues of
are ,\ 1 = 2, ,\ 2 = 1, ,\ 3 = 3, and ,\4 = - 2, by Theorem 4. 15. Indeed, the characteristic
polynomial is just (2 - ,\)( 1 - ,\)(3 - ,\)( - 2 - A).
Note that diagonal matrices are a special case of Theorem 4.15. In fact, a diagonal
matrix is both upper and lower triangular.
Eigenvalues capture much important information about the behavior of a matrix.
Once we know the eigenvalues of a matrix, we can deduce a great many things without
doing any more work. The next theorem is one of the most important in this regard.
Theorem 4 . 1 6
A
square matrix A is invertible if and only if 0 is not an eigenvalue of A.
Let A be a square matrix. By Theorem 4.6, A is invertible if and only if
<let A * 0. But <let A * 0 is equivalent to <let (A - O J ) * 0, which says that 0 is not a
root of the characteristic equation of A (i.e., 0 is not an eigenvalue of A).
Proof
We can now extend the Fundamental Theorem of Invertible Matrices to include
results we have proved in this chapter.
296
Chapter 4 Eigenvalues and Eigenvectors
Theorem 4 . 1 1
The Fundamental Theorem of Invertible Matrices: Version 3
Let A be an n X n matrix. The following statements are equivalent:
a.
b.
c.
d.
e.
f.
g.
h.
i.
j.
k.
A is invertible.
A x = b has a unique solution for every b in IJ�r.
A x = 0 has only the trivial solution.
The reduced row echelon form of A is In A is a product of elementary matrices.
rank(A) = n
nullity(A) = 0
The column vectors of A are linearly independent.
The column vectors of A span IJ�r.
The column vectors of A form a basis for IJ�r.
The row vectors of A are linearly independent.
1. The row vectors of A span !R n .
m. The row vectors of A form a basis for !R n .
n. det A * 0
o. 0 is not an eigenvalue of A.
The equivalence (a) � (n) is Theorem 4.6, and we just proved (a) � (o) in
Theorem 4. 16.
Proof
There are nice formulas for the eigenvalues of the powers and inverses of a matrix.
Theorem 4 . 1 8
Let A be a square matrix with eigenvalue A and corresponding eigenvector x.
a. For any positive integer n, A n is an eigenvalue of A n with corresponding
eigenvector x.
b. IfA is invertible, then 1 /A is an eigenvalue ofA - i with corresponding eigenvector x.
c. If A is invertible, then for any integer n, A n is an eigenvalue of A n with corre­
sponding eigenvector x.
Proof
We are given that A x = Ax.
(a) We proceed by induction on n. For n = 1 , the result is just what has been given.
Assume the result is true for n = k. That is, assume that, for some positive integer k,
A k x = A k x. We must now prove the result for n = k + 1 . But
A k + 1 x = A (A kx) = A(Akx)
by the induction hypothesis. Using property ( d) ofTheorem 3.3, we have
1
A(Akx) = A k (Ax) = Ak (Ax) = A k +I x
Thus, A k + l x = A k + x, as required. By induction, the result is true for all integers
n 2: 1 .
(b) You are asked to prove this property in Exercise 13.
(c) You are asked to prove this property in Exercise 14.
The next example shows one application of this theorem.
Exa m p l e 4 . 2 1
Compute
Solution
[o
2
Section 4.3 Eigenvalues and Eigenvectors of
] []
1 10 5
.
1 1
Let A
[�]
=
[� � ]
eigenvalues of A are A 1
and v2
�
=
=
and x
=
[�l
n
X
n
Matrices
291
�
[_ ]
then what we want to find is A 1 0 x. T e
- 1 and A 2 = 2, with corresponding eigenvectors v 1
. That is,
=
1
(Check this.) Since {v1 , v2 } forms a basis for IR 2 (why?), we can write x as a linear com­
bination of v1 and v2 . Indeed, as is easily checked, x 3v1 + 2v2 .
Therefore, using Theorem 4.1 8(a), we have
10
A 1 0x A 10 ( 3v1 + 2v2 ) 3(A 10v 1 ) + 2 (A v2 )
3(A: 0 ) v1 + 2 ( A� 0 ) v2
1
3 + 211
205 1
3 ( - 1 ) 10
+ 2 ( 2 1 0)
1
2
-1
2
-3 + 2
4093
=
=
=
=
=
[
l]
[]
=
] [ ]
[
=
This is certainly a lot easier than computing A 1 0 first; in fact, there are no matrix
multiplications at all!
When it can be used, the method of Example 4.2 1 is quite general. We summarize
it as the following theorem, which you are asked to prove in Exercise 42.
Theorem 4 . 1 9
Suppose the n X n matrix A has eigenvectors v1 , v2 ,
, vm with corresponding
eigenvalues A 1 , A 2 , . . . , A m . If x is a vector in !R n that can be expressed as a linear
combination of these eigenvectors-say,
•
.
.
then, for any integer k,
Warning
The catch here is the "if" in the second sentence. There is absolutely
no guarantee that such a linear combination is possible. The best possible situation
would be if there were a basis of !R n consisting of eigenvectors of A; we will explore
this possibility further in the next section. As a step in that direction, however, we
have the following theorem, which states that eigenvectors corresponding to distinct
eigenvalues are linearly independent.
Theorem 4 . 2 0
Let A be an n X n matrix and let A 1 , A 2 , . . . , A m be distinct eigenvalues of A with cor­
responding eigenvectors v1 , v2 , , vm . Then v1 , v2 , , vm are linearly independent.
.
.
•
.
.
•
Chapter 4 Eigenvalues and Eigenvectors
298
The proof is indirect. We will assume that v1 , v2 , , vm are linearly dependent
and show that this assumption leads to a contradiction.
If v1 , v2 , , vm are linearly dependent, then one of these vectors must be express­
ible as a linear combination of the previous ones. Let vk + 1 be the first of the vectors V;
that can be so expressed. In other words, v1 , v2 ,
, vk are linearly independent, but
there are scalars c 1 , c2 , , ck such that
(1)
Proof
•
•
.
.
.
•
•
•
.
.
.
•
Multiplying both sides of Equation ( 1 ) by A from the left and using the fact that Av; =
A;V; for each i, we have
A k+ i Vk+ i = Avk+ 1 = A(c 1v1 + c2v2 + · · · + ckvk )
= C 1AV1 + C2AV2 + . . . + ckAvk
= C 1 A 1 V1 + C2 A 2V2 + . . . + ck ,\ kvk
Now we multiply both sides of Equation ( 1 ) by A k + l to get
(2)
(3)
When we subtract Equation (3) from Equation (2), we obtain
0 = c 1 (,\ 1 - A k+1 )v1 + c 2 (A 2 - A k+ 1 )v2 + · · · + ck (,\ k - A k+1 )vk
The linear independence of v 1 , v2 , , vk implies that
•
.
•
c 1 (,\ 1 - A k+1 ) = c2 (A 2 - A k+1 ) = · · · = ck (,\ k - A k+1 ) = 0
Since the eigenvalues A; are all distinct, the terms in parentheses (A; - ,\ k + 1) ,
i = 1 , . . . , k, are all nonzero. Hence, c 1 = c2 = · · · = ck = 0. This implies that
vk+1 = c 1 v1 + c2v2 + · · · + ckvk = Ov1 + Ov2 + · · · + Ovk = 0
which is impossible, since the eigenvector Vk + 1 cannot be zero. Thus, we have a
contradiction, which means that our assumption that v1 , v2 , . . . , vm are linearly
dependent is false. It follows that v1 , v2 , . . . , vm must be linearly independent.
..
I
Exercises 4 . 3
In Exercises 1 - 12, compute (a) the characteristic poly nomial
of A, (b) the eigenvalues of A, (c) a basis for each eigenspace
ofA, and (d) the algebraic and geometric multiplicity of
each eigenvalue.
2
1 3
1. A
2. A =
-1 0
-2 6
=
3. A
�
5. A
�
[ ]
[� - � :J
[ - � - : :J
4. A
�
6. A
�
[ l]
[� : �]
[� -� :J
Section 4.3 Eigenvalues and Eigenvectors of
13. Prove Theorem 4. 1 8(b).
14. Prove Theorem 4. 1 8(c). [Hint: Combine the proofs of
parts (a) and (b) and see the fourth Remark following
Theorem 3.9 (page 1 69).]
[_�]
[�]
In Exercises 1 5 and 1 6, A is a 2 X 2 matrix with eigenvec­
tors v1 =
and v2 =
corresponding to eigenvalues
[ � ].
A 1 = t and A 2 = 2, respectively, and x =
15. Find A 1 0x.
16. Find Ak x. What happens as k becomes large (i.e., k ---+ oo)?
In Exercises 1 7 and 1 8, A is a 3 X 3 matrix with eigenvectors
v,
�
[H [J
v,
�
and v,
�
[ :J
wm,pond;ng to
eigenvalues A 1 = -L A2 = t , and A 3 = 1 , respectively, and
Fm
17. Find A 2 0 x.
18. Find Ak x. What happens as k becomes large (i.e., k ---+ oo)?
19. (a) Show that, for any square matrix A, A T and A have
the same characteristic polynomial and hence the
same eigenvalues.
(b) Give an example of a 2 X 2 matrix A for which A T
and A have different eigenspaces.
20. Let A be a nilpotent matrix (that is, A m = 0 for some
m > 1 ) . Show that A = 0 is the only eigenvalue of A.
21. Let A be an idempotent matrix (that is, A 2 = A). Show that
A = O and A = 1 are the only possible eigenvalues ofA.
22. If v is an eigenvector of A with corresponding eigen value A and c is a scalar, show that v is an eigenvector
of A - cI with corresponding eigenvalue A - c.
23. ( a) Find the eigenvalues and eigenspaces of
A
=
[� �]
(b) Using Theorem 4. 1 8 and Exercise 22, find the eigen­
values and eigenspaces of A - i , A - 2I, and A + 2I.
24. Let A and B be n X n matrices with eigenvalues A and
IL, respectively.
( a) Give an example to show that A + IL need not be
an eigenvalue of A + B.
(b) Give an example to show that AIL need not be an
eigenvalue of AB.
n
X
n
Matrices
299
(c) Suppose A and IL correspond to the same eigen­
vector x. Show that, in this case, A + IL is an eigen value of A + B and AIL is an eigenvalue of AB.
25. If A and B are two row equivalent matrices, do they
necessarily have the same eigenvalues? Either prove
that they do or give a counterexample.
Let p(x) be the poly nomial
The companion matrix ofp(x) is the n X n matrix
- an - I - an - 2
C(p) =
1
0
0
0
- a 1 - ao
0
0
0
0
0
0
0
0
(4)
26. Find the companion matrix ofp(x) = x 2 - 7x + 12 and
then find the characteristic polynomial of C( p).
27. Find the companion matrix of p(x) = x 3 + 3x 2 4x + 1 2 and then find the characteristic polynomial
of C(p).
28. ( a) Show that the companion matrix C (p) ofp (x) =
x 2 + ax + b has characteristic polynomial
A 2 + aA + b.
(b) Show that if A is an eigenvalue of the companion
matrix C(p) in part (a), then
of C ( p) corresponding to A.
[ �]
is an eigenvector
29. (a) Show that the companion matrix C(p) of p (x) =
x3 + ax 2 + bx + c has characteristic polynomial
- (A 3 + aA 2 + bA + c).
(b) Show that if A is an eigenvalue of the companion
[� l
2
matrix C(p) in part (a), then
of C ( p) corresponding to A.
is an eigenvector
1
30. Construct a nontriangular 2 X 2 matrix with eigen­
values 2 and 5. [Hint: Use Exercise 28.]
31. Construct a nontriangular 3 X 3 matrix with eigen­
values - 2, 1 , and 3. [Hint: Use Exercise 29.]
32. ( a) Use mathematical induction to prove that, for
n 2, theI companion matrix C(p) of p (x) = x n +
a n _ 1 x n - + · · · + a 1 x + a 0 has characteristic
polynomial ( - 1 ) np (A). [Hint: Expand by cofactors
along the last column. You may find it helpful to
introduce the polynomial q ( x) = (p ( x) - a 0 ) /x .]
2
Chapter 4 Eigenvalues and Eigenvectors
300
(b) Show that if A is an eigenvalue of the companion
matrix C(p) in Equation (4), then an eigenvector
corresponding to A is given by
An - I
,t n - 2
Ifp(x) = x n + a n _ 1 x n - ! +
+ a 1 x + a0 and A is a
square matrix, we can define a square matrix p(A) by
+ a 1A + a0I
p(A) = A n + a n _ 1A n - i +
An important theorem in advanced linear algebra says that
if cA (,\) is the characteristic poly nomial of the matrix A,
then cA (A) = 0 (in words, every matrix satisfies its charac­
teristic equation). This is the celebrated Cayley-Hamilton
Theorem, named after Arthur Cayley ( 1 82 1 - 1 895),
pictured below, and Sir William Rowan Hamilton (see
page 2). Cayley proved this theorem in 1 858. Hamilton
discovered it, independently, in his work on quaternions, a
generalization of the complex numbers.
·
·
·
·
·
·
and
A3 =
=
=
=
AA 2 = A( - aA - bl)
- aA 2 - bA
- a( - aA - bl) - bA
(a 2 - b )A + ab I
It is easy to see that by continuing in this fashion we can
express any positive power of A as a linear combination of I and A. From A 2 + aA + bl = 0, we also obtain
A(A + al ) = - bl, so
1
a
A - 1 = - -A - - I
b
b
provided b * 0.
35. For the matrix A in Exercise 33, use the Cayley­
Hamilton Theorem to compute A 2 , A 3 , and A 4 by
expressing each as a linear combination of I and A.
36. For the matrix A in Exercise 34, use the Cayley­
Hamilton Theorem to compute A 3 and A 4 by express­
ing each as a linear combination of I, A, and A 2 .
37. For the matrix A in Exercise 33, use the Cayley­
Hamilton Theorem to compute A - i and A - 2 by
expressing each as a linear combination of I and A.
38. For the matrix A in Exercise 34, use the Cayley­
Hamilton Theorem to compute A - i and A - 2 by
expressing each as a linear combination of I, A, and A 2 .
39. Show that if the square matrix A can be partitioned as
[-ij-i-;-]
A =
c
j
[ ]
1 -1
.
2
3
That is, find the characteristic polynomial cA ( ,\) of A
and show that cA(A) = 0.
34. Verify the Cayley-Hamilton Theorem for
33. Verify the Cayley-Hamilton Theorem for A =
A�
[� � :J
The Cayley-Hamilton Theorem can be used to calculate
powers and inverses of matrices. For example, ifA is a 2 X 2
matrix with characteristic poly nomial cA (,\) = ,\ 2 + a,\ + b,
then A 2 + aA + bl = 0, so
A 2 = - aA - bl
where P and S are square matrices, then the character­
istic polynomial of A is cA (,\) = cp(A) c5(,\). [Hint: Use
Exercise 69 in Section 4.2.]
40. Let A 1 , A 2 , . . . , A n be a complete set of eigenvalues (rep­
etitions included) of the n X n matrix A. Prove that
det (A ) = A 1 A 2 A n and
tr (A ) = A 1 + A 2 +
+ An
•
•
·
·
·
·
[Hint: The characteristic polynomial of A factors as
det (A - AI) = ( - l) n (,\ - ,\ 1 ) (,\ - A 2 ) (,\ - A n )
Find the constant term and the coefficient of A n - ! on
•
•
•
the left and right sides of this equation.]
41. Let A and B be n X n matrices. Prove that the sum of all
the eigenvalues of A + B is the sum of all the eigenval­
ues of A and B individually. Prove that the product of all
the eigenvalues of AB is the product of all the eigenval­
ues of A and B individually. (Compare this exercise with
Exercise 24.)
42. Prove Theorem 4. 19.
Section 4.4 Similarity and Diagonalization
Writi n g Proiect
301
The History of Eigenvalues
Like much oflinear algebra, the way the topic of eigenvalues is taught today does not
correspond to its historical development. Eigenvalues arose out of problems in sys­
tems of differential equations before the concept of a matrix was even formulated.
Write a report on the historical development of eigenvalues. Describe the types
of mathematical problems in which they originally arose. Who were some of the
key mathematicians involved with these problems? How did the terminology for
eigenvalues change over time?
1 . Thomas Hawkins, Cauchy and the Spectral Theory of Matrices, Historia Math­
ematica 2 ( 1 975), pp. 1 -29.
2. Victor J. Katz, A History of Mathematics: An Introduction (Third Edition) (Read­
ing, MA: Addison Wesley Longman, 2008) .
3. Morris Kline, Mathematical Thought from Ancient to Modern Times (Oxford:
Oxford University Press, 1 972) .
S i m i l a rilv and D i a u o n alizalion
As you saw in the last section, triangular and diagonal matrices are nice in the sense
that their eigenvalues are transparently displayed. It would be pleasant if we could
relate a given square matrix to a triangular or diagonal one in such a way that they
had exactly the same eigenvalues. Of course, we already know one procedure for con verting a square matrix into triangular form-namely, Gaussian elimination. Unfor­
tunately, this process does not preserve the eigenvalues of the matrix. In this section,
we consider a different sort of transformation of a matrix that does behave well with
respect to eigenvalues.
Similar Malrices
D e fi n it i o n Let A and B be n x n matrices. We say that A i s similar to B if
there is an invertible n X n matrix P such that P - 1AP = B. If A is similar to B,
we write A � B.
R e m a rks
If A � B, we can write, equivalently, that A = PBP - 1 or AP = PB.
•
Similarity is a relation on square matrices in the same sense that "less than or
equal to" is a relation on the integers. Note that there is a direction (or order) implicit
in the definition. Just as a :::; b does not necessarily imply b :::; a, we should not assume
that A � B implies B � A. (In fact, this is true, as we will prove in the next theorem,
but it does not follow immediately from the definition.)
•
The matrix P depends on A and B. It is not unique for a given pair of similar ma­
trices A and B. To see this, simply take A = B = I, in which case I � I, since p - i IP = I
for any invertible matrix P.
•
302
Chapter 4 Eigenvalues and Eigenvectors
Exa m p l e 4 . 2 2
Let A =
[ J [ J
[ � - � J [ � -l J [ l J [ 1 -l J [ O J
[ -l J
1
2
1
and B =
0 -1
-2
1
-
0
. Then A � B , since
-1
3
-1
-1
-
1
1
1
-2 - 1
1
. (Note that it is not necessary to compute P - 1 .
1
1
See the first Remark before Example 4.22. )
Thus, AP = PB with P =
Theorem 4 . 2 1
Let A, B, and C be n X n matrices.
4
a. A � A
b. If A � B, then B � A.
c. If A � B and B � C, then A � C.
(a) This property follows from the fact that r 1AI = A.
(b) If A � B, then P - 1AP = B for some invertible matrix P. As noted in the first
Remark on the previous page, this is equivalent to PBP - 1 = A. Setting Q = P - 1 , we
have Q - 1 BQ = (P - 1 ) - 1 BP - 1 = PBP - 1 = A. Therefore, by definition, B � A.
(c) You are asked to prove property (c) in Exercise 30.
Proof
Remark
Any relation satisfying the three properties of Theorem 4.2 1 is called
an equivalence relation. Equivalence relations arise frequently in mathematics, and
objects that are related via an equivalence relation usually share important properties.
We are about to see that this is true of similar matrices.
Theorem 4 . 2 2
Let A and B be n X n matrices with A � B. Then
a.
b.
c.
d.
e.
f.
g.
det A = det B
A is invertible if and only if B is invertible.
A and B have the same rank.
A and B have the same characteristic polynomial.
A and B have the same eigenvalues.
A m � B m for all integers m 2::: 0.
If A is invertible, then A m � B m for all integers m .
Proof We prove (a) and (d) and leave the remaining properties as exercises. If A � B,
then P - 1AP = B for some invertible matrix P.
(a) Taking determinants of both sides, we have
det B = det ( P - 1AP) = ( det P - 1 ) ( det A ) ( det P)
1
= -- ( det A ) ( det P) = det A
det P
(d) The characteristic polynomial of B is
det ( B - AI) = det (p - i AP - AI)
= det ( P - 1 AP - AP- 1 IP)
( )
Section 4.4 Similarity and Diagonalization
303
det ( P - 1 AP - P - 1 ( AI) P)
det ( P - 1 (A - AI) P) = det (A - AI)
with the last step following as in (a) . Thus, det(B - i\I) = det(A - i\I); that is, the
characteristic polynomials of B and A are the same.
Remark
�]
�]
Two matrices may have properties (a) through ( e) (and more) in common
and yet still not be similar. For example, A =
[�
and B =
[�
both have de­
terminant 1 and rank 2, are invertible, and have characteristic polynomial ( 1 - i\) 2
and eigenvalues i\ 1 = i\ 2 = 1 . But A is not similar to B, since P - 1 AP = P - 1 IP =
I * B for any invertible matrix P.
Theorem 4.22 is most useful in showing that two matrices are not similar, since
A and B cannot be similar if any of properties (a) through (e) fails.
Exa m p l e 4 . 2 3
[ � �] and B = [� � ] are not similar, since det A = - 3 but det B = 3.
1 3
1
1
] are not similar, since the characteristic polyno]
and B = [
(b) A = [
2 2
3 -1
(a) A =
mial of A is i\ 2 - 3i\ - 4 while that of B is i\ 2 - 4. (Check this.) Note that A and B do
have the same determinant and rank, however.
Diagonalization
The best possible situation is when a square matrix is similar to a diagonal matrix.
As you are about to see, whether a matrix is diagonalizable is closely related to the
eigenvalues and eigenvectors of the matrix.
An n x n matrix A is diagonalizable if there is a diagonal matrix
D such that A is similar to D -that is, if there is an invertible n X n matrix P such
that P - 1AP = D.
D e fi n it i o n
Exa m p l e 4 . 2 4
[l ]
[l ]
[
3
3
4 0
A=
is diagonalizable since, if P =
and D =
, then
0 -1
2 2
1 -2
1
P - AP = D, as can be easily checked. (Actually, it is faster to check the equivalent
statement AP = PD, since it does not require finding P - 1 .)
]
-+
Example 4.24 begs the question of where matrices P and D came from. Observe
that the diagonal entries 4 and - 1 of D are the eigenvalues of A, since they are the
roots of its characteristic polynomial, which we found in Example 4.23(b ). The origin
of matrix P is less obvious, but, as we are about to demonstrate, its entries are obtained
from the eigenvectors of A. Theorem 4.23 makes this connection precise.
304
Chapter 4 Eigenvalues and Eigenvectors
Theorem 4 . 2 3
Let A b e an n X n matrix. Then A is diagonalizable if and only if A has n linearly
independent eigenvectors.
More precisely, there exist an invertible matrix P and a diagonal matrix D such
that P - 1AP = D if and only if the columns of P are n linearly independent eigen­
vectors of A and the diagonal entries of D are the eigenvalues of A corresponding
to the eigenvectors in P in the same order.
Suppose first that A is similar to the diagonal matrix D via P - 1AP = D or,
equivalently, AP = PD. Let the columns of P be p1 , p2 ,
, P n and let the diagonal
entries of D be A 1 , A 2 , , A n . Then
A, 0
Az
A [ p 1 Pz
(1)
[ p l Pz
p. ]
Pn l
Proof
•
.
or
.
.
.
.
[ A p 1 A pz
.
.
A p n ] = [ A 1P1 A z pz
.
[
.
.
•
!
]J
0
A nPn ]
(2)
where the right-hand side is just the column-row representation of the product PD.
Equating columns, we have
A pn = A nPn
A p 1 = A 1P1 , A p z = A z pz ,
which proves that the column vectors of P are eigenvectors of A whose corresponding
eigenvalues are the diagonal entries of D in the same order. Since P is invertible, its col­
umns are linearly independent, by the Fundamental Theorem of lnvertible Matrices.
Conversely, if A has n linearly independent eigenvectors p1 , p2 ,
, P n with cor­
responding eigenvalues A 1 , A 2 , , A n , respectively, then
A pn = A nPn
A p 1 = A 1P1 , A p z = A z pz ,
This implies Equation ( 2) above, which is equivalent to Equation ( 1 ) . Consequently,
if we take P to be the n X n matrix with columns p1 , p2 , . . . , P n ' then Equation ( 1 )
becomes AP = PD. Since the columns of P are linearly independent, the Fundamental
Theorem of lnvertible Matrices implies that P is invertible, so P - 1AP = D ; that is, A
·
·
·
,
.
.
•
•
.
•
·
·
·
,
is diagonalizable.
Exa m p l e 4 . 2 5
If possible, find a matrix P that diagonalizes
A
� [� -� :J
We studied this matrix in Example 4. 1 8, where we discovered that it has
eigenvalues A 1 = A 2 = 1 and A 3 = 2. The eigenspaces have the following bases:
Solulion
Fod , � A, � 1 , E , has hasis
Fod, � 2, E, has basis
[n
[ :J
Section 4.4 Similarity and Diagonalization
305
Since all other eigenvectors are just multiples of one of these two basis vectors, there
cannot be three linearly independent eigenvectors. By Theorem 4.23, therefore, A is
not diagonalizable.
Exa m p l e 4 . 2 6
4
[� - � � =i l
If possible, find a matrix P that diagonalizes
A
Solution
This is the matrix of Example 4. 19. There, we found that the eigenvalues of
A are A 1 = A 2 = 0 and ,\ 3 = - 2, with the following bases for the eigenspaces:
Fod ,
Fod,
� �
�
A,
�m �m
[� - :J
0, E 0 h" b"i' p,
- 2, E_, h" b"i' P ;
ond p,
It is straightforward to check that these three vectors are linearly independent. Thus,
if we take
then P is invertible. Furthermore,
as can be easily checked. (If you are checking by hand, it is much easier to check the
equivalent equation AP = PD.)
R e m a rks
When there are enough eigenvectors, they can be placed into the columns of P
in any order. However, the eigenvalues will come up on the diagonal of D in the same
order as their corresponding eigenvectors in P. For example, if we had chosen
•
then we would have found
306
Chapter 4 Eigenvalues and Eigenvectors
In Example 4.26, you were asked to check that the eigenvectors p1, p 2 , and p 3
were linearly independent. Was it necessary to check this? We knew that { p1, p2 } was
linearly independent, since it was a basis for the eigenspace E0 We also knew that the
sets { p1, p 3 } and { p2 , p 3 } were linearly independent, by Theorem 4.20. But we could not
conclude from this information that { p 1 , p2 , p 3 } was linearly independent. The next
theorem, however, guarantees that linear independence is preserved when the bases of
different eigenspaces are combined.
•
•
Theorem 4 . 2 4
Let A be an n X n matrix and let A 1 , A 2 , , A k be distinct eigenvalues of A. If B; is
a basis for the eigenspace E;v then B B1 U B 2 U · · · UB k (i.e., the total collection
of basis vectors for all of the eigenspaces) is linearly independent.
•
•
•
=
Proof
Let Bi
=
{v;1 , v;2 ,
•
.
•
, v; n ) for i
=
1, . . . , k. We have to show that
is linearly independent. Suppose some nontrivial linear combination of these vectors
is the zero vector-say,
( C 1 1 V1 1 + . . ' + C 1 n,V1 n ) + ( C2 1 V2 1 + . ' . + C2 n2V2 n) + . ' . + ( ck! Vk l + . . . + Ckn,Vkn) 0
(3)
Denoting the sums in parentheses by x1 , x2 , . . . xk, we can write Equation (3) as
x 1 + x2 +
(4)
+ xk 0
=
·
�
·
·
=
Now each x, is in EA, (why?) and so either is an eigenvector corresponding to A; or
is 0. But, since the eigenvalues A; are distinct, if any of the factors X; is an eigenvector,
they are linearly independent, by Theorem 4.20. Yet Equation ( 4) is a linear depen­
dence relationship; this is a contradiction. We conclude that Equation (3) must be
trivial; that is, all of its coefficients are zero. Hence, B is linearly independent.
There is one case in which diagonalizability is automatic: an n X n matrix with
n distinct eigenvalues.
Theorem 4 . 2 5
If A is an n X n matrix with n distinct eigenvalues, then A is diagonalizable.
Let v1, v2 , . . . , vn be eigenvectors corresponding to the n distinct eigenvalues
of A. (Why could there not be more than n such eigenvectors?) By Theorem 4.20,
v1 , v2 , . . . , vn are linearly independent, so, by Theorem 4.23, A is diagonalizable.
Proof
�
Exa m p l e 4 . 2 1
The matrix
has eigenvalues A 1 2, A 2 5, and A 3 - 1 , by Theorem 4.15. Since these are three
distinct eigenvalues for a 3 X 3 matrix, A is diagonalizable, by Theorem 4.25. (If we actu­
1
ally require a matrix P such that p - AP is diagonal, we must still compute bases for the
eigenspaces, as in Example 4. 1 9 and Example 4.26 above.)
=
=
=
4
Section 4.4 Similarity and Diagonalization
301
The final theorem of this section is an important result that characterizes diagonaliz­
able matrices in terms of the two notions of multiplicity that were introduced following
Example 4. 1 8. It gives precise conditions under which an n X n matrix can be diago­
nalized, even when it has fewer than n eigenvalues, as in Example 4.26. We first prove a
lemma that holds whether or not a matrix is diagonalizable.
Le m m a 4 . 2 6
A
If is an n X n matrix, then the geometric multiplicity of each eigenvalue is less
than or equal to its algebraic multiplicity.
A
A
Q
Suppose 1 is an eigenvalue of with geometric multiplicity p; that is,
dim EA , = p. Specifically, let EA , have basis B 1 = {v1 , v2 , , vp }. Let be any invertible
n X n matrix having v1 , v2 , , vp as its first p columns - say,
Proof
Q ( Y1
.
•
=
or, as a partitioned matrix,
·
·
Yp
·
Let
C
U
Yp + l
Q [U VJ
=
where is p X n.
Since the columns of
also have
.
.
•
.
. . .
Yn
J
i
are eigenvectors corresponding to A 1 ,
AU A U.
= 1
CU= CV= DU= DV
�!.l:!!._l-�!.l:Y] LA[�_11-D�_f!U:_�-D�!.l:AVYJ [�-1�!'.L!DAV
��-�-]
Q- 1AQ [�lD ] A [ U :' V J [-DAU!DAV
1
(
)
(
)
det
Q
A
Q
A
P
V
A
AI
A
1
Q- 1AQ - AI) A,
Q- 1AQ,
A1
A
A ... , Ak.
A
A
from which we obtain
=
IP ,
0, and
0,
=
= In- p · Therefore,
=
=
By Exercise 69 in Section 4.2, it follows that
=
We
det ( DA
0
I)
(5)
But det (
is the characteristic polynomial of
which is the same
as the characteristic polynomial of by Theorem 4.22(d) . Thus, Equation (5) implies
that the algebraic multiplicity of is at least p, its geometric multiplicity.
Theorem 4 . 2 1
The Diagonalization Theorem
Let be an n X n matrix whose distinct eigenvalues are 1 , A 2 ,
statements are equivalent:
The following
a. is diagonalizable.
b. The union B of the bases of the eigenspaces of (as in Theorem 4.24) contains
n vectors.
c. The algebraic multiplicity of each eigenvalue equals its geometric multiplicity.
308
Chapter 4 Eigenvalues and Eigenvectors
a)=? (ebm) I4.f2A3.isIfdiagonalof thieszable eie,gtenvect
hen it ohasrs corlirneearspondly intdependent
eigenvec­
ttohrsen, bycont(Theor
o
t
h
e
ei
g
enval
ueare A;ly,
a
i
n
s
at
l
e
as
t
vect
o
r
s
.
(
W
e
al
r
e
ady
know
t
h
at
t
h
es
e
vect
o
r
s
ar
e
l
i
n
ng that micontghtaprinesventat letashemt vectfromorbeis. nButg a, basby Theor
is for emis4.th2at4,
tinhdependent
eyis amilingearht lnoty; tihnesdependent
ponlanyitt.h) iThus,
(multbip)l=?icit(yc)ofLetA; bethe geometBy setLemma
ricinmul4.tihence,
p26,licid;ty ioft contA; beforaidn;s==exactdi1,m...ly andvectNowloerst t.ashesualmegebrthaatic
property (b) holds. Then we also have
iplicities ofy, the
eiButgenvalIt foluloeswsofthAati·sd·j1·ust tdh2e degr=· · ·eesiofndcekth=tehchare suamcteofristth·iec· pol·algyebrnomiaiwhic amullcofhtA-namel
implies that
(6)
Uscanindeduce
g Lemmathat4.each26 agaisummand
n, we knowin Equatthat ion-(6)d;is2:zer0 ofor; that=is1,, ...= d; from
whi1, ...ch we
fo
r
=
ity has andd1 thde2 geometridck mul= tip1 licity d; are
equal(c)for==?eachvect(a) Ioeifrtgsh,envalwhie alcguhebrearA;aeicoflinmulA,earttliyhpenliincdependent
Thus,etmhes4.e2ar3.e
linearly independent eigenvectors of A, and A is ,dibyagonalTheorizeablm e4., 2by4.Theor
n
Proof
�
B;
n;
n;
n;
B
B
n
!R n ;
:s m;
m;.
m1 + m 2 +
+ mk
n
Exa m p l e 4 . 2 8
n
EA ;
, k.
i
n,
+ mk
+
EA ;
+
+
m1 + m 2 +
+ m k,
m;
i
B
n
m;
+
+
n.
m;
· · ·
, k,
m + m2 +
+
, k.
· · ·
(a) The matrix A = [ 2� - �5 4� i from Example 4. 1 8 has two distinct eigenvalues,
A2 1but= Ageomet
2 = 1 andric mulA3 t=ipl2.icSiitynce1, Athies einotgenvaldiagonalue A1iz=ablAe2, by= 1thhase DialaggonalebraiiczatmuliontipTheo­licity
rem. (See also Example-4.25.)
(b) The matrix A = [ �l 0� - �1 ] fromExample 4. 1 9also hastwodistincteigenvaltipluicesit,yA2,1 =andA2 t=he0eiandgenvalA3 =ue--2.2Thehas eialggenval
uc eand0 hasgeomet
algebrriacicmulandtigeomet
ricThusmul­,
ebr
a
i
p
l
i
c
i
t
y
1.
tfihnidis matngsriinxExampl
diagonale 4.i2z6.abl) e, by the Diagonalization Theorem. (This agrees with our
tionWeof tconcl
he powerude sthofis asectmatiornixwi. th an application of diagonalization to the computa­
Compute A10 if A = [ � � ].
In Example 4.2 1, we found that this matrix ]has eigenvalues]. A1 = - 1
and A2 = 2, with corresponding eigenvectors = [ _ � and = [ � It follows
-
is
Exa m p l e 4 . 2 9
i
Solution
v1
v2
Section 4.4 Similarity and Diagonalization
309
P-(fro1AmPany= Done, wherof ae number of theorems in this section) that A1 is diagonalizable and
l J and D = [ - 0 2O J
2
Solvi1.ng for A, we have A = PDP- 1 and, by Theorem 4.22(f), An = PDnP - 1 for all
Since
[ ( -01 r 2On J
we have
An = PDnp- 1 = [ _ � � J [ ( -01 r �n J [ _ � � r l
[ _ � � J [ ( -Ol )n 20n J [iI - iJ
2( - 1 )" + 2" H r' + 2· l
[ 2( - 1 )•: + 2•+• r - 1r: + 2·+•
Sisent ce=we10wertoefionlndy asked for A10, this is more than we needed. But now we can simply
10 ( - 1 ) 1 + 2 10 2 1
�
)
(
+
1
2
2
AlO = [ 2( - 1 ) 1 + 2 1 :3 + 2 " ] � [68234 68343 ]
3
n 2:
n
(- !)'
I
Exercises 4 . 4
A
A = [! � l = [� �J
2. A = [ - 42 6l J , [ 3 - l J
1
3.A = [ : 02 ! J .B � [- : 403 � ]
4.A � [ : - 21 - : J B � [ : 011 � ]
In Exercises 1 -4, show that and B are not similar matrices.
1.
B
B=
-5
7
P- 1AP = D.
AA
[ - 21 - � J [ � - � J [ � � J = [� �J
1 - 11
1
0
0
[! 1 -I
1 1 �m 0 J
0
0
� [: 0 -� l
In Exercises 5-7, a diagonalization of the matrix is given
in the form
List the eigenvalues of and bases
for the corresponding eigenspaces.
5.
6.
l
6
-2
3
-3
Chapter 4 Eigenvalues and Eigenvectors
310
0
3
[ - ! -�I23 03 �m _ : ]
02
� [ � 0 - �l
A
A = -3
8.A = [ � � ]
A = [ �]
0
IO.A = [ � 03 � ] A = [ : il
2
0
12.A = [ � 20 : J 13. A = H 0 il
02 00
03 02
14.A = 0 3 ] 15. A � 0 - 2 ]
ll 0 0 ; l� 0 0 j
[ -- 43 :r
[ - � �r
[ 4 - r6 [ � �r
"
�
r
�
[� n
[i 0
0
[ � - 2 ff
[: 0 J
A
A = [� �]
24.A = [ � �]
0
26.A = [ � � ]
A = [i 0 �]
I
8
3
4
3
-3
7.
1
-1
-3
In Exercises 8-15, determine whether is diagonalizable
and, if so, find an invertible
matrix P and a diagonal
matrix D such that P 1 P D.
-
9.
-1
11.
1
In Exercises 1 6-23, use the method of Example 4.29 to
compute the indicated power of the matrix.
1 7.
1 6.
18.
1 9.
-1
21.
20.
22.
1
1
-1
-1
23.
In Exercises 24-29, find all (real) values ofkfor which is
diagonalizable.
25.
27.
1
Pr
o
ve
Theor
e
m
4.
2
(
c
)
.
l
31. PrProoveve Theor
e
m
4.
2
2(
b
)
.
Theoreemm 4.4.222(2(ce))..
PrProoveve Theor
Theor
e
m
4.
2
2(
f
)
.
PrIfAoveandTheorB areemin4.ver22(tibgl)e. matrices, show that AB and
BAProvearethsiatmiiflaAr. and Bare similar matrices, then
tr(A) Sect= tri(oBn) .3.2.] Find a way to use Exercise 45
from
A
B
= B.
38.A = [ � - � l B = [ � � ]
39.A = [ : -- 23 ],B = [ -l 4l ]
2
40.A = [ : - 20 : J B � [ : 22 -- 4S i
2
0
2
[
3
41. A = [ : 0 � ] ,B = : 45 - � l
PrProoveve tthhatat iiffAA iiss disimagonal
ilar toizB,ablteh,ensoAisTAisT.similar to BT.
matrix. Prove that ifA is diago­
nalLetProiveAzablbetheat,ansoifinAisverAis taibdileagonal
eimatgenvalrix isucale letdhena A is of thiezablforemmatA rix wi(tShuchonlya one
eivectLetgenvalAorands iufesandB. bePronlovey tihfABatmatA=andriBA.cesB, eachhavewithtehsamedisteiingctences. PruoesveofthAatandthe Balarge­e
brtLetheaiAscamulme.andtiBplbeicitsiiems ioflarthmate eirgienval
30.
32.
33.
34.
35.
36.
37.
[Hint:
In general, it is difficult to show that two matrices are simi­
lar. However, if two similar matrices are diagonalizable, the
task becomes easier. In Exercises 38-41, show that and
are similar by showing that they are similar to the same
diagonal
matrix. Then find an invertible matrix P such that
P - 1AP
-6
1
-1
-1
42.
43.
44.
-
45.
46.
47.
A,
i.
scalar matrix.)
nXn
=
AI.
n
Section 4.5 Iterative Methods for Computing Eigenvalues
48.
ove that it ivects notorposs sible toifinnIRd6tshurchee ltihnatearly
metLetthe sraicme.andmulBtibeplicsiiShow
tmieislaofrtmaththate,reiiifcgesBenval. Prouvees tofhatthtandhene geo­everB arye iPrndependent
and
eivectgenvect
o
r
of
Bi
s
of
t
h
e
for
m
for
s
o
me
ei
g
en­
Ihf e eiisgdiensagonal
izable, whatandare the dimensions of
o
r
of
t
p
aces
Preverovey eithgatenvalif uies ofa diaigonal
ierzableormattrhixensuchistihdatem­ 52.LetA = [ : �] .
s
ei
t
h
potLetentbe(thaatniilsp,otent matrix (that is, for some Prove that is diagonalizable if
and is not diagonalizable if
Pr
o
ve
t
h
at
i
f
i
s
di
a
gonal
i
z
abl
e
,
t
h
en
mus
t
beSupposthe zere tohatmatrisixa. matrix with characteristic
Find two examples totdemons
trateorthmayat ifnot be
h
en
may
diagonalizable.
( (
polynomial
A
50.
51.
v A.]
A
A
m > 1).
(a)
A
= P - 1AP,
p - 1v
[Hint:
49.
311
A
A 2 = A).
0
1,
A
(a)
Am = 0
A
A 6X6
cA (,\) = ( 1 + ,\) 1
v1 , v2 , v3
Av1 = v1 , Av2 = v2 , Av3 = v3 .
(b) A
E2 ?
E_ 1 , E 1 ,
A
(b)
-
4bc > 0
4 bc <
,\) 2 2 - ,\) 3 •
(a - d) 2 +
(a - d) 2 +
A
0.
(a - d) 2 + 4bc = 0,
A
Iterative M e t h o d s t o r C o m p ut i n g E i g e n v a l u e s
Inematician
1824, theNielsNorwegian
math­
Henri
k
Abel
(1802-1829)
proved thatpolynomial
a general
fiequation
fth-degreeis not(quintic)
that
is,inthere
isofnoitformula
for its
roots
terms
s
coefficients
that uses onlsubtraction,
y the operations
ofi­
addition,
mul
t
i
p
l
cati
on,Indivision,
andwrittaki
ning 1830
nth
roots.
a
paper
ten
and
published
posthumously
in
1846,
the
French
mathematician
Evariste
Galois (1811-1832)
gave
alished
moreconditions
complete
theory
that
estab­
under
which
an
arbitrary
polynomial
equation
can
bewassolinstrumental
ved by radicals.in establishing
Galois's work
the branchhis approach
of algebratocalled
equations is now knownpolynomial
as
solvable by radicals;
group
theory;
Galois
theory.
Atto stohlivsepoithnetchar, theaonlcteyrismettic hequatod weion.haveHowever
for comput
iengartehesevereigenvalal pruoesbleofmsa wimatthritxhiiss
,
t
h
er
metit depends
hod thatonrenderthe comput
it impraatctioincalofina aldetl butermsimnantall exampl
esis. Thea verfiyrsttiprme-oblconsemuismithnatg
,
whi
c
h
pra polocesynomi
s for alalrequat
ge matiorn,ices.andThetherseecondare noprforoblmemulaiss tforhatstohlveicharng polactyenomiristicaequat
ioinonsis
l
equat
ofquadrdegrateiec forhigmherulathandan its(panalolynomi
al.sThusof degr, weeesare forcanded to can be solvedeigusenvaling uthese
o
gues)
t praalctariecalquiprteoblseensms.itivUnfor
tuundoff
nately,ermetrorhandods arfore tapprhereoforxime unratineglitahbleer. oots of a
polin mosyInomi
e
t
o
r
o
cteristoirc fipolrstyandnomithaenl alustogetinghterhisandeigenvect
take aodir tffoefirenntd
apprthe coronach,streead,spapprondiweobypas
xingmeiatgsienval
ntgheancharue.eigIaenvect
one such method that is based onn athsiismsplecte iioten,rawetivewitelchniexplque.ore several variations on
2, 3,
4
4
approximate
Thethat power
metgenvalhodueapplthatiesistloaranger in absmatolutreixvalthuate hasthanaall of the other eigenvalues.
i
s
,
an
ei
Onandthuee.otthherenhand,is athmate domirix nwiantth
eieiForggenval
envalexampluuese, es,inifcea matrixandhas eihasgenvalno 2domiues n2ant eigenval
meta sheodquenceproceedsof vectiteorrastitvhelatyconver
to produceges toa tshequence
ofpondiscalnargseithgatenvec­con­
vertor gThees tthopower
and
e
cor
r
e
s
sume metthathtod.he matrix is
diagonaliezable. The following theorForemsiismtplheicbasity,isweforwithl easpower
The Power Method
nXn
= 1 -41 >
- 4, - 3, 3,
4
4
v1 ,
,\ 1
l -31
dominant eigenvector.
dominant eigenvalue A 1 -
- 4, - 3,
I l l.
131
1,
3,
-4
A
312
Chapter 4 Eigenvalues and Eigenvectors
Theorem 4 . 2 8
an o vectdioargonalsuichzabltheatmattherisxequence
with domiof vectnantoreis xgenval
u
e
Then
t
h
er
e
exiLetsts abenonzer
defi
n
ed
by
k
,
approaches a dominant eigenvector of
We may assume that the eigenvalues2: of have2: · ·been· 2: labeled so that
be),ththeeycorforersmpondia basngiseifogrenvectConsors. Siequent
nce ly, we...can, vwrn aritee linearaslya
ilLetinndependent
(
w
hy?
ear combination of these eigenvectors-say,
Nowx1
and,
gener
a
l
y
,
fo
r
k
2:
As we saw in Example
( (�:)\ · · · ( ��y )
wherThee wefacthavethusated thise tfacthe domi
that nant eigenvalue means that each of the fractions
is les than in absolute value. Thus,
all go to zero as k � It folxk lows that� ask�
Now,
and
x
i
s
appr
o
achi
n
g
a
mul
t
i
p
l
e
of
(
t
h
at
i
s
,
an
s
i
n
ce
v1
k
eiongenvect
(This is thine trheequidirreedcticondi
tdomihne antinioteiiraglcorrenvect
vectesoproondir Int gmusto t have a nonzero component
on oftitohne
Appr
Theoroeximmate the dominant eigenvector of [ � � ] using the method of
We wil take [ � ] as the initial vector. Then
[ � � ] [� ] [ � ]
[� �] [�] [�]
We continue in this fashion to obtain the values ofxk in Table
nXn
A
A1.
x0
x1 = AXo , x2 = Ax1 x3 = Ax2 , . . . , xk = Axk _ 1 ,
•
•
•
A.
Proof
_.
I A1 I
v1, v2 , . . . , v"
> I Az I
A
I A3 I
I A" I
v1, v2 ,
!R n .
x0
Xo = C 1 V 1 + C2V2 + . . . + C nVn
= Ax0, x2 = Ax1 = A (Ax0) = A 2 x0, x3 = Ax2 = A (A 2 x0) = A 3 x0,
xk = A k Xo
4.2 1 ,
1
A kXo = C 1 A �v1 + C2 A ;v2 + . . . + cn A �vn
= A i c 1 v1 + c 2
= A k Xo
v1 *
x0 :
v1.)
0,
A � c 1 v1
(2)
oo
nonzero
A1) provided c 1 * 0.
c1
A=
4.28.
Solution
(1)
1
oo.
Exa m p l e 4 . 3 0
v"
A1 * 0 .
A1
A 2 / A1, A 3 / A1, . . . , A n / A1,
A1 * 0
+ c"
2+
x0 =
x 1 = Ax0 =
x2 = Ax1 =
4. 1 .
Section 4.5 Iterative Methods for Computing Eigenvalues
313
[1.!0�2] [0.!:99]
2.05 1.98
[1.1701710 1]
2.0 1
Ta b l e 4 . 1
xk [ � ] 0.[ � ]0 1.[ �5]0 0.[ :8J3 [1.�1�0]
rlkk 1.050 3.00 1.67 2.20
k
0
2
1
3
4
[0.�9�5]
1. 9 1
5
6
7
8
y
7
6
5
4
3
2
o
x
0-/-t----2 --t- 3---1 --t-----t
---r---r-4
5
6
7
Xo
Figure 4 . 1 4
�
e 4.n14antsheiowsgenvect
whatoisr happeni
ngdigeomet
rnic1.all(yW. Wehy?knowSee Exerthatcitshee46.eig)ensTherpacee­
foforretFi,hietgiurdomi
wi
l
have
m
ens
i
o
in sTheas thfioughrst fewthieteirtaetreastexskarareeconver
showngalinogngonwiththe
the diresctaiolinsnetthheyroughdeterthmeinore.igIitnappear
lweinesewhosek, wee dineedrectionlon yvectobsorerivse[t�h].atTotheconfiratiormr tofhatthtehifisrisst tthoethdomie secondnant eicomponent
genvector
k line in the body of Table 4. 1 gives
ofthesxkegetvalsuveres, yandcloyouse tocan1 asseke iclnecrarelasesy th.atTherk isseicond
n
deed
appr
o
achi
n
g
1.
We
deduce
t
h
at
a
dominant eigenvector of is [ � ].
Oncenantweeighaveenvalfound
a domiapprnoantacheigisenvect
ore,rvhowe thcanat ifweanfixnkd itsheapprcororxiesmpondiatelynga
domi
u
e?
One
t
o
obs
dominant eigenvector of for the dominant eigenvalue A1, then
Iaskt folilnocrwseastheats. tTablhe reat4.io1 lgik ofvesthtehefivalrst ucomponent
ofyouxk+can1 tosteheatthofat xthk eywilarappre approachoach­A1
es
of
l
,
and
k
ing 2, which is the dominant eigenvalue.
IR 2 .
A
A
314
Chapter 4 Eigenvalues and Eigenvectors
e ivers aydrlaarwback
srsof. Totheavoiiterd­
attheissTherdragetwback,
gwe cane vermulytoquitthipcelklymetyeachandhodicanteofraExampl
causte byessoiegmenifisccantalThearrotcomponent
undoff
er
r
o
hatwirleduces
the magni
­
tdomiude ofnantits eicomponent
s
.
Si
n
ce
s
c
al
a
r
mul
t
i
p
l
e
s
of
t
h
e
i
t
e
r
a
t
e
s
s
t
i
l
conver
g
e
t
o
a
oach byis accept
wayseachto accom­
pla ish itvect. Oneor)gienvect
s. Anto noreasorm,ietalhr imetiszeappreachhod-and
dithveidoneianblgweeit. Ther
bywil eusare(e-ii.vare.s, titooousmake
iterabyte
di
v
i
d
e
each
largest component
nowtthhee maxicomponent
1. Thimums metabsolwihodthuittsehecalvalmaxiluee,d wemumwilabsolrThusepluacet,eivalf byue,denotso the(1ats ththeecomponent
of witihs
is notWehinilgutsotrdo,ate tshinisceapproach1. wiHence,th the calculations from Example For there
4.30:
xk
xk
II xd l
xk
unit
mk
xk
scaling.
Yk
xk
= / m k) xk .
m0 =
xk
4.30.
x0,
We then compute [ � ] as before, but now we scale with 1 to get
m =2
x1 =
Now the calculations change. We take
and scale to get
C�5 )[ � ·5 ] [ �.67 ]
The next few calculations are summarized in Table 4.2.
], a
You
can
now
see
cl
e
ar
l
y
t
h
at
t
h
e
s
e
quence
of
vect
o
r
s
i
s
conver
g
i
n
g
t
o
[
�
sdomipondinantng domieigenvectnantoeir.gMorenvaleoverue ,\,1 the sequence of scalars converges to the corre­
=
Y2 =
Yk
mk
= 2.
Ta b l e 4 . 2
k
xk
Yk
mk
[ � ] [ � ] [ � ·5 ] [ �·67 ] [ 1.1.6837 ] [ �·9 1 ] [ 1.1.9951 ] [ �·98 ] [ 1.1.9998 ]
[ � ] [ � · 5 ] [ �. 67 ] [ � . 83 ] [ �. 9 1 ] [ � · 95 ] [ �. 9 8 ] [ � · 99 ] [ � .99 ]
1.5
1
1.8
1.99
1.95
0
1
2
2
3
2
5
4
3
2
6
7
2
8
Section 4.5 Iterative Methods for Computing Eigenvalues
315
This method, called the
is summarized below.
Levaltue be a diagonalizable matrix with a corresponding dominant eigen­
Let x0 =thy0e folbeloanywiningisttiaelpsvectforkor i=n whos... :e largest component is
Repeat
(((cba))) SetyComput
Let mk k=bee( lxt/mhke=component
of xk with the largest absolute value.
.
)
x
k
k
k converor. ges to the dominant eigenvalue and Yk converForgmoses tot achoidomicesnantof x0ei,gmenvect
4
eiUsgeenvectthe power
or of method to approximate the domi5 nant- eigenvalue and a dominant
= [ -� 0l
Taking as our initial vector
power method,
The Power Method
A
1.
2.
nX n
,\ 1 .
!R n
1.
1, 2,
AYk - 1·
,\ 1
Exa m p l e 4 . 3 1
A
-2
12
-2
6
- 12
1
Solution
we compute the entries in Table 4.3.
0.
5
0
l
You can see that the vectors Yk are approaching [ - 0.50 and the scalars mk are
tapprhe domioachinnantg eigThienvals suueggesof ts that they are, respectively, a dominant eigenvector and
1
16.
A.
the domimetnhantod
eigenvectIf tohre in(iit.iea.l, vectif o=r x00 inhasthae zerprooofcomponent
of Theoremin the dirthecteniotnheofpower
R e m a rks
•
v1
4.28),
c
Ta b l e 4 . 3
xk [: l [ = : l [ l [ - .00 l [ . 5 l [ - 80..0053 l [ - 8..00 1 - [ - 800.00 ]
8
Yk [ : l [ = � �; l [ - 0.�40 l [ - 0.�s5o2 l [ - �s0.50o l [ - 0.�s50o l [ - 0.� 5s0o l [ - �s0.50o l
mk
. 5 .05
k
0
1
1
I
1
6
2
3
4
5
6
7
- 9 33
- 1 9.33
1 1 .67
8 62
1 7.3 1
9
8 12
16 2
- 8.20
16
8 4
1 6.0 1
8 1
1 6.00
8
6
- 1 9.33
1 7. 3 1
16 2
16
1 6.01
1 6.00
316
Chapter 4 Eigenvalues and Eigenvectors
John
BaronciRayl
eigh,madewasmajor
a Britishcontribu­
physi
s
t
who
tions
toInthe1871,fieldsheofgaveacoustics
and
optics.
the
first
correctandexplanation
ofdiscovered
why the skytheis
blue,
in
1895,
he
gasvedargon,
for whiPrize
ch discovery
heinert
recei
the
Nobel
inRoyal
1904.
Rayl
e
i
g
h
was
president
of
the
Society from
1905 toof1908
and
became
chancellor
Cambridge
Uni
veersiightyquotients
in 1908. inHeanused1873 paper
Rayl
onbookvibrating systems and later in his
William Strutt ( 1 842- 1 9 1 9),
The Theory of Sound.
ee tsoubsa domi
,undoff
it is quierterloikr elwiylthpratodurduceinang thxek
calwiwiltchulnotaanonzer
ticonver
on ofothgcomponent
equentninantitthereeiadigteenvect
sre, ctatisoonomerof. However
poiv1. Thent ropower
metherodrowirs actl thuenallysthelartpt!o)
convergThee to power
a multimetple ofhodv1.s(tTilhiwors is kones wheninstatnceherwher
e
r
o
undoff
e is acertain condidomitionns.antDeteigenval
ue,
orbe evenfoundwhenin mosthet moder
matrixnistenotxtbdiooksagonalon inumer
zable, iunder
a
i
l
s
may
cal analgesyrsaispi. d(SlyeetoExera domicisesnant eigen­
For
s
o
me
mat
r
i
c
es
t
h
e
power
met
h
od
conver
vectof Theoror, whiemle for otrehvealers sthwhy.ekconverSincegencel ,\2 /,\mayk 1 2:bel qui,\3 /,\te s1lo2':w.· A· ·car2:elful,\n /,\look1 , atif lt,\h2e/,\proof1 is
clthoensesthoowszero,thtathenxk(=,\2A/,\kx01 ) wi, l appr, (,\n /o,\ach1) ,\wiicl1val1 lraapprpidloyachtoo.zero rapidly. Equation
der Example byThethe sevent
eigenvalh iuteesraaretion we, shoulandd haveso
,\cl2o/s,\Ase1to=anfour4 il16-udseci=tramtioaln,-pSiconslanceceiaccur
actyo. Thiestismisatexacte thelydomiwhatnantwe saw.eigenvalue ,\1 of a ma­
Ther
e
i
s
an
al
t
e
r
n
at
i
v
e
wa
y
trix A in conjunction wi(tAhxth) e· xpower(,\metx) ·hxod. FiAr(sxt,·obsx) erve that ifAx = ,\1x, then
= x·x1 = 1x·x = Ai
x·x
tsThehymmet
e itexprerarteeicss matxiok,ntrhiec(essxu),ccest=he(sRaylAivx)e Rayle·ix)gh/equot(ixgh· x)quotientis calimetentleshdoda(xisk)aboutshoultdwapprice asoachfasAst,\as1we. tIhncomput
fae sccalt, ifornge
factor method. (See Exercises 17-20.)
Thewhatpower
metwehoddo canif wehelwantp us apprthe otoxihmerateiegtenval
he ues? Foreitugnatenvalelyu, ethoferae matare rsiexver, butal
s
h
oul
d
variTheations of the power methodustheats tcanhe obsbe applervatieid.on that, if ,\ is an eigenvalue of A,
tThushen ,,\if-,\1 iiss tanhe eidomigenvalnantue ofeigAenval- ueforof A,anythsecaleiagrenval(ExeruescofiseA -in,\Sectif wiiolnbe
A22 -- ,\A11,, and,\3 -fr,\om1, .th. i.s,valAn u-e we,\1. canWe ficannd t,\h2en. Repeat
applyinthgethpower
metcesshwiodl talo lcomput
e
i
s
pr
o
o
w
us
t
o
,\comput
e all of the eigenvalues.
Use th]e shifted power method to compute the second eigenvalue of the matrix A
[ � � from Example
method toIn Example we foundA -that =,\1 - To fi]nd ,\2, we apply the power
[� �
We take x0 = [ � ], but other choices wil also work. The calculations are summarized
in Table
repeated
•
2 1 -24.)
•
4.28
.
(2)
•
1 6 4,
4.3 1 .
0.25 7 = 0.00006,
0.25.
/
•
2,
•
R
Rayleigh quotient.
R
The Sh illed Power Melhod and lhe Inverse Power Melhod
dominant
shifted power method
a
Exa m p l e 4 . 3 2
al
a
22
4.3).
0,
=
4.30.
4.30,
Solulion
=
21
4.4.
2.
_
Section 4.5 Iterative Methods for Computing Eigenvalues
311
Ta b l e 4 . 4
[ � ] [ - � ] [ - �·5 ] [ - �·5 ] [ _ �·5 ]
Yk [ � ] [ -2�.5 ] [ - �.5 ] [ - �.5 ] [ - �.5 ]
teratiuonse of. Ther:._+e­
foreOur, A2 -choice of-x0 shaso produced the ei2g-envalue isafthteersonlecondy tweiogienval
k
1
0
2
3
4
-3
-3
-3
xk
1
mk
A 1 = 3,
A2 = A 1 - 3 =
-3
3 = -1
A
m preigoenvalpertyu(eb) of Theor
eforme, if wethapplat ify thiseipower
nvertibmetle wihtodh eitogenvall, uites
tdomihRecalennantl eifrl gohasenval
Ther
e
u
e
wi
l
be
t
h
e
(
i
n
magni
t
u
de)
ei
g
enval
u
e
ofmethod,To except
use thisthat in step 2(a) we computweefollow the sYkame- l · (sItnepsprasactiince,thwee power
don't
actforuallusy icomput
e
expl
i
c
i
t
l
y
;
i
n
s
t
e
ad,
we
s
o
l
v
e
t
h
e
equi
v
al
e
nt
equat
i
o
n
ng Gaussian elimination. This turns out to be faster.)
A-
A,
A.
xk
Exa m p l e 4 . 3 3
4. 1 8
1 /A.
A
A-
recip rocal of the smallest
inverse power method,
xk = A - 1
i
A-
Axk = Yk - i
Use the] inverse power method to compute the second eigenvalue of the matrix
[ � � from Example
We start, as in Example with [ � ]. To solve
we use row reduction:
A=
4.30.
Xo = y0 =
4.30,
Solution
Ax1 = y0,
Thus, [ � ] ,so [ � ] · Then we get from
[2 I J [ I J
0·5 ], and, by scaling, we get [ ] Continuing, we get the
Hence,
[
valsmualeslesstheiowngenvalinuTable ofe is thwhere recieptrhoecalvalofues m(kwarhiechconver
ging toThis agrThusees ,withthe
i
s
al
s
o
our previous finding in Example
x1 =
y1 =
x2
Ax2 = y1:
1 o 0.5
1 1 o
�
[A I Yi ] =
0 1
0 1 - 0.5
x2 =
- 0.5
4.5,
A
1
y2 =
.
-1
4.32.
-1
- 1).
- 1.
Chapter 4 Eigenvalues and Eigenvectors
318
Ta b l e 4 . 5
k
[ � ] [ �] [ - �:� ]
[ �] [ � ] [ _�O].S
0
1
2
[ - 0.l .SS ]
[ - �..S33 ]
l
3
[ - o0..s83 ]
[--0.�.863]
[ - o0..9sS ]
[ -- 0.�.9SS2 ]
[ - o1..s1 ]
[ --�1..41S ]
6
5
4
[ - o1..0s2 ]
[ -- �1..0429 ]
[ - o0..9s9 ]
[ -- 0.�.9S9l ]
7
[ - o1..s0 1 ]
[ -- �1..0SO1 ]
8
9
mostiversoned.atiIltecanof tbehe usvareidanttosfiofndthane power
metmathiodonifors one theiatgenval
combiune,esprtohvie tdwedo
jweTheusthavement
appr
o
xi
n othuerewor,\ ofds, tifhaatscalis clarosesits gitoven,
the If ,\ aisclanoseeiapprgenvaloxiumeatofion tando thwiatl eifignenval
d,\, tthheenueie.gIenval
- is inver(SeetibExerle ifciseis4S.not) Ifan
eiisgclenvalose tuoe,\of, thenand1/(1 /,\(,\- - wiils anbe eia gdomienvalnuante ofeigenval
of tude thanInthfacte next, if
ieigsenvalucle,osseo t(oas,\not, theend i1n/(t,\he-thirdwiRemar
l be k follbiowiggerng uiExampl
ne magni
e 4.3 1) the conver­
gence wil be very rapid.
Use the shifted inverse power method to approximate th6e eigenvalue of
]
s
[ - �2 -122 - 1210
that is closest to S.
Shifting, we have
- SI= [ =:- 2 - �2 -� �s l
Now we apply the inverse power method with
T h e Sh illed Inverse Power Method
any
shifted inverse power method
A
a)
a)
A
a very
Exa m p l e 4 . 3 4
a*
A al
(A - aI ) - 1 .
a)
much
A
a
a
(A - aI ) - 1 .
A= -
Solulion
A
We solve
[A
for
6
s
s
- SI [ -- 42 - 2 - 12s :]--� [ : �0 �1 =- 0.�::39� ]
(A - SI )x1 = y0
I
x1 :
y0 ] =
7
a.
a
Section 4.5 Iterative Methods for Computing Eigenvalues
319
Ta b l e 4 . 6
-[ - 0410.69 : [ -- 0470.89 : [ -- 0490.95 ] [ -- 0500.98 ] [ -- 0500.99 ] [ -- 0501.00 ]
-0590.35 - 0.053 - 0510. 8 - 0.0509 - 0.05050 - 0.05050 [ 0.1.0501 ] [ 0.1.0500 ] [ 0.1.0500 l [ 0.1.0500 ] [ 0.1.0500 ] [ 0.1.0500
- 0.69 - 0.89 - 0.95 - 0.98 - 0.99 - 1.00
This gives
-[ - 0.0.6881 ] , - 0.88, and
1- -0.88 [ -- 0.0.8681 l [ 0.1 69 ]
- 0.39
- 0.39 0.45
Wethatcontthe ieingueenvalin tuheisoffasAhiclonosteostobtto a5inisthappre valoxiuesmatineTablly 5 e 4.16/, from whi5 ch1/we( - deduce
1)
which, in fact, is exact.
4
and itservar5, iweantswirel prdiessentcussonlanoty oneherapprmethoodachbasto ethdeoncomput
a­
tfactionoTheofrizeiatpower
gioenvaln ofmetua esmat.hIodnrixChapt
t
h
e
For a morical emetcomplhodset. e treatment ofthis topic, you can consult
almost any textbook on .numer
�
I
n
t
h
i
s
sect
i
o
n,
we
have
di
s
c
us
s
e
d
sever
a
l
var
i
a
t
i
o
ns
on
t
h
e
power
met
h
od
for
ap­
pr
o
xi
m
at
i
n
g
t
h
e
ei
g
enval
u
es
of
a
mat
r
i
x
.
Al
l
of
t
h
es
e
met
h
ods
ar
e
i
t
e
r
a
t
i
v
e,
and
t
h
e
WeRussian
owe this
theorem
to
the
s
p
eed
wi
t
h
whi
c
h
t
h
ey
conver
g
e
depends
on
t
h
e
choi
c
e
of
i
n
i
t
i
a
l
vect
o
r
.
I
f
onl
y
we
had
mathematician
sidedinmakeformaatjiuodin"ciabout
theceloofcatthioeninofitiathl vecte eigoenval
uperes ofhapsa gispveeden matup rthixe,
S.statedGerschgorin
(1901-1933),
whove tshoenmewe"incoul
o
us
choi
r
and
i
t
i
n
1931.
I
t
di
d
not
recei
muchresurrected
attention unti
l 1949,Taussky­
when it convergence of the iterative process.
was
by
Olga
ion ofuesthofe eia g(renvaleal oruescomplof anyex)
Todd in a note she published in the matFor
rn X nixmat.tunatrixelaly,l tlhieereinsiids eathwaye unitoonesofsttiamnteatcis erthcultathaetrhldieocateiskgsenval
in the complex plane.
LetA
[
a
;
]bea(
r
e
alorcompl
e
x)nX
nmat
r
i
x
,
a
ndl
e
tr
;
d
enot
e
j
tish,ers; um of tl ahue· IabsTheolute values of the off-diaisgonalthe cientrculrieasr idin stkhe ithinrotwheofcomplA; theatx
plane wijt1'hicenter a;; and radius r;. That is,
k
xk
Yk
mk
061
[: J [ -- 0.0.3898 ]
069
[: ] [ 0.1.005 - 0.88
0
1
2
5
4
3
44
4
6
7
4
4
X1
=
m1
=
=
+
m7
=
+
= 4,
QR
Gerschgori n 's Theorem
American Mathematica/ Monthly.
Gerschgorin's Disk Theorem
Definition
= 2:
=
ith Gerschgorin disk
D;
320
Chapter 4 Eigenvalues and Eigenvectors
Olga
Taussky-Todd
(1906-1995)
was born
in Olmiitz
indoctorate
the Austro-Hungarian
Empirefrom the
(now
Olmuac
in
the
Czech
Republic).
She
recei
v
ed
her
i
n
number
theory
Uni
versity ofinVienna
inwhere
1930. Duri
nvesti
g World
WartheII,problsheeworked
for thein theNational
Physi
cal
Laboratory
London,
she
i
n
g
ated
m
of
flutter
wi
n
gs
of
super­
sonic
aircraft.on theAlthough
the problem
involmatrix.
ved diffTaussky-Todd
erential equations,
the stabiGerschgori
lity of an aircraft
depended
eigenvalues
of
a
related
remembered
n's
Theorem
from
her
graduate
studies
in
Vienna
and
was
able
to
use
i
t
to
simplify
the
otherwise
laborious
computations
needed
to determine
thein 1947,
eigenvalues
relevant
to thesheflutter
problem.
Taussky-Todd
moved
to
the
United
States
and
ten
years
later
became
theover
first
woman
appointed
to
the
California
Insti
t
ute
of
Technol
o
gy.
In
her
career,
she
produced
publications
and receivnow
ed numerous
the200 branch
of mathematics
known asawards.
matrixShetheory.was instrumental in the development of
Exa m p l e 4 . 3 5
Sketch the Gerschgorin disks and the eigenvalues fo-r the following matrices:
(a)A = [ � - � J
(b)A = [ � � ]
adii 1and2,
respectively(a. )TheThechartwoactGererissctihgorc poliynnomidisksaarl ofe centA ise,\r2ed at,\2-and8, s-o3thwie teihgrenval
ues are
,\ = ( - 1 v12 - 4( - 8) /2 2.37, - 3.37
diFi(bsg)kurs.Thee 4.t1w5oshGerowsschgorthat itnhedieiskgsenvalare centues earreedcontat 1aandined3wiwiththinrtahdiei two3 Ger= s3chgorand i2,n
respectively. The characteristic polynomial ofA is ,\2 - 4,\ 9, so the eigenvalues are
(4 V( - 4)2 - 4(9) /2 = 2 iVs 2 2.23i, 2 - 2.23i
Figure 4. 1 6 plots the location of the eigenvalues relative to the Gerschgorin disks.
Solulion
+
±
=
+
A=
±
±
Im
4
-4
Figure 4 . 1 5
=
+
I- I
Section 4.5 Iterative Methods for Computing Eigenvalues
Im
321
4
-4
Figure 4 . 1 6
GerAsschgorExamplin diesks. Thesunextggesttsh, eortheemeigverenvalifieusesthofat tahimats is rsiox. are contained within its
4.35
Theorem 4 . 2 9
Gerschgorin's Disk Theorem
wiLetthinbea anGerschgor(rienaldiorskcompl
. ex) matrix. Then every eigenvalue of is contained
respondihenceng einonzer
genvecto.or(Why?Let) Thenbe
the entrLety tofhe itbehwirantohwteiofhgewhienvallargchesuites absof oluwitetvalh corue-and
nXn
A
A
Proof
_....
Ax = Ax,
A
A
x
x.
2: a x
1 �1 ij J
n
Rearranging, we have
=
x;
Ax;
becaus
Appendie x weTakiobtanign absolute values and using properties of absolute value (see
X; * 0.
C),
IA
-
a;1x1
2:
i
*
J
a;; I = X;
I fi1 aij xJ I
l x; I
:S
2: la;1 x1 1
J*i
l x; I
2: lau l l x1 I
J*i
l x; I
:S
2: la;1 I = r;
J*i
becausThies establishes forjthat thei.eigenvalue is contained within the Gerschgorin disk
centered at with radius
l x1 1
:S
a ;;
l x; I
*
r;.
A
322
Chapter 4 Eigenvalues and Eigenvectors
Ther
e
i
s
a
cor
r
e
s
p
ondi
n
g
vers
i
o
n
of
t
h
e
precedi
n
g
t
h
eorem
for
Gers
c
hgor
i
n
disks whosIt cane rbeadisiharowne thethsatumifofofthethoffe Ger-diasgonal
entn dirisekssinartehediistjhoint fromofthe other
c
hgor
i
haenr, ifexacta sinlgly e dieisgkenvalis diusjesoinartefrcontom tahienedothwiertdihisnks,thtehunien iotnmusof tthcontese aindiexactsks. Ilny
pardionesktiseic, ulgtenval
utheatofinthExampl
e matrixe.4.Exampl
e i4.s 3not5(acont) il uasitnraedtesinthaisGer. schgorin disk; that is,
Not
e
3
5(
a
)
,
0
tful0hiatswhennotthe anmatappleirgiixenval
ied tosuliaenrvergofertmatibHence,
lerbyicesTheor, wibecaustheoutme4.tanyh1e6.GersfurtThihschgorerobscomput
eirnvatdiisokantsiicanosn,parwebetidetccanulaerrdeduce
lmy iusnede­
directly from the entries of the matrix.
R e m a rks
•
•
k
•
k
A.
A
Exa m p l e 4 . 3 6
column A.
k
Consider the matrix [ 2� 06 �8 l Gerschgorin's Theorem tel s us that the eigenvalandues2, ofrespectareivcontely. aSeeinedFiwigurthein4.tlh7(reae).diBecaus
sks cente tehreedfiatrst2,di6,skandi s di8swijoitnht rfradiomi 1,th1,e
otremher4.t2w9.o,Becaus
it mustecontthe acharin exactactelryisonetic poleigyenvalnomiuae,l ofby thhase secondreal coeffi
Remarcikentafst,eirfTheo­
i
t
has
complxex rootHences (i.e.t,heiergeenvalis a uunies qofue realtheieygenval
mustuoccur
ineenconj1 andugate3,paiandrs.th(SeeeuniAp­on
pendi
e
bet
w
o (possibly complex) eigenvalues whose real parts
lofietbethOne wotteenhhere ot5twandhoerdihand,
10.sks conttheaifinrssttwRemar
kdiafsktsercentTheoreredemat4.2,296,tandel s us8 withtath rtahdie isame, 1,
tandhree eiregsenval
u
es
of
are
cont
a
i
n
ed
i
n
Figrealure )4.eil 7(genvalb) . Thesue. eCombi
disks narinegmutthesuale lryesdiusltjos,inwet, sodeduce
each con­that
taihasns athsirnpegleectereiv(alaendleiy.gSeehence
(Compute the actualenvaleigenvalues, oneues ofin eachto verofitfhyethinist.e)rvals [ 1 , 3] , [5, 7], and [7.5, 8.5] .
A=
·
A
A),
D.)
t,
A
Im
Im
4
Figure 4 . 1 1
�
A
A
-4
A
4
(a)
-4
(b)
. '�
1
Section 4.5 Iterative Methods for Computing Eigenvalues
Exercises 4 . 5
In Exercises 1 -4, a matrix A is given along with an iterate
x5 , produced as in Example 4.30.
(a) Use these data to approximate a dominant eigenvector
whose first component is and a corresponding dominant
eigenvalue. (Use three-decimal-place accuracy.)
(b) Compare your approximate eigenvalue in part (a) with
the actual dominant eigenvalue.
1
1. A =
2. A =
3. A =
4. A =
[ l 42 ] ' Xs [ 111094443 ]
[ _ � - 41 ] , Xs [ - 78113904]
[ � l1 ] ' Xs [ 144 ]
[ 1.5 0.3.05 ] ' 5 [ 60.625 ]
6. A =
7. A =
=
89
x =
2.0
239.500
[ _ � -1 03 ] ,xs = [ -11.3.060167 ]
[ � - � l X1 0 [ � : ! � � ]
1
�
X
,
�
[-� 4-12 g [ �:��3.�4115
1- 1 - 31 1 , X10 [ - 2.1.290714 1
=
=
0
2
1 0.000
In Exercises 9-14, use the power method to approximate
the dominant eigenvalue and eigenvector of A. Use the given
initial vector x0 , the specified number of iterations k, and
three-decimal-place accuracy.
10. A =
12. A =
13. A =
=
=
9. A =
11. A =
=
5
In Exercises 5-8, a matrix A is given along with an iterate
xk > produced using the power method, as in Example 4.31.
(a) Approximate the dominant eigenvalue and eigenvector
by computing the corresponding m k and Yk· (b) Verify that
you have approximated an eigenvalue and an eigenvector
ofA by comparing Ayk with m kYk ·
5. A =
323
[ 14 l � l
[ - 6 - �l
5
8
Xo
=
Xo
[�l
[ �l 6
k=
=
- 0.5
=
' Xo
=
0
=
5
In Exercises 15 and 1 6, use the power method to approxi­
mate the dominant eigenvalue and eigenvector ofA to
two-decimal-place accuracy. Choose any initial vector you
like (but keep the first Remark after Example 4.31 in mind!)
and apply the method until the digit in the second decimal
place of the iterates stops changing.
16. A =
60 - 6 1
[ -6 6 1
12
2
-2
2
Rayleigh quotients are described on page 316. In Exercises 1 7-20, to see how the Rayleigh quotient method ap­
proximates the dominant eigenvalue more rapidly than the
ordinary power method, compute the successive Rayleigh
quotients R(x;) for i =
for the matrix A in the given
exercise.
18. Exercise 2
17. Exercise
20. Exercise
19. Exercise
1, ... , k
1113
114
The matrices in Exercises 21 -24 either are not diagonaliz­
able or do not have a dominant eigenvalue (or both). Apply
the power method anyway with the given initial vector x0,
performing eight iterations in each case. Compute the exact
eigenvalues and eigenvectors and explain what is happening.
[� � ],x0 [ � ] [ _� � ],x0 [ � ]
� [ � � +• � m
� [� � n � � m
21. A =
23. A
k=5
=
[ � � l Xo [ �l k 6
[ 3.1. 55 1.5 ] [ l ] ' k 6
9[ 4 14
8 -4
: n � � uk � 6
24. A
=
22. A =
=
Chapter 4 Eigenvalues and Eigenvectors
324
In Exercises 25-28, the power method does not converge to
the dominant eigenvalue and eigenvector. Verify this, using
the given initial vector x0 • Compute the exact eigenvalues
and eigenvectors and explain what is happening.
[ = � �l [ � ]
[ _� �l [ � ]
�
_
n ; n [: J
- [: �: � ] · � - [: ]
25. A =
Xo =
26. A =
Xo =
27. A
42. p(x) = x 2 - x - 3, a = 2
43. p(x) = x 3 - 2x 2 + 1 , a = 0
44. p(x) = x 3 - 5x 2 + x + l, a = 5
45. A
A
x. a * A
a
A,
1 /(A - a)
be anor eigIenvalf ue ofandwitishnotcorranespeiondigenvalngue of
eiLetgsenvect
is an eiogrenval(Wuhye ofmust
wibetihnhverowcortrtiehbsatlpeondi?) ng eigenvect
Igensf haspacea domiis one-nantdeiimgenvalensiounale . , prove that the ei­
46. A
_
_
x.
(A - al ) - 1
A - al
A1
EA,
� In Exercises 47-50, draw the Gerschgorin disks for the given
matrix.
2& A
1
-i
4
2i 1 +
48.
47.
In Exercises 29-32, apply the shifted power method to
-2i
0
approximate the second eigenvalue of the matrix A in the
2
4 3i
-2
given exercise. Use the given initial vector x0 , k iterations,
0
0
l
+
i
and three-decimal-place accuracy.
49.
2;
5 + 6i
1 + i
-i
30.
10
29.
9
1
2i
-2i
-5 - Si
31.
32.
I2 0
In Exercises 33-36, apply the inverse power method to
4 !4
approximate, for the matrix A in the given exercise, the ei­
50.
I 6
6
genvalue that is smallest in magnitude. Use the given initial
0 8I
vector x0 , k iterations, and three-decimal-place accuracy.
51.
strictly diagonally dominant
34.
10
33.
9
36.
Exercise
Exercise
Exmi" � - [ _ :J
Exercise
7.
k-5
ExerExercciissee
ExerExercciissee
�i
tinhethsuatmroofw.th(SeeeabsSectoluitoenvaluesUseof thGerse remaichgorniinng's entDisrkies
Theor
e
m
t
o
pr
o
ve
t
h
at
a
s
t
r
i
c
t
l
y
di
a
gonal
l
y
domi
n
ant
matafterriTheor
x musetmbe invertible. See the third Remark
Itfhe sius msan of thematabsolrixu,tleetvalues denot
of theertohwse maxiof mtumhat ofis,
c� ) (See Section Prove that
iLetf ,\ is beananeigeienvalgenvalueuofe of athsentochastic matrix
Apply
Exer(see Sectcise i52onto Prove that
Prove that the eigenvalues of [ � � � �1 are
alcllosreeald ,inandtervloalcatone teachhe reofal tlihnese.e eigenvalues within a
2.5.)
14
In Exercises 37-40, use the shifted inverse power method
to approximate, for the matrix A in the given exercise, the
eigenvalue closest to a.
38.
12, a = 0
37.
9, a = 0
39.
7, a = 5
40.
13, a = - 2
Exercise 32 in Section 4.3 demonstrates that every poly­
nomial is (plus or minus) the characteristic poly nomial of
its own companion matrix. Therefore, the roots of a poly­
nomial p are the eigenvalues of C (p). Hence, we can use
the methods of this section to approximate the roots of any
poly nomial when exact results are not readily available. In
Exercises 41 -44, apply the shifted inverse power method to
the companion matrix C ( p) ofp to approximate the root of
p closest to a to three decimal places.
41. p (x) = x 2 + 2x - 2, a = 0
[�
[
l
[ ; �]
Aabssquaroluteevalmatueriofx ieachs diagonal entry is greater thifanthe
ExerExercciissee 14
ExerExercciissee 13
35.
u !]
�
nXn
52. A
ll A ll = �'!,�
1
53.
A
3.7).
A r.]
54.
[Hint:
4.29.]
11A 11
A;
l a iJ I .
A,
7 .2.)
I A I ::; II A II .
I A I ::; 1 . [Hint:
A=
A
! 0 3 2
0 0 � 7
'
Section 4.6 Applications and the Perron-Frobenius Theorem
325
A p p l icati o n s a n d the Perro n - Fro b e n i u s T h e o re m
genvaleruses. and eigenvectors.
WeIn thbegiis snectbyiorn,eviwesitiwingl sexplomeoapplre seivercatiaolnsapplfroimcatprionseviofouseichapt
sSectitiisoitnohne(sttroachasnsinittritooicduced
)n matmatrrMariixcesofkasaovMarsochaiciaktnovesdandchaiwithmade
tn, hthem.enseverInhasparal obsatiscteuleradyvarat, iwesotnsatobseaboutvectervoedtrhethtThatratan­if
eiis,gtenvalhere uise.aWevectaroer nowsuchintahposat ition toThiprosveistequihis factvale. nt to saying that has as an
If Pis the transition matrix of a Markov chain, then is an eigenvalue of
Recal
l
t
h
at
ever
y
t
r
a
ns
i
t
i
o
n
mat
r
i
x
i
s
s
t
o
chas
t
i
c
;
hence,
each
of
i
t
s
col
u
mns
sciusmse toin SectTherionefore, iTakif isnag rtoransw vectposoesr, conswe haveisting of l s, then (See Exer­
whiExercchisiempliiens Sectthation is an eiandgenvecthaveor ofthe samewitheicorgenvalrespuondies, snog eiisgenval
ueeigen­By
al
s
o
an
value of
is truue.e Foris most transitthioatnismat, if rices, theneigenvaluWee needsat­
itshfieeIfolsn lfactowi,nmuch
andg twothmordefie einegienval
are posie­,
tive, and a sq] uare matrixtiios ns:callAedmatrix i]sifcalsolmeed power ofifitalisl ofpositistientve. rFories exampl
[ � � is positive but B [ � � is not. However, B is regular, since B2
[ ! � ] is positive.
Leta. be an transition matrix with eigenvalue
b. If Pis regular and then
As
i
n
Theor
e
m
t
h
e
t
r
i
c
k
t
o
pr
o
vi
n
g
t
h
i
s
t
h
eor
e
m
i
s
t
o
us
e
t
h
e
fact
t
h
at
has(a) tLethe sabemeaneigeienvalgenvectuesoasr of corresponding to and let be the component of
withethktthhecomponent
largest absols ofutethvale equat
ue ioThenn we havefor
Comparing
···
Markov Chains
3.7
P
P
Theorem 4 . 3 0
x.
Px = x.
x
P
nXn
1
P.
1
Proof
1.
13
3.7.)
jr
19
P.
jP = j .
n
j
pYf =
( j Pf =
pT
pT
4.3, P
I A I ::=::: 1
f
1 dominant;
1
every
A * 1,
I A I < 1.
1.
A
positive
regular
Theorem 4 . 3 1
nXn
P
IAI
=
=
A=
1
::=:::
1
A.
A * 1,
4.30,
Proof
x
IAI < 1 .
pT
P.
pT
m.
lx;I
::=:::
A
l xk l = m
PTx = Ax,
Plk X1 + P2 k X2 +
xk
i = 1, 2, . . . , n.
+ Pnk Xn = Axk
x
326
Chapter 4 Eigenvalues and Eigenvectors
(Remember that the rows ofpT are the columns ofP.) Taking absolute values, we obtain
I A l m = I A I l xk l = I Axk l = l p lk x 1 + P 2 k X2 + · · · + Pnk xn l
:S l p lk x l I + I P 2 k X2 I + · · · + I Pnk Xn I
= P1k l x, I + P2 k l x2 I + · · · + Pnk l xn l
:S Plk m + P 2 k m + · · · + Pnk m
=
+ P2 k + · · · + Pnk)m = m
(1)
(p,k
Thecomesfirsfrtoimnequalthe factity folthatlowsthefrroomwstofhepTrT isaunglm etoInequalThusit,y in andAfttheerladistviequal
ibyty
d
i
n
g
we
have
as
des
i
r
e
d.
ove tPhe(aequind tvhalereentforime PplTi)catis iaon:posIfitive matrtihxen. If Fithrenst, weall ofshtowhe
ti(nhbequal
at) Weit isitwitireuls eiprnwhenEquat
ions are actually equalities. In particular,
Equivalently,
itive, p;k in Equatfor ion must beAlszero, o, and this canforhappen only if
TherNow,esforinforcee, eachPis possummand
thermoraree, weposgetitivequal
itlyarine tnegat
he Triivae;ngline Iotnhequaler woritydisn,
ithefp;andkX; 'sonlallyhaveif alltofhetsamehe suFurmmands
e
or
al
sign. This implies that
IAI :s 1,
m,
IR,
IAlm :s m.
1.
IAI = 1 ,
( 1)
A = 1.
IAI = 1,
(2)
>0
lx;I = m
i = 1 , 2,
IR
i = 1, 2,
(2)
m - lx;I 2: 0
i = 1, 2,
. . . , n.
. . . , n.
wher
e.m Thus, in either case, the eigenspace
l s, as inspTheor
r
of pButTecorj,iussreaisnrpogondiwthvectenprgootroofofofisTheor
an(
j
)
em This handl
we seeesththate jcasT ekwher
pT {e P is{,posand,itivkcompar
­
ing Icomponent
s
,
we
fi
n
d
t
h
at
e.
kt prandovedofkPit+hatsarpose eiitgivenvale-suay,esPofTher.pItkfolandeforlowspek,+thk(atresppect+ I musivelyt,
albysoTheorbef Pipossermeitgulive.ar(W, twehhy?enhave)soSimenjucespower
which implies that since is impossible if
ChaptWeercanInnowExamplexplaein someweofsatwhethbehavi
at for othr eoftrMaransiktiovonchaimatnrisxthat we observed in
p [ ]
and ]initial state vector [ ] , the state vectors xk converge to the vector
[ , a steady state vector for P (i.e., We are going to prove that for regular
4.30.
n
A EA =
A = 1.
�
. . . , n.
4. 18,
3.
A 1
A
A = 1,
4.30,
A=0
Ak = Ak + l = 1 .
IAI = 1 .
=A
1,
A A - 1) = 0,
3.64,
=
Xo
0.4
0.6
=
=
0.6
0.4
0.7 0.2
0.3 0.8
x=
Px = x) .
Section 4.6 Applications and the Perron-Frobenius Theorem
321
Marthe sktovate chaivectnosr,s txhkissataliwsfaysy xkhappens.
Indeed,
wetigwiatel whatprovehappens
much morto teh.eRecalpowerl tshat
k
P
x0.
Let
'
s
i
n
ves
as P becomes large.
The transition matrix P [ 0.0.73 0.0.82 ] has characteristic equation
0 det(P 1 0.70.3 0.80.2 I 2 - l .5A + 0.5
- 0.5)
so3i1,ts weeigenval
uesin aradvance
e thandat would0.5be. (Nanoteiegtenval
hat, thuankse andtothTheor
emseig4.enval30 andue
4.woul
knew
e
ot
h
er
eigensdpbeaceslesartehan in absolute value. However, we stil needed to compute The
span( [ � ] ) and E0.5 span( [ � ] )
So, taking [ 3 ], we know that [ 0 0.O 5 ] From the method
used in Example 4.29 in Section 4.4, we have
[ 3 ][0 0(0.5 ] [ 3 ] - !
Now, ask� (0.5)k � 0, so
and � [ � � ] [ � � ] [ � ] [ 0.0.46 0.0.46 ]
(Observe that the columns of thias ]"limit matrix" are identical and each is a steady
state vector for P.) Now let [ b be any initial probability vector (i.e., a + b
Then
� [ 0.0.46 0.0.46] [ ab] [0.0.46aa ++ 0.0.46bb] [0.0.46 ]
Not only does this explain what we saw in Example 3.64,0.4it also tel s us that the state
veoto;s wil oonmge to the steody stote vedoc x [ 0.6] foe any choke of x,!4
e is notorhalinwgaysspoccur
ecial about
Exampl
etr4.ans37.itioThen matnextrices.theorBeforemeswehowscanthpratesentthis
ttyhpee tTherhofeorbehavi
s
wi
t
h
r
e
gul
a
r
em, we need the following lemma.
LeteigenvalP beuaeregular nhas aln gtrebransaiitciomuln mattiprliicxi.tyIfP is diagonalizable, then the dominant
pk
=
Exa m p l e 4 . 3 1
=
-
=
-A
AI) =
A1 = 1
-A
1
1
=A
A2 =
A2 .)
E1 =
Q=
2
= (A - l)(A
=
1
-1
Q - 1 PQ =
2
pk = QDk Q - i =
_
1
l lk
)k
-1
oo,
pk
= D.
2
1
-1
1 -l
-1
_
= 1).
Xo =
Xk
= p k·v�
·-u
=
=
�
x,
Lem m a 4 . 3 2
A1 = 1
X
=
1.
328
Chapter 4 Eigenvalues and Eigenvectors
eigenvalricuesmuloftPipandlicity 1 araseantheeisame.
FruoemoftPher.prSionofceofPiTheor
em 4.i3z1abl(be),,
Aso1 i=s P1rhas,Theby geomet
g
enval
s
di
a
gonal
Thereeform.e, the eigenvalue A1 = 1 has algebraic
multiplicityExer1, bycitshee41DiiangonalSectiiozatn i4.on4.Theor
Let P bemata rriexgulL awhosr e colturamnsnsitiarone matidentriixc.alThen, eachasequalk ---+ to thek sapprameovectachesor anx.
This vector xis a steady state probability vector for P.
seorimplemifyisthtreupre, however
oof1 , we wi, wil tconshoutidtheris onlassyumptthe icason.e wher1 e P is diagonaliz­
ableWe. ThediToathgonal
ize Pas Q- PQ =Dor, equivalently, P = QDQ- , where
pT
Proof
Theorem 4 . 3 3
n
See and J. L. Snell (Newby J.York:
G.
Kemeny
Springer-Verlag, 1976).
Finite Markov Chains
X
n
n
n
oo,
p
Proof
0
From Theor
e
ms
4.
3
0
and
4.
3
1,
we
know
t
h
at
each
ei
g
enval
u
e
ei
t
h
er
i
s
1
or
sat
i
s
fi
e
s
k
ask---+matrix-sappray, D*-each
oaches1 1or0of whosfor e=di1a, gonal
... entIt folriesloiwss 1or0.that DThusap­,
prpko=achesQD1. Hence,
akQdi-a1 gonal
approaches L = QD*Q- . We write
lim = L
Observe that PL = P lim lim = lim = L
tTher
hneedat eacheonlforye,ofobseachtheesrvcolee tcoluhmnatu,mnsifofj iLisisthsaeanroeiwgvectenvectorvectwiortofhorP(1icorr.se,.,thLeensipsondia stonchasg to tAic1 mat= 1.riTox), sweee
j
L
=
j
l
i
m
l
i
m
j
P
=
l
i
m
j
=
j
snowinceipmkplisieasstthoatchasL itsicstmatochasrixt,ibyc. Exercise 14 in Section 3.7. Exercise 13 in Section 3.7
umn oforLs ofis
jPuforstWeLe;mi,nneedwherg a basonle isyisofsthhowe itwithhattshtatvndarh1ecorcoldruebassmnspiondis vectofnLgortar.oLete i1dvent1,1.vi2cWr,al.... iThete, vnibeth eicolgenvect
for scalars c1, c2, , cw Then, by Theorem 4. 1 9,
By---+Lemma0 as k4.---+32, for 1 for1. It fol1,loswso, bythatTheorem 4.3 l (b), 1 for 1. Hence,
= lim = c 1 v1
I A; I <
Wethe notion
are takiofngasome
liNevertheless,
berties with
l
i
m
i
t
.
these
steps
shouldproofsbe infoltuilotwivelfrom
y
clear.
Rigorous
themayproperties
oflimits,inwhia calcu­
ch you
have
encountered
lus course.withRather
than getofside­
tracked
a
discussion
limits, we wil omit the proofs.matrix
X
oo,
A7
A;
i
k---+ 'X!
pk
, n.
pk
+
pk i
pp k
=
probability
k---+ 'X!
e;
=
k
k---+ oc·
•
Aj
oo,
k---+ oc·
A
!R n ,
.
AJ
pk
n
=
.
-=F
j -=F
I Aj l <
j -=F
Le ;
k---+ oc,
P k e;
j -=F
Section 4.6 Applications and the Perron-Frobenius Theorem
329
Isnhownotherthworat thdes,colcoluumnsmn ofofLLiarseanproeibabigenvectlity vector cororrse, sspoondiLe; ins gthteo A1 mulButtweiplehavex of
val1lwhosof thee colcomponent
umns ofsLsarume itdoentiSicaln,ceeachthisequalis truetoforthieachs vectcolorux.mn of L, it implies that
SiofntceheLMaris akstovochaschaitinc. matThatrixis,,weLijcanreprinesteentrpsretthiet pras othbabie lity of being in
Theas
facttshtaetenextthathaviexampl
thengcolstueamnsritleudstoffrraotLmes.arsetaitdeentical says that the
Recall the rat in a box from Example The transition matrix was
= 1.
i
unique
1.
long range tran­
Remark
sition matrix
i,
Exa m p l e 4 . 3 8
j, if the transitions were to continue indefinitely.
starting state does not matter,
3.65.
We determined that the steady state probability vector was
Hence, the powers of appr[ � oach � i [
]
L
frmentom whiandch we canofseeitstthimatethine eachrat wiofl the other twspoendcompartofmentits tsi.me in compart­
discussioofn tofhereingulitiaalr sMartatek. ovThechaipronofs byis preasoivilynadapt
g thatetdhteostcoveready
sthtaeteWecasvecteconcloforsxtuaitdese ivectnourdependent
o
r
s
whos
e
component
s
s
u
m
t
o
an
ar
b
i
t
r
a
r
y
cons
t
a
nt
s
a
y
,
I
n
the exercises, you are asked to prove some other properties of regular Markov chains.
gular em traThen,
nsitionformatanyrixi,nwiititahl xprtohbabie steladyity vectstateorprx0obabi, thelisteyquence
vector
forofLetiterbeaasteaisnrxeTheor
k approaches x.
Let
P
�
=
1
�
� � �
8
8
-
0.250 0.250 0.250
0.375 0.375 0.375
0.375 0.375 0.375
eventually
3 7 .5%
25%
s.
Theorem 4 . 3 4
P
P,
Proof
nXn
4.33.
330
Chapter 4 Eigenvalues and Eigenvectors
kx0, we must show that lim Pkx0 = x.kNow,
wher
e
=
Si
n
ce
x
=
P
X
k
n
byTherTheoreforee,m the long range transition matrix is L = [ x x x] and lim p = L.
lim pkx0 = ( lim p )x = Lx0
x1 + x2 +
· · ·
4.33,
1.
+
···
k
k-'>GO
k-tGO
k-toc-
k--+oo
0
== X1X XXz2X XnX)nxX= X
WeSectiroentu3.rn7.tIontExampl
he Lesleie3.model
67 in thofat popul
sectioan,tiweon sgrawowtthhat, forwhitchhe weLeslfiiersmatt explrixored in
[�s L �]
iterates of the population vectors began to approach a multiple of the vector
+
(xl +
+ '''+
+ ···+
Pooulalion G rowlh
r
�
populiattiisonstaeventble, suinalcely tended
rfolInatliotoowiherngworyear.dMors,artehegieovervthenr,ebyoncee agetclhiasssteasteofisthreisached,
he ratioups forin tthhee
[�5 00.25
l . 5x
s arofethstiisl popul
in theatriaotniowhen it1.has5 =reached itObss steeadyrvestthaatte.1.5 repre­
sandentWesththe ecancomponent
recognithezsetetadyhat sxitasteangroeiwtgenvect
oirs aof L correeispgondienvalnugetooftL,heandeigen­an
valeigenvect
ue =or1.nowcor5. Thus,
h
r
a
t
e
thibeens eigenvalreached.ue reWeprecansentcomput
s the e thesesidizesreofctlyt,hwie ageth­
cloutas havies whenng totihteerresasttpeeondiadyas wesntagditetodhasbefor
e.
wth rate and the corresponding ratios between the age clas es
forFindthtehLese stWeeliadye matneedstartixetogrLfioabove.
L. The characteristic polnydnomiall posal ofitivLeieis genvalues and corresponding eigenvectors of
det(L = 00.5 0.25 0
0.375
18:6: 1
IF
4
27 : 9 :
growth rate
,\
Exa m p l e 4 . 3 9
18:6: 1 .
positive
relative
Solution
- ,\
-
AI)
4
-A
3
- ,\
- ,\ 3
+
2 ,\
+
Section 4.6 Applications and the Perron-Frobenius Theorem
331
sinog,wewemushavet solve 2,\ 0. 5 0 or, equivalently,
0. Factor­
(2,\
) 0
1
x Since- t1.he1,setcondhe onlfacty posor hasitiveonlroyotthofe rtohotis sequation i s - 0.1.159.
andThe(See corAppendi
row reductrespioondin: ng eigenvect1.o5rs are in the null space of - l .50I, whi18ch we find by
O
- 1.5II O J [ �-5 - 0.1.255 - 01.5
01 0 00 J
Thu,, if [ :: ] i"n eigenvectoc '°""ponding to 1.5, it "ti'fi" I
and That is,
- ,\ 3 +
8,\ 3 - 16,\ - 3 =
+ 37 =
- 3)(4,\ 2 + 6,\ +
=
D.)
( - 3 - Vs)/4
L
4
=
[L
( - 3 + Vs)/4
A=�=
=
3
=
3
-6
x�
A�
x, � Bx,
x2 = 6x3 •
Hence,
t
h
e
s
t
e
ady
s
t
a
t
e
gr
o
wt
h
r
a
t
e
i
s
1.
,
and
when
t
h
i
s
r
a
t
e
has
been
r
e
ached,
t
h
e
age
5
clas es are in the ratio 18: 1, as we saw before.
ddatwee forhavethedonesteadyif stahadte grhadowtmorh raetet:htanhe
unioneqIposnueExampl
positivietieiveeg4.enval
ei3g9,envaltuheerorueewasnone?
of onlButWey onewhatwercandiwoul
e also appars werentleyposfortituivnate, ewhithatchtalherloewedwasusa
corto rreelaspteondithesnegcomponent
eigenvectors altol tofhewhossize eofcomponent
atandiona icors notresaccipondidentngaeil; gthenvect
at is, or witLesh poslitehimatteivpopul
re component
ix hasatioexactn. Wesl.ycanoneprposoveitivtheateigthenvalis situue­
Recall that the form of a Leslie matrix is
0
0
0
0
00 0 0 00 00
000
0
Since thoe(oenttherriews ise,rethpreepopul
sent sautrivoinvalwoulprodbabirapilidtliyesdi, wee outwi)l. Weassuwimel althsato asthseyumearethalatl
nonzer
atand,leasagait onen, tofhethpopul
e birtahtiparon awoulmetedrsdie iouts nonzer
o
(
o
t
h
er
w
i
s
e,
t
h
er
e
woul
d
be
no
bi
r
t
h
s
now prove the assertion we made above as). Wia thteorh theesm.e standing assumptions, we can
a uniques. positive eigenvalue and a corresponding eigen­
vectEveroyrLeswitlhieposmatitrivixe hascomponent
6:
L.
L
every
b 1 b2 b 3
51
52
L=
53
(3)
Sn - I
sj
b;
Theorem 4 . 3 5
bn - 1 bn
332
Chapter 4 Eigenvalues and Eigenvectors
Proof
Let L be as in Equation The characteristic polynomial of Lis
(3).
= ( - l ) nj(A)
rs(Yuorotouvsivareofal prasokbabiedSitnloicetipresatovelareastehtipossoneinitiExerofve,tthcheiesebicoeffirth parcTheientametseiofgenval
ers uchange
iess posof Litsivarigeneandtexactheraleforllyofonce.e tthhee
x D), therefore, has exactly one positive
rByootDescar
.ByLetdirusetctecals'cals lRulictuleatofion,Sigwens can(Appendi
check that an eigenvector corresponding to is
16.)
f(A).
b;
(A)
f
f(A)
sj
A 1.
A1
s1 /A 1
S 1 S 2 I Af
S 1 S 2S3 /A i
pos(Youitiarve.e asked to prove this in Exercise Clearly, all of the components of are
biruthe
parofaImetnLifactsers, morande is trtuhe.atarWieis,posteverh tithiyvee,otaddiihterttui(orrnnalesaloutrorequicompl
thratementtheex)uniteihgatqenvalue posueitiveofeiLgenval
satproisoffieiss
(
I
t
i
s
beyond
t
h
e
s
c
ope
of
t
h
i
s
book
t
o
pr
o
ve
t
h
i
s
r
e
s
u
l
t
,
but
a
par
t
i
a
l
i.n) edThiins explExeraciinses whyforweregetaderconver
s who garencee famitoliaasrtwieadyth tshteatale gvectebraorofwhencomplweexinum­
bertouthelspopul
ation vectors: It is just the power method working for us! terate
x1
18.)
b;
two consecutive
b;+ i
A1
dominant;
I A I < A1.
A
27
Ieingtenvalhe prueeviofousinttweroesapplt wasicposatioinstiv,eMarandkovdomichainantns .andMorLeseoverlie mat, therrieceswas, wea corrsawetshpatond­the
heoremof
guartinhge oneseiantgenvect
eeswethhaveoatr twihbeenitshwiposlconsbeitivtiehdeercomponent
casing.e Thefor afislr.asrItgt veretuclrsanisosnoutofofmattthhiatsritcaheseorr,emarineclmukidiabls nforeg tmany
matFiricres.st, we need some terminology and notation. Let's agree to refer to a vectposoitrivase
ifweall wioflitwrs component
sf are positfoivre.alForl andtwo (Similamatr defiricnesitions wil applandy
i
t
e
i
for to be theandmatsorixon.of)tThus
he absol, a posuteivaltiveuesvectofotrhe entsatirsifieessof Let us define
The Perro n -Frobenius Theorem
Oskar Perron
(1880-1975)whowasdida
German
mathematician
work
in manyanalysis,
fields of mathemat­
ics,equations,
including
differentiandal
al
g
ebra,
geometry,
number
theory.inPerron'
s Theorem
was
published
1907
in
on continued fractions. a paper
positive
B = [b ii ] ,
A > B, A :s B,
IA I = [ la;j l l
A 2: B a ii 2: b ij
i
j.
mX n
x
A = [a ii ]
x > 0.
A.
Section 4.6 Applications and the Perron-Frobenius Theorem
Theorem 4 . 3 6
333
Perron's Theorem
prLeta. operbe>tie0as:positive matrix. Then has a real eigenvalue with the following
b.c. If hasis anya corotrheserpondieigenvalng posueioftive eitghenvect
o
r.
en
Iofntauitively,posweitcanive matsee whyrix thThee fircorrst tweospsonditatementng mats shrioulx trdansbefotrrume.atConsion mapsider tthhee casfirset
quadr
e pltoaneactpronoperthelyiminagesto itsweelf,getsin,cethaleyl neces
component
sconver
are posgeittiovware. Ifdwesomere­
peatray iendlathntyealoffilrostwthquadr
s
a
r
i
l
y
a
nt
(
F
i
g
ur
e
A
di
r
e
ct
i
o
n
vect
o
r
for
t
h
i
s
ray
wi
l
be
a
pos
i
t
i
v
e
vectleavesor thewhiraychfixmused. tInbeotmapped
ins,to some posiwitivteh mulandtiple ofbotithseposlf (siatyiv, e. since
h
er
wor
d
oalvectl o>rs thus2, we foneedr someonlyscconsalar iderWhen tvecthis happens
,
tChapt
hen erForwes2omewilnonzer
for
o
r
s
I
n
vectors iinve vect(tohres on this unit
isnpthero ae",gtenerherealwiizedl beelseleiaptsoimaxihatd:'mSo,mapsumasvalthruaeengessetof ofoversalulchtunihethtnonnegat
Denote this number by and the corresponding unitatvecto2r by x1.(See Figure
nXn
A
A1
A1
A
A
A,
2X2
4. 18).
x,
Proof
7,
I A I :::; A1.
A.
A
A(kx)
A(kx)
A
k
Ax = A 1 x,
x, Ax Ax
x
A
y
y
Figure 4 . 1 8
y
Figure 4 . 1 9
x
A1),
A1
A.
O;
A1
y
A1
Ax
A
unit
x.
n
!R
unit sphere)
Ax.
4. 19.)
y
334
Chapter 4 Eigenvalues and Eigenvectors
obtaWein now show that Ax1 = A1x1. If not, then Ax1 > A1x1, and, applying A again, we
wher
Exerthereceiwistehle beinequalsButometithAyen2is>yprA1=eses(rulv/ched,l Athx1siatnl ce)AyAAx1isisAposa2yuni.itThiivte.vects (contSeeorExerrtahdiatcctsissaettihsfie efactands AytSecth>at A1iA1onywas, so
tA1hex1maxi; thatmisum, A1valis anue wieigtenval
h thisupre ofoperA.ty. Consequently, it must be the case that Ax1 =
Now
A
i
s
pos
i
t
i
v
e
and
x1
i
s
pos
i
t
i
v
e,
s
o
A1
x
1
=
Ax1
>
Thi
s
means
t
h
at
A1
>
andTox1 >provewhi(c)c,hsucompl
etAesisthanye protoofherof(r(eaal) andor compl
(b) . ex) eigenvalue of A with cor­
ppos
e
responding eigenvector z. Then Az = Az, and, taking absolute values, we have
e uniinequalt vectityorfoul oiwsn thfreomdirtehcteioTrniaofngllzel Iinsequalalso iposty. (itSieeve Exerand csiasetisfies
SiAuwhernceeIAlzthlul e>. Bymitdhtdlheemaxi
mality ofAJrom the first part ofthis proof, we must have IAI ::::: A1.
I A1 forric, anymuleitigpenval
AWe IwinA1lfact. notIt ,imorspralosveeoitsththeresucase.e factIet ttuhsrat.nsA1outhasthalatgA1ebrisadomi
ic, andnanthence, so IAgeomet
licity ue
tain nonnegat
iveematmatrrixi­.
cesA squar. PerFrorbenieon'mats uTheor
sridix Ad esimos calincanledbe gener
The raelsiuzifedlt, srufrebjquiomectrepostsoasittoeivmechnie topercceralmcondi
t
i
o
n
on
t
h
the same permutation of the columns, A can be writ en in bloutckatforionmofasthe rows and
[ � �]
wher
matriex andsuch tharate square. Equivalently, A is reducible if there is some permutation
A [ � �]
(See page For example, the matrix
A=
is reducible, since interchanging rows and and then columns and produces
40
36.)
3.7
2:
0.
0,
0
(4)
40.)
0,
2:
<
i=
1.
1912.
reducible
B
P
D
P PT =
1 87.)
0
2
2
0
0
2
4
1
6
1
0
1
7
0
0
3
5 5
3 0
2
7 2
3
1
7 2: 1 3 0
�----�-J � ?.
0 0i2 1
0 0!6 2
0 0! 1 7
__
____
____
?.
3
1
2
3
Section 4.6 Applications and the Perron-Frobenius Theorem
335
(This is just where
PAPT,
0
0
0
P= 1
0
0
0
0
0
�
Theorem 4 . 3 1
0
A
Ak > 0
irreducible.
primitive.
Letwith thbee anfolliorwieducing prbloepernonnegat
ive n n matrix. Then has a real eigenvalue
t
i
e
s:
b.c.a. If hasis aanycorroteshperondieignenvalg posuietivofe eigenvect
or. If is primitive, then this
t
h
en
iInf equal
strict.ue of such that ,\ then is a (complex) root of the
d. equat
,\ isioaninty,\eiisgenval
e. has algebraic,\mul� tiplicity
rested readerive matcanricfiesndoramatprorofix analof thyesisPer. Theron-eiFgenval
robeniuues Theor
eemn calinlemany
tTheextsionntenonnegat
i
s
oft
dartihlye
of
and
a
cor
r
e
s
p
ondi
n
g
ei
g
envect
o
r
(
w
hi
c
h
i
s
neces
s
unique) is called the
of
The e, after the first two arteermthse, eachnumbernews itnertmheisseobtquenceained by summing the two ter...ms,
wher
plpreetcediely defing int.edIf weby tdenot
he equate thieonsnth Fibonacci number
and, forbyn then this sequence is com­
The Perron-Frobenius Theorem
X
A1 > 0
A1
A
A
A,
n
-
A1
Matrix Analysis
0
0
0
0
A
A
See C. R. Johnson (Cambridge,
by R. A. Horn
and
England:
Press, 1985).Cambridge University
0
0
0
0
0
0
0
CheckA stqhuaris!)e matrix that is not reducible is called If for some
ttranshenitionismatcalrleixd, by definitiForon. exampl
e
,
ever
y
r
e
gul
a
r
Mar
k
ov
chai
n
has
a
pr
i
m
i
t
i
v
e
ir educible. (Do you see why? TrIyt sihs owinotnharg thdetocontshowraposthatitivevere ofythprisi.m) itive matrix is
k,
__..
1
1
=
0.
A
IAI
IAI
=
::=:::
A
A1.
1,
A1
A
1.
A1
Perron root
A,
probability
Perron eigenvector A.
linear Recurrence Relalions
0,
Fibonacci numbers
Jo = 0, f1 = 1,
Jn = fn - 1 + fn - 2
1, 1, 2, 3, 5, 8, 13, 21,
fn ,
2: 2,
ThiFibonacci
s last equatnumbers
ion is, anbutexampl
e ofwial lconsinearidrerecurlinrearencerercurelatrieoncen. Werelwiatilonsretsuormewhat
n to the
fi
r
s
t
we
more general y.
336
Chapter 4 Eigenvalues and Eigenvectors
Leonardo
of Pisa"son(1170-1250),
pictured
left,a number
is betterofknown
by hisbooks,
nickname,
Fibonacci,
whi
c
h
means
of
Bonaccio:'
He
wrote
i
m
portant
many
ofappears
which have
survived,
including
and
The
Fibonacci
sequence
assurrounded
the solutionontoallasides
problem
inwall. How many pairs
certainof rabbits
man putcana pairbe produced
of rabbitsfrom
in a place
by
a
thatthe
pair
in
a
year
if
i
t
is
supposed
that
every
month
each
pair
begets
a
new
pair
which
from
month onbybecomes
ve?" The name
was given to the terms
ofsecond
this sequence
the Frenchproducti
mathematician
Edouard Lucas (1842-1891).
Liber abaci
Liber quadratorum.
Liber abaci: "A
Fibonacci numbers
be a sequence ofnumbers that is defined
as follows: Let
where wherare e scalars. are
Forscalaalrls. k,
If ionstihne equatareiorenfeinr ed tios ascaltlheed a
equat
of the recurrence. The
Thus, the Fibonacci numbers satisfy a linear recurrence relation of order
the entiht, ttherenmthine raecurrecurrencerencerelraetliaotniohasn, weordrerequik. re the
rmordbuter tnoofo defitineritmniaelbeforcondi
- k)IThetfh, itnenumber
tfiirosnst teisrmtheofortdheersofequence
the recurbercalenceledrelatiWeon. could
I
t
i
s
not
neces
s
a
r
y
t
h
at
t
h
e
start at It isorposanywher
e
el
s
e.
curbyralencelowirnelgatanionsextbyra,alisloowilatnedg
tcoeffi
he coefciefintci,ewhintsscibhtlomaye tbeo havealso evenbe a funct
rmorathereiogener
tn.hanAnasclexampl
allinaearrs andreewoul
d be the recurrence
We wil not consider such recurrences here.
rsConseecurquence.irdeerncethreelsaetiquence
on defined by tforhe initial Wrcondiite toutionsthe first five termsandof tthhies
late the nextWetharreeegitevrmenst. hWee fihaverst two terms. We use the recurrence relation to calcu­
D e fi n it i o n
(xn ) = (x0, x1, x2 , . . . )
a0, a 1 , . . . , a k - l
1 . x0 = a0, x 1 = a 1 , . . . , xk - l = a k _ 1 ,
2.
c 1 , c2 ,
n 2: Xn = c 1 xn - l + c2xn _ 2 + · · · + ckxn - k'
ck * 0,
(2)
( 1)
.
•
.
, ck
linear recurrence relation of order k.
initial conditions
2.
Remarks
(n
•
•
•
•
Exa m p l e 4 . 4 0
x0.
x1
C;
functions
(xn )
xn = 5xn - l - 6xn - l
n 2: 2.
x 1 = 1, x2 = 5
Solulion
X3 = 5 X2 - 6X1 = 5 • 5 - 6 • 1 = 19
X4 = 5X3 - 6X2 = 5 . 19 - 6 . 5 = 65
X5 = 5X4 - 6X3 = 5 · 65 - 6 · 19 = 2 1 1
1, 5 , 19, 65, 2 1 1,
so the sequence begins
....
Section 4.6 Applications and the Perron-Frobenius Theorem
331
pltoeapplCleyarttlhyhe,enirfetwecurhe rwerapprenceeoiachrnetleatrusieosneteddthinteri,mesay,eswoul. Itthdwoule belOOdrtahbethterenirmtceediofifoweusth,ecoulsisnecequence
wed findwoulanin Exam­
d have
forcurmreulncea forrelaXtnioasn.aWefunctwiilonil ofustraWete threefprerotceso fisnwidinthgtshuechseaquence
formulfraoasm Exampltehe re­
To begin, we rewrite the recurrence relation as a matrix equation. Let
A [� -�J
and introduce vectors xn [ :: J for 2. Thus, x2 [:: J [ � l X3 [:: J
[ � l x4 [::J [ �: l and so on. Now observe that, for 2, we have
AXn- [ 5 - 6 J [Xn-2Xn-I J [ 5Xn-IXn-- 16xn-2 J [Xn-Xn 1 J Xn
NotLesliiceematthatritches.is iAss thine stahmeose tcasypeesof, weequatcaniowrn weite encountered with Markov chains and
We Thenowcharuse tahcteetreichnistic equat
que ofioExampl
n ofA2 ise to compute the powers ofA.
A
6
5,\
hat thieoneigfolenvallowsuesthatareof,\t1he r3eandcurre,\nce2 r2.el(aNtiootn.iceIftwehatwrtheitfore thme
ofrfreocurtmherwhiecharncechaasctweeXrnfiisn-tdic5tequat
the same!) The correspn-ondin6g ein-genspacesit isarappare ent that the coefficients are exactly
E3 span( [ � ]) and E2 span( [ � ])
SettingP [ � � J , we knowthatP- 1AP D [ � �J. ThenA PDP - 1 and
Ak PDk 1 [ 3 2 J [ 3k 2OkJ [ 3 2 J - 1
[ 3 2 J [ 3k 2OkJ [ - 23 J
- 2-(32k(+3kI)) 33((22kk)+ 1) J
It now follows that
4.40,
98
explicit
solving
n.
4.40.
=
=
1
=
n
x
=
=
2:
=
n
i= 1
=
=
0
2:
=
4.29
+ =
=
x
I
+ x
z
=
=
0,
=
=
=
=
=
0
p-
=
1 1
0
1 1
0
=
=
1 1
1
-1
+
+
=
338
Chapter 4 Eigenvalues and Eigenvectors
�
frinonm which we readtooffvertihfye sthoatluttihoins xforn mul3na-gi2vnes. (tThoe check
ourermwors thkat, wewecoulcalcdulplatuedg
s
a
me
t
usinObsg theervreecurthatrenceXn isrealalitnioearn. Trcombiy it! )nation of powers of the eigenvalues. This is nec­
essariicliyt]t.hUse casinge asthilsongobseras tvhateieion,genvalwe canues aresavediourstinsctelv[esas Theor
eworm 4.k.38(Oncea) wiwel makehave
explcomput
s
o
me
ed the eigenvalues ,\1 3 and ,\ we can immediately write
where c1 and are to be determined. Using the initial conditions, we have
when n and
when n We now solve the system
for cThi1 ands is thtoe obtmetahinod1we wiandl use in practThusice. ,Wexn now3n -il u2snt,rasatebeforits use. e to find an
explicit formula for the Fibonacci numbers.
Solve the FiWrbonacci
r
e
cur
r
e
ncef0
,
f
andf
f
for
n
f
O
1
2
n
n
n
l
f
f
f
we
s
e
e
t
h
at
t
h
e
char
a
ct
e
r
i
s
­
i
t
i
n
g
t
h
e
r
e
cur
r
e
nce
as
2
n
l
n
n
tic equation is ,\ ,\ so the eigenvalues are
and
Itthefolforlomws from the discussion above that the solution to the recurrence relation has
for sUsomeingstchaleairnsitci1aandl conditions, we find
fo
and
in 1 is and
Hence, an explicit
forSolmviulnga forfor theandnth Fibweonacciobtanumber
=
= 1, 2, . . . , 5
=
c2
2 = 2,
1 = x 1 = c 1 3 1 + c2 2 1 = 3c 1 + 2c2
=1
5 = X2 = C 1 3 2 + C2 2 2 = 9c 1 + 4c2
= 2.
c2
Exa m p l e 4 . 4 1
c =1
=
Solulion
X
= 1,
=
+
= 0,
2 - - 1 = 0,
A1 =
Jacques
Binet (1786-1856)
made
contributions
to
matrix
theory,
number theory,
physics, theandruleas­
tronomy.
He
discovered
forBinet'matrix
multiplication
in 1812.
s
formula
for
the
Fibonacci
numbers
isishedactualit inly due
tohow­
Euler,
who
publ
1765;
ever, it washisforgotten
unti
l Binet
published
version
in
1843.
LikeheCauchy,
Binet
was
ay royal
ist,
and
lost
his
uni
v
ersi
t
posi­
tion
when
Charles
abdicated
in
1830.
He
recei
v
ed
many
honors
for hisinwork,
including
his elec­des
tion,
1843,
to
the
Academie
Sciences.
3c 1 + 2c2 = 1
9c 1 + 4c2 = 5
c2 = - 1 .
=
1 + Vs
2
A2 =
0=
= C 1 A � + C2 A � = C 1 + C2
----
2: 2.
1 - Vs
2
---
c2 .
c1
c = 1 / Vs
c2 ,
l
_
Jn =
Vs
( 1 +2 vs) n
-
c2
- 1/ Vs.
(
1 - vs ) n
Vs
2
l
_
(5)
4
Section 4.6 Applications and the Perron-Frobenius Theorem
Fornumber
mula is ayetremarthe Fikablbonacci
e formnumber
ula, becauss aree alitl iisntdefiegersned! Triny tplerumggis ofngtihnea few
valuisesknown
for toassee how the terms cancel out to leave the integer valuesfn- Formula
lineduwores arkes alforl dianystinctse.condWhenorthderereliisneara repeatrecured­
reiegnceenvalThereluamete,tiotnhhodewhostewechnie ashaveqsueocimusjautsetdtouteibegenval
ince thizeesdibotagonalh sitiuzatatiioons.n method we
used may no longer work. The next tmodiheorefimed,susmmar
LetLeta,\ -X,\n1b=and=ax0.n,\-2l be bxthne-eiz begenvala recuruesrofencethereaslatsioocinattheatd ichars sataisctfieerdisbyticaequatsequenceion ,\(2xn-) .
a.b. IIff AA11 = AA22, =A,thentXhnen=Xcn1A=7 c1Acn2A� cfor2n,\some
n for ssocmealarsscalc1aandrs c1cand2. c2•
In either case, c1 and c2 can be determined using the initial conditions.
Axn_ 1, wher(a) Gener
e alizing our disXcussion above, we can write the recurrence as xn =
xn = [Xn-n 1 ] and A = [ � �]
Sifonr ceExerA chasise distinct eigenvalues, it can be diagonalized. The rest of the details are left
(bxb)n-Wez orwi, equil shvowtalenthlatxy, n = c1,\n c2n,\n satisfies the recurrence relation Xn = axn- l
if ,\2 - a,\ - b = 0. Since
sXubs-tiaxtution intbxo Equat= i(ocnAn yicelnd,\s n) - a (c An- I c ( ),\n- l)
n n- I - n-2 -1 b (c1An-2z c2( - 12),\n-2) 2 - l
= C1 (An --a,\2)n,\- 1 - b,\n-2) C2(n,\n - l )An- 1
== cc1,\Ann-2-2((,\02)- a,\c n-,\bn)-2(0c)2n,\nc-2,\(,\n2-2-(aa,\,\ - 2bb) ) c2An-2(a,\ 2b)
1= c An-2(a,\ 2b2 ) 2
2
Buta/2,, susiniceng,\thies aquadrdoublateicroforotmof,\ula2. -Consa,\ e-quentb = ly0,, wea,\ mus+ 2bt have= a2/a22 4b2b == 0-and4b/2,\ =
2b = 0, so
tional
n
(5)
Theorem 4 . 3 8
339
(5)
Vs
irra­
Vs
Binet's formula.
+
*
+
+
Proof
53.
+
+
(6)
( 6)
+
+
+
n
+
n - Z)
- b(n
n
- a (n
+
+
+
+
+
+
+
+
+
+
340
Chapter 4 Eigenvalues and Eigenvectors
is a Suppos
unique esothluetiinonitiforal condic1 andtions (arSeee Exerrandcise 54.) Then, in either (a) or (b) there
Solve the rTheecurrcharenceactreelartisiotinc equation is and- Xn n- l -whi9xnc-hl hasfor 3 as a
n
n
n
doubl
e
r
o
ot
.
By
Theor
e
m
4.
3
8(
b
)
,
we
mus
t
have
X
3
(
c
n
)
3
.
3
2
n
Since c1 and ( c2)3, we find that so
Therelattioechnins. Wequesstaoutte, lwiintedhoutin Theor
38 canalberesextulte. nded to higher order recurrence
proof,etmhe4.gener
LettXhnat isamsat- iiXsfin-edl byama-lsXequence
Xn-m bee athreecurasrsoenceciaterdelachartionaofcteorrisdteric
n-l ( n ) . Suppos
polynomial
where
factThenorxs nashas the form
X (
x0 =
c2 .
Exa m p l e 4 . 4 2
x0 = 1, x1 = 6,
A2
Solulion
1 = x0 =
Theorem 4 . 3 9
m
x1 = s .
6 = x1 = c 1 +
=
+
+
(,\ - A 1 ) m 1 (A - A 2 ) m2
•
•
· · ·
•
= 6x
n 2 2.
A=
6,\ + 9 = 0,
= c1 + c2 n = c 1 +
c2 = 1,
+ a0
x
(,\ - ,\ k ) m',
m1 + m 2 +
· · ·
+ m k = m.
n - C 11 1t n1 + C 1 2 n 1t n1 + C 13 n 2 fl1n + . . . + C 1 m 1 n m 1 - I fl nJ ) + . . .
+ (ck1 AJ: + ck2 nA J: + ck3 n 2 AJ: + · · · + ck m, n m, - I A J: )
_
\
\
\
\
Ientn calialcequat
ulus, iyouon ofletarhne forthatmif ( wher) is aediffiseraentconsiabltaentfunct, thenionthsate gener
isfyinagl sola diufftieorn­
is specitfiheed,unithqen,ue
bysisolusutibsontittouttiwher
nhegdiffe ereiisntnaitahconsleequatgenertantioa. nlIfstoanhlatuitnsioaittn,iiasfilweecondis fithnedtiitnohintatial conditiHence,
on is
Suppose we have differentiable functions oft-say, , Xn-that satisfy a
{
We can write this system in matrix form as where
[x{�(tt)) l , and
( t)
Now we can use matrix methods to help us find the solution.
svs1ems of Linear Differenlial Equalions
x = Ce kt,
x=x t
x' = kx,
C
t=0
k
x ( O ) = x0
C = x0•
x = x e kt
o
n
system of differential equations
x1, x2 ,
x = a 11 x 1 + a 1 2x2 + · · · + a 1n xn
x� = a z1 X 1 + a z 2X2 + · · · + a zn Xn
x' = Ax,
x' ( t) =
x�
x�
•
•
•
Section 4.6 Applications and the Perron-Frobenius Theorem
341
a usefulions:observation. Suppose we want to solve the following sys­
temFiofrsdit,ffweeremakential equat
2
Each equation can be solved separately, as above, to give
wher
a e andcoefficarieentconsmattraintx s. Notice that, in matrix form, our equation has
X{ = X 1
x; = 5X2
X1 = C1 e 21
Xz = C2 e s t
C1
diagonal
C2
A=
tion. cThiients
smatanduggesrtihxet,sieifthgposenval
at,sforiblueanes. 2arandbitraryoccursysteim,n thwee exponent
should stiaarltsby diandagonalofizitnhge tshoelucoeffi
Solve the following system of differential equations:
5
Exa m p l e 4 . 4 3
[� �]
x' = Ax
e 21
e 51
Here the coefficient matrix is [31 22 ] , and we find[ 2th]at the eigenval[ -u1 es]
arreespective4lyand. Therefor-e,1, wiisthdicoragonalrespizondiablen, gandeigenvect
,
o
r
s
and
1
3
the matrix that does the job is
A=
Solution
,\ 1 =
,\ 2 =
A
P = [v1 v2 ] =
We know that
P - 1AP =
[� - � ]
[40 - OJ1 = D
v1 =
P
v2 =
Let to get(so that or, equiandvsaluebsnttliytu, te these results into the original equation
This is just the system
whose general solution is
or
x = Py
x ' = Ax
x ' = Py' )
Py' = APy
y' = P - 1APy = Dy
4
Y1 = C 1 e 1
Y2 = C 2 e - t
342
Chapter 4 Eigenvalues and Eigenvectors
To find we just compute
x,
x = Py =
[�
(Check that these values satisfy the
gisoven system.) and
Observe that we could also expres the solution in Example as
[] [ ]
This itzeablchnie.qTheue genernext athlieorzeseeasm,ilwhosy to e proofsyisstelemsft aswheran eexertheccoeffi
cmmar
ient matizesrixthies
disitaugonal
i
s
e
,
s
u
ation.
Let be an diagonalizable matrix and let
be such that
x1 = 2C 1e 4 t - C2 e - t
x2 = 3C 1e 4 1 + C2 e - t.
4.43
Remark
2
-1
x = C I e41 + C 2 e -1 l = C 1 e41v1 + C2 e -1v2
3
n
Theorem 4 . 4 0
A
n
X
X
n
P = [ v1 v2
n
•
•
•
v" ]
Then the general solution to the system is
Theecosnextysteexampl
ereinasvolonablves ea tbio oaslosgiumecal tmodel
ingrwhiowtchhrtawteoofspeachecies slpivecie iens de­the
same
m.
I
t
i
s
h
at
t
h
e
pends
ations.sim(plOfe courby igsnore, thinergethareese.ot)her factors that govern
growtIf hon, butthandewesizwieslofkeepdenotourpopul
model
wo populis ofattihoensforatmtime then and
are their rates of growtehthate tsiimzese ofOurthe tmodel
0
x' = Ax
x = C 1 e A 1 t v1 + C 2 e Air v2 +
·
·
·
+ C " e A"1 V"
both
x�(t)
Exa m p l e 4 . 4 4
x 1 (t)
x 2 ( t)
t.
x{(t) = ax 1 (t) + bx 2 (t)
x;(t) = cx 1 (t) + dx2 (t)
a, b, c,
d
t,
x{(t)
where the coefficients and depend on the conditions.
Raccoons
andr, andsquisprace.els iLetnhabithte traccoon
he same andecosysqstuiemr eandl populcompetationse wiatthtieachme otyearhers forbe
food,
wat
e
giravteenisby and butreswhen
pectivesqlyui. Irnetlhs earabse prenceesentof, tsquihe compet
r els, tihtieornaccoon
grheorwtac­h
s
l
o
ws
t
wth rat.eIntothe absence of raccoonsThe, thsquie grorwtelhpopul
atitohne sisqsuiimr ielal rpopul
ly affeacttioend
bycoon
tis hegrroaccoons
r
a
t
e
of
e populisation growth rate for squir els when
the ecosystem wiandth rtahccoons
Suppostheeytharate isnhiartiainl gy
r ( t)
s ( t),
r'( t) = 2.5r ( t),
r' ( t) = 2.5r ( t) - s ( t) .
s'( t) = 2.5s ( t) ,
s'( t) = - 0 . 25r ( t) + 2.Ss(t).
t
Section 4.6 Applications and the Perron-Frobenius Theorem
343
raccoons
and squir els in the ecosystem. Determine what happens to
tthheresee tarweo popul
a
t
i
o
ns.
Our system is where
[ ] and [ -I.O J
The- ei] genvalues of] are and with corresponding eigenvectors
[ � and [ � By Theorem the general solution to our system is
60
Solution
60
x' = Ax,
2.5
r (t)
Ax = x(t) =
s (t)
-0.25
A2 = 2,
A A1 = 3
.
4.40,
v2 =
x(t) = C 1 e 3 tv1 + C2 e 2 tv2 = C1 e 3 r
2.5
v1 =
[ - 21 ] + C2e 2r [ 21 ]
(7)
.tionThe iwenitihaveal population vector is [ : ���] [ :� ], so, set ing in Equa­
Solving this equation, we find and Hence,
(7),
C1 = 15
x(t) = 15e3 r
t=0
C2 = 45.
[ - � ] + 45e 2t [ � ]
Ficcoongurepopulshatowsion
tdifrhoeemsgroutwhiaphsafchtofewer tahfilesitnedletwmoro functe thanions,yearand. you(Canandcanyouseedetcleerarmlyintehat the rawhen
it dies out?)
.+
We
now
cons
i
d
er
a
s
i
m
i
l
a
r
exampl
e
,
i
n
whi
c
h
one
s
p
eci
e
s
i
s
a
s
o
ur
c
e
of
food
forl
tbehedrotahsteric. alSuchly overa model
i
s
cal
l
e
d
a
Once
agai
n
,
our
model
wi
simplified in order to il ustrate its main features.
r(t) = - 30e 3 1 + 90e 2 1
�
=
x( 0) =
s(t) = 15e 3 1 + 45e 2 1.
1
exactly
predator-prey model.
Raccoon and squirrel populations
Figure 4 . 2 0
4.20
344
Chapter 4 Eigenvalues and Eigenvectors
Exa m p l e 4 . 4 5
Robionlynsours andce worof food.ms cohabi
t
an
ecos
y
s
t
e
m.
The
r
o
bi
n
s
eat
t
h
e
wor
m
s
,
whi
c
h
ar
e
t
h
ei
r
Theivelyr, oandbinthande equatwormionspopulgoveratnioinnsg atthetigrmoewtyearh ofsthareetwdenoto populed bya­
and
r
e
s
p
ect
tions are
Ithf eintiwtioalpopul
y roabitionnss andover timwore. ms occupy the ecosystem, determine the behavior of
fiandrst thinginwethnote twicoe equat
aboutiotns.his exampl
e eilsyt, hwee prcanesenceget rofidthofe texthemra
conswithtaantsims,plThee change
For
t
u
nat
and ofvariablSubses. Itfitweutilnegt into Equations andwe have then
r ( t)
t
w(t),
r'(t) = w (t) - 12
w '(t) = - r (t) + 10
6
(8)
20
Solulion
- 12
r'(t) = x'(t)
10,
w'(t) = y'(t).
r(t) = x(t) + 10
w(t) = y(t) + 12,
(8),
x'(t) = y (t)
(9)
y'(t) = -x(t)
which is] easier to work with. Equations have the form where
[ . Our new initial conditions are
and
so Proceedi[ -ng:].as in the last example, we find the eigenvalues and eigenvectors of
2 ewhix rootch hass, whinocrhealareroAot1 s. WhatandsAh2ould weThedo?
TheWe havecharanoctechoiristiccepolbutynomito usaeltihseAcompl
=
=
correspemonding oureigenvect
Theor
solutioorns hasare althseoforcomplm ex-namely, v1 [�] and v2 [ -�l By
x' = Ax,
(9)
O 1
-1 0
x(O) = r (O) - 10 = 6 - 10 = - 4
A =
y (O) = w (O) - 12 = 2 0 - 12 = 8
x(O) =
A.
+ 1,
i
=
4.40,
From
x( O ) =
- i.
=
[ - : ] ,we get
whose solution is C1 =
- 2 - 4i
and C2 =
So the solution to system is
(9)
- 2 + 4i.
of thisesxonumber
lution? Robis! Fearnsleands lyworpromceedis innhabig, wet a applreal yworEullde­r's
yetformWhat
ourulasoarluetiowen itnovolmakeves compl
cos sin
e ;1 =
t+i
t
Section 4.6 Applications and the Perron-Frobenius Theorem
345
IN.Si\NC.T.
T\<?sERS AAE
�n ·tum n .
CALV I N A N D H O B B E S © 1 988 Watte rson. R e p r i nted with p e r m i s s i o n o f U N IVERSAL P R E S S SYN D I CATE. A l l r i g hts reserved
(Appendix to get cos ( - t) + i sin ( - t) cos t - i sin t. Substituting, we have
x(t) ( - 2 - 4i) (cost + isin t) l�J + ( - 2 + 4i) (cost - isin t) � �J
(
)
)
+
4s
i
n
t
+
i
2s
i
n
t
4cost
]
[ ( -(42cost
)
(
)
cost
+
2s
i
n
t
+
i
2cost
+
4s
i
n
t
+24ssinint)t)++i (i2(4coscostt -+42ssinint)t) J
+ l ( -(42cost
cos
t
+
[ - 4coscost t++4 sisnint t]
Thiterms sgiofvesourx(tor) igi-nal4 cost+
variables,swein tconcland y(udet) thatcost+ 4 sin t. Putting everything in
r(
t
)
x(
t
)
+
10
4
cost
+
s
i
n
t
+
10
w( t) y( t) + 12 cost + 4 sin t + 12
and
e - it =
C)
=
=
_
=
8
8
=
=8
8
=
=
=
= 8
8
25
20
15
10
5
15
0
2
4
6
8
10
Robin and worm populations
Figure 4 . 2 1
12
14
30
16
Figure 4 . 2 2
346
Chapter 4 Eigenvalues and Eigenvectors
Sotheourtwosolpopulutionatiisornsealosafctielraaltel!perTheiodigrcaalphsly. ofAs theandrobin populin Fiagturioen increshasesow, tthhate
wortheimr number
populastisotnarstttaortdecls to idecrne aseaswele, lbut. Asasthethpre erodatbinors's onldisyappear
food,stohure worce dimmpopul
inishesa­,
tandion tbegihe cyclns etorerpeatecovers it.seAslf. Thiits foods oscisluappltiony iisntcryepiasescal ,ofsoexampl
does tehseinrowhibinchpopulthe eiatgien­on,
valuPlesoart iencompl
enx.s, worms, and time on separate axes, as in Figure clearly
g
r
o
bi
reveals the cyclic nature of the two populations.
secta diioffnerbyentlioaokiblenfunctg at whation oftwe, havethendonethe generfromaal sdiolffuetrioentn ofpoithnet
ofordviiWenearw.yconclIfdiffuerdeentthiailissequat
ionwe have beenis consid,erwhering havee is tahscale forarm. The syste,mssoofif
lweinearsimdiplffyeplreontwedial equat
i
o
ns
ahead witwher
hout ethiniskianvectg, weor.miButghtwhatbe teonmptearedthtocouldeduced thisthmean?
at the
sOnolutthioenrwoul
d
be
i
g
ht
h
and
s
i
d
e,
we
have
t
h
e
r
a
i
s
e
d
t
o
t
h
e
power
of
a
Thi
s
appearLets'stsotabert nonsby consensied,eryetingyouthewiexprl seees tihoatn therIeniscala cwayulus,toyoumakelearsennstheatoftiht.e func­
tion has a power series expansion
2! 3!
that converges for every real number By analogy, let us define
2! 3!
Theit converrightg-eshandfor anysideriesaljumatst defirixnedSoi nnowterms ofispower
siofx, callanded thitecan be shownofthat
a
mat
r
But how can we compute or For diagonal matrices, it is easy.
Compute for [ ].
From the definition, we have
2! 3!
] [ ] [
]
[
r(t)
w ( t)
4.2 1
4.22,
x = x(t)
x' = ax x = c e a1
x = ce A 1,
c
x' = Ax
c
number e
eA.
ex
matrix.
xz x 3
ex = 1 + x + - + - + . . .
x.
Az A3
eA = I + A + - + - + · · ·
eA
Exa m p l e 4 . 4 6
eD t
D=
A.
e A 1?
A,
eA
exponential
4 0
0 -1
Solution
e
Dt
=
I + Dt +
(Dt) 2
--
+
(Dt) 3
--
+· · ·
(4t) 3
(4t) 2 0
0
+ _i,2
+
· 0 ( - t) 2
31
0
-t
1 + (4t) + t C4t) 2 + t, C4t) 3 + · · ·
o
1 + C - t) + ti C - t) 2 + t, C - t) 3 + · · ·
0
et
e� t
I
[; ]
The matrix exponential is also nice if is diagonalizable.
A
A.
Section 4.6 Applications and the Perron-Frobenius Theorem
Exa m p l e 4 . 4 1
Compute eAforA = [� � ].
In Example 4.43, we found the eigenvalues of-A],to be ,\1 = and ,\2 =
with corresponding2eigenvectors = [ � ] and v2 = [ � respectively. Hence, with
P = k V2 ] k=-[1 3 -l ],wehaveP- 1AP = D = [ 0 0 ].sinceA=PDP - 1, we
have A = PD P , so
eA = I A A2!1 A3!3
= PIP- 1 PDP- 1 -2! PD 2P- 1 -3! PD 3P- 1
= P(I D -D2!z -D3!3 )
0- 1 ] [ 23 ]
e
[ 2e3e44 - 3e3e-- 11 2e3e44 - 2e2e-- 11 ]
are nowial"insoalupostionitiofon to=showwasthatnotoursobolfardoff(andaftseeremiall!ngly foolish) guess at
an "Weexponent
Let A abel solanution todithaegonalsysteimzable=matrixis wi=th eA1eigcenval, wherueesc,\is1,an,\2,arbit,raryThenconsttahnte
gener
vector. If an initial condition is specified, then c =
Let P diagonalize A. Then A PDP - 1, and, as in Example
1
Hence,
we
need
t
o
check
t
h
at
i
s
s
a
t
i
s
fi
e
d
by
Pe01
P
c. Now, everything
is constant except for e01, so
( 1 0)
If
4
Solution
v1
[vi
4
1
+-+-+
+
+
+
+
·
·
1
1
+
+
+
·
-1
+
·
·
·
+
·
·
·
p- i
- 1 -I
1
_!_
5
+
+
x'
Theorem 4 . 4 1
341
n
X
n
Ax
•
x' Ax x
x ( O)
x'
=
Ax
.
Aw
x ( O).
=
Proof
.
4.47,
x=
- 1,
348
Chapter 4 Eigenvalues and Eigenvectors
then
Taking derivatives, we have
[I
e Dt =
0
e A2 t
0
0
0
0
). , ]
0
0
d ( A,t
e )
dt
0
� n·, JJ
� [ ! �r·' J,]
0
A z e A,t
0
0
A1
0
e A,t
0
An
0
0
Substituting this result into Equation we obtain
as reThequirlead.st statement follows easily from the fact that if
then
since (Why?)
, Theorateimon ofmatis rtirxueexponent
even if ialsisfornotnondidiagonalagonalizablizable, ebutmatwericwies lrenotquirpres othvee
tInhisfact. Comput
in more advanced linear
algebrIdeala telyxt, tsh. is shortofdiagrmates iroixn, hasa tospiercvtedhattomayil usbetrafound
t
e
t
h
e
power
of
mat
h
emat
i
c
s
t
o
gener
imporatlaintze andtoolsthine valmanyue ofapplcreicatativioensthiofnkilinng.earMatalgrebrix exponent
a, both thieorals etutircnaloutandtoapplbe veried.y
Wechaiconcl
er as weofbegan
itt-ionbygrloookiwthngarate dynami
csalofsystems. Markov
ns andudethtehiLess chaptliEache model
popul
a
exampl
e
can be described by a matrix equation of the form
( 10),
x' = PDe Dtp - 1 c = PDP - 1 Pe Dtp - 1 c = (PDP - 1 )(Pe Dtp - 1 )c = Ae A tc = Ax
x = x ( t) = e A tc,
x(O) = e A · Oc = e0c = le = c
e0 = I.
see A. J. Insel, and
byL.ForE.S.example,
H.Spence
Friedberg,
(Engle1979).
wood Cliffs,
NJ: Prentice-Hall,
Linear Algebra
4.41
A
Jordan normal form
Discrele Linear ovnamical svs1ems
discrete linear
dynamical systems.
Section 4.6 Applications and the Perron-Frobenius Theorem
349
wher
ate oftohre ofsystthemeseatsy"tsitmemse" isandrelateids atosqthuare eie gmatenvalrix­.
Asuesweande thaveheiegvectenvect
seen,or othrrseeofcorlong-thdes ttmatehremsrtibehavi
x oxiThematpower
methuodes explandoeiitsgenvect
the iteroartsi,veandnatuthree
ofPersruochn-Fdynami
c
al
s
y
s
t
e
ms
t
o
appr
e
ei
g
enval
inforcmieatntiomatn aboutrix thisenonnegat
long-termivbehavi
or
ofaWhen
discrertoebeniliins earaus2Theor
dynami2 matemcalgirisxvy,esswetsepmecicanwhosaldesizedeccoeffi
e.
reibalelythane evolinfiuntiitoencolofleactdynami
cal syisotensm.
geomet
r
i
c
al
l
y
.
The
equat
i
o
n
i
s
r
i
o
n
of
equat
Beginning with an initial vector we have:
xk
A
k
A.
A
A
X
xk + I = Axk
XcJ,
X1 = A Xo
x2 = Ax1
x3 = Ax2
wiNotThel eisetdthentatify each vect} oisr calin laetdraajectory withofittshheade systseom.th(atForwegrcanaphiplcoalt purit aspaospoies,nwet.)
Letin the traj[e0.0ct5orie0.0s 8wi] .thForthethfole dynami
c
al
s
y
s
t
e
m
pl
o
t
t
h
e
fi
r
s
t
fi
v
e
poi
n
t
s
lowing initial vectors:
(a) [ Os ]
1.0 25 ]
2.0 5 ]
(
a
)
We
comput
e
[
[
0.0 625 ] , [ 0.0 3 125 ] . These are plot ed in Figure 4.23, and the points are
[connect
marked ed to higandhlight itnheFitgraurjeect4.o2ry3.. Similar calculations produce the trajectories
{ Xo , x1, x2 ,
.
•
trajectory
.
xk = A k x0 .
Exa m p l e 4 . 4 8
xk + 1 = Axk ,
A =
Xo =
x 1 = A Xo =
Solution
, x2 = Ax 1 =
X4 = Ax3 =
( b), (c),
(d)
y
X4
-2
(a)
Xj
X3 X 2
2
Xo
4
6
x
-2
-4
(b)
Fioure 4 . 2 3
, x3 = Ax2 =
350
Chapter 4 Eigenvalues and Eigenvectors
everysttraandjectwhyory conver
in thInisExampl
case. Wee can under
t] his i]s gsoesfrtoom TheTheororeimgin is calTheled anmatrix A in
Exampl, respeectivelhasy. (Ceiheckgenvectthios.r)sAccor
[ � anddingl[ �y, focorrr anyespiondinitianl vectg to oitrs eigenvalues and
4.48,
0.
attractor
4. 19.
4.48
�
0.5
0.8
we have
Because
zero askem gets ltahrgate,becausappreo0.aches8 is thefordomiany­
choinantceie gofx0envalbot•hIuneaddiof A,tandion, wiwel knowapprapprofroachachomaTheor
mul
t
i
p
l
e
of
t
h
e
cor
r
e
s
p
ondi
n
g
ei
g
envect
o
r
cient of corresponding to [ � ]) . In other words,
[al�l ]trasajelctongorieass except th(toshee coeffi
y-axis, as Figure shows.that begin on the x-axis (where wil approach the
(0.5) k
(0.8) k
xk
c2
*0
Xo
c2
4.23
Exa m p l e 4 . 4 9
= O)
Discuss the behavior of the dynamical system xk+i Axk corresponding to the
]
matrix A [
The eigenvalues of A are and 0.8 with corresponding eigenvectors [ � ]
andwe have[ - � ],respectively. (Checkthis.) Henceforaninitialvector Xo [ � ] [ � l
xk Ak ( )k [ 11 ] (0.8)k [ - 11 ]
Once agaithen ttrhaejeorctoigriynwiislanapprattroaachctort,hbecaus
ehxrkoughapprtohachese originfowir tanyh dichoirectcioenofvectoIrf
e
l
i
n
e
t
[ - � ]. Several such trajectories are shown in Figure The vectors where
arineg tonrajethcteolriyneintthhriosughcasethfole lorowsigitnhiwis ltihnediinrteocttihoenorvectiginor. [ � ], and the correspond­
=
0.65 -0.15
=
.
-0.15 0.65
0.5
Solution
�
0
xk
4.28
=
=
c2
Xo
= c1
0.5
+
c,
+
c2
-
c2
0
* 0,
4.24.
Xo ·
Xo
c2
=0
Section 4.6 Applications and the Perron-Frobenius Theorem
351
y
Figure 4 . 2 4
Exa m p l e 4 . 5 0
folDislocwiussngthmate behavi
rices: or of the dynamical systems corresponding to the
(a) [ � �] (b) [ �.5 �-S J
(], a) The eigenvalues of are 5 and with correspondi] ng ei-genvect], ors [ � ]
and [ � respectively. Hence for an initial vector [ � [ � we have
xk + 1
A
A
=
Axk
=
A
Solution
=
3
Xo = c 1
+
c2
AsBecausbecomes
l
a
r
g
e,
s
o
do
bot
h
and
Hence,
t
e
nds
away
fr
o
m
t
h
e
or
i
g
i
n
.
5
],
e
t
h
e
domi
n
ant
ei
g
enval
u
e
of
has
cor
r
e
s
p
ondi
n
g
ei
g
envect
o
r
al
l
t
r
a
j
e
c­
5
[
�
tories for which 0 wil eventually end up in the first or the third quadrant. Tr-lajec­].
tSeeorieFis gwiurteh 4.25(a0).start and stay on the line - whose direction vector is [ 1
(b) In this-example, the eigenvalues are 1.5 and 0.5 with corresponding eigenvectors
[ � ] and [ � ], respectively. Hence,
k
k
c1
c1 =
3k_
xk
-=fa
y =
x
352
Chapter 4 Eigenvalues and Eigenvectors
y
y
20
- 20
(a)
Figure 4 . 2 5
If c1 = 0, then xk = c2(0.5)k [ - � ] [ � ] ask ---+ But if c 0, then
---+
1
cxi.
-=fa
and such trnjeoto;ies osymptotkally apprnooh the line See Egme 4.25(b) 4
etu4.de50(becaus
a), all epoints th1atforstabotrt outh eineargenvalthueesor; iginis become
increasinglIny
lExampl
argIeninExampl
magni
a
cal
l
e
d
e
4.
i
s
cal
l
e
d
a
becaus
e
t
h
e
or
i
g
i
n
at
t
r
a
ct
s
poi
n
t
s
i
n
s
o
me
5
0(
b
direThectionsnextandexampl
repelsepoishnowsts inwhatothercandirhappen
ections.whenIn thitshcase eieg, envalues1 andof a real 2 1. 2
matrix are complex (and hence conjugates of one another) .
y �
), 0
Exa m p l e 4 . 5 1
l,t l >
saddle point
x.
0
l -\1 1 <
repeller.
l -\2 1 >
X
Plcorotretshpeonditrajnectg otorythbegie folnloniwingngwimatth rices=: [ ! ] for the dynamical systems xk+ 1 = Axk
(a) A = [ 0.0.55 - 0.0.55 ] (b) A = [ 0.0.62 - 1.1.42 ]
ure 4.e2as6((a)b)andappear(b),srteospfoectl oivwelany. Notellipettihcatal
(ora)biits. a trajTheectortryajsepctiroalriinegs arinetoshtownhe oriingiFin,gwher
Xo
Solulion
Section 4.6 Applications and the Perron-Frobenius Theorem
y
y
-4
-
(a)
Figure 4 . 2 6
Theorem 4 . 4 2
353
1
0
(b)
ExamplThee 4.folSllo(wia)n. g theorem explains the spiral behavior of the trajectory in
LetbothAzer=o,[tabhen-Aab Jcan. Thebe facteigenvaloreduases of A are = a ± bi, and if a and b are not
A = [ ab - abJ = [ r Or J [ cs�ms()() -cos()sin () J
where r = I A I = Va2 b2 and() is the principal argument of a bi.
TheA = eit(g2enval
u
es
of
A
ar
e
(
)
a
±
=
2
a
±
2\/b2\/=l
=
a
±
b
i
=
a
±
bi
t
l
l
by Exercise 35(b) inaSect-iobn 4. 1 . Fia/rgure 4.-2b/r7 displaysra O bic, r,s()and()-.sIint fol() lows that
A = [ b a J = r [ b/r a/rJ [ r J [ sm� () cos() J
Geomet
r
i
c
al
l
y
,
Theor
e
m
4.
4
2
i
m
pl
i
e
s
t
h
at
when
A
=
[: �b J
thce�sl()inear- sitnra()nsfothrrmoughationthe angle A() folilsowedthe bycompos
i
t
i
o
n
of
a
r
o
t
a
t
i
o
n
=
r
O
a scaluiesng are= [ = 0.r5J ±wiOth.Sifac-so
[tosrmr() (Figurcos()e 4.J 28). In Example 4.S l (a), the eigenval
r = I A I = \./2/2 1, and hence the trajectories all spiral inward toward
The next theorem shows that, in general, whena a-realb 2 2 matrix has complex
eigenvalues, it is similar to a matrix of the form [ b a J . For a complex vector
x = [ :J [ ::�:J = [ : J [ �}
,\
0
+
+
Proof
�)
+
=
Im
0
-=F 0,
Remark
a + bi
r
e
Figure 4 . 2 1
T(x) = x
S
=
b
R
0.707
,\
<
X
+
0
0.
354
Chapter 4 Eigenvalues and Eigenvectors
y
ln /
Sca i g
Ax
=
SRx
Rx
/
Rotation
rotation followed by a scaling
Figure 4 . 2 8
A
we define the real part, Re x, and the imaginary part, Im x, of x to be
Rex [ :] [ :::] and Im x [ �] [ :::]
andandLet corrbe easpreondial ng eimatgenvectrix wiorthx.aThencompltheexmateigenvalrix ue A[ Re x Im x](wheris inevertible
=
=
Theorem 4 . 4 3
A
2X2
P=
= a - bi
b i= 0)
Let x Axso that Rex and Im x From x Ax, we have
Equating real and imaginary parts, we obtandain
Now[ so
] [ ]
to s,htowhenthitatwoulandd folarelowlinthearatly indepen­for
dentsomeTo. I(fnshonzerowandthoatcompl
weriseinnoteverx) slticinbalearlae,rilyt iisbecaus
nenough
dependent
e neither nor Thus,
x
Now, because is real, x Ax implies that
Ax
Ax
Butso is an eigenvector corresponding to the other eigenvalue A
= u + vi
=u
= v.
A =
Au + Avi = Ax = = (a - bi) (u + vi)
= au + avi - bui + bv = (au + bv) + ( - bu + av)i
Proof
Au = au + bv
Av = - bu + av
P = [u l vJ ,
a -b
a -b
= [ul vl
= [au + bv l - bu + av] = [Au l Av] = A [ u l v l
P
b a
b a
= AP
P
u v
u
v
v = ku
k,
u v is 0.
= u + vi = u + kui = ( 1 + ki)u
A
A =
Ax = Ax = Ax = =
x = u - vi
= a + bi.
x = ( 1 + ki)u = ( 1 - ki)u
Section 4.6 Applications and the Perron-Frobenius Theorem
355
becaus
muleigenvect
tipleesoofriss coraandrerealstphvectondiereforonr.geHence,
artoedimulstitnhtctiepleieiesggenvect
ofenvaloneuoesanotrsmusxhandert.beThiliofsniearsAimlaryposienbotdependent
sibhlenonzer
becausbyoe
Theorems.)4.20. (This theorem is valid over the complex numbers as well as the real
number
inverThitibslecont. It nowradicfoltiolonwsimtplhaties that and are linearly independent and hence P is
Theor0.2 em- 1.4.24]3 serves to explain Example 4.5 l (b). The eigenvalues of
A = [ 0.6 1.4 are 0.8 0.6i. For ,.\ = 0.8 - 0.6i, a corresponding eigenvector is
Frhaveom Theorem 4.43, it follows that for P = - 01 ] and C = [0.0.68 - 0.0.68 ] , we
A = PCP- 1 and P- 1AP = C
For the given dynamical system 1 = Axk, we perform a change of variable. Let
Then
so
NowChast
calthe sorysitgeimn byhesameei
=TheorCyemkg4.enval
sim42.pluyesasA(
rotates wthhy?e poi) andnts i0.n 8ever0.y6triaje=ct1.Thus,t
ory in a cihedynami
rcle about­
cttroarnsyfoofrmthateiodynami
c=al Axsyst=emPCPin Exampl
e t4.ra5nsl (fob)r,mweatioitn­
1
(
)
ercanatiTobeveltdetyhapploughtermyitnofheeasalinttrhearaejecompos
n
T
x
x
.
The
ofvariaofblevar(xitaobly)e (,yfolbacklowedto x)by. Wethe
rwiotlatencount
ion deteerrmtihnieds idbyeaC,agaifolnlowedinittihobyen applofthae irchange
ecatveriosne tchange
os"grinaphiSectnigonquadr6.3. aInticExerequatcisieons74 iofn
SectSectiioonn 5.5.55,and,you wimorl sehgener
a
l
y
,
as
"
c
hange
of
bas
i
as itToappearsummars to ibeze frthoen:mowIFifgaturhreateal4.t2h2e6(trb2a)jmat.ectorriyx AinhasExamplcomple 4.ex5ei1 (gbenval) is inudeedes,.\ =an ellipsbie,,
thenis athe trajectories of stphieradynami
cald isfy,.\stem 1 i=s aAxk spiral inwardandif l,.\ie on a1
l
out
w
ar
closed orbit if ,.\ = 1 is an
u
x
u
u
v
±:
xk +
�
l
Yk +i
X
(O
1 1
spiral attractor) ,
(O
±:
l
a
1 1
>
orbital center) .
xk + I
(O
spiral repeller) ,
1 1
±:
<
Vi gnette
Ranking Sp o r t s Te a m s and S e arch ing
t h e I nternet
Inrankanythecompet
istiorve tsepamsorts. Count
league,inigt iwis notns andneceslossaersilyaloanestroveraightlofooksrwarthde proces
slittoy
pl
a
yer
pos
s
i
b
i
tanothat honeer tteeamam maymay haveaccumulfewaerteviacltaorgrieesnumber
of vithcemtoriagaies againstnssttroweakng tetamseams. Whi, whiclhe
but
al
l
of
ofanotthhesere?tShoul
eams dispoibetnttesrs?cHow
shtoulakend weintocompar
e t?wPoio tnetamss agaithnatst?never play one
or
e
d
be
account
Despite thaceeseandcomplmuch-exitiaents,itchipeatraenkid fenatguofreatinhltehteesmediand asp. Fororts exampl
teams hase, thbecome
avarcommonpl
erseandare
itenniouss plannual
r
a
nki
n
gs
of
U.
S
.
col
l
e
ge
foot
b
al
l
and
bas
k
et
b
al
l
t
e
ams
,
and
gol
f
er
ustheedprtooblpraeyeromducebys arusesuialnchsgotrhraaenkinkedidneasgsi,nfrbuttoemrnweatthiioscannalchaptlgaiy. Therenr.somee arienmany
sight incopyrto howightteodapprschemesoach
let's revi. sWiit Exampl
elos es arFiveeretecornnidsedplianyerthseplforaymoneof
anota digTohraerphesitnainbla whiirsohund-tchheabasrodibiricenctitdoeea,durnament
n
s
and
edgerifrx omtherefore
to indihascates that1 plifaplyerayerdefeatdefesatpls playerayer
Theandcorhasrespondi0notghaderwacency
mat
ise.
3.69.
aiJ
j
j
=
A
5
i j
aiJ
2
4
3
A=
i
=
0
0
1
0
0
1
0
0
0
0
j.
i
0
1
0 1 0
0 0 1
1 0 0
We iwoulndicatd elsiktehatto plasasyerociate isa rraanked
nking morwie thhigplhlayyerthaninplsauyerch a wayFor tthhatis
r; > r1
356
i
r;
i
j.
purpose, let's require that theand'sthbeenprorogbabianizleittiehse (rtahnkiat nisgs, in a for all and
r1 + r2 + r3 + r4 + r5 = 1 )
0 :s r; :s 1
i,
ranking vector
r
;
d be pre,oplporayertionaldefeatto thede plsuamyerofs
tFurhe trhaandnkiermnorgsseo,oflweetth'swant
eiplnsaisyert ths atdefeatplayered byi's plraankiyerng Forshoulexampl
he consthteantfoloflowiprnoporg sytsitoenalm:ity. Writing out similar equations for the other
plwherayeres prisotduces
2, 4,
i.
5,
1
a
r 1 = a(r2 + r4 + r5 )
r2 = a(r3 + r4 + rs )
r3 = a(r 1 + r4 )
Observe that we can write this system in matrix form as
or
'1
r2
f3 =
f4
f5
ll'
0
0
1
0
0
1
0
0
0
0
0
1
0 1 0
0 0 1
1 0 0
'1
r2
f3
f4
f5
r = aAr
Equiis anvaleiegntenvect
ly, weosrecore thratespthondie rannkig tnogtvecthe mator rixmust satisfy In other words,
Fur
t
h
er
m
or
e
,
i
s
a
pr
i
m
i
t
i
v
e
nonnegat
i
v
e
mat
r
i
x
,
s
o
t
h
e
Per
r
o
nF
r
o
beni
u
s
Theo­
vectremoguarr turannts eoutes tthoatbethere is a ranking vector In this example, the ranking
r
r
A
1
Ar = -r.
a
A!
unique
r.
0.29
0.27
r = 0.22
0.08
0.14
so weBywoulmodidfryainkng tthhee plmatayerrixs in tiht eisorposdersible to take into account many of the com­
plseerxivedtietsomentindiciatonede oneinustehfule openiappronachg partoatgrheaph.problHowever
em of ra, nkithins gsitmeams.ple example has
A,
1, 2, 3, 5, 4.
351
The
same
i
d
ea
can
be
us
e
d
t
o
under
s
t
a
nd
how
an
I
n
t
e
r
n
et
s
e
ar
c
h
engi
n
e
s
u
ch
as
Googlefulesworites kwouls. Oldderoftseenarbechburengiiendesamong
used toir reelteuvantrn thonese res.uMuch
lts of ascsreoarl cinhg was often
Usneeded
sisuneeded.
lts to uncover
accordwhating toyoutheiwerr liekelloyokirelnegvance.for. ByThuscont, raametst, Googl
hod fore reratunkirnns gsewebsarchirtees­
, we nowthehavesituatwebsion, onlitesylinownkinang tedgeo onefran­om
ottoherjIni.nsWetdieadccanatofesoncettheamsat websagaiplaniyitusenegliaonendiksgtranotoaph(ohrtoerremodel
tdieamgraph,digrianph,comiinncomig dirnegctdiedreedges
cted edgesare goodarefebadr(sthtoey(t)hwebsieyndiincidiatteecj.atliSone kslowhersfreos)me,asforotforhterhtehsIeintestps)eror.nIettns
tinhgsis sofetalinl g,thwee webswantitetshtehatrankilinkngtoof website to be proportional to sum of the rank­
Using the digraph on page to represent just five websites, we have
formatexampl
e
.
I
t
i
s
eas
y
t
o
s
e
e
t
h
at
we
now
want
t
o
us
e
t
h
e
of
t
h
e
adj
a
cency
Thereforor eof, the rIannkithnisgexampl
vector er,muswe obtt saatiisnfy r -rand wil
thusribex ofthtehPere dirgornaph.eigenvect
.
and r
so aGoogl
searche actthatualtulyrnuss eups atvarhesieantfivofe stihteesmetwoulhodd ldesist cthriembedihern theeandordcomput
er es the rank­
ing vector via an iterative method very similar to the power method (Section
unordered.
ordered
i
i
i
i.
356
r4 = a(r 1 + r2 + r3 )
A r.
0 0
0
T
A = 0 1
1
1
1
0
0
1
0
0
0
0
0
0
0
1
0
0
transpose
1
AT = a
0 14
0.08
= 0.22
0.27
0.29
5, 4, 3, 1, 2.
4.5).
358
I
Section 4.6 Applications and the Perron-Frobenius Theorem
Exercises 4 . 6
17.
Markov Chains
Which of the stochastic matrices in Exercises 1-6 are regular?
1.
3.
5.
[� �]
[i � ]
O[ 0.J5 0 � ' ]
0.4 0 0.5
2.
4.
1
6.
[ � !]
0
[! 0 �]
[ �".5' 00 � ]
10.
P - 1 LP
18.
[l [
[i iJ
02[ 0.6 0.0.31 00.4
]
0.
2
0.
2
0.
6
eovstchaieadynsitsatunie prqoue.bability vectUseorTheo­of a
rremPregulove4.art3h3MaratorthkTheor
em 4.34.]
3
2
I
6
8. P =
4
In Exercises 1 1 -14, calculate the positive eigenvalue and a
corresponding positive eigenvector of the Leslie matrix L.
1
11. L =
12. L =
15.
16.
[ 0.5
[ 0.O 5 � ]
7
5
L [�s 00.5 � ] [i 0 � ]
If a Lesisltiheematsigrniixfihascancea uniforqtuehe pospopulitivaetieiongenvalif ue
what
VermatirfiyxtLihatntEquat
he charioanct(e3r)isistic polynomial of the Leslie
()
det(L Use matalohngemattheiclaalstincolductumn.ion] and expand
14. L
�
2
3
A1 < l ? A1 = 1 ?
S1
A1,
A1 > 1 ?
cL A = ( - l ) " (A " - b 1 A " - 1 - b 2s 1 A " - 2 - b 3s 1 s2 A " - 3
- · · · - bns 1 s2 · · · Sn - I )
[Hint:
- Al)
S1S2
S1S2 · · · sn - 1
L. [Hint:
L
A1
S1 / A 1
S1S2 / A i
S1S2S3 / A i
l
[Hint:
�
1
1
Population Growth
13.
s;
l
7. P =
9. p =
If all ofthe survival rates are nonzero, let
0
0
0
00 0 0 00
00 0
Comput
tSectic polionynomi4.e 3.] al ofand use itRefeto firndtotExerhe charciseact32eriins­
Verify that an eigenvector of corresponding to is
P=
In Exercises 7-9, P is the transition matrix of a regular
Markov chain. Find the long range transition matrix L of P.
359
SectionCombi
4.3 andneExerExercciissee4617iabove
n Sectiwionth4.Exer4.] cise 32 in
[Hint:
GAs In Exercises 1 9-21, compute the steady state growth rate of
the population with the Leslie matrix L from the given
exercise. Then use Exercise 18 to help find the corresponding
distribution of the age classes.
19.
20.
21.
GAs 22.
ExerExercciissee 4039 iinn SectSectiioonn 3.3.77
ExerManycisspe eci44eisnofSectsealionhave3.7 suffered from commercial
huntand meating. They
havefur trbeenade, ikin lparedtforicultahrei, rresduced
kin, blsuobberme ,
.
The
pointaoftioextnsiarenctideclon. iToday,
tshhe
grssteoalecksatpopul
esduet thartteoiatoovernss totofisealtshheinpopul
n
e
of
fi
habiby fitsahter, enty owners
anglement. Someing,smarepolalsilnuhaveetiodebrn,beendiiss,tuanddeclrbanceculareldiofng
endangereTabld specie 4.e7s;giotvheserthspeecibiretsharande carsuerfulvivlyal rates
managed.
forclasthees. nor[Thetherdatnafuraresealbas,edidvonided iE.ntoYor2-kyearandageR.
HarNorttlheery,n"PFurup Product
i
o
n
Fol
l
o
wi
n
g
Har
v
es
t
of
Femal
e
Seals(1981)
;' , pp. 84-90.]
A.
Aquatic Science, 38
J.
Canadian Journal ofFisheries and
Chapter 4 Eigenvalues and Eigenvectors
360
(b)
sShowents that = if and only if ,\1 = Let(This repre­
g(,\) = ,\ ,\2 ,\3
"
,\
Show
t
h
at
,\
i
s
an
ei
g
enval
u
e
of
i
f
and
onl
y
i
f
g(As,\s)umi= n1.g] that there is a unique positive eigen­
lvalpopul
atiuoen,\aits1i,odecrsnhowiseiasnthcrinategasanding. if andifonlandyonlif thyeifpopu­
the
r 1
1.
zero population growth.) [Hint:
bnS1S2 · . . Sn - 1
b l bzS1 b 3S1S
- + - + --2 + . . . + ---L
(c)
A sustainable harvesting policy is a procedure that allows a
certain fraction of a population (represented by a population
distribution vector x) to be harvested so that the population
returns to x after one time interval (where a time interval
is the length of one age class). If h is the fraction of each
age class that is harvested, then we can express the harvest­
ing procedure mathematically as follows: If we start with
a population vector x, after one time interval we have Lx;
harvesting removes hLx, leaving
Ta b l e 4 . 1
Birth Rate
0.00
0.02
0.70
1 .53
1 .67
1 .65
1 .56
1 .45
1 .22
0.91
0.70
0.22
0.00
Age (years)
0-2
2-4
4-6
6-8
8- 10
10-12
12-14
14- 1 6
16- 1 8
18-20
20-22
22-24
24-26
r< 1
r> 1
Survival Rate
0.91
0.88
0.85
0.80
0.74
0.67
0.59
0.49
0.38
0.27
0.17
0.15
0.00
Lx - hLx
=
( 1 - h)Lx
=
mattIhf ath,\1riixs theanduni1/,\qiues 1t.hpose siutisvteaieinablgenvale haruveesoft araLestiol,ipre ove
Fiandnd cartheibsuoustaininExerable charisevestinraSecttio forionthe wood­
Cons
t
r
u
ct
t
h
e
Les
l
i
e
mat
r
i
x
for
t
h
es
e
dat
a
and
l
comput
eigenvalor.ue and a corre­
Useduceing tthheedatcaraibinouExerhercdisaccor
e idninSectg toioyourn answer
sInpondithe lnoegngtposheruposin,tivwhatietieivegenvect
r
each age clas and whatperwicentl thaegegrofowtsealh sratwielbe?be in Findtorothipargeinsutals(talae)ivel.nVerablafeitfyeharrtonehvatesttthimreapopul
etioinfoterravtthalioe.nserealtuinrns to its
Exerseal popul
cise ati(oCnsonswhenervatoverionifistsshhaveing hashadretduced
o harvteshet
lable foodof stasurvpplatiyon.to)the point where the seals are
The =
of a population is defined as avaiiLetn danger
be,\1.aShowLesliethmatat ifri,\x iwis anyth aotuniherque(reposal oritivcom­e eigen­
val
u
e
wher
rth raatteison.and the are the
plr(ecx)os eigenvalsinu()e)ofand tshuenbst1i,\tu1 te i,\t 1in. to the Wrequatiteio,\=n
survExplivealthraeaintewhysarfore ththecanebipopul
bers iborntenrptroetaesdinasgltehfeemalavereaoverge g(Theor,\) =em andas inthparentta(bke) ofExer
ciparse t ofUsbote hDesidMoies. vThere's
number
of
daught
e
t
h
e
r
e
al
her lifetime.
Triangle Inequality should prove useful.]
(a)
L
Sustainability requires that
(1 - h)Lx
24.
cAs
x
L
h
=1-
25. (a)
44
44
(b)
3.7.
3.7,
(b)
26.
Exercise 23 shows that the long-run behavior of a popula­
tion can be determined directly from the entries of its Leslie
matrix.
23.
net reproduction rate
� 27.
r b 1 + b 2s 1 + b 3 s 1 s2 + · · · + b n s 1 s2 · · · s n - l
b;
sj
(a)
r
22.
L
() + i
1,
L,
:s
[Hint:
23.
Section 4.6 Applications and the Perron-Frobenius Theorem
In Exercises 28-31, find the Perron root and the correspond­
ing Perron eigenvector ofA.
2[ 1 01 ]
[ : � i]
28. A =
30. A
�
29. A =
31. A
�
(
1[ 2 03 ]
[: 0 �l
�
35. A =
36. (a)
A
A
[ !0 �1 �0 �0l 0
G,
G
connected
(b)
37.
( a)
(b)
G
A.
A
A,
[Hint:
l cA I = l c l I A I
(b) I A + B l ::; I A I + I B I
I A x l :s I A I l x l
(d) I A B I ::; I A I I B I
a a
41.
2X2
A = " ,2 ·
a z1 a z 2
a1 2 =
a2 1 =
42. A
(I A - 1 0.
I A
v1
A.
( a)
< 1 < [Hint:
22
( a)
(c)
Prove that a matrix 0.[ ] red 'ble
iLetf and beonlaynonnegat
if 0ivore, ir educible matrix such
tAh1atand- beistihneverPertirbolne androot and- Per) ron2eigenvect
Let or
of Prove that 0 A 1. Apply Exercise
iDeduce
n Sectiofrno4.m3 (anda) thTheor
e
m
4.
1
8(
b
)
.
]
at
IS
Linear Recurrence Relations
In Exercises 43-46, write out the first six terms of the
sequence defined by the recurrence relation with the given
initial conditions.
43. Xo = Xn = 2xn - I
n
44. a , =
a n = a n _ , /2 n 2
45. y0 = O, y 1 = yn = Yn - I - Yn - z n 2
46. bo = b , = b n = 2bn - I + b n - 2 n 2
1128,,
1,
0,0,
( a)
=
A
[Hint:
G
l1,,
for fo2r 12
forfor 22
In Exercises 47-52, solve the recurrence relation with the
given initial conditions.
47. Xo = X 1 = Xn = 3Xn - I + Xn - 2 n 2
48. Xo = X 1 = Xn = 4Xn - I - 3Xn - 2 n 2
k-regular
G
UC!
v1 > Av, .
(b)
A.
38.
[Hint:
i
00 01 0 00 0 0 0 0 0
10 00 1 01 01 01 00 01 00 01
0
0
0
1
1
0
0
0
0
showed.
tI(hfAatgrisaphtihs eiirsadjreduciacencyble imatf andif trhixeronlofe yiasigraf pataphishconnect
everWhicyhpaiofrthofevergratphsices.i)n Section 4.0 havebetanween
ir iemduciitivbeladje adjacency
acencymatmatrixri?x? Which have a
pr
Let Showbe atbihatpartiitsenotgraphprimwiitthivadje. acency matrix
Show tUsehat iExerf A iscanise ei80genval
ueioofn 3.7sando is par- A.tition
i
n
Sect
anthiseipargenvecttitioniorngforofA soUstheatthitisisparcompat
ibnlge twio fithnd
t
i
t
i
o
ni
an
ei
g
envect
o
r
for
A.
]
Atex.grLetaph isbeakcalledregular graiph.f k edges meet at each verEV
ShowA thatk astheanadjeiagcencenvalyumate. rix ofAdapt
hasTheor
em 4.30.]
34. A =
39.
A
In Exercise 40, the absolute value of a matrix A = [ a j is
defined to be the matrix IA I = [ l a u l
40. A
B nXn
x
IJ�r,
c
It can be shown that a nonnegative n X n matrix is irreduc­
ible if and only if I + A) n - t > 0. In Exercises 32-35, use
this criterion to determine whether the matrix A is irreduc­
ible. If A is reducible, find a permutation of its rows and
columns that puts A into the block form
33. A
Showuestarhate alifl leiss prthiankmitivine,absthenoluthteevalothuere. eigen­
valAdapt
Theor
e
m
4.
3
1.
]
iExpln ligahtinofthExere resucilstessof36-38your andexploSectratiioonn i4.n5Sect. ion 4.0
]
].
Leta scalaandr. Provebethe follomatwinrgicmates, raixvectinequalor inities: and
(b)
The Perron-Frobenius Theorem
361
52.
53.
5,1,
4 forfor 22
Theolutrieocurn agrrenceees wiretlahttiohne ansin ExerwertcoisExere 45.cShowise 45.that your
sCompl
that if tehtee rthecure prroenceof ofreTheor
lationem 4.38(a) by showinhasg
Xn = axn - + bxn - 2
i
Chapter 4 Eigenvalues and Eigenvectors
362
diofstthinectforeimgenvalues ,\2, then the solution wil be
e 4.40 works in
generShowaShow
l.] thatthforat tanyhe metchoihodce ofofiExampl
nitiandal condic2 cantiobens
rand
t
h
e
s
c
al
a
r
s
found,
as
s
t
a
t
e
d
i
n
Theor
e
m
4.
3
8(
a
)
and
(
b
)
.
iInfitthiaeleicondigenvalltiounses areand ,\20,are dist1,inscthowandththate
( )(,\7 - ,\�).
asThesociFiabtonacci
ed matrriexcurequatr]enceion l l wherhase the
and [ 1 0 ]
tWiionthtf0o pro0veandfthat1 1, use mat] hematical induc­
[
Usforinalgl par2t 1.(a), prove that ( - 1 )"
This is Gicalolvanni
ed Domenico Cas­
afsfointreialr(1625-1712)
tlhe 2ast1.ro[nomer
sinsi wasXIVbor, movedn in Iitnaly1669but,
onto Frthaence,inviwhere
tation.heofCasLouibecame
di
r
ect
o
r
of
t
h
e
Par
i
s
ObserevdattohreyFr. HeenchbecameversioanFrofenchhis name:
citizenJeandan­
adopt
DomininiqteueresCasts otsihneri. Matthanhematastroinomy.
cs was Casonesofinihi'ss
many
ed in 1680ofinScia epapernces isnub­Paris.]
miAnIdentt 8eidtyto8wastchecker
hepublRoyalibsoarhAcademy
itno form
Figureth4.e259(a)1and3 recttdhaecannglpieebecesindiFiresgasurectseeembl4.d 2as9(edsbh)own.
,\1 *
v
[Hint:
54. (a)
x0 =
x1
(b)
,\1
Xn
55.
=
A 1 - Az
Xn
(a)
=
Ltn- 1
n
Therectaarngleae'ofs artheeasiqs uar65 seqiuars 64e suniquartse! Wher
units,ebutdidtthhee
exthaveratsoquardo wie cometh thefrFiobm?onacci seWhat
does
t
h
i
s
Youof 1 have2 tialessupplandyoneofthkireendkiofn1ds of1ttiquence?
lieles:, astwsoh]kiownndsin
Figure 4.30.
[Hint:
56.
X
X
D CJ
ways toe,cover a
tthhatthesofe tdii5.leffs.eForrentexampl
1FigurLeter4.ect3bea1 nglshthowseewinumber
Fi(Dnoesd make any sense? If so, what is it?)
ence rteiolanstio, snofortlve t"h'e
UsrSetecurinupgreancesandecondretl2atasiorotndherienirnpareicurtital(rbcondi
against the data in part (a) . ) . Check your answer
Figure 4 . 3 0
Jn + I fn
Jn fn - 1
fn + ifn - 1 - fn2
n
(c)
(b)
=
=
=
Cassini's Identity,
__.v"
......
Figure 4 . 2 9
=
+ fn - 2
xn = AXn _ 1,
A=
I fn
A"
(b)
x1
Xo =
f,, = fn - I
=
_......
c1
= s,
�
v
�
1.._...
...- ..
X
�
(a)
(b)
(c)
n
tn
t1, • • • , t5•
t0
t3 =
t1
X
X
\
v
v
......
�
\'
\
\
The five ways to tile a 1 X 3 rectangle
Figure 4 . 3 1
\
'
(a)
57.
Youto coverhaveaa2supplyreofcta1ngle2. Letdominbeoesthwie tnumber
h whichof
diFiffgurereent4.3ways2 showsto coverthat the re3.ctangle. For example,
X
X
n
dn
d3 =
Section 4.6 Applications and the Perron-Frobenius Theorem
Em
/ l �
[8 BJ ITIJ
The
three ways to cover a 2 X 3 rectangle with 1 X 2
dominoes
Figure 4 . 3 2
Fi
n
d
,
..-... Set(Doesup a smake
anyordserensreecur? Ifresnceo, whatrelatiisoint?for)
e
cond
ragaiUsecurinngrset ncethande rdatelaatiasionnparthine tparin(ait)ti.a(lbcondi
) . Checktionsyour, solansve twhere
In Example find eigenvectors and corresponding to, Aver1 =ify and A2 = 2 . With
xk = ] fo:mula (2) in Section That is,
show that, for some scalar c1,
hm Ax1 = C1V1
(a)
(b)
(c)
d1,
d0
.
•
d5.
•
d1
58.
dw
d2
4.41,
I fk
Ltk- 1
v1
1 + Vs
v2
1 - Vs
4.5.
. kk
k->oo
svstems of L i n e a r D i f f e r e n t i a l E q u a t i o n s
llili_
In Exercises 59-64, find the general solution to the given
system of differential equations. Then find the specific
solution that satisfies the initial conditions. (Consider
all functions to be functions oft.)
=
59. = + 3y,
5
y'
+ 2y,
60.
1
1
y'
+ 2y,
61.
+
1
EV
62.
63.
x' = 2xx yx((OO)) = 0
x' == 2x- x- y, xy((OO)) ==
x;x� == XX11 - XXzz,, XXz1 ((O0)) == 0
YiY� == Y1Y1 - YzYz,, YzY1 ((0O)) == 1
X1 == X - yx((OO)) == 0
= x z (O) =
1
+
y'
z'
y
+
+ y,
Z,
Z,
-1
363
x' == Xx - 3z, yx((OO)) == 32
(
)
=
3x
=
z
O
Aa petscireintdiissthpl. Iancesitiatlwy,othsterraeinars eof400bactofXeriaand, X and500 ofY.in
nottThehe number
ftewedo bacton seacheofritahotcompet
he sterra. iInfesxfoat=rtifoodme andanddayss,p=tacehe grbutoarwtdoeh
rates of the two popul
a
t
i
o
ns
ar
e
gi
v
en
by
t
h
e
s
y
s
t
e
m
x' == - 0.l .22xx Detatioensrmbyinesowhatlvinghappens
tmo thofesdiefftweroentpopu­ial
lequat
t
h
e
s
y
s
t
e
Explbyletoirionensgx(th.eOeff) ect ofandchangiy(O) n=g thDese inictriaibl popul
ahap­tions
e
what
pens
t
o
t
h
e
popul
a
t
i
o
ns
i
n
t
e
r
m
s
of
and
TwoThat sips,ecineietsh, Xerandspecieslivcane insuarvive on itsreownlatioandnshieachp.
depends
valand. Initi=al y, thareree the
arsizees ofofXthone popul
andthe ot10ahtofY.ieronsforIatfixttsi=msuervimont
rates of the two populx' =at-io0.nsSxare given byhs,tthheesgrysotewtmh
=
Determine what happens to these two populations.
=
=
x'
=
x
x = [;]
x(O). x = =
= [[ _� l ] ,b = [ -- 3010 ] ,x(O) = [ 2030 ]
= - 1 ] ' - [ 40O J ' x(o) - [ 3010 ]
Letconsxid=er the be a twice-differentiable function and
x"
0
Show= x althlatowstheEquatchangeionof vartioablbeeswri=t ex'n asanda sys­
tem of two linear differential equations in and
64.
65.
+
y'
z'
+
2y +
Z,
z,
4
x (t)
(b)
=a
Y,
66.
15
t
b.
a
symbiotic
x (t)
t
y'
y y (t)
0.2y
+ I.Sy
y'
(a)
Y,
b.
y y (t)
+ 0.4y
0.4x - 0.2y
In Exercises 67 and 68, species X preys on species Y. The
sizes of the populations are represented by x x (t) and
y y (t) . The growth rate of each population is governed
by the system of differential equations
A + b, where
and b is a constant vector. Determine what happens
to the two populations for the given A and b and initial
conditions
(First show that there are constants a and b
such that the substitutions
u + a and y v + b convert
the system into an equivalent one with no constant terms.)
67. A
1
1
b
-1 -1
69.
x ( t)
second order differential equation
+ ax' + bx =
( a)
y
z
1
1)
(
y
68. A
(1 1)
z.
Chapter 4 Eigenvalues and Eigenvectors
364
the
sShowShowytshteatmtthhinaterparethiest achar(achange
) ias ct,\erofistivarc equatiableisotnhatofconver
ts
the nth _ into acsiyentstematm ofrix liisntearhe compani
differentoianl equat
ions whosoftehe
coeffi
mat
r
i
x
pol[Theynominotaatilp(on ,\) denot,\" es th_e,\kt-h derivative of See
Exercompaniciseson matriinx.Sect] ion for the definition of a
(b)
70.
2 + a,\ + b = 0.
order differential equation
n
(
x ) + an 1x(n I ) +
+ a 1 x' + a o = 0
n
C( p)
I
n
=
+ an 1
+ + a1A + a0•
x (k)
x.
26-32
4.3
·
·
·
·
·
·
In Exercises 71 and 72, use Exercise 69 to find the general
solution of the given equation.
71. x " - 5x' + 6x = 0
72. x " + 4x' + 3x = 0
In Exercises 73-76, solve the system of differential equations
in the given exercise using Theorem 4.41.
74.
60
59
73.
76.
64
75.
63
ExerExercciissee
ExerExercciissee
D i s c r e t e linear D v n a m i c a l svstems
In Exercises 77-84, consider the dynamical system
xk + 1 = Axk.
1
(a) Compute and plot Xo' x1 , x2 , x3 x0 = .
1
1
(b) Compute and plot Xo' x1 , x2 , x3 Xo =
0
(c) Using eigenvalues and eigenvectors, classify the origin as
an attractor, repeller, saddle point, or none of these.
(d) Sketch several typical trajectories of the system.
2
0.5 - 0.5
78. A =
77. A =
0 3
0
0.5
for [ ]
for [ ] .
[
l]
[
Chapter Review
adjalgebroinat iofc mula mattiprliixci,ty of an
charchareigenval
aactcteerriiusstte,iicc polequatynomiion, al,
cofact
Crdetaemerrmoir'nsexpans
Rulant,e, ion,
Kev Defi nitions and Concepts
276
294
266
274-275
263-265
292
292
]
79. A =
[ _�
[
[
-�]
-�]
1 .5
-1
0.2 0.4
83. A =
- 0.2 0.8
81. A =
[
]
296
294
[ -� -�J
[ ]
[O ]
0.1 0.9
0.5 0.5
- 1 .5
84. A =
1 .2 3.6
82. A =
]
In Exercises 85-88, the given matrix is of the form
a -b
A=
. In each case, A can be factored as the
b a
product of a scaling matrix and a rotation matrix. Find
the scalingfactor r and the angle e of rotation. Sketch the
first four points of the trajectory for the dynamical system
and classify the origin as a spi­
xk + 1 = Axk with x0 =
ral attractor, spiral repeller, or orbital center.
1
0 0.5
85. A =
86. A =
1
- 0.5 0
V3
1
- V3;2 - 1 /2
88. A =
87. A = V3
1
1/2
- V3;2
In Exercises 89-92, find an invertible matrix P and a maa -b
trix C of the form C =
such that A = PCP i .
b a
Sketch the first six points of the trajectory for the dynamical
system xk + 1 = Axk with Xo =
and classify the origin as
a spiral attractor, spiral repeller, or orbital center.
[�]
[ -�]
[
89. A =
91. A =
303
]
0.1 - 0.2
0.1 0.3
]
[
[
]
[
[ ]
[ � -�]
'
dieigaenval
gonaluize,able matrix,
eieiggenvect
or,
ens
p
ace,
Fundament
al Theorem of Invertible
Mat
r
i
c
es
,
geomet
r
i
c
mul
t
i
p
l
i
c
i
t
y
of
an
eigenvalue,
254
254
256
80. A =
[�]
90. A =
92. A =
-
[ _� � ]
[o - 1 ]
1 V3
GersGerscchgorhgoriinn'sdiDisks,k Theorem,
LaplacemetExpanshod (iaondn Theor
em,
power
i
t
s
prvaroperianttiess),of determinants,
similar matrices,
319
3 1 1-319
269-274
301
321
266
]
Chapter Review
MarFork eachall sofquartheefolmatlowiricnesg A,statdetement( -A)s tr=ue-or<lefat A.lse:
(b) <lIfeAt (andBA).Bare matrices, then det(AB) =
IfarAe tandhe sameBarebut in dimatfferreintcesorwhosders,etcolhenumns
( If<lIfeAt Biiss=tihnever-onl<lteitybA.leie,gtenval
hen detue(ofA -a sq=uar<leetmatAT.rix A,
tTwohen Aeigienvect
s the zeroros matcorrreisxp. onding to the same
eiIfganenvalue musmattribex haslineardilystdependent
. ues, then
i
n
ct
ei
g
enval
iItfmusan t be dimatagonalrix iiszabldiaegonal
. izable, then it must
haveSimilardimatstinrictceseihavegenvalthueessa. me eigenvectors.
A and rBoarweecheltwoon form,matthrenicesAwiis tshimthilearsatmeo B.
rIfeduced
LetA = [ 3 3 � ] .
Comput
e
<l
e
t
A
by
cofact
o
r
expans
i
o
n
al
o
ng
any
rComput
ow or cole <luemn.t A by first reducing A to triangular
form.
If = 3, find
Let<let AB and= -B�.beFind <let matC forricteshewiintdihc<lateet dAmat= 2riandx C:
1 (b)C=A2B(3AT)
C=
(
A
B)
prIfAoveis tahsatkew-<letsAymmet
= ric matrix and2 is odd,
Find all values of for which 2 =
x
A
x = [ � ] ,A = [! �]
]
H J [ �: - 32
Review Questions
1.
(a)
(c)
d)
n
X
n
n
X
n
i)
0
(e)
(f)
(g)
n
X
n
(h)
n
n
X
n
(i)
n
(j)
2.
n
X
n
1
5
7 9 11
(a)
(h)
3.
4.
5.
6.
a b c
d e f
g h
4X4
( a)
0.
k
3d 2e - 4f f
3a 2b - 4c c .
3g 2h - 4i
nXn
-1
1 k
1
4 k2
n
0.
In Questions 7 and 8, show that is an eigenvector of and
find the corresponding eigenvalue.
7.
s. F
A�
- 60 - 45
15
18
- 40
365
LetA = [ - : - : -- 2� ] .
ristic upolesyofnomiA. al of A.
(b) FiFiFinnnddd alathbasleofcharitshforeaeicteachgeenval
Detnot edirmagonal
ine whetizablheer, explAofistahdiineaeiwhygonalgensnotipzacesabl. IfeA.ofIfisA.Adiisago­
nalimatzablrei,xfindsuanchitnhverat tPib-l1eAmatP =rix P and a diago­
nal
I-fA2, i3,s aand3 3fidindagonal
izable matrix with eigenvalues
<l
e
t
A.
IfA is a 2 2 matrix with eigenvalues A1 = t, =
and cor-responding eigenvectors v1 = [ � ], = [ - �l
findA s [ � J .
IsfatAisfiys a diagonalprizoablveethmatat Arinxapprandoalachesl of ittsheiegzerenvalo ma­ues
trix as gets large.
A
B.
A
B,
P
P- 1 AP= B.
= [! � ] ,B = [ � � ]
14.A = [ � � ] ,B = [ � � ]
[� i :J. B [� � �]
Let A = [ � �] . Find all values of for which:
(b) AAA hashashas noaneigenval
eiregalenvaleiugesenvalu3e andwiutesh .algebraic multiplicity 2.
IIffAa sq=uarA,e whatmatriarex Athashe postwosibequalle eigrenvalows, uwhyes ofmusA? t A
haveIf xis anaseionegenvectof itsoeir gofenvalA wiutesh ?eigenvalue = 3, show
tthhate corxisrealspsoondian neiggenvect
oruofe?A - SA 2I. What is
ei
g
enval
Ivectf A oisr sofimA,ilasrhowto BthwiatthP -P1-x1iAs Pan=eiBand
genvectxiosr ofanB.eigen­
9.
0
0
( a)
(c)
(d)
X
10.
D
D.
4,
X
11.
A2
- 1,
Vz
12.
IAI
n
< 1,
In Questions 13-15, determine, with reasons, whether is
similar to If � give an invertible matrix such that
13. A
15. A �
�
k
16.
( a)
17.
18.
19.
20.
( c)
- 1.
3
0
2
+
A
O rthogona l ity
. . . that sprightly Scot of Scots, Douglas,
that runs a-horseback up a hill
perpendicular-
-Wil iam Shakespeare
Act II, Scene IV
Henry IV, Part I
5.0
I n t ro d u ctio n : S h a d ows o n a Wa l l
r, we wierl extandendthentheagainotnioinnofChaptortheogonal
priol jnow,ectiowen thhaveat wediencoun­
tonlIenretyhdiprsfiorchaptsjtectinieoChapt
r
Unt
sbscuspsaceed
n
ont
o
a
s
i
n
gl
e
vect
o
r
(
o
r
,
equi
v
al
e
nt
l
y
,
t
h
e
oned
i
m
ens
i
o
nal
s
u
spannedas forbyprtohjeatctvection oontr) .oIna plthaisneseictn ion,Fiwegurwiel seeshifowswe whatcan fihappens
nd the anal, forogousexamplfor­e,
mulwhen
ona tawwalo-dli.mAenssimioinallar sprcroecesen,s soccur
s when
a
tputhreee-r moni
dparimaenslteolri.oliLatnalghteobjrraiysnectthcriisesatchapt
diesaplsaehyedradow
on
u
ch
as
a
com­
,2 weat wiwhatl conswe ialderreadythesknow
e ideasaboutin fulprl generojectaiolinsty.. In
To
begi
n
,
l
e
t
'
s
t
a
ke
anot
h
er
l
o
ok
Section we showed that, in IR , the standard matrix of a projection onto the line
through the origin with direction vector = [ �:] is
)]
]=[
Hence, the projShowectionthofattPhecanvectbeorwrv iontt eno itnhitshleinequie is vjualsetntPv.form
P = [ cos()cos2si()n () cos()sin2si()n () ]
(What does() reShowpresetnthatherP ecan?) also be writ en in the form P = where is a
vector in the dirUsectiniognProfoblem find P and then find the projection of v = [ _! ]
onto the lines with the following unit direction vectors:
(a) " = [ �:�J (b) " = [fl (c) " = [ -n
rm P e=nt) . show that (a) = P (i.e., Pis symmetric)
and (b) P2 = PUs(i.ein.,gPithseidfoempot
1
3.
IR 3 .
5.1
3.6,
d
dl !(dl + d i ) d , d2!(d l + di
d, d2!(df + d i ) d i !(df + d i )
d, d2
di
Problem 1
Shadows on a wall are projections
Figure 5 . 1
�
uuT,
Problem 2
d.
Problem 3
Problem 4
366
2,
uuT,
pT
u
unit
Section 5.0 Introduction: Shadows on a Wall
361
projects vectorsExplis thaeincolwhy,umnifspaceis aof projection matrix, the line onto which it
consider projections onto planes through the
origNowFiing. urWeewe5.wi2wilshexpll owsmoveoronee severinwayto atloapprprandooceed.aches.
norv mal vectoforr sandomeifsvcalisara vector in tIhf enis a plpraneojgpth(vro) ughis a vectthe ororiignin insuchwithtath
P
Problem 5
2X2
P.
IR 3
-c
n
=p
IR 3 ,
n
<!J'
c.
p=
n
v
<!J'
IR 3
- en
Projection onto a plane
Figure 5 . 2
v
=p
Us
i
n
g
t
h
e
fact
t
h
at
i
s
or
t
h
ogonal
t
o
ever
y
vect
o
r
i
n
s
o
l
v
e
forUsteothfiendmetanhexprod ofesPrioonblforem itnotfiernmdsthofe vprandojection of
Problem 6
-c
n
Problem 1
c
<!/',
p
n
n.
6
onto the planes wi(a)th the yfollowing equat(b)ions: (c) 3y
finding theethpre oprjeoctjeictonioofn ofa vectv ontoor ontinotoa
plthaeneAnotis shofuerggesitapprs tpredoojbyachectFiiotgnsourthontee5.pro3.otWehbleedimcanreofctdecompos
irson. Accor
vectordsinforgly, letThisandworksbeonldiyreictf itohne
divectrectorisonforvectwiorstharthe eorprthoogonal
uni
t
vect
o
perty that and
x+ +z=
0
x - 2z
= 0
2x
<!/'.
sum
<!J'
ll u 1 ll
=
ll u2 ll
=
u 1 • u2
1
= 0
IV
I
I
I
I
'"""
I
I
U2
:
.- -:::�
.L.:...-
Figure 5 . 3
p = pI
+ P2
-
+z
= 0
<!J'
u1
u2
368
Chapter 5 Orthogonality
By Problem 2, the projections of onto andand are
gioves tIhteisprenough
ojectiontoofshowonttohat we-need to showis
trorhesattphectogonal-ivelyt. oTobotshhowistandhoratthuogonal
t
2. (Why?) and
Show
t
h
at
ive formunioft vectthe dotors.pr] oduct, together with the fact that
andUseItthfoleareallotwseorrnthfratogonal
projection ontootmhe Prsuobsblpeacem andof thespcomment
anned bysorprthecediogonalng iunit thtatvecttheormats randix of thies
Usortehogonal
the sameunitRepeat
andvectusoresPrinoblandtheemgivenasuspliinnadine.g ctat)heedforbelmoulw.a(FforirstP, vergivienfy tbyhatEquatandion are
(a) x z with [ - �j�] and [ � ]
2
(b) x - 2z with [ 1:5] and [ � ]
(c) 2x - z with [ - ] and [ 2 ]
ties (a) and (b) ofShow
PrShowobltthheatmat athpreomatjectrioixn Pmatofriax prgivoenjectbyionEquatontoioan pl( anesatiinsfies prcanoperbe­
expres ed as
for some 2 matShowrix that if PiShows the tmathat rEquat
iaonprojeicts ianonoutonteropra oplductane expans
itohn.en]
i
x
of
i
n
rank(InPt)his chapt
2. er, we wil look at the concepts of orthogonality and orthogonal pro­
jgener
ectionaliiznedgrandeatetrhdetat tahiley. Wehavewimany
l see tihmatporthteantideasapplinictratoduced
ions. in this section can be
v
v
( p1 + p2 )
u1
u2
p 1 = u 1 u fv
p 1 + p2
p2
= u2 ufv
!JP.
u1
u1 (v
Problem 8
•
u2
!JP,
v
v
( p1 + p2 )
- ( p1 + p2 )) = 0 u2 (v - (p1 + p2 )) = 0. [Hint:
xTy = x y,
u1
•
·
8
!JP
IR 3
u1
u2
( 1)
7,
Problem 9
v
u1
+y+
=
3y +
u2 ,
=
0
u1
1/\/6
0
u1 =
l/ Vs
=
Uz
u1 =
0
u1 =
u2 =
l/ v'3
l/ v'3
l/ v'3
4.
3X
A. [Hint:
=
•
l/
- 1/ V2
0
/\/6
1 / \/6
- 1 / \/6
u2 =
IR 3
p
Problem 12
( 1).
1)
Problem 10
Problem 11
=
u2
=
AA T
( 1)
IR 3 ,
O rth o g o n a l ilV i n IR n
tsIwntaondarthvectis sdeobasctrsiotion,s setwes wioflvectgenerorsofa. lIinzedoiteashenygnottsooiwor,owen kofwiwiorltthsh:eogonal
eFirthstat, anyittwyotofwprovectodipersotirtnsiectisnvectmakeofrrsothmine
!R n
{ e1, e2 , . . . , en } !R n
Section 5. 1 Orthogonality in
lffi 11
369
terhteiesets learade usortthoogonal
. iSecond,
eachhogonalvectbasor ienstandhe setortishonora unimt vectal basore. Thess-concept
e two prs tohp­at
t
h
e
not
o
n
of
or
t
we wil be able to fruitful y apply to a variety of applications.
O rth ogonal and Orth onormal Sels of Veclors
rths e set are ort,hogonalin -itshcalat lies,difan
if
all pairs of distinActsvectet ofovectrs inowhenever
i j for i, j
tThehe fisrtsatndarexampld base iils ustrates, th,ere arofe manyis anotorherthposogonalsibilsiteite,sas. is any subset of it. As
Show that is an orthogonal set in if
Definilion
{v1 , v2 ,
vi · vj
=
z
•
•
.
!R n
vd
*
0
{e 1 , e2 ,
Exa m p l e 5 . 1
•
.
en }
•
orthogonal set
=
1, 2, . . . , k
!R n
IR 3
{v1 , v2 , v3 }
is true, sinceWe must show that every pair of vectors from this set is orthogonal. This
Solution
• Vz
Vz • V3
V 1 • V3
V1
y
x
An orthogonal set of vectors
Figure 5 . 4
=
=
2(0) + 1 ( 1 ) + ( - 1 ) ( 1) = 0
0( 1) + 1 ( - 1) + ( 1 ) ( 1) = 0
2( 1) + 1 ( - 1) + ( - 1) ( 1) = 0
Geomet
r
i
c
al
l
y
,
t
h
e
vect
o
r
s
i
n
Exampl
e
ar
e
mut
u
al
l
y
per
p
endi
c
ul
a
r
,
as
Figure shows.
4
of thesamairilynlinadvant
of work,iasngTheor
with eormthogonalshowsset. s of vectors is that
theyOneare neces
early iangesdependent
is an or.thogonal set of nonzero vectors in then these vectors
arIfe linearly in, dependent
If are scalars such that
then
or, equivalently,
t, all oftothe dot products in Equation
are zerSinoce, except Thusis ,anEquatorthiogonal
on rseeduces
=
5.1
5.4
5.1
Theorem 5 . 1
{v1 , v2 ,
Proof
.
•
c1,
!R n ,
vd
•
.
•
.
, ck
c 1 v1 +
(c 1 v1 + · · · + ckvk ) · vi
+ ckvk = 0,
= 0 · V; = 0
· · ·
c 1 (v1 · vi ) + · · · + ci (vi · vi ) + · · · + ck (vk · vi )
{v1 , v2 , . . . , vk }
V; · vi .
( 1)
c; (vi · v;) = 0
=
0
(1)
(1)
310
Chapter 5 Orthogonality
The fact tsheatt.
tNow,his is true for allbecause imbypliehypot
s thathesis. So we musistahavelinearly independent
tloyTheor
em . weForknowexamplthate, iwef a setcanofimvectmediorsatieslyordeduce
thogonalth,atit
itsheautthormatee vecticalThanks
loyrlsininearExampl
independent
dependent. diContrectrlays!t this approach
with the work needed to estaeblish tarheie rlilninearearlyinindependence
for a subspace W of is a basis of W that is
an orthogonal seAnt.
The vectors
V; V;
•
*0
V; * 0
i = 1, . . . , k
C; =
{v1 , v2 , . . . , vk }
0.
5.1,
Remark
5.1
Definition
Exa m p l e 5 . 2
!R n
orthogonal basis
e arevectortohrogonal
and,formhence,
lifornearly byindependent
. Sinceal anyTheorthereme
lfroflinoearmnverExampl
ly tiinbdependent
s
i
n
a
bas
i
s
t
h
e
Fundament
le Matrices, it follows that is an orthogonal basis for 4
onlvecty tohre orttohmake
ogonal vectors anandorthogonalwere
gibasvenis forand youOneInwerExampl
ewayasktoededototfihinssudippossattohireredmember
theeatExplin oratthioen:crTheos product
ofoducttwo
vectin Chaptors erand Henceis orwethogonal
t
o
each
of
t
h
em.
(
S
Cros
s
Pr
may take
5.1
IR 3
5.2,
Remark
IR 3 .
v1
v2
1 .)
IR 3 ,
{v1 , v2 , v3 }
v3
IR 3 .
IR 3 ,
v1
{v1 , v2 , v3 }
Note that the resulting vector is a multiple ofthe vector in Example as it must be.
Find an orthogonal basis for the subspace W of given by
v3
Exa m p l e 5 . 3
v2
5.2,
IR 3
Sect
i
o
n
gi
v
es
a
gener
a
l
pr
o
cedur
e
for
pr
o
bl
e
ms
of
t
h
i
s
s
o
rt
.
For
now,
tweofhevectwiorligofiirnsnofdintthhee orforFrtohmmogonalthe equatbasisiobyn ofbruthtee forplacne,e. Thewe havesubspace Wis a plsoaneWtconshroughists
5.3
Solution
IR 3 .
x=y
-
2z,
Section 5. 1 Orthogonality in
IR"
311
It follow' that [ �] and [ - �] "e a b";' foe W, but they Me mthogo­
nalthes. eIt. suffices to find another nonzero vector in W that is orthogonal to either one of
Suppo'e [ �] ;, a vedm ;n W that ;, mthogonal to Then ssyinstceem is in the plane W. Since we also have Solving the linear
we find that and (Check this.) Thus, any nonzero vector of the form
u
�
v�
nol
w�
x y + 2z � 0,
u.
w
x + y = 0.
u w = 0,
·
x y + 2z = 0
=0
x+y
�
x = -z
y = z.
w
wil do. To b"pedfie, we wold toke [ - :J- It to ehe<k that ;, an
orthogonal set in Wand, hence, an orthogonal basis for W, since dim W
base.iIsnideed,
s that tthheercoore is adifornatmesulofa
aforvecttAnothesorewihcoorerthadvant
rdeinspatecteasge,tasoofstuhworche folaklibasonwig iwins garthtehaneoreasoryemtthoogonal
comput
establishes.
vectLet or in W. Then betheanuniorqtuehogonal
scalarsbasis for a ssuubschptacehat W of and let be any
are given by
for ...
;""Y
w�
{u, w}
= 2.
Theorem 5 . 2
{v1 , v2 , . . . , vk }
!R n
c 1 , . . . , ck
W ' V;
C; = -V; ' V;
i = 1,
w
,k
Sincesuch that is a basis for W,(frweomknowTheortheatmthere arToe uniesqtaueblisschaltahres
formula for we take the dot product of this linear combination with to obtain
Proof
c 1 , . . . , ck
{v1 , v2 , . . . , vk }
w = c 1 v1 +
C;,
· · ·
+ ckvk
3.29).
w · V; = (c 1 v1 + · · · + ckvk ) • V;
= c 1 (v1 · v;) + · · · + c ; (v; • v;) + · · · + ck (vk • v; )
= C; (V; ' v; )
vj V; = 0
* i.
V; V;,
V; * 0, V; V; * 0.
dessinceired result. for j Since
·
•
V;
Dividing by we obtain the
•
312
Chapter 5 Orthogonality
Exa m p l e 5 . 4
[
FiofnExampl
d the coores dinandates of � ] with respectto the orthogonal basis
Using Theorem we compute
w =
5.1
5.2.
B
{v1 , v2 , v3 }
=
5.2,
Solulion
�
3
w · V1
c] =
--
=
C2
=
--
=
C3
=
--
=
V1 • V1
w · V2
V2 " V2
w · V3
V3 • V3
2+
4 +
0+
0+
11+
2
1
2
1
2
1
-3
+
+
+
+
+
1
3
1
3
1
1
6
5
2
2
3
=
-
=
-
Thus,
(Checkionthasis.) With the notation introduced in Section 3.5, we can also write the above
equat
to findbastheess. e
coorAsCompar
dinnotateesdediattrhetctehelprybegiandocedurnyouninegshiofnoulExampl
dthisstsaertcteitoo5.n,appr4thwieetciothahtterheethprworeovalperkutreyeofquiofortrhteehdogonal
iornthogonal
is thatityeach, weshavetandarthde basfollioswivectngodefir is naiunition.t vector. Combining thissprtaondarpertdybaswitihs
sthetatofisunian torvectthonoroArs.smeAntalofsevectt. ors in is anfor a subspace ioff it is ians aorbasthiogonal
s of
If The fact thatiseachan orthionors a unimalt vectset oofr vectis equiorsv,altheennt to for
j
and
It follows that we can summarize the statement that is orthonormal as
{ o iiff jj
Show that is an orthonormal set in if
[ ] and [ ]
!R n
!R n
Definition
orthonormal set
W
orthonormal basis
Remark
i
-=fa
ll q; ll
S = { q1 , . . . , qd
1.
Exa m p l e 5 . 5
qi =
q; qj = 0
q ; q; = 1 .
=
·
S
1
i*
i=
IR 3
S = { q1 , q2 }
l / v'3
- l / v'3
l / v'3
W
·
q;
=
q; . qj
!R n
q2
=
l / v'6
2/ v'6
l / v'6
Solution
Section 5. 1 Orthogonality in
We check that
313
lffi "
q1 • q2 = 1 / \/18 - 2/ \/18 + 1 / \/18 = 0
q i . q i = 1 /3 + 1 /3 + 1 /3 = 1
q2 · q2 = 1 /6 + 4/6 + 1 /6 = 1
Exa m p l e 5 . 6
simplIfywenorhavemalanizeoreachthogonal
vectors. et, we can easily obtain an orthonormal set from it: We
Construct an orthonormal basis for from the vectors in Example
malize themSintocegetwe already know that and are an orthogonal basis, we nor­
IR 3
v1, v2 ,
Solution
v3
[
l
[
]
; �
Then is an orthonormal basis for
Since any, byorTheor
thonoremm5.al1s.eItfofwevecthaveorsanis,orinthparonorticmulalar,basortihs,ogonal
,eimt is5.l2inbecomes
early independent
Theor
even simpler.
anyLet vector in Thenbe an orthonormal basis for a subspace of and let be
and this representation is unique.
Apply Theorem 5.2 and use the fact that for
q3 = 11 1r3 =
3
l
-1
1
l / V3
= - 110
l / V3
IR 3 .
{ q 1, q2 , q3 }
Theorem 5 . 3
5. 1 .
{ q 1, q2 , . . . , qk}
W
W.
IR "
w
w = ( w · q1 ) q1 + ( w · qz ) qz + · · · + ( w · qk) qk
q; q; = 1
Proof
·
i = 1, . . . , k.
iclesswhos
younowMatrwiexami
eenie.n eSectcoliuomnsn formSuchanmatortrhiconores havemalsseteveraralisate tfrraectquentive prlyoipern appltiesi,cwhiatiocnsh ,weas
Orthogonal Malrices
5.5.
314
Chapter 5 Orthogonality
Theorem 5 . 4
The columns of an n matrix form an orthonormal set if and only if
We need to show that
{ Q iiff ii jj
Letentry ofdenote tihs ethitehdotcoluproduct
mn of of(athnd,e ithence,
tofhe ithandrowthofe )th colSiunmnce thofe ( i, jit)
h
r
o
w
follows that
by tNow
he defithneitcolionuofmnsmatrforix mulm antiporlicthatonorion. mal set if and only if
{ o iiff ii j
which, by Equation holds if and only if if i j
- { Q if i j
ThisIfcompl
e
t
e
s
t
h
e
pr
o
of
.
the matrix in Theorem is a matrix, it has a special name.
called an An n n matrix whose columns form an orthonormal set is
The most important fact about orthogonal matrices is given by the next theorem.
A square matrix is orthogonal if and only if
only if Byis Theor
inverteibmle and is orthogonalby Theorif andemonly if This is true if and
Show that the following matrices are orthogonal and find their inverses:
and [ cossin -cossin ]
the staandndard basis vectors for which are
clearly orthTheonorcolmalumns. Hence,of ariseorjuthstogonal
[� � � ]
m
Q TQ = In "
X
Q
Proof
(Q TQ )ij =
q;
*
1
=
Q
Q TQ
(Q TQ) ;j =
Q
Q T) .
QT
q; . qi
=
Q,
q; · qj
(2)
*J
=
1
(2),
is an "Ortho­
unfortu­
nate
bi
t
of
terminology.
normalterm,
matributx" would
clstandard.
early be a
better
i
t
is
not
Moreover,
there
is winothtermorthonor­
for a
nonsquare
matrix
mal columns.
Orthogonal matrix
Theorem 5 . 5
( Q TQ )ij
Q
Definition
x
orthogonal matrix.
1
square
Q
Q - 1 = Q T.
Q
Proof
Q
Exa m p l e 5 . 1
5.4
*
=
_
5.4, Q 1
Q- = Q T,
3.13.
B
Solulion
A
A
A - ' � A' �
=
{J
{J
Q TQ = I.
{J
{J
IR 3 ,
Section 5. 1 Orthogonality in
For B, we checkcos()diresctinly()thatcos() - sin ()
BrB = [ - sin () cos () ] [ sin () cos() ]
[ - sin ()coscos2 sicosn2 ()() sin () - cos s()insi2n cossin2 ()() cos () ] [ � � ] = I
Therefore, B is orthogonal, by Theorem and
B = B = [ - sin () cos()sin () ]
ational,matanyrix, a matper­rix
obtmutaaintiedonbymatperMatrixmirutsixorintghitnogonal
hExampl
e colu(mnsseee Exerofisanancisidexampl
ente ityMatematofrairxiperxB. Iimns uttgener
tthrarnsoughformthateianglon (ke nown
() in IRas2. Anyan rotation hasin geomet
the prorpery) . tThey thatnextihteistmathaeorriexmofshaowsrotattihoatn
ation matis anricisesometare rchary. Oracttheogonal
prteverheseseeyrprvoreotdotperhogonal
prtieos.ductmats.rIinx facttrans, orfotrhmogonal
rized bymateithrericesonealsofo
Leta. QQisbeoranthogonalmat. rix. The following statements are equivalent:
IR". n !R".
= l xl · yforforever
b.c. lQxQx·l Qy=x
every xiynxandyi
(mn)c)::::}vect(b)::::}ors i(na)IR. To", thdoensox, ·wey =wixlryneed. to make use
of(a)::::}the fact(cWe) tAshwiatsluimefprxoandvethatthyatQarei(sa)::::}or(colthuogonal
r
.
Then
Q
Q
=I
,
and
we
have
(
Qx
·
Qy
=
Q
xfQy
=
xrQrQy
=
xrly
=
xry
=
x
·
y
and yi=n IRl x".l Then,
taking y = x,
.
we((bc)::::})::::}have((baQx)) AsAs· Qxssuumeme= xtthh·atatx, Qxprsool·perOQyxtyl ==(bx)v'· holyQx·fodrs everQxandy=lextVX:-X
Using Exercise in Sectx·yio=n i ( l andx prr lo2per- tyl x(b-), wer l 2)havedenote the ith column of Q.
== ii (( ll QQx(x Qy)rl l22 -- l l QQ(xx -- yQ)rl 2l )2)
=
Qx·Qy
for alNowl x andif yisinthIRe"i.th[Tshitasndarshowsd basthiats vect(b) o::::}r, t(hc)en.] Qe;. Consequently,
= Qe; . Qej = = { o iiff =
Thus, the columns of Q form an orthonormal set, so Q is an orthogonal matrix.
() +
() +
() +
() +
_1
isometry
isos
metron
Theorem 5 . 6
5.5,
COS ()
T
A
Remark
The wordpreserving;'lisince
terallyitmeans
"length
is
deri
v
ed
from
the
Greek
roots
("equal") and ("measure").
315
lffi "
5. 7
nXn
25).
length-preserving
isometry
nXn
Proof
63
q;
1 .2
+
+
+
e;
q; . %
q;
e; . ej
=
1
i*j
i j
Chapter 5 Orthogonality
316
matorritchesonorandmalBisents-Exampl
eth5.ei7r, you mayIn notfacti,ceeverthaty
notorthLooki
onlogonaly dongmatattheitrhrixecolorhastuhmnstogonal
for
m
s
o
do
his property, as the next theorem shows.
If Q is an orthogonal matrix, then its rows form an orthonormal set.
From Theorem 5.5(,QweT) know that Q- 1 QQT. Ther( QT) Yefore,
T
sQ-for
o QT ismananororththogonal
mat
r
i
x
.
Thus
,
t
h
e
col
u
mns
of
Q
-which are just the rows of
onor
m
al
s
e
t
.
The
fi
n
al
t
h
eorem
i
n
t
h
i
s
s
e
ct
i
o
n
l
i
s
t
s
s
o
me
ot
h
er
propert
i
e
s
of
or
t
h
ogonal
matrices.
Leta. QQ-be1 isanortorhogonal
thogonal. matrix.
b.c. <lIfet Qis an eigenvalue of Q, then 1.
d. If Q1 and Q2 are orthogonal matrices, then so is Q1 Q2.
ve property (c) and leave the proofs of the remaining properties
as(c)exerLetcWeisebes.wianl preigoenval
and, using Theorem 5.6 (ube), ofweQhavewith corresponding eigenvector Then Q
Since 0, this implies that 1.
o1 - 01 ]
Pr
o
per
t
y
(
c
)
hol
d
s
even
fo
r
compl
e
x
ei
g
enval
u
es
.
The
mat
r
i
x
[
is orthogonal with eigenvalues and - both of which have absolute value 1.
A
Theorem 5 . 1
Proof
Theorem 5 . 8
-1
=
n
X
rows.
co- 1 ) - 1
=
=
=
= ::t:: l
A
IAI =
n
Proof
A
v.
JJ v JJ
=
JJ v JJ *
JJ Q v JJ
IAI =
Remark
..
I
i
Exercises 5 . 1
In Exercises 1 -6, determine which sets of vectors are
orthogonal.
I.
,_
nHn [ - �l 2. uirn m
[ Jf :H - !l ·· [ :H - H Ul
i,
=
JJ Av JJ
=
I A I JJ v JJ
v = AV,
Section 5. 1 Orthogonality in
In Exercises 7-10, show that the given vectorsform an ortho­
gonal basis for IR 2 or IR 3 • Then use Theorem 5.2 to express
w as a linear combination of these basis vectors. Give the
coordinate vector [w]8 of w with respect to the basis
B = {v1 , vJ of !R 2 or B = v1 , v2 , v3 of!R 3 •
7. v1
=
8 . v1
=
9. v, �
10. v, �
[ � ] [ � J [ �]
[ � ] , [ �J [ � ]
[ +' [ l; [ - :} [: l
[ l, [ -J, [ _n m
_
=
, v2
v2
=
w=
-
w=
�
_
_
�
�
�
w�
13.
[il [ -iJ
[il [ -n
[!J [ !J [ J l . [ l l
• -i H ! U
27.
]
]
[ [ l[ [ l
18.
19.
20.
[o ]
[
u -! :i
cos[ cose 2sien e -sicosn ee - cos- sien2sien e
sin e
cos e l
2
l
[ -l 2 ]
1
I
-2
I
2
I
2
!
0
!
I
2
I
2
I
-2
Q
(b)
In Exercises 1 6-21, determine whether the given matrix is
orthogonal. If it is, find its inverse.
16.
Q
2X2
IR 2 .
Qx
x
x
y,
y
2X2
b
v'3/2
0
0
- v'3/6
0
v6/3
'
'
1 / v6
v'3/6
1 / \/2
- 1 / v6
1 / \/2
- v'3/6
1 / \/2 1 / \/2
17.
- 1 / \/2 1 / \/2
PrProoveve Theor
e
m
S
(
a
)
.
5.
Theor
e
m
S
(
b
)
.
5.
PrProoveve tTheor
e
m
S
(
d
)
.
5.
h
at
ever
y
per
m
ut
a
t
i
o
n
mat
r
i
x
i
s
or
t
h
ogonal
.
Iobtf aiins edan byortrheogonal
matng trhiex,rprowsoveofthatisanyalsomatrix
ar
r
a
ngi
orLetthogonal
.
beprovectve bethoatrans itnhore tanglhogonal
Ifeebetis twheeenanglmate betandrixwandQyeenilsealtandsoande.
s provesmatthatricteshearlineear transformatioinsn defianfaedctby
ort(hTatthihogonal
iPrs otrveuethinatgeneran oratlh.)ogonal matrix must have
the form a b
[ - ] or [ ]
where [ : ] is a unit vector.
Using matpartri(xa)i,ssofhowthethformat every orthogonal
cos[ sin ee -cossin ee ] or [ cossin ee - cossin ee ]
wher
ethat evere y orthogonal matrix corre­
Show
spondsthtato eianthorerthaogonal
rotation or a rmateflerctixioncor­in
Show
resreflpeondsctiontiona roitfa<ltieotn in if <let and a
28. (a)
-
-1
0
]
angle-preserving IR 2 ,
12 '
1 /2
1 /2
15 '
- 1 /2 '
1 /2
1 / v6
0
1 / v6
1 / \/2
1 / \/2 - 1 / v6
1 / \/2
0
22.
23.
24.
25.
26. Q
w�
In Exercises 1 1 - 1 5, determine whether the given orthogo­
nal set of vectors is orthonormal. If it is not, normalize the
vectors to form an orthonormal set.
11.
0
2/3
- 2/3
1 /3
311
IR"
a
2X2
(c)
]
a b
b -a
0 :::::: < 21T.
2X2
2X2
(d)
IR 2
IR 2
Q = - 1.
IR 2 .
Q
Q=1
In Exercises 29-32, use Exercise 28 to determine whether
the given orthogonal matrix represents a rotation or a
reflection. If it is a rotation, give the angle of rotation; if it is
a reflection, give the line of reflection.
[ 1 / \/2
- 1 /2
31. [
V3/2
29.
1 / \/2 - 1 / \/2
1 / \/2
V3/2
1 /2
]
]
30.
[
- 1 /2 V3/2
- V3/2 - 1 /2
]
Chapter 5 Orthogonality
318
33.
Let Prandove thbeat r orthr)ogonal matrices.
Usethenpart (a)itos notproivenverthtatib,lief. <let <let
Let be a unit vector in !Rn. Partition as
A
B
(a)
n
A(A +
(b)
34.
X
n
x
B B =
A+B
A + B.
A+
x,
B=
0,
x
35.
36.
37.
Let
Q
=
[-�1- �- - - C- - I�-xJyy- - - r]
yi I
n
I Ax I
(a)
B={
•
> m,
= ll x ll
•
x
m
X
A
n
•
x
+ (x
x y = (x
+
+ (x
·
·
·
Parseval's Identity. )
Prquiocvek metthathodisfoorrtfihnogonal
ding an. (Torhiths onorprocedurmal base giivsesfora !Rn
Q
wifretquent
h a prleysuscriebfuledifinrapplst vecticatorions.a)construction that is
triangulmatrairxmat. rix is orthogonal,
tPrPrhenooveveittthhmusatat iitffbean aupperdiagonal
t
h
en
t
h
er
e
i
s
no
mat
r
i
x
n
sLetuch that v1, , vn} beforan alorlthonorin !R m. al basis for !Rn.
Prove that, for any and yin !Rn,
· · v1) (y · v1) · v2) (y · v2) · vn) (y · vn)
(This idoesdentiParsty isecalvalle'sdIdentity imply about the
What
relationship between the dot products · and
(b)
-
-
x y
[xla · [y] a?
O rth o g o n a l C o m p l e m ents a n d
O rth o g o n a l Proieclio n s
red in Chapt
eermentThes, andno­
ttIihnoetnhprofisosjaeectctnoriioomn,n alofwevectonegenerovectr atoliozareplonttwaneooconcept
wianotl hbeersextwithatlendedgiweveencount
troiseorttohetogonal
compl
he concept of orthogonal
projection onto a subspace.
Apasnorsesmthalrovectughotrhne ortoiagipln,atnehenisiort isthaogonal
to everW ofy vectasorisinsptan(hatnpl) .aHence,
ne. If thwee plhaveane
s
u
bs
p
ace
vecttwo osur bsofptaceshe otofher. Thiwitshisththeepridoeaperbehity thnatd teverhe foly vectlowionrgofdefionenitiisoorn.thogonal to every
n
n
Let
Wbe
a
s
u
bs
p
ace
of
!R
.
We
say
t
h
at
a
vect
o
r
v
i
n
!R
irss that are
W
i
f
v
i
s
or
t
h
ogonal
t
o
ever
y
vect
o
r
i
n
W.
The
s
e
t
of
al
l
vect
o
orthogonal to Wis calW_j_led th{vie n !Rn : v ·w for all wiW,n denot
e
d
W_j_
.
That
i
s
,
W}
diorIfctWihulogonal
asr atoplWatneo(everit.eh.r,oyparughvectalteholretwioro tighnienW;norinhence,
malandvectee oisrW_Ltthoe W).liMorne, tthheenoverroughever, Wtyhconsevectoriigositrnsvperpen­
on e is
ofFigturhose e vectil uosrtsrwatetshtathiars seitoruatthioogonal
n. to every v on hence, we also have W e_L.
1.
w _j_
is pronounced perp:'
"w
O rth ogonal Complements
IR 3 ,
IR 3
w
w
DefiniliOD
nal to
I
e
and
orthogonal complement of
Figure 5 . 5
e
=
w_j_
orthogo­
=
w
=
e _j_
Exa m p l e 5 . 8
=
0
IR 3
=
5.5
€;
precisely
=
Section 5.2 Orthogonal Complements and Orthogonal Projections
319
thogonalementcomplof theementcomplofementa subsofpacea sutbsurpnaceed wasout tthoebeorian­gi­
otnalhersIunbssExampl
upbsace.pace.These Alseoprt, htohepereorcompl
(b) ofATheorofemsets APrandoperttconsiieess ar(ics)etsandtrofuet(hidnei)gener
wir common
l alasloandbeelaruseement
eprfulo.ved(sR. Seeecalas prAppendi
l thoatperthtieexs A.(a)) and
Leta. beis aa ssuubsbsppaceace of!Rof !Rnn..
b.c.
d. If span(W1, . . . , Wk), then Vis in if and only if V W; for all
be a sc(ala)ar.SiThennce w u ·forw allvwi· wn ifos rinall w inLet u and v be in and let
Therefore, (u v) ·w u·w v·w
so uWe valissoinhave
(
)
·w
c
u
·w
fr(bo)mWewhiwichl weprosveeetthhiats prcuoperis itny as CorIt ofoll alorwsy that is a subspace of !Rn.
((dc)) YouYou araree asaskkededtoto prprooveve tthhiiss prprooperperttyyiinn ExerExercciissee
assoWeciatecand winowth anexpres matsomerixfundament
. al relationships involving the subspaces
LetA isAthbee nulanl space ofmatA, randix. Thenthe ortthheogonal
orthogonalcomplcompl
e
ment
of
t
h
e
r
o
w
s
p
ace
of
of the column space ofA
is the null space(ofrowAT(A: ) .L null(A) and (cole(ment
A) .L null(AT)
n
I
f
x
i
s
a
vect
o
r
i
n
!R
everin nuly lr(oAw),ofsoA.weButhavethiess itsabltr,uistehhenedif andtxhiesonlfiinrsyt(rioidfw(entAxAit)y. Toifwhiandprocvehonlitshyequieifsexcondvials eornttidhtentogonal
o xitbeiy, wentog
simply replace A by AT and use the fact that row(AT ) col (A) .
rix hascompl
four suement
bspaces:s inro!Rw(n, Aand), nulthle(Ala)s, tcoltw(Ao )ar, ande ornulthogonal
l (A T ) .
TheThus,first tanwo are ormatthogonal
5.8,
tion
Theorem 5 . 9
5.9.
nB
intersec­
B
W
w.L
( W.L ) .L = W
w n w.L = {o}
W=
W.L
·
i = 1, . . . , k.
0
Proof
c
·
=
0
=
+
+
=
0
W
+
=
0+0
=
W .L .
(cu)
=
=
c (O)
5.12.
Theorem 5 . 1 0
m X
23.
24.
n
n
=
=
Proof
=
.L
0,
=
m X
0
W.L
W.L .
m X
=
n
0
W.L
W.L .
W, 0
=
=
0
380
Chapter 5 Orthogonality
null(AT)
null(A)
.
/
/
�.
row(A)
col(A)
[ffi m
The four fundamental subspaces
Figure 5 . 6
compl
e
ment
s
i
n
The
mat
r
i
x
A
defi
n
es
a
l
i
n
ear
t
r
a
ns
fo
r
m
at
i
o
n
fr
o
m
into Figurwhose
rlaungestraitsescolth(esA)e. iMordeaseoverschemat, thisictalralyns. foThesrmeatfourion sesundsbspnulacesl (Aar)etocalleind
e
i
the
of the matrix A.
Find bases for the four fundamental subspaces of
A=
]
I
[
;
and verify Theorem
column space,In Exampl
and nulelsspace ofA. Weandfound twehat comput
row(A) e=d sbaspan(esufor1, u2t,hu3e )r,owherw speace,
, U3
=
u2 =
Also, null (A) = span(x1, x2 ), where
= , =
Toeachshxowj, whithatch(riosw(anAeas) 1y- =exernulcils(eA. )(,Withyis enough
is this sutoffischiowentt?h) at every U; is orthogonal to
�m.
�m
�m.
X
m
n
5.6
mX
fundamental subspaces
Exa m p l e 5 . 9
0
�n
n
3
-1 0
2 1 -2
1 6 1
5.10.
3.45, 3.47,
Solulion
ll 1
[1 0
0 -1],
X1
�
3.48,
[O
2 0 3]
-1
-2
1
0
0
1
-3
0
-4
Xz
=
[0 0 0
4]
Section 5.2 Orthogonal Complements and Orthogonal Projections
381
The column space ofA is col (A) span(a1, a2, a3), where
=
We stil need to compute the null space ofAr. Row reduction produces
-3
63
[Ar 3 6
6 3
So,It foliflowsisthinat the null space of Ar, then y1 -y4, y2 - 6y4, and y3 - 3y4.
null(A') { [ =:�]} sp� ( [ =m
and it is easy to check that this vector is orthogonal to a1, a2, and a3.
The method of Example is easily adapted to other situations.
Let be the subspace of spanned by
-3 ,
,
3
Find a basis for
space of The subspace spanned by w1, w2, and w3 is the same as the column
A = -3
3
2
4 0
1 -1 2 1 0
0
0
1
1 -2 1 0
-1
0
i O]
y
�
1
0
0
0
0
0 0 1 0
0
0
0
0
0 0 0 0
0 0 0 0
=
=
�
�
5.9
Exa m p l e 5 . 1 0
IR 5
W
1
5
0
5
W1 =
Wz =
-1
1
2
-2
W3 =
W_j_ .
Solution
W
1 -1 0
1 -1
5 2 4
0 -2 - 1
5
5
0
-1
4
-1
5
382
Chapter 5 Orthogonality
T
Ther
e
for
e
,
by
Theor
e
m
Wl
_
=
(
c
ol
(
A
)
J
_
=
nul
l
(
A
), and we may proceed as in
the previous example. We compute
IA' o J [ - :
J [
J
= and = It
folHence,lows thiats in W1-if and only if =
=
= span
and these two vectors form a basis for WJ_ .
Recall that, in the projection of a vector (vu•vonto)a nonzero vector u is given by
proj0(v) = u·u u
Furcantdecompos
hermore, tehve asvector perp0(v) = v - proj0(v) is orthogonal to proj0(v), and we
as shIfownwe leint WFig=ursepan(u), then w = proju(v) is in Wand w1- = perpu(v) is in W1- . We
a way ofto"dW-namel
ecomposiyn,g"v v=inwto thwle _s.uWem ofnowtwogenervectoarlisz, eonethifrs iodmeaWand
tthhere oteforhere haveorthogonal
to
basdefiins edforasW. ForLetanyWvectbe aorsuvbsinpace ofthe and let {u1, ... , uk} be anvorthogonalWis
proj w(v) (UU11••U1 )u, ( Uk"Uk·Uk)uk
The
v
W is the vector
perpw (v) = v - projw (v)
tsoensr Each
(eo)r., Therequisummand
veforaleent, lwiyi,ntthhthetheone-edefinotdnaiitmtiiooensnn ofofiotprnalheojprswu(ebsvcedi)piaces nalgssodefipaannedprniotjioectn,byioweint-ontcaninowroura siitneprgleevivec­ous
5.10,
l
-3 5 0 5 O
2 -2 3 0
-1 4 -1 5 0
�
y
- 3y4 - 4y5 , y2
y1
wl_
�
1 0 0 3
0 1 0 1
0 0 1 0
-y4 - 3y5 ,
-4
-3
-2
0
-3
-1
0
1
0
- 3y4 - 4y5
-y4 - 3y5
- 2y5
Y4
Ys
4 O
3 0
2 0
y3 - 2y5 .
O rth ogonal Proieclions
v
IR 2 ,
--
u
v proj0(v) + perp0(v)
Figure 5 . 1
=
5.7.
!R n .
+
Definition
!R n
!R n ,
=
orthogonal projection of
V
--
component of orthogonal to
+ ···+
V
--
onto
Section 5.2 Orthogonal Complements and Orthogonal Projections
383
w
Figure 5.8
P = P1
+ Pz
ofurSineitces prthiolejuvectescttriaootnsress tonthiarsoseione-toruatthdiogonal
ionmwiensthi,otWnalhe orsutsbshpogonal
panaces(u1, tuprh2at)o,jparecteimutoprn ofouj alwvl(yontv)or, opt1hWiogonalsprthoej.0sFi1u(vmg)­,
andAsp2 a spprecioja0l,(casv) .e of the definition ofproj w(v), we now also have a nice geometric
erminolbasogy,is
ti{nvhtat1e, rvtp2h,reoretate,imovd,nstofathteTheor
senthateimf is inInthteersmubss ofpaceourWprofesent whinotcahtiohasn andorthtogonal
U;
5.8
=
=
=
=
5.2.
!R n ,
w
•
.
•
Thus,
oforthogonal projections onto mutually orthogo­
nal one-The ddefiisimdecompos
ensnitioionnalaboveesdubsinspteoacesemsa suoftmo W.depend
on
t
h
e
choi
c
e
of
or
t
h
ogonal
bas
i
s
;
t
h
at
i
s
,
differentFor" prnow,oj w(lve)t'sandbe
perconta dipffew(ntervewint) . tForbash antiusexampl
nat{u;e,ly..., eth,. iu£s }isfornotWthwoule casde,appear
as we wito lgisvoeona "prove.
w
Exa m p l e 5 . 1 1
Let W be tbe plone in with equation
and let v � [ : ] . Find tbe
orthogonal projection ofv onto Wand the component of v orthogonal to W.
In Example we found an orthogonal basis for W. Taking
II'
Solution
we have
x
-
y
+ 2z � 0,
5.3,
ll1
" V
=
U1 • U1 =
2
2
Uz • V =
Uz • Uz =
-2
3
-
384
Chapter 5 Orthogonality
Therefore,
proj w (v) = ( ) + ( )
� i[ � l - f : J Ul
and perpw(v) = v - proj w (v) = [ - �3 ] [ -i� ] [ -�; ]
nce it sattoisW,fiessitnhcee equat
is equalIt isleasy easy tyotsoeseetehatthatprperojwp(vw() vis)iins orW,thsiogonal
it is aioscnalofartmulhe pltiapne.le ofIt
thrnornal vootm [ - : l to W. igmo
u1 · v
-- u 1
U 1 . U1
( S" F
u2 · v
U U
z. z
u2
5.9.)
perpw(v)
v projw(v) + perpw(v)
Figure 5 . 9
=
Theorem 5 . 1 1
respTheect tnexto a suthbseorpaceemandshowsits torhatthweogonalcan alcomplwaysefimentnd a.decomposition of a vector with
ubspacew_j_ofin W_j_andsuchletthvatbe a vector in Then there are unique
vectLet oWrsbewian sWand
v = w + w_j_
We need to show two things: that such a decomposition and that it is
for Let
orthogonal basis
w =Toprowsjhwow+(vw_j_)exiands=telprence,t ow_j_j wwe(=v)perchoos
p+ perw(epvanw) .(Then
v) = proj w(v) + (v - proj w(v) = v
The Orthogonal Decomposition Theorem
IW
Proof
unique.
!R n .
exists
{u1, . . . , uk}
W
Section 5.2 Orthogonal Complements and Orthogonal Projections
385
Clu1,ear. .ly. ,, wuk. Toprshojoww (vth) atis wlin. W,is insinWlce. ,ititiiss aenough
linear combi
nattihoatn wlof.thiseorbasthiogonal
s vectotros
t
o
s
h
ow
each of the basis vectors U;, by Theorem 5.9(d) . We compute
U; wl. u;· per(v -pwpr(vo)j w(v)
u;· (v - (:1.";Ju1 - - (�k.";Juk)
Uu;·; " u;V ) (u' · u)'
UU11·"Vu1 ) (u' · u ) - (-u-·' v - (-u-·v - - (-uui' ·-· uvi ) (u-· u.)
that wl. is in Wl. and completes the existence
parsincetToofu;sthhujowe prtohof.e founir quenesThis ofs tprhiosvesdecompos
ppose wew_Lhave anotW1 +hwt,er de­so
composition v W1 +wt, where w1 is in Wandwtitioin,s inletwl's s.uThenw+
But sinthceatwth-is common
w1 is in Wand
these earme 5.su9bs(cp) ]aces. Thus,), we
know
vectorwtis in-Wwl. isWlin. Wl. (b[ecaus
usingeTheor
=
•
=
=
U;
•
=
···
=
=
t
0-···
z
-···
-···
z
z
-· · ·- 0
= u; · v - u; · v = 0
·
=0
j -=F i.
=
=
n
= {O}
the Exampl
subspacee 5.of1IR1 3ilgiuvsentrabytedththeeplOranethwiogonalth equatDecompos
ion - iti+on Theoretm.he orWhenthogonalW is
dernmpo,;tion of v [ -! ] wdh mped to W ;, w + w", whece
w pmj w(v) Ul Md w" pecpw(v) H l
s of thve)ordothogonal
decomposon thiteiochoin guarce aofnteoresththogonal
at the defibasinsi.tiTheons
ofOrtprhTheogonal
ojwuni(v)qDecompos
anduenesperpw(
not
depend
i
t
i
o
n
Theor
e
m
al
s
o
al
l
o
ws
us
t
o
pr
o
ve
pr
o
per
t
y
(
b
)
of
Theo­
rem 5.e9m.. We state that property here as a corol ary to the Orthogonal Decomposition
Theor
If Wis a subspace of then
x y
�
�
�
C o r o l l a rv 5 . 1 2
.
!R n ,
�
2z = 0,
�
�
386
Chapter 5 Orthogonality
f w is inWWand(W_Lxis) _Lin. NowW_L ,lteht env bewinx(W_L ) _LBut. BythTheor
is nowemimplieswethcanat wwris iitne
(W_Lv w) _L . IHence,
w_Lv w_Lfor (u(wnique)w_Lvect) wo_Lrs wiwn Wand
w_L
i
n
W_L
.
But
now
_L w_L w_L w_L w_L w_L w_L
w
and,so w_Lsince tTherhe reeverforsee, vincluwsionw_Lis alsow,trandue, wethusconclvisuindeW.thThiat (sWsh_Lows) _L thW,at (asWr_Le)qui_L red.W
There eims al5.so1 3.a nice relationship between the dimensions of Wand W_L , expres ed in
Theor
If W is a subspace of then dim W dim W _L
anworthogonal
onalWe clbasaimLetistforhatw_LBi...s. Thenan, ukor} betdihmogonal
kbasandis fodibasrmiw_Ls for WandLet lBet {v1, ......, v1,} Ukbe,vanl, ...orth, v1og-}.
We first note thatU,; 'siVnj ce eachforiis in Wand.. .,kandj
each vj is in W_L ,
Thus,, iBf viiss ana vectortohrogonal
shete and,Orthhence,
iDecompos
s linearly iintidependent
, bytelTheor
emat v
Next
i
n
t
ogonal
o
n
Theor
e
m
s
us
t
h
wtion ofw_Lthfoe rvectsomeorswiU; andn Wandw_L canw_LbeinwrW_Lit .enSinascea wlincanear combi
be wrint atenioasn ofa litnheare vectcombiors vnj,a­v
andcan beso wris aitbaseniass fora linearItcombi
follodiwsnmatthiwoatnkofdithme vectw_Ldimors in orB. Therefore, B spans also
As
a
l
o
vel
y
bonus
,
when
we
appl
y
t
h
i
s
r
e
s
u
l
t
t
o
t
h
e
fundament
a
l
s
u
bs
p
aces
of
a
matCororilxa, rwey get a quick proof of the Rank Theorem (Theorem restated here as
C
Proof
= +
0= ·
= 0.
Theorem 5 . 1 3
=
·
+
·
= +
=
=
+
·
= 0.
5. 1 1,
=0+
·
=
·
C
!R n ,
=
+
Proof
{u 1 ,
=
!R n .
u;
=0
n
= l.
= {u 1 ,
= 1,
= 1, . . . , l
5. 1 .
=
!R n ,
+
!R n .
!R n
!R n ,
+l=
+
=
n
3.26),
5. 14.
C o r o l l a rv 5 . 1 4
=
·
The Rank Theorem
IfA is an matrix, thenrank(A) nullity(A)
em
so diNotm eWInthTheor
rat awenk(emgetA) anda counttdiakemeW_LrWpart riodnulw(entlAiitty)y.(Then
Aby) . tTheakiW_LnrgesuWlnult folll(coloAws.)(,Aby) Theor
[and therefore
W_L null (AT )J :
rank(A) nullity(A
mX
n
+
Proof
=
5.13,
=
=
n
=
=
5.10,
=
=
+
r)
=m
Section 5.2 Orthogonal Complements and Orthogonal Projections
381
have, weilhaveustranotted essometablisofhedthtehatadvanteveryagessubsofpaceworkinganwiorth­
orthogonal
thSectogonaliobasnsbasis, enorsand. However
ven a mete hodThesfor conse is turuesctarineg tshuechsuabjbasectisof(etxcept
parsecttiioculn. ar exampleshave, suchweasgiExampl
he nextin
5.1
5.2
has
5.3).
I
Exercises 5 . 2
In Exercises 1 -6, find the orthogonal complement W_j_ of W
and give a basis for W_j_ .
1. W =
2. W =
3. w �
4. W �
5. W �
6.
W�
{ [;]
{ [;]
{ [�l
{ [:J
{ [�l
{ [�]
:
:
}
3x + 4y = 0 }
2x - y = 0
x+y-z� 0
2x - y + 3z
x
x
�
�
I,
�
y
�
t, y
)
�
0
- I, z
�
)
�
+. z
3t
)
�
21
)
In Exercises 7 and 8, find bases for the row space and null
space of A. Verify that every vector in row(A) is orthogonal
to every vector in null(A).
7. A =
[ � : _! ]
[ -� � -� � !l
-1 -1
8. A =
1
2 2 -2 0 1
-3 - 1 3 4 5
In Exercises 9 and 1 0, find bases for the column space of
A and the null space of A T for the given exercise. Verify
that every vector in col(A) is orthogonal to every vector
in null(A T).
10.
8
9.
7
Exercise
In Exercises 1 1 - 1 4, let W be the subspace spanned by the
given vectors. Find a basis for W_j_ .
Exercise
14.
W1
4
6
= -1
1
-1
, Wz
=
2
0
-3
, W3
2
2
= 2
-1
2
In Exercises 1 5- 1 8, find the orthogonal projection of v onto
the subspace W spanned by the vectors u;. (You may assume
that the vectors U; are orthogonal.)
Chapter 5 Orthogonality
388
t
h
at
v
=
I
s
i
t
neces
s
a
r
i
l
y
t
r
u
e
t
h
at
i
s
i
n
W
J
.
?
EiLetth{erv1pr, ...ove, vthn}atbeit anis torrutehorogonalfind abascountis foerre!Rxampl
eand l.et
v
n
v = [ - � l W = span( [�] )
Wpan(= vspk+an(l' ...v1, ,...vn),?vEik)t·hIers itprnecesove tsharatilyit tirsuteruthe atorWlfin.d=a
scount
e
r
e
xampl
e
.
J
W
([
J
)
[
:
!Rn, x
n!R . W
Pr
o
ve
t
h
at
xi
s
i
n
W
i
f
and
onl
y
i
f
pr
o
j
w
(
x
)
=
x.
(
)
H
to W if and only if
[ � ]. [ - : ]
[prPrProoojvevew(ttxhh)atat=prxiosjorw(tphrogonal
o
j
w
(
x
)
=
pr
o
j
w
(
x
)
.
Letlet xSbe= a{vvect1, ...or ,invk!R} nbe. an orthonormal set i n !Rn, and
[ -; l W ( [J [ -m
Prol vex l 2th2:at lx ·v1 l2 lx ·v l 2 lx ·vkl2
23. PrProoveve Theor
e
m
9
(
c
)
.
5.
2
(
T
hi
s
i
n
equal
i
t
y
i
s
cal
l
e
d
Theor
e
m
9
(
d
)
.
5.
LetthatW andbe a subsarepaceorthofogonal
!Rn andvectv aovectrs wiotrhin !Rinn. Wand
Suppose onlProyveiftxihats iBesn sspean(l's ISn).equality is an equality if and
In Exercises 1 9-22, find the orthogonal decomposition of
with respect to W
� 'P"•
20. v �
22.
+ w' .
w'
26.
19.
21.
w
In Exercises 27-29, let
a vector in
27.
28.
0.
29.
30.
w � 'P""
F
� 'P""
F
be a subspace of
(a)
+
24.
25.
and let be
+···+
Bessel's Inequality. )
(b)
w
w'
w
T h e G r a m - S c h m i d t Process
and t h e OR Facto rizati o n
tmosIhnonorthtiuss mseefulalct)iobasofn,aliwesl formatpranyerisxentfactsubsaospriimacezatpliofeons.met!Rn.hThiod sformetconshodtrwiuctl itnhgenanleoradthusogonal
to one(oofr torhe­
n
Weis towoulbegidnliwiketthoanbe arablbietrtaorfiynbasd anis or{xt1h, ogonal
bas
i
s
for
a
s
u
bs
p
ace
W
of
!R
.zThee" it ionedea
.
.
.
,
xd
for
Wand
t
o
"
o
r
t
h
ogonal
i
vectExamplor ate 5.a 3ti.me. We wil il ustrate the basic construction with the subspace W from
Let W = span(x1, x2 ), where
The Gram-Schmidl Process
Exa m p l e 5 . 1 2
Construct an orthogonal basis for W.
1, we getto ax1se(Fcondigurevect5. 1o0)r.that is orthogonal to it by taking
the componentStartiofngxwi2 orththxogonal
Solulion
Section 5.3 The Gram-Schmidt Process and the Factorization
QR
389
w
Constructing orthogonal to
Figure 5.10
v2
x1
Algebraically, we set v1 = x1 , so
v2 = perpxJ xz ) =
- proj x, ( xz )
x2
( )
= x2 - � xI
X 1 • xi
Then {v1 , v2 } is an orthogonal set of vectors in W. Hence, {v1 , v2 } is a linearly independent set and therefore a basis for W, since dim W = 2.
Remark
Theorem 5 . 1 5
[il
Observe that this method depends on the order of the original basis
mtocs. Jn Exrunpl' 5 . 1 2, if we had taken x ,
�
[ -fl
4
�
and x,
�
we would have
obtained a different orthogonal basis for W. (Verify this.)
The generalization of this method to more than two vectors begins as in
Example 5 . 1 2 . Then the process is to iteratively construct the components of subse­
quent vectors orthogonal to all of the vectors that have already been constructed. The
method is known as the Gram-Schmidt Process.
The Gram-Schmidt Process
Let {x 1 , . . . , xd be a basis for a subspace W of !R n and define the following:
_ ( : ) _ ...
xk
Vz Vz
v2
vz
Then for each i = 1 , . . . , k, {v1 , . . . , v; } is an orthogonal basis for W;. In particular,
{v1 , , vk } is an orthogonal basis for W.
.
•
.
390
Chapter 5 Orthogonality
Jorgen
PedersenwasGram
(1850-1916)
a Danishwhoactuary
(insurance
stati
s
tician)
was
isurement.
nterested inHethefirstscience
of
mea­
published
the
process
that
bears
his
name
in
an 1883Schmidt
paper on(1876-1959)
least squares.was
Erhard
astudiGerman
mathematician
whod
e
d
under
the
great
Davi
Hilbert
and is considered
one of
ofmathematics
the founders
of
the
branch
known as functional
analysis.
His
contribution
to thein a
Gram-Schmidt
Process
came
integral
inof1907thewhipaper
cmethod
h heonwrotemore
outexpltheequations,
details
i
c
itly than
Gram had done.
Stated succinctly, Theorem 5 . 1 5 says that every subspace of ll�r has an orthogonal
basis, and it gives an algorithm for constructing such a basis.
Proof
We will prove by induction that, for each i = 1, . . . , k, {v1 , . . . , v;} is an or­
thogonal basis for W;.
Since v1 = x1 , clearly {v i } is an (orthogonal) basis for W1 = span(x 1 ) . Now assume
that, for some i < k, {v1 , . . . , v;} is an orthogonal basis for W;. Then
V;+ J
=
X;+ J
_(
) _(
VI • X;+ 1
V1
V1 . V1
)
V2 • X;+ 1
Vz
Vz . Vz
_ ... _ (
)
V; • X;+ 1
V;
V; . V;
By the induction hypothesis, {v1 , . . . , v;} is an orthogonal basis for span(x1 , . . . , x;) =
W;. Hence,
V;+ 1
=
X;+ 1 -
proj w, C x;+ 1 ) = perp w, C X;+ 1 )
So, by the Orthogonal Decomposition Theorem, V;+ 1 is orthogonal to W;. By
definition, v1 , . . . , v; are linear combinations of x1 , . . . , X; and, hence, are in W;.
Therefore, {v1 , . . . , V;+ d is an orthogonal set of vectors in W;+ i ·
Moreover, V;+ i * 0, since otherwise X;+ 1 = proj w, C X; + 1 ) , which in turn implies
that X;+ 1 is in W;. But this is impossible, since W; = span(x 1 , . . . , X;) and {x 1 , . . . , X;+ i }
is linearly independent. (Why?) We conclude that {v1 , . . . , V;+ d is a set of i + ! linearly independent vectors in W;+ i · Consequently, {v1 , . . . , V;+ d is a basis for W;+ 1 ,
since dim W;+ 1 = i + 1 . This completes the proof.
If we require an orthonormal basis for W, we simply need to normalize the
orthogonal vectors produced by the Gram-Schmidt Process. That is, for each i, we
replace V; by the unit vector q; = ( l / ll v; ll )v; .
Exa m p l e 5 . 1 3
Apply the Gram-Schmidt Process to construct an orthonormal basis for the subspace
W = span(x1 , x2 , x3 ) of IR 4 , where
First we note that {x1 , x2 , x3 } is a linearly independent set, so it forms a basis
for W. We begin by setting v1 = x1 . Next, we compute the component of x2 orthogo­
nal to W1 = span(v1 ) :
Solulion
Section 5.3 The Gram-Schmidt Process and the Factorization
QR
391
For hand calculations, it is a good idea to "scale" v2 at this point to eliminate fractions.
When we are finished, we can rescale the orthogonal set we are constructing to obtain
an orthonormal set; thus, we can replace each V; by any convenient scalar multiple
without affecting the final result. Accordingly, we replace v2 by
We now find the component of x3 orthogonal to
W2 span ( x1 , xi ) span ( v1 , v2 )
=
=
=
span ( v1 , v� )
using the orthogonal basis {v1 , v�}:
Agllin, we mrnle and "'";
�
[
� 2 � !l
v,
We now have an orthogonal basis {v1 , v�, v� } for W. (Check to make sure that
these vectors are orthogonal.) To obtain an orthonormal basis, we normalize each
vector:
q3
=
C: )
v
1 � 11 �
=
-1� 1
[
( �) 2
=
[ -11/�V6V6]
2/V6
Then {q 1 , q2 , q3 } is an orthonormal basis for W.
[ -V6/�/66 ]
V6/3
392
Chapter 5 Orthogonality
One of the important uses of the Gram-Schmidt Process is to construct an orthogo­
nal basis that contains a specified vector. The next example illustrates this application.
Exa m p l e 5 . 1 4
Find an orthogonal basis for IR 3 that contains the vector
Solulion
�
We first find any basis for IR 3 containing v1 • If we take
then {v1 , x2 , x3 } is clearly a basis for IR 3 . (Why?) We now apply the Gram-Schmidt
Process to this basis to obtain
and finally
Then {v1 , v�, v� } is an orthogonal basis for IR 3 that contains v1 .
Similarly, given a unit vector, we can find an orthonormal basis that contains it by
using the preceding method and then normalizing the resulting orthogonal vectors.
R e m a r k When the Gram-Schmidt Process is implemented on a computer, there
is almost always some roundoff error, leading to a loss of orthogonality in the vec­
tors q;. To avoid this loss of orthogonality, some modifications are usually made. The
vectors V; are normalized as soon as they are computed, rather than at the end, to
give the vectors q;, and as each q; is computed, the remaining vectors xj are modified
to be orthogonal to q;. This procedure is known as the Modified Gram-Schmidt
Process. In practice, however, a version of the QR factorization is used to compute
orthonormal bases.
The OB Facto rization
2
If A is an m X n matrix with linearly independent columns (requiring that m n),
then applying the Gram-Schmidt Process to these columns yields a very useful fac­
torization of A into the product of a matrix Q with orthonormal columns and an
Section 5.3 The Gram-Schmidt Process and the Factorization
QR
393
upper triangular matrix R. This is the QR factorization, and it has applications to the
numerical approximation of eigenvalues, which we explore at the end of this section,
and to the problem of least squares approximation, which we discuss in Chapter 7.
To see how the QR factorization arises, let a 1 , . . . , an be the (linearly independent)
columns of A and let q 1 , . . . , qn be the orthonormal vectors obtained by applying the
Gram-Schmidt Process to A with normalizations. From Theorem 5.15, we know that,
for each i = 1, . . . , n,
W; = span ( a 1 , . . . , a; ) = span ( q 1 , . . . , q; )
Therefore, there are scalars r 1 ;, r2 ;, • . • , r;; such that
That is,
a, = r l l q,
az = r 1 2 q 1 + Yzz qz
which can be written in matrix form as
r 1n
r� n
0
rnn
l
= QR
Clearly, the matrix Q has orthonormal columns. It is also the case that the diago­
nal entries of R are all nonzero. To see this, observe that if r;; = 0, then a; is a linear
combination of q 1 , . . . , q; _ 1 and, hence, is in W; _ 1 . But then a; would be a linear com­
bination of a 1 , . . . , a; _ 1 , which is impossible, since a 1 , . . . , a; are linearly independent.
We conclude that r;; -=fa 0 for i = 1, . . . , n. Since R is upper triangular, it follows that it
must be invertible. (See Exercise 23. )
We have proved the following theorem.
Theorem 5 . 1 6
The QR Factorization
Let A be an m X n matrix with linearly independent columns. Then A can be fac­
tored as A = QR, where Q is an m X n matrix with orthonormal columns and R is
an invertible upper triangular matrix.
R e m a rks
We can also arrange for the diagonal entries of R to be positive. If any r;; < 0,
simply replace q; by - q; and r;; by - r;;.
•
The requirement that A have linearly independent columns is a necessary one.
To prove this, suppose that A is an m X n matrix that has a QR factorization, as in The­
orem 5.16. Then, since R is invertible, we have Q = AR - 1 • Hence, rank(Q) = rank(A),
by Exercise 61 in Section 3.5. But rank( Q) = n , since its columns are orthonormal and,
therefore, linearly independent. So rank(A) = n too, and consequently the columns of
A are linearly independent, by the Fundamental Theorem.
•
Chapter 5 Orthogonality
394
=
=
•
The QR factorization can be extended to arbitrary matrices in a slightly
modified form. If A is m X n, it is possible to find a sequence of orthogonal matrices
Q1, , Om - I such that Om - I · · · Q2 Q1A1 is an upper triangular m X n matrix R. Then
A QR, where Q ( O m - I · · · Q2 Q1) - is an orthogonal matrix. We will examine this
approach in Exploration: The Modified QR Factorization.
.
Exa m p l e 5 . 1 5
•
.
=A [ - � � �1
Find a QR factorization of
-1 0 1
1
2
The columns of A are just the vectors from Example 5.13. The orthonormal
basis for col(A) produced by the Gram-Schmidt Process was
Solulion
q1
= [ = : ;: 1 ,
=
=
qz
Q
q3
V6/3
Vs/ 10
1/2
so
= [ ��:: 1 , = [ - ��:1
=[
v: 1
=
= =IR =
l =:
l
[
�
l
[ q i qz q 3 ]
1/2 3 Vs/10 - /6
- 1/2 3 Vs/10
- 1/2 Vs/10 V6/6
1/2 Vs/10 V6/3
From Theorem 5.16, A QR for some upper triangular matrix R. To find R, we
use the fact that Q has orthonormal columns and, hence, Q TQ I. Therefore,
=
We compute
R
..
I
Exercises 5 . 3
Q'A
[=
Q TA
1/2
3 Vs/ 10
- V6/6
1
Vs
0
� [:
In Exercises 1 -4, the given vectors form a basis for IR 2 or IR 3 •
Apply the Gram-Schmidt Process to obtain an orthogonal
basis. Then normalize this basis to obtain an orthonormal
basis.
2. X1
3.
..
Q TQR
R
- 1/2 - 1 /2
1 /2
3 Vs/10 Vs/10 \/5/ 10
V6/6 V6/3
0
1 /2
3 Vs/2
V6/2
= [ _�l
[�]
[ x [ x
� J ,� n ,�m
Xz =
1 2
1
0
1
Section 5.3 The Gram-Schmidt Process and the Factorization
QR
Exercises 5 and 6, the given vectors form a basis for a
subspace W of IR 3 or IR 4. Apply the Gram-Schmidt Process
to obtain an orthogonal basis for W
**
*]
In Exercises 13 and 1 4, fill in the missing entries of Q
to make Q an orthogonal matrix.
1/ \/2 1 / \/3
13. Q
0
1 / \/3
- 1/ \/2 1 / \/3
1/2 2/ Vi4
1/2 l/ Vi4
14. Q
0
1/2
1/2 - 3/ Vi4
=
In
395
=
[
[
**
** **: ]
In Exercises 1 5 and 1 6, find a QR factorization of the
matrix in the given exercise.
16. Exercise 10
15. Exercise 9
In Exercises 7 and 8, find the orthogonal decomposition of v
with respect to the subspace W
7. v �
8.
F
[ -:}
[ i}
w "' i• Emd" ,
[: � �]
[:J
1 1 . Find an orthogonal basis for IR 3 that contains the
veoto'
12. Find an orthogonal basis for IR 4 that contains the
vectors
] ['
i -! ]
=
=
W as in Exmise 6
Use the Gram-Schmidt Process to find an orthogonal basis
for the column spaces of the matrices in Exercises 9 and 1 0.
9.
In Exercises 1 7 and 1 8, the columns of Q were obtained by
apply ing the Gram-Schmidt Process to the columns ofA.
Find the upper triangular matrix R such that A QR.
l
8 2
3
�
17. A
7 -1 ' Q
3
2
1
-2
--3 3
18. A
=
u
]
=
[ -j - � [ �
3
4
,Q
=
:3
1 /v'6 1/
2/ V6
- 1 V6 1 / \/3
1 / \/3 _
19. If A is an orthogonal matrix, find a QR factorization
of A.
20. Prove that A is invertible if and only if A QR, where
Q is orthogonal and R is upper triangular with nonzero
=
entries on its diagonal.
In Exercises 21 and 22, use the method suggested by
Exercise 20 to compute A i for the matrix A in the given
exercise.
22. Exercise 15
21. Exercise 9
23. Let A b e an m X n matrix with linearly independent
-
columns. Give an alternative proof that the upper
triangular matrix R in a QR factorization of A must
be invertible, using property (c) of the Fundamental
Theorem.
24. Let A be an m X n matrix with linearly independent
columns and let A = QR be a QR factorization of A.
Show that A and Q have the same column space.
Exp lorations
T h e M o difi e d
QR
0
Factorizat i o n
When the matrix A does not have linearly independent columns, the Gram-Schmidt
Process as we have stated it does not work and so cannot be used to develop a gen­
eralized QR factorization of A. There is a modification of the Gram-Schmidt Process
that can be used, but instead we will explore a method that converts A into upper
triangular form one column at a time, using a sequence of orthogonal matrices. The
method is analogous to that of LU factorization, in which the matrix L is formed
using a sequence of elementary matrices.
The first thing we need is the "orthogonal analogue" of an elementary matrix;
that is, we need to know how to construct an orthogonal matrix Q that will trans­
form a given column of A-call it x-into the corresponding column of R-call it y.
I Qx ll
By Theorem 5.6, it will be necessary that ll x ll
Figure 5. 1 1 suggests a
way to proceed: We can reflect x in a line perpendicular to x y. If
= = l r l-.
y
figure 5 . 1 1
- uj_ = [ �: ]
u
--2d21dd22 ] = I - 2uu .
is the unit vector in the direction of x
y, then
-
is orthogonal to
u,
and
we can use Exercise 26 in Section 3.6 to find the standard matrix Q of the reflection
in the line through the origin in the direction of J_ .
1 . Show that Q =
2.
Compute Q for
[ l--2d,2dd212
u = [fl = [ : J , r = [ � ]
u
= I - 2uur
T
1
(b) x
( a)
We can generalize the definition of Q as follows. If is any unit vector in !R n , we
define an n X n matrix Q as
Q
396
Such a matrix is called a Householder matrix (or an elementary reflector) .
3. Prove that every Householder matrix Q satisfies the following properties:
(a) Q is symmetric.
(b) Q is orthogonal.
(c) Q 2 = I
4. Prove that if Q is a Householder matrix corresponding to the unit vector
u, then
-v if v is in span ( u )
Qv =
v if v · u = 0
AlstononeHouseholder
(1904-1993)
was
of
the
pioneers
in theHefield
ofwasnumerical
linear
algebra.
the firstoftoalgpresent
a forsystemati
c
treatment
ori
t
hms
sol
v
i
n
g
problems
intovolintroducing
ving linear systems.
Inwidaddition
the
el
y
used
Householder
trans­
formations
thatfirstbeartohisadvocate
name, hethe
was
one
of
the
systematic
algebra. Hisuse1964of norms
book in linear is
considered a classic.
The Theory
of Matrices in Numerical Analysis
5.
Compute Q foe u
�
{
[ �]
-
<md wdfy Prnblem' 3 ond 4.
6. Let x * y with ll x ll = llrll and set u = ( 1/ ll x - rll ) ( x - y). Prove that the
corresponding Householder matrix Q satisfies Q x = y. [Hint: Apply Exercise 57 in
Section 1 .2 to the result in Problem 4.]
7. Find Q and verify Problem 6 for
We are now ready to perform the triangularization of an m X n matrix A, column
by column.
8. Let x be the first column of A and let
Show that if Q 1 is the Householder matrix given by Problem 6, then Q 1 A is a matrix
with the block form
where A 1 is (m - l) X ( n - 1 ).
If we repeat Problem 8 on the matrix A 1 , we use a Householder matrix P2 such that
where A 2 is ( m - 2 ) X ( n - 2 ).
9. Set Q =
2
[ � :J.
Show that Q 2 is an orthogonal matrix and that
391
10. Show that we can continue in this fashion to find a sequence of orthogonal
matrices Qi, . . . , Q m - i such that Q m - i Q2 QiA = R is an upper triangular m X n
matrix (i.e., r;1 = 0 if i > j).
1 1 . Deduce that A = QR with Q = Qi Q2 Q m - i orthogonal.
1 2. Use the method of this exploration to find a QR factorization of
· · ·
· · ·
3
3
-4
-1
-5
Ap proxi m ating Eigenval u e s
with t h e
SeeLoan,G. H. Golub and C. F. Van
(Baltimore:Press,
Johns1983).
Hopkins
University
Matrix Computations
QR
Algo rith m
One of the best (and most widely used) methods for numerically approximating the
eigenvalues of a matrix makes use of the QR factorization. The purpose of this ex­
ploration is to introduce this method, the QR algorithm, and to show it at work in a
few examples. For a more complete treatment of this topic, consult any good text on
numerical linear algebra. (You will find it helpful to use a CAS to perform the calcula­
tions in the problems below.)
Given a square matrix A, the first step is to factor it as A = QR (using whichever
method is appropriate). Then we define Ai = RQ.
1. First prove that Ai is similar to A. Then prove that Ai has the same eigen­
values as A.
2.
If A =
[ � �]
, find A 1 and verify that it has the same eigenvalues as A.
Continuing the algorithm, we factor A i as Ai = QiR 1 and set A 2 = RiQi. Then we
factor A 2 = Q2 R 2 and set A 3 = R 2 Q2 , and so on. That is, for k 2: 1 , we compute Ak =
QkRk and then set Ak i = RkQk.
+
3. Prove that Ak is similar to A for all k 2: 1.
4. Continuing Problem 2, compute A 2 , A 3 , A 4 , and A5, using two-decimal-place
accuracy. What do you notice?
It can be shown that if the eigenvalues of A are all real and have distinct absolute
values, then the matrices Ak approach an upper triangular matrix U.
5. What will be true of the diagonal entries of this matrix U?
6. Approximate the eigenvalues of the following matrices by applying the QR
algorithm. Use two-decimal-place accuracy and perform at least five iterations.
( a)
( c)
[� � ]
[� �]
�[ � - � ] [ � � ]
[ ]
(b)
-4
0
1
-
( d)
-2
2
4
2
2
7. Apply the QR algorithm to the matrix A =
-1
Why?
398
3
-2
. What happens?
8. Shift the eigenvalues of the matrix in Problem 7 by replacing A with
B = A + 0.9I. Apply the QR algorithm to B and then shift back by subtracting 0.9
from the (approximate) eigenvalues of B. Verify that this method approximates the
eigenvalues of A.
9. Let Q0 = Q and R 0 = R. First show that
QO Q I . . . Qk - lA k = AQ0Q 1 . . . Qk - 1
for all k 2: 1 . Then show that
[Hint: Repeatedly use the same approach used for the first equation, working from
the "inside out:'] Finally, deduce that ( Q0Q 1 Qk) (Rk R 1 R0) is the QR factorization
of A k+ i .
· · ·
· · ·
399
400
Chapter 5 Orthogonality
Orthogonal Diagonalizalion ot svmmetric Matrices
[o ]
We saw in Chapter 4 that a square matrix with real entries will not necessarily have real
-1
eigenvalues. Indeed, the matrix
has complex eigenvalues i and - i. We also
1
0
discovered that not all square matrices are diagonalizable. The situation changes
dramatically if we restrict our attention to real symmetric matrices. As we will show
in this section, all of the eigenvalues of a real symmetric matrix are real, and such a
matrix is always diagonalizable.
Recall that a symmetric matrix is one that equals its own transpose. Let's begin by
studying the diagonalization process for a symmetric 2 X 2 matrix.
Exa m p l e 5 . 1 6
If possible, diagonalize the matrix A =
[
]
1
2
.
2 -2
The characteristic polynomial of A is A 2 + A - 6 = ( A + 3 )(A - 2), from
which we see that A has eigenvalues A1 = - 3 and A 2 = 2. Solving for the correspond­
ing eigenvectors, we find
Solulion
v1
[ -� �]
=
[ _�]
and
v2
=
[�]
respectively. So A is diagonalizable, and if we set P = [ v1
p - 1AP =
v2 ] ,
then we know that
= D.
However, we can do better. Observe that v1 and v2 are orthogonal. So, if we nor­
malize them to get the unit eigenvectors
[
l / Vs
U1 = -2/ Vs
and then take
]
and
[
u2
=
[ 2/l/ VsVs]
]
l / Vs 2/ Vs
-2/ Vs l / Vs
we have Q - 1AQ = D also. But now Q is an orthogonal matrix, since { u1, u2 } is an
1
orthonormal set of vectors. Therefore, Q - = Q T, and we have Q TAQ = D. (Note that
Q = [ u 1 Uz ] =
checking is easy, since computing Q - 1 only involves taking a transpose!)
The situation in Example 5.16 is the one that interests us. It is important enough
to warrant a new definition.
D efi n iii 0 n
A square matrix A is orthogonally diagonalizable if there exists an
orthogonal matrix Q and a diagonal matrix D such that Q TAQ = D.
We are interested in finding conditions under which a matrix is orthogonally
diagonalizable. Theorem 5 . 1 7 shows us where to look.
Section 5.4 Orthogonal Diagonalization of Symmetric Matrices
Theorem 5 . 1 1
401
If A is orthogonally diagonalizable, then A is symmetric.
Proof
If A is orthogonally diagonalizable, then there exists an orthogonal ma­
trix Q and a diagonal matrix D such that Q TAQ = D. Since Q - 1 = Q T, we have Q TQ =
I = QQ T, so
But then
since every diagonal matrix is symmetric. Hence, A is symmetric.
Theorem 5 . 1 7 shows that the orthogonally diagonalizable matrices are
all to be found among the symmetric matrices. It does not say that every symmetric
matrix must be orthogonally diagonalizable. However, it is a remarkable fact that this
indeed is true! Finding a proof for this amazing result will occupy us for much of the
rest of this section.
Remark
We next prove that we don't need to worry about complex eigenvalues when work­
ing with symmetric matrices with real entries.
Theorem 5 . 1 8
If A is a real symmetric matrix, then the eigenvalues of A are real.
Recall that the complex conjugate of a complex number z = a + bi is the number
a - bi (see Appendix C). To show that z is real, we need to show that b = 0. One
way to do this is to show that z = z, for then bi = - bi (or 2bi = O), from which it
follows that b = 0.
We can also extend the notion of complex conjugate to vectors and matrices by,
for example, defining A to be the matrix whose entries are the complex conjugates of
the entries of A; that is, if A = [a;) , then A = [ au ] . The rules for complex conjugation
extend easily to matrices; in particular, we have AB = AB for compatible matrices
A and B.
z=
Proof Suppose that A is an eigenvalue of A with corresponding eigenvector v. Then
- Av = Av, and, taking complex conjugates, we have Av = Av. But then
Av = Av = Av = Av = Av
since A is real. Taking transposes and using the fact that A is symmetric, we have
vTA = vTA T = (Avf = ( Avf = Avr
Therefore,
A ( vrv) = vr( Av) = vr(Av) = ( vrA ) v = ( Avr ) v = A( vrv)
[ � l __ [ � l
or ( A - A) ( vrv) = o.
Now if v =
a,
.
b, i
an + bn i
, then v -
a,
.
b,i
an - bni
, so
402
Chapter 5 Orthogonality
since v * 0 (because it is an eigenvector). We conclude that A - A = 0, or A = A.
Hence, A is real.
Theorem 4.20 showed that, for any square matrix, eigenvectors corresponding
to distinct eigenvalues are linearly independent. For symmetric matrices, something
stronger is true: Such eigenvectors are orthogonal.
Theorem 5 . 1 9
If A is a symmetric matrix, then any two eigenvectors corresponding to distinct
eigenvalues of A are orthogonal.
Let v1 and v2 be eigenvectors corresponding to the distinct eigenvalues
A 1 * A 2 so that Av1 = A 1 v1 and Av2 = A 2v2 . Using A T = A and the fact that x y = xTy
for any two vectors x and y in !R n , we have
Proof
·
(vfA T) Vz
( vfA ) v2 = vf (Av2 )
A 2 ( vfv2 ) = A 2 ( v1 • v2 )
= vi ( A 2v2 )
Hence, (A 1 - A 2 ) (v 1 v2 ) = 0. But A 1 - A 2 * 0, so v1 v2 = 0, as we wished to show.
•
Exa m p l e 5 . 11
•
Verify the result of Theorem 5.19 for
The characteristic polynomial of A is - A 3 + 6A 2 - 9A + 4 = - (A - 4)
2
(A - 1 ) , from which it follows that the eigenvalues of A are A 1 = 4 and A 2 = 1 . The
corresponding eigenspaces are
Solulion
�
·
(Check this.) We easily verify that
from which it follows that every vector in £4 is orthogonal to every vector in £ 1 .
(Why?)
Roman
Note thot
[ - �] [ - i ]
•
� !.
same eigenvalue need not be orthogonal.
Thos, 'igenvedo;s conesponding to the
Section 5.4 Orthogonal Diagonalization of Symmetric Matrices
403
We can now prove the main result of this section. It is called the Spectral Theo­
rem, since the set of eigenvalues of a matrix is sometimes called the spectrum of the
matrix. (Technically, we should call Theorem 5.20 the Real Spectral Theorem, since
there is a corresponding result for matrices with complex entries.)
Theorem 5 . 2 0
The Spectral Theorem
Let A be an n X n real matrix. Then A is symmetric if and only if it is orthogonally
diagonalizable.
We have already proved the "if" part as Theorem 5 . 1 7. To prove the "only if"
implication, we proceed by induction on n. For n = 1 , there is nothing to do, since a
1 X 1 matrix is already in diagonal form. Now assume that every k X k real symmet­
ric matrix with real eigenvalues is orthogonally diagonalizable. Let n = k + 1 and let
A be an n X n real symmetric matrix with real eigenvalues.
Let A 1 be one of the eigenvalues of A and let v 1 be a corresponding eigenvector.
Then v1 is a real vector (why?) and we can assume that v1 is a unit vector, since
otherwise we can normalize it and we will still have an eigenvector corresponding
to A 1 . Using the Gram-Schmidt Process, we can extend v1 to an orthonormal basis
{v1 , v2 , . . . , vn } of !R n . Now we form the matrix
Q l = [v1 Vz vn J
Proof
is a Latinatomswordvibrate,
meaning
"image:'
When
they
emit
l
i
g
ht.
And
when
l
i
g
ht
passes
through
a prism,-ait spreads
out
irainbow
nto a spectrum
band
of
colors.correspond
Vibrationto the
frequencies
eigenvalues
of aascertain
operator
and
are
visible
bright
lines
in the
spectrum
ofl
i
g
ht
that
is
emitted
fromseea prism.
Thus, we ofcantheliter-atom
ally
the
eigenvalues
inson,itsitspectrum,
and forthatthistherea-word
is appropriate
to be applofiaed
tomatrix
the set(orhasofoperator).
allcome
eigenvalues
Spectrum
�
.
.
.
Then Q 1 is orthogonal, and
spectrum
In a lecture he delivered atconsidered
the Universilinear
ty of operators
Gottingenacting
in 1905,onthecertain
German
mathematician
infini
tte-dimensional
vector
spaces.
Out
of
this
lecture
arose
the
notion
of
a
quadratic
form
in
i
n
fini
elmean
y manya
variables,
and
i
t
was
i
n
this
context
that
Hi
l
b
ert
first
used
the
term
to
complete
setmade
of eigmajor
envalcontributions
ues. The spacestoinmany
question
areof mathematics,
now called among them integral
Hilbert
areas
equations,International
number theory,
geometry,
and the foundations
ofHilbert
mathematics.
Inaddress
1900, entitled
at the
Second
Congress
of
Mathematicians
in
Paris,
gave
an
"Thefundamental
Problems ofimportance
Mathematics:'
In theit, hecoming
challenged
mathematicians
to solve 23haveproblems
ofsolved-some
during
century.
Many
of
the
problems
been
were
proved
true,
others
false-and
some
may
never
be
sol
v
ed.
Nevertheless,
mathematical
community and is often regarded as the most
iHilbert'
nfluentis aspeech
l speechenergized
ever giventheabout
mathematics.
David Hilbert ( 1 862- 1 943)
spectrum
Hilbert spaces.
404
Chapter 5 Orthogonality
since vf (A 1 v1 ) = A 1 ( vfv1 ) = A 1 (v1 · v1 ) = A 1 and vf (A 1 v1 ) = A 1 ( vfv1 ) = A 1 (v; · v1 ) = 0
for i of- 1, because {v1 , v2 , , vn } is an orthonormal set.
But
•
.
.
so B is symmetric. Therefore, B has the block form
�
..-...
and A 1 is symmetric. Furthermore, B is similar to A (why?), so the characteristic poly­
nomial of B is equal to the characteristic polynomial of A, by Theorem 4.22. By
Exercise 39 in Section 4.3, the characteristic polynomial of A 1 divides the character­
istic polynomial of A. It follows that the eigenvalues of A 1 are also eigenvalues of A
and, hence, are real. We also see that A 1 has real entries. (Why?) Thus, A 1 is a k X k
real symmetric matrix with real eigenvalues, so the induction hypothesis applies to it.
Hence, there is an orthogonal matrix P2 such that PiA 1 P2 is a diagonal matrix-say,
D 1 . Now let
Then Q2 is an orthogonal (k + l)X(k + 1) matrix, and therefore so is Q = Q 1 Q2 .
Consequently,
Q rAQ = ( Q 1 Q 2 fA( Q 1 Q z ) = ( Q IQ f ) A ( Q 1 Q z ) = Q I( Q fAQ 1 )Q z = Q IB Q 2
2
which is a diagonal matrix. This completes the induction step, and we conclude that,
for all n 1, an n X n real symmetric matrix with real eigenvalues is orthogonally
diagonalizable.
Exa m p l e 5 . 1 8
Orthogonally diagonalize the matrix
This is the matrix from Example 5.17. We have already found that the
eigenspaces of A are
Solution
Section 5.4 Orthogonal Diagonalization of Symmetric Matrices
[ -�] r -il
n l [! ]
405
We need three orthonormal eigenvectors. First, we apply the Gram-Schmidt Process to
and
to obtain
and
The new wctm, which h., been rnmtructed to be orthogoml to
,._...
(why?) and
'°
;, o,thogonal to
[
[:l
[ -�l
;, ''ill ;n E ,
Thu,, we haw th'" mutually mthogonal
]
vectors, and all we need to do is normalize them and construct a matrix Q with these
vectors as its columns. We find that
Q
=
1/ v3 - 1/ v2 - 1 / v6
l/ v3
0
2/ v6
v2
v3
l/
l/
- 1 / v6
and it is straightforward to verify that
Q 'AQ
�
[� � �]
The Spectral Theorem allows us to write a real symmetric matrix A in the form
A = QDQ T, where Q is orthogonal and D is diagonal. The diagonal entries of D
are just the eigenvalues of A, and if the columns of Q are the orthonormal vectors
q1, . . . , qn , then, using the column-row representation of the product, we have
This is called the spectral decomposition of A. Each of the terms A;q; qT is a rank 1
matrix, by Exercise 62 in Section 3.5, and q;qT is actually the matrix of the projec­
tion onto the subspace spanned by q;. (See Exercise 2 5.) For this reason, the spectral
decomposition
is sometimes referred to as the projection form of the Spectral Theorem.
406
Chapter 5 Orthogonality
Exa m p l e 5 . 1 9
Find the spectral decomposition of the matrix A from Example 5 . 1 8.
Solulion
[
From Example 5 . 1 8, we have:
qi =
Therefore,
]
l / VJ
l / VJ ,
l / VJ
q3 =
l / VJ] =
[
1 /3
1 /3
1 /3
[ ]
- 1 / \/6
2/ \/6
- 1 / \/6
]
�]
1 /3 1 /3
1 /3 1 /3
1 /3 1 /3
[��
[
1 /2
-1 2 0
-1 2
1 /2
1 /6 - 1 /3
2/3
- 1 /3
1 /6 - 1 /3
so
=4
[ttt ttt tttl [
+
1 /6
- 1 /3
1 /6
]
! 0
0 0
-! 0
which can be easily verified.
In this example, ,\ 2 = ,\ 3 , so we could combine the last two terms A 2 q2 qi + ,\ 3 q3 qr
to get
The rank 2 matrix q2 qi + q3 qr is the matrix of a projection onto the two-dimensional
subspace (i.e., the plane) spanned by qi and q3 . (See Exercise 26.)
4
Observe that the spectral decomposition expresses a symmetric matrix A explic­
itly in terms of its eigenvalues and eigenvectors. This gives us a way of constructing a
matrix with given eigenvalues and (orthonormal) eigenvectors.
Exa m p l e 5 . 2 0
Find a 2 X 2 matrixwith eigenvalues ,\ 1 = 3 and ,\ 2 = - 2 and corresponding eigenvectors
Section 5.4 Orthogonal Diagonalization of Symmetric Matrices
401
We begin by normalizing the vectors to obtain an orthonormal basis
{q 1 , q2 }, with
Solution
Now, we compute the matrix A whose spectral decomposition is
A = A 1 q 1 q f + A 2 qz qi
[i] [ i]
!1
[ ] [ -�]
[-J �]
-2
[t � ]
=3
!s
= 3 1 2 2516 - 2
25
=
�
I
3. A =
5. A =
7. A =
9. A =
16
25
12
-
25
l
[ -� t
25
It is easy to check that A has the desired properties. (Do this.)
Exercises 5 . 4
Orthogonally diagonalize the matrices in Exercises 1 - 1 0
by finding a n orthogonal matrix Q and a diagonal
matrix D such that Q TAQ = D.
I. A =
25
-
[� :J
[ V21 V2]
[� : :]
u : -�]
O
2. A =
4. A =
6. A =
[-1 -1]
[9 ]
[ � ! �]
[� � ;]
3
3
-2
-2
6
[ � � � ! ] [ � � � �:
8. A =
I 0. A =
1 1 . If b * 0 , orthogonally diagonalize A =
12. If b oF 0, orthogonally diogonolire A =
[ ab
]
[� � n
b
.
a
13. Let A and B be orthogonally diagonalizable n X n
matrices and let c be a scalar. Use the Spectral
Theorem to prove that the following matrices are
orthogonally diagonalizable:
(c) A 2
(a) A + B
(b) cA
14. If A is an invertible matrix that is orthogonally diago­
nalizable, show that A- 1 is orthogonally diagonalizable.
15. If A and B are orthogonally diagonalizable and AB =
BA, show that AB is orthogonally diagonalizable.
16. If A is a symmetric matrix, show that every eigenvalue
of A is nonnegative if and only if A = B 2 for some
symmetric matrix B.
In Exercises 1 7-20, find a spectral decomposition of the
matrix in the given exercise.
408
Chapter 5 Orthogonality
17. Exercise 1
19. Exercise 5
18. Exercise 2
20. Exercise 8
In Exercises 21 and 22, find a symmetric 2 X 2 matrix with
eigenvalues A 1 and A 2 and corresponding orthogonal
eigenvectors v1 and v2 •
2 1 . A 1 = - l , A 2 = 2 , v1 =
22. A 1 = 3, A 2 = - 3, v1 =
[�] [_�]
[�] [ -�]
, v2 =
·
, v2 =
In Exercises 23 and 24, find a symmetric 3 X 3 matrix with
eigenvalues A 1 , A 2 , and ,\ 3 and corresponding orthogonal
eigenvectors v1 , v2 , and v3 .
23 A, � 1 , A, � 2, A, � 3, v, �
25. Let q be a unit vector in !R n and let W be the subspace
spanned by q. Show that the orthogonal projection of a
vector v onto W (as defined in Sections 1 .2 and 5.2) is
given by
proj w ( v) = ( qq T) v
and that the matrix of this projection is thus qq T.
[Hint: Remember that, for x and y in !R n , x y = xTy.]
26. Let {q 1 , . . . , qd be an orthonormal set of vectors in !R n
and let W be the subspace spanned by this set.
(a) Show that the matrix of the orthogonal projection
onto W is given by
p = q 1 qf + . . . + qkql
[}, [ - : ] .
�
(b) Show that the projection matrix P in part (a) is
symmetric and satisfies P 2 = P.
(c) Let Q = [ q 1 · · · qk ] be the n X k matrix whose
columns are the orthonormal basis vectors of W.
Show that P = QQ T and deduce that rank(P) = k.
27. Let A be an n X n real matrix, all of whose eigenvalues
are real. Prove that there exist an orthogonal matrix Q
and an upper triangular matrix T such that Q TAQ = T.
This very useful result is known as Schur's Triangular­
ization Theorem. [Hint: Adapt the proof of the Spec­
tral Theorem.]
28. Let A be a nilpotent matrix (see Exercise 56 in Sec­
tion 4.2). Prove that there is an orthogonal matrix Q
such that Q T AQ is upper triangular with zeros on its
diagonal. [Hint: Use Exercise 27.]
A p p l icati o n s
Quad ratic Forms
An expression of the form
ax 2 + by 2 + cxy
is called a quadratic form in x and y. Similarly,
ax 2 + by 2 + cz 2 + dxy + exz + fyz
is a quadratic form in x, y, and z. In words, a quadratic form is a sum of terms, each of
which has total degree two in the variables. Therefore, 5 x 2 - 3y 2 + 2xy is a quadratic
form, but x 2 + y 2 + x is not.
We can represent quadratic forms using matrices as follows:
ax 2 + by 2 + cxy = [x y]
[ ; � J [;J
c 2
c 2
Section 5.5 Applications
and
ax 2 + by 2 + cz 2 + dxy + exz + fyz
�
=
[�
d/2 e/2
[x y z] d 2 b f/2
e/2 f/2 c
409
][ ]
x
y
z
(Verify these.) Each has the form xTAx, where the matrix A is symmetric. This obser­
vation leads us to the following general definition.
D e fi n it i o n
form
A
quadratic form in
n variables is a function f : !R n
---+
IR of the
where A is a symmetric n X n matrix and x is in !R n . We refer to A as the matrix
associated with f
Exa m p l e 5 . 2 1
What is the quadratic form with associated matrix A
Solution
If x
=
[:J
=
[
]
2 -3
?
-3 5
then
Observe that the off-diagonal entries a1 2 a 2 1 -3 of A are combined to give
the coefficient - 6 of x1x2 • This is true generally. We can expand a quadratic form in
n variables xTAx as follows:
=
xTAx
=
=
a 11 x� + a 22x� + · · · + a nnx� + 2: 2a ;jxi xj
i<j
Thus, if i * j, the coefficient of X; Xj is 2a ij .
Exa m p l e 5 . 2 2
Find the matrix associated with the quadratic form
f (x 1 , x 2 , x3 )
=
2x � - xi + 5xf + 6x1 x 2 - 3x 1 x3
The coefficients of the squared terms x;2 go on the diagonal as a;;, and the
coefficients of the cross-product terms x;xj are split between a ij and aji· This gives
Solution
3
-1
0
-�]
410
Chapter 5 Orthogonality
so
f(x ,. x,, x, )
�
[x, x, x, J
as you can easily check.
[ _� - ! -�J[ ::l
In the case of a quadratic form f(x, y) in two variables, the graph of z = j(x, y) is
a surface in IR 3 . Some examples are shown in Figure 5 . 1 2 .
Observe that the effect o f holding x o r y constant i s t o take a cross section of
the graph parallel to the yz or xz planes, respectively. For the graphs in Figure 5 . 1 2,
all of these cross sections are easy to identify. For example, in Figure 5 . 1 2 (a), the
cross sections we get by holding x or y constant are all parabolas opening upward,
so f(x, y) 0 for all values of x and y. In Figure 5 . 1 2 (c), holding x constant gives
parabolas opening downward and holding y constant gives parabolas opening upward,
producing a saddle point.
2
z
z
x
x
(a) z = 2x2 + 3y2
y
y
(b)
z=
- 2x2 - 3y 2
z
z
x
y
(c) z = 2x2 - 3y2
Graphs of quadratic forms f
Figure 5 . 1 2
(x, y)
(d)
z = 2x2
Section 5. 5 Applications
411
What makes this type of analysis quite easy is the fact that these quadratic forms
have no cross-product terms. The matrix associated with such a quadratic form is a
diagonal matrix. For example,
2x2 - 3y2 = [x y ]
[ � - � J [;J
In general, the matrix of a quadratic form is a symmetric matrix, and we saw in Sec­
tion 5.4 that such matrices can always be diagonalized. We will now use this fact to
show that, for every quadratic form, we can eliminate the cross-product terms by
means of a suitable change of variable.
Let f(x) = xTAx be a quadratic form in n variables, with A a symmetric n X n
matrix. By the Spectral Theorem, there is an orthogonal matrix Q that diagonalizes A;
that is, Q TAQ = D, where D is a diagonal matrix displaying the eigenvalues of A. We
now set
X = Qy Or, equivalently, y = Q - 1 X = Q TX
Substitution into the quadratic form yields
xTAx =
=
=
(Q yfA(Q y)
yTQ TAQy
yTDy
which is a quadratic form without cross-product terms, since D is diagonal. Further­
more, if the eigenvalues of A are A1, . . . , A n , then Q can be chosen so that
If y = [y l
becomes
Yn ] T, then, with respect to these new variables, the quadratic form
yTDy = A 1Y12 + · · . + A ny;
This process is called diagonalizing a quadratic form. We have just proved the fol­
lowing theorem, known as the Principal Axes Theorem. (The reason for this name
will become clear in the next subsection.)
Theorem 5 . 2 1
The Principal Axes Theorem
Every quadratic form can be diagonalized. Specifically, if A is the n X n symmet­
ric matrix associated with the quadratic form xTAx and if Q is an orthogonal
matrix such that Q TAQ = D is a diagonal matrix, then the change of variable
x = Qy transforms the quadratic form xTAx into the quadratic form yTDy,
which has no cross-product terms. If the eigenvalues of A are A1,
, A n and
y = [y 1
Yn f , then
•
xTAx = yTDy = A 1y � + · · · + A ny;
.
.
412
Chapter 5 Orthogonality
Exa m p l e 5 . 2 3
5
Find a change o f variable that transforms the quadratic form
f (x1 , x2 ) = x f + 4x1 x2 + 2x�
into one with no cross-product terms.
Solulion
The matrix off is
A=
6 2/Vs1.
l /Vs]
[ - 2/Vs
- [ l /Vs]
6
/Vs
o
l
]
[ 2/Vs
[
2/Vs
1]
/Vs
l [:: ] [;:]
[ � �] [;:] 6
with eigenvalues A1 = and A 2 =
(Check this.) If we set
Corresponding unit eigenvectors are
and
qI
�
[� �]
q2
=
and D =
Q=
then Q TAQ = D. The change of variable x = Qy, where
x=
converts f into
0
and y =
f (y) = f(y 1 , yi } = [ Y1 Y2 l
= y f + Yi
The original quadratic form xTAx and the new one yTDy (referred to in the Princi­
pal Axes Theorem) are equal in the following sense. In Example 5.23, suppose we want
[ - � ].
- 1 5 - 1 - 1 2 11
1
/Vs
/Vs
l
l
]
]
]
[]
[
[
[ 2/Vs
2/Vs
/Vs
/Vs
l 6 6( 1/Vs)2 /Vs)2 55/5 11
to evaluate f( x) = xTAx at x =
j(
, 3) = (
We have
)2 + 4(
) ( 3) + ( 3)2 =
In terms of the new variables,
so
Y 1 = y = Q TX =
y2
f(yl , yi } = y f + y� =
3
+ (-7
=
=
-7
=
exactly as before.
The Principal Axes Theorem has some interesting and important consequences.
We will consider two of these. The first relates to the possible values that a quadratic
form can take on.
2.1.
5.
D e fi n i t i o n
A quadratic form f(x) = xTAx is classified as one of the following:
positive de.finite iff( x) > 0 for all x -=fa 0
positive semidefinite iff( x) ::=:: 0 for all x
3. negative de.finite ifj(x) < 0 for all x -=fa 0
4. negative semidefinite iff( x) :s 0 for all x
indefinite ifj(x) takes on both positive and negative values
Section 5. 5 Applications
413
A symmetric matrix A is called positive definite, positive semidefinite, nega­
tive definite, negative semidefinite, or indefinite if the associated quadratic form
f( x) = xTAx has the corresponding property.
The quadratic forms in parts (a), (b ), (c), and (d) of Figure 5 . 1 2 are positive definite,
negative definite, indefinite, and positive semidefinite, respectively. The Principal Axes
Theorem makes it easy to tell if a quadratic form has one of these properties.
Theorem 5 . 2 2
Let A be an n X n symmetric matrix. The quadratic form f( x) = xTAx is
a.
b.
c.
d.
e.
positive definite if and only if all of the eigenvalues of A are positive.
positive semidefinite if and only if all of the eigenvalues of A are nonnegative.
negative definite if and only if all of the eigenvalues of A are negative.
negative semidefinite if and only if all of the eigenvalues of A are non positive.
indefinite if and only if A has both positive and negative eigenvalues.
You are asked to prove Theorem 5.22 in Exercise 27.
Exa m p l e 5 . 2 4
Classify f(x, y, z) = 3.x2 + 3y2 + 3z2 - 2xy - 2xz - 2yz as positive definite, negative
definite, indefinite, or none of these.
Solution
-�
-[ � = �
The matrix associated with f is
]
3
-1 -1
which has eigenvalues 1 , 4, and 4. (Verify this.) Since all of these eigenvalues are posi­
tive,f is a positive definite quadratic form.
If a quadratic form f( x) = xTAx is positive definite, then, since f(O) = 0, the
minimum value of f(x) is 0 and it occurs at the origin. Similarly, a negative definite
quadratic form has a maximum at the origin. Thus, Theorem 5.22 allows us to solve
certain types of maxima/minima problems easily, without resorting to calculus. A type
of problem that falls into this category is the constrained optimization problem.
It is often important to know the maximum or minimum values of a quadratic
form subject to certain constraints. (Such problems arise not only in mathematics
but also in statistics, physics, engineering, and economics.) We will be interested in
finding the extreme values of f( x) = xTAx subject to the constraint that I x ii = 1 .
In the case of a quadratic form in two variables, we can visualize what the problem
means. The graph of z = f(x, y) is a surface in IR 3 , and the constraint ll x ll = 1 restricts
the point (x, y) to the unit circle in the xy-plane. Thus, we are considering those
points that lie simultaneously on the surface and on the unit cylinder perpendicular
to the xy plane. These points form a curve lying on the surface, and we want the high­
est and lowest points on this curve. Figure 5 . 1 3 shows this situation for the quadratic
form and corresponding surface in Figure 5 . 1 2 (c).
414
Chapter 5 Orthogonality
z
y
The
intersection of 1
cylinder
Figure 5 . 1 3
x2 + y2
=
z =
2x2 - 3y2
with the
In this case, the maximum and minimum values off(x, y) = 2x 2 - 3y 2 (the high­
est and lowest points on the curve of intersection) are 2 and - 3, respectively, which
are just the eigenvalues of the associated matrix. Theorem 5.23 shows that this is
always the case.
Theorem 5 . 2 3
Let f(x) = xTAx be a quadratic form with associated n X n symmetric matrix A.
Let the eigenvalues of A be A 1 :::::: A 2 :::::: :::::: A w Then the following are true, subject
to the constraint II xii = 1 :
· · ·
a . A i 2: f(x) 2: A n
b. The maximum value of f(x) is A 1 , and it occurs when x is a unit eigenvector
corresponding to A 1 .
c. The minimum value of f(x) is A n , and it occurs when x is a unit eigenvector
corresponding to Aw
As usual, we begin by orthogonally diagonalizing A. Accordingly, let Q be an
orthogonal matrix such that Q TAQ is the diagonal matrix
Proof
Then, by the Principal Axes Theorem, the change of variable x = Qy gives xTAx =
yTDy. Now note that y = Q Tx implies that
since Q T = Q - 1 . Hence, using x x = xTx, we see that llrll = Wr = � =
ll x ll = 1 . Thus, if x is a unit vector, so is the corresponding y, and the values of xTAx
and yTDy are the same.
·
Section 5. 5 Applications
415
(a) To prove property (a), we observe that if y = [y 1 · · · Yn ] T, then
f ( x) = xTAx = yTD y
= A1Yl + A zyi + · · · + A ny;
:::; A 1y f + A 1Yi + · · · + A 1y;
= A 1 (y f + Yi + · · · + y; )
= A 1 llrll 2
= A1
Thus, f(x) s A 1 for all x such that ll x ll = The proof that f(x) 2: A n is similar.
(See Exercise 37.)
(b) If q 1 is a unit eigenvector corresponding to A 1 , then Aq 1 = A 1 q 1 and
1.
f ( q 1 ) = q fAq 1 = q fA 1 Q 1 = A 1 ( q fq 1 ) = A 1
This shows that the quadratic form actually takes on the value A 1 , and so, by prop­
erty (a), it is the maximum value off(x) and it occurs when x = q 1 .
(c) You are asked to prove this property in Exercise 38.
Exa m p l e 5 . 2 5
2xi
1,
Find the maximum and minimum values of the quadratic form f(x 1 , x 2 ) = 5x i +
4x 1 x2 +
subject to the constraint xi + xi = and determine values of x 1 and x 2
for which each of these occurs.
1,
In Example 5.23, we found thatf has the associated eigenvalues A 1 = 6 and
with corresponding unit eigenvectors
Solution
A2 =
-_ [ 2/Vs
1/Vs]
1 1/Vs
Qi
1/Vs]
-_ [ - 2/Vs
2/Vs 1/Vs.
- 2/Vs.
and Qz
Therefore, the maximum value off is 6 when x 1 =
and x2 =
The mini­
mum value off is when x 1 =
and x2 =
(Observe that these extreme
values occur twice-in opposite directions-since - q 1 and - q2 are also unit eigen­
vectors for A 1 and A 2 , respectively.)
Graphing Quad ratic Equations
The general form of a quadratic equation in two variables x and y is
ax 2 + by 2 + cxy + dx + ey + f = 0
where at least one of a, b, and c is nonzero. The graphs of such quadratic equations are
called conic sections (or conics), since they can be obtained by taking cross sections
of a (double) cone (i.e., slicing it with a plane) . The most important of the conic sec­
tions are the ellipses (with circles as a special case), hyperbolas, and parabolas. These
are called the nondegenerate conics. Figure 5 . 1 4 shows how they arise.
It is also possible for a cross section of a cone to result in a single point, a straight
line, or a pair of lines. These are called degenerate conics. (See Exercises 59-64.)
The graph of a nondegenerate conic is said to be in standard position relative to
the coordinate axes if its equation can be expressed in one of the forms in Figure 5 . 1 5.
416
Chapter 5 Orthogonality
Ellipse
Circle
Parabola
Hyperbola
The nondegenerate conics
Figure 5 . 1 4
x
2
Ellipse or Circle: 2 +
a
y2
2
b
= l; a , b > 0
y
y
y
b
a
-b
-a
b
-b
a<b
a>b
a = b
Hyperbola
y
y
b
-b
x2
a2
_
.i_ - 1 , a, b > 0
b2
_
r2
b
_
x2
- 1 , a, b > 0
a2
Parabola
y
y
y
y
2
y = ax , a > 0
2
y = ax , a < 0
x = ay 2 , a > 0
x = ay 2 , a < 0
Nondegenerate conics in standard position
Figure 5 . 1 5
Section 5. 5 Applications
Exa m p l e 5 . 2 6
411
If possible, write each of the following quadratic equations in the form of a conic in
standard position and identify the resulting graph.
(a) 4x 2 + 9y 2 = 36
(c) 4x 2 - 9y = 0
(b) 4x 2 - 9y 2 + 1 = 0
Solulion
(a) The equation 4x 2 + 9y 2 = 36 can be written in the form
x2 y z
+
9
4
-
-
=
1
so its graph is an ellipse intersecting the x-axis at ( ± 3, 0) and the y-axis at (O, ± 2) .
(b) Th e equation 4x2 - 9y2 + 1 = 0 can b e written i n the form
y2 x2
1- 1 = 1
9
4
so its graph is a hyperbola, opening up and down, intersecting the y-axis at (O, ± t) .
(c) Th e equation 4x2 - 9y = 0 can b e written i n the form
4
y = -x 2
9
so its graph is a parabola opening upward.
If a quadratic equation contains too many terms to be written in one of the forms
in Figure 5 . 1 5, then its graph is not in standard position. When there are additional
terms but no xy term, the graph of the conic has been translated out of standard
position.
Exa m p l e 5 . 2 1
Identify and graph the conic whose equation is
x 2 + 2y 2 - 6x + Sy + 9
Solulion
=
0
We begin by grouping the x and y terms separately to get
( x 2 - 6x) + ( 2y 2 + Sy) - 9
=
or
( x 2 - 6x) + 2 (y 2 + 4y)
=
-9
Next, we complete the squares on the two expressions in parentheses to obtain
or
( x 2 - 6x + 9 ) + 2 (y 2 + 4y + 4 )
=
-9 + 9 + S
( x - 3 ) 2 + 2 (y + 2 ) 2
=
S
We now make the substitutions x'
tion into
( x' ) 2 + 2 (y' ) 2
=
x - 3 and y'
=
S
or
=
(x ' ) 2
--
s
y + 2, turning the above equa­
+
(y ' ) 2
4
--
= I
418
Chapter 5 Orthogonality
This is the equation of an ellipse in standard position in the x 'y ' coordinate system,
intersecting the x' -axis at ( ± 2 \/2, O) and the y'-axis at (O, ± 2) . The origin in the x'y'
coordinate system is at x = 3, y = - 2, so the ellipse has been translated out of stan­
dard position 3 units to the right and 2 units down. Its graph is shown in Figure 5 . 1 6.
y'
y
2
x
-2
-2
-
x
'
4
translated ellipse
Figure 5 . 1 6
A
If a quadratic equation contains a cross-product term, then it represents a conic
that has been rotated.
Exa m p l e 5 . 2 8
Identify and graph the conic whose equation is
5x 2 + 4xy + 2y 2 = 6
The left-hand side of the equation is a quadratic form, so we can write it in
matrix form as xTAx = 6, where
Solulion
A=
[� �]
In Example 5.23, we found that the eigenvalues of A are 6 and 1 , and a matrix Q that
orthogonally diagonalizes A is
Q=
.......
[
l / Vs
2/ Vs
l / Vs - 2 / Vs
]
Observe that <let Q = - 1 . In this example, we will interchange the columns o f this
matrix to make the determinant equal to + 1 . Then Q will be the matrix of a rotation,
by Exercise 28 in Section 5. 1 . It is always possible to rearrange the columns of an
orthogonal matrix Q to make its determinant equal to + 1 . (Why?) We set
Q=
instead, so that
[
l / Vs 2/ Vs
- 2/ Vs l / Vs
]
Section 5. 5 Applications
419
The change ofvariable x = Qx' converts the given equation into the form (x' ) TDx' = 6
by means of a rotation. If x' =
[; : ]
x'
, then this equation is just
(x') 2 + 6(y') 2 = 6 or
y
(x') 2
+ (y') 2 = 1
6
which represents an ellipse in the y' coordinate system.
3
To graph this ellipse, we need to know which vectors play the roles of e; =
[x' �]
and e� =
of the
in the new coordinate system. (These two vectors locate the positions
and y' axes.) But, from x = Qx' , we have
Qe; =
and
1
Qez =
x'
A rotated ellipse
Figure 5 . 1 1
[�]
l /Vs 2/Vs
l
]
[ - 2/Vs
[
l /Vs ]
l /Vs 2/Vs
[ - 2/Vs
l /Vs] [ Q ]
0
1
=
l /Vs]
[ - 2/Vs
[ 2/Vs
l /Vs]
These are just the columns q 1 and q2 o f Q , which are the eigenvectors o f A ! The fact
that these are orthonormal vectors agrees perfectly with the fact that the change of
variable is just a rotation. The graph is shown in Figure 5 . 1 7.
4
You can now see why the Principal Axes Theorem is so named. If a real symmet­
ric matrix A arises as the coefficient matrix of a quadratic equation, the eigenvectors
of A give the directions of the principal axes of the corresponding graph.
It is possible for the graph of a conic to be both rotated and translated out of stan­
dard position, as illustrated in Example 5.29.
Exa m p l e 5 . 2 9
5x 4xy 2 - -x
Vs28 - Vs4 4
4
[ � � ] [ - � -�]
5.28
/Vs 2/Vs
l- 2/Vs
]
[
/Vs
l
5.28,
x'
l /Vs 2/Vs
]
[ - Vs28 - Vs4 ] [ - 2/Vs
[
l /Vs ] - 4x' - 12y'
Identify and graph the conic whose equation is
2+
+ y2
-
y+ =0
Solution
The strategy is to eliminate the cross-product term first. In matrix form,
the equation is xTAx + Bx + = 0, where
and B =
A=
The cross-product term comes from the quadratic form xTAx, which we diagonalize
as in Example
by setting x = Qx' , where
Q=
Then, as in Example
xTAx = ( x' fDx' = (x') 2 + 6(y') 2
But now we also have
B x = BQx =
I
y'
=
420
Chapter 5 Orthogonality
y
\
2
-2
y'
-1
y"
x
-1
-2
�
-4
x
'
x
"
Figure 5 . 1 8
Thus, in terms of x' and y', the given equation becomes
(x') 2 + 6(y ') 2 - 4x' - 12y ' + 4 = 0
To bring the conic represented by this equation into standard position, we need
to translate the x'y' axes. We do so by completing the squares, as in Example 5.27.
We have
or
((x') 2 - 4x' + 4) + 6((y ') 2 - 2y ' + 1 ) = -4 + 4 + 6 = 6
(x' - 2) 2 + 6(y ' - 1) 2 = 6
This gives us the translation equations
x " = x' - 2 and y " = y ' In the x "y " coordinate system, the equation is simply
(x " ) 2 + 6(y ") 2 = 6
4
which is the equation of an ellipse (as in Example 5.28). We can sketch this ellipse by
first rotating and then translating. The resulting graph is shown in Figure 5.18.
The general form of a quadratic equation in three variables x, y, and z is
ax 2 + by 2 + cz 2 + dxy + exz + fyz + gx + hy + iz + j = 0
where at least one of a, b, . . . , f is nonzero. The graph of such a quadratic equation
is called a quadric surface (or quadric) . Once again, to recognize a quadric we need
x2 y2 z2
c
z
Ellipsoid: :::2 + b2 + :::2
a
=
Section 5. 5 Applications
x--,,2 y2 z2
c
z
Hyperboloid of one sheet:
1
a "'
+ b2
-
:::2
421
=
I
y
x
x2 -+--ty22 -z22
c
Hyperboloid of two sheets : -;;,
z
y
x
Elliptic paraboloid:
z
x
Quadric surfaces
Figure 5 . 1 9
=
z x-;;,2 yb22
=
-
Elliptic cone:
I
z
z2 x-;;,2 y2
y
x
Hyperbolic paraboloid:
+
z
y
+ b2
=
z x2 y2
=
-;;, - p
422
Chapter 5 Orthogonality
t o put it into standard position. Some quadrics i n standard position are shown in
Figure 5.19; others are obtained by permuting the variables.
Exa m p l e 5 . 3 0
Identify the quadric surface whose equation is
5x 2 + l ly 2 + 2z 2 + 16xy + 20xz - 4yz
Solulion
]
=
36
The equation can be written in matrix form as xTAx = 36, where
8 10
1 1 -2
-2 2
We find the eigenvalues of A to be 18, 9, and - 9, with corresponding orthogonal
eigenvectors
respectively. We normalize them to obtain
and form the orthogonal matrix
=
[ ! ! =l l
=
-
Note that in order for Q to be the matrix of a rotation, we require <let Q 1 , which
is true in this case. (Otherwise, <let Q - 1 , and swapping two columns changes the
sign of the determinant.) Therefore,
and, with the change of variable x = Qx', we get xTAx = (x' )Dx' = 36, so
18(x') 2 + 9(y') 2 - 9(z') 2
=
(x') 2 (y ') 2
+
36 or
2
4
--
--
-
(z') 2
4
--
=
1
From Figure 5.19, we recognize this equation as the equation of a hyperboloid of one
sheet. The x', y', and z' axes are in the directions of the eigenvectors q 1, q2 , and q3 ,
respectively. The graph is shown in Figure 5.20.
Section 5. 5 Applications
423
z
z'
Anonstandard
hyperboloidposition
of one sheet in
Figure 5.20
I
We can also identify and graph quadrics that have been translated out of standard
position using the "complete-the-squares method" of Examples 5.27 and 5.29. You
will be asked to do so in the exercises.
Exercises 5 . 5
Q u a d ratic F o r m s
In Exercises 1 -6, evaluate the quadratic form f(x) = xTAx
for the given A and x.
1. A =
2. A =
3. A =
4. A =
5. A =
6. A =
[ � ! l [;]
[ � � l [::]
[ _ � ! l- [ ! ]
u -} � [�]
}
�
:]
[
u
[ : :Jx � m
x=
x=
-
0
2
0
2
2
0
x=
In Exercises 7-12, find the symmetric matrix A associated
with the given quadratic form.
7. xl + 2xf + 6x 1 x2
8. X1 X2
9. 3x 2 - 3xy - y2
10. x � - x � + 8X 1 X2 - 6XzX3
1 1 . 5xl - xi + 2x� + 2x 1 x2 - 4x 1 x3 + 4x2x3
12. 2x 2 3y 2 + z 2 - 4xz
-
Diagonalize the quadratic forms in Exercises 13-18 by
finding an orthogonal matrix Q such that the change of
variable x = Qy transforms the given form into one with no
cross-product terms. Give Q and the new quadratic form.
13. 2x l + 5x� 4x 1 x2
14. x 2 + 8xy + y 2
15. 7xl + x i + xj + 8X 1 X2 + 8X 1 X3 - 16XzX3
16. x l + x i + 3xj - 4x 1 x2
17. x 2 + z 2 - 2xy + 2yz
18. 2xy + 2xz + 2yz
-
424
Chapter 5 Orthogonality
Classify each of the quadratic forms in Exercises 1 9-26 as
positive definite, positive semidefinite, negative definite,
negative semidefinite, or indefinite.
20. x f + xi - 2x, x2
19. x f + 2x}
2
2
21. - 2x - 2y + 2xy 22. x 2 + y 2 + 4xy
23. 2x f + 2x i + 2x� + 2x1 x2 + 2x1 x3 + 2x2x3
24. x f + x i + x� + 2x1 x3 25. x i + x � - x � + 4x 1 x2
26. -x 2 - y 2 - z 2 - 2xy - 2xz - 2yz
27. Prove Theorem 5.22.
28. Let A =
[ � �]
be a symmetric 2 X 2 matrix. Prove
that A is positive definite if and only if a > 0 and
<let A > 0. [Hint: ax 2 + 2bxy + dy 2 =
( � y (d - �2 )y 2.]
a x+ y
+
29. Let B be an invertible matrix. Show that A = B TB is
positive definite.
30. Let A be a positive definite symmetric matrix. Show
that there exists an invertible matrix B such that A =
B TB . [Hint: Use the Spectral Theorem to write A =
QDQ T. Then show that D can be factored as c Tc for
some invertible matrix C.]
31. Let A and B be positive definite symmetric n X n
matrices and let c be a positive scalar. Show that the
following matrices are positive definite.
(a) cA
(b) A 2
(c) A + B
_
,
(d) A (First show that A is necessarily invertible.)
32. Let A be a positive definite symmetric matrix. Show
that there is a positive definite symmetric matrix B
such that A = B 2 • (Such a matrix B is called a square
root of A.)
In Exercises 33-36, find the maximum and minimum val­
ues of the quadratic form f(x) in the given exercise, subject
to the constraint ll x ll = 1, and determine the values of x for
which these occur.
33. Exercise 20
34. Exercise 22
35. Exercise 23
36. Exercise 24
37. Finish the proof of Theorem 5.23 (a) .
38. Prove Theorem 5.23 ( c).
G r a p h i n g Q u a d ratic E q u a t i o n s
In Exercises 39-44, identify the graph of the given equation.
39. x 2 + 5y 2 = 25
40. x 2 - y 2 - 4 = 0
41. x 2 - y - 1 = 0
42. 2x 2 + y 2 - 8 = 0
44. x = - 2y 2
43. 3x 2 = y 2 - 1
In Exercises 45-50, use a translation of axes to put the conic
in standard position. Identify the graph, give its equation in
the translated coordinate system, and sketch the curve.
45. x 2 + y 2 - 4x - 4y + 4 = 0
46. 4x 2 + 2y 2 - Sx + l2y + 6 = 0
47. 9x 2 - 4y 2 - 4y = 37 48. x 2 + lOx - 3y = - 13
49. 2y 2 + 4x + Sy = 0
50. 2y 2 - 3x 2 - 1 8x - 20y + 1 1 = 0
In Exercises 51 -54, use a rotation of axes to put the conic in
standard position. Identify the graph, give its equation in the
rotated coordinate system, and sketch the curve.
51. x 2 + xy + y 2 = 6 52. 4x 2 + lOxy + 4y 2 = 9
53. 4x 2 + 6xy - 4y 2 = 5 54. 3x 2 - 2xy + 3y 2 = 8
In Exercises 55-58, identify the conic with the given equa­
tion and give its equation in standard form.
55. 3x 2 - 4xy + 3y 2 - 28 v'2x + 22 Vly + 84 = 0
56. 6x 2 - 4xy + 9y 2 - 20x - lOy - 5 = 0
57. 2xy + 2 \/2x - 1 = 0
58. x 2 - 2xy + y 2 + 4 V2x - 4 = 0
Sometimes the graph of a quadratic equation is a straight
line, a pair of straight lines, or a single point. We refer to
such a graph as a degenerate conic. It is also possible that
the equation is not satisfied for any values of the variables,
in which case there is no graph at all and we refer to the
conic as an imaginary conic. In Exercises 59-64, identify
the conic with the given equation as either degenerate or
imaginary and, where possible, sketch the graph.
59. x 2 - y 2 = 0
60. x 2 + 2y 2 + 2 = 0
62. x 2 + 2xy + y 2 = 0
61. 3x 2 + y 2 = 0
63. x 2 - 2xy + y 2 + 2 Vlx - 2 Vly = 0
64. 2x 2 + 2xy + 2y 2 + 2 \/2x - 2 Vly + 6 = 0
65. Let A be a symmetric 2 X 2 matrix and let k be a
scalar. Prove that the graph of the quadratic equation
xTA x = k is
(a) a hyperbola if k * 0 and <let A < 0
(b) an ellipse, circle, or imaginary conic if k *
0 and
det A > 0
(c) a pair of straight lines or an imaginary conic if
k * 0 and <let A = 0
(d) a pair of straight lines or a single point if k = 0
and det A * 0
(e) a straight line if k = 0 and <let A = 0
[Hint: Use the Principal Axes Theorem.]
Chapter Review
In Exercises 66-73, identify the quadric with the given
equation and give its equation in standard form.
66. 4x 2 + 4y 2 + 4z 2 + 4xy + 4xz + 4yz = 8
67. x 2 + y 2 + z 2 - 4yz = 1
68. -x 2 - y 2 - z 2 + 4xy + 4xz + 4yz = 12
69. 2xy + z = 0
70. 16x 2 + 100y 2 + 9 z 2 - 24xz - 60x - 80z = 0
71. x 2 + y 2 - 2z 2 + 4xy - 2xz + 2yz - x + y + z = 0
72. 10x 2 + 25y 2 + 10z 2 - 40xz + 20 \/2x + soy +
20 \/2z = 1 5
73. l lx 2 + l ly 2 + 14z 2 + 2xy + 8xz - 8yz - 12x +
12y + 12z = 6
425
74. Let A be a real 2 X 2 matrix with complex eigenvalues
A = a :±: bi such that b =F 0 and I A I = 1 . Prove that
every trajectory of the dynamical system xk + i = Axk
lies on an ellipse. [Hint: Theorem 4.43 shows that if v
is an eigenvector corresponding to A = a - bi, then
the matrix P = [ Re v Im v] is invertible and
- -1
A=P
P . Set B = (PPT) - 1 . Show that the
[: �J
quadratic xTBx = k defines an ellipse for all k > 0,
and prove that if x lies on this ellipse, so does Ax.]
Chapter Review
Kev Defi nitions and Concepts
fundamental subspaces
of a matrix, 380
Gram-Schmidt Process, 389
orthogonal basis, 370
orthogonal complement
of a subspace, 378
orthogonal matrix, 374
orthogonal projection, 382
orthogonal set of vectors, 369
Orthogonal Decomposition
Theorem, 384
orthogonally diagonalizable
matrix, 400
orthonormal basis, 372
Review Questions
1 . Mark each of the following statements true or false:
(a) Every orthonormal set of vectors is linearly
independent.
(b) Every nonzero subspace of u;g n has an orthogonal
basis.
(c) If A is a square matrix with orthonormal rows,
(d)
(e)
(f)
(g)
(h)
(i)
then A is an orthogonal matrix.
Every orthogonal matrix is invertible.
If A is a matrix with det A = 1 , then A is an
orthogonal matrix.
If A is an m X n matrix such that (row(A))_j_ = u;g n ,
then A must be the zero matrix.
If W is a subspace of u;g n and v is a vector in u;g n such
that projw(v) = 0, then v must be the zero vector.
If A is a symmetric, orthogonal matrix, then A 2 = I.
Every orthogonally diagonalizable matrix is invertible.
orthonormal set of vectors, 372
properties of orthogonal
matrices, 374-376
QR factorization, 393
Rank Theorem, 386
spectral decomposition, 405
Spectral Theorem, 403
(j) Given any n real numbers A 1 , , A n , there exists
a symmetric n X n matrix with A 1 , , A n as its
eigenvalues.
2. Find all values of a and b such that
.
•
.
.
\ [H [ J [fl )
.
i< an mthogonal <et of vedm.
3. Find the coordinate vector [ v] of v =
respect to the orthogonal basis
B
•
8
[ -�]
� ml [ J [ - � l) or n'
2
with
Chapter 5 Orthogonality
426
4. The coordinate vector of a vector v with respect to an
-3
orthonormal basis B = {v1 , v2 } of lR 2 is [v] 8 =
.
1/2
3/5
If v1 =
, find all possible vectors v.
4/ 5
[ ]
[ ]
[ - �j�
[ � :]
5. Show that
�
]
- [ ! J · � - [ iJ· � - [ rJ
15. ( a) Apply the Gram-Schmidt Process to
2
is an
4/7 Vs - 1 5/7 Vs 2/7 Vs
orthogonal matrix.
6. If
��
2 7
with respect to
1 2
�
is an orthogonal matrix, find all possible
values of a, b, and c.
7. If Q is an orthogonal n X n matrix and {v1 , , vk } is
an orthonormal set in !R n , prove that { Q v1 , , Q vk }
is an orthonormal set.
8. If Q is an n X n matrix such that the angles
L ( Q x, Q y) and L ( x, y) are equal for all vectors x and
y in IR " , prove that Q is an orthogonal matrix.
•
.
•
•
to find an orthogonal basis for W = span{x 1 , x2 , xJ.
(b) Use the result of part (a) to find a QR factorization
•
In Questions 9-12, find a basis for W _j_ .
9. W is the line in IR 2 with general equation
2x - Sy = 0
10. W is the line in IR 3 with parametric equations
x=t
y = 2t
z = -t
A- [ ! � ]
m [J
- ml· - }
A [ � �l
A. A.
of
.
16. Find an orthogonal basis for IR 4 that contains the
vecto"
'" d
17. Find an orthogonal basis for the subspace
w
18. Let
+ x, + x, + x.
=
2
-1 1
0 om'
-
·
2
( a) Orthogonally diagonalize
(b) Give the spectral decomposition of
A [-�
-� -�
� -�]
13. Find bases for each of the four fundamental subspaces of
=
1
2
3 -5
4
8
6 -1
14. Find the orthogonal decomposition of
v
- [ -�]
9
7
19. Find a symmetric matrix with eigenvalues ,\ 1 = ,\ 2 = 1,
A 3 = - 2 and eigenspaces
20. If {v1 , v2 ,
•
prove that
ues c 1 , c2 ,
•
A
.
•
.
•
, v" } is an orthonormal basis for !R n and
is a symmetric matrix with eigenval­
, e n and corresponding eigenvectors
Vector Spaces
6.0
Algebra is generous; she often gives
more than is asked of her.
-Jean le Rond d'Alembert
In Carl B. Boyer
Wiley,
( 1 7 1 7- 1 783)
A History of Mathematics
1 968, p. 48 1
I n t ro d u ctio n : Fib o n acci i n (Vecto r> S p a c e
The Fibonacci sequence was introduced in Section 4.6. It i s the sequence
0, 1, 1, 2, 3, 5, 8, 13, . . .
of nonnegative integers with the property that after the first two terms, each term
is the sum of the two terms preceding it. Thus 0 + 1 = 1, 1 + 1 = 2, 1 + 2 = 3,
2 + 3 = 5, and so on.
If we denote the terms of the Fibonacci sequence by f0,f1,f2 ,
, then the entire
sequence is completely determined by specifying that
Jo = O,f1 = 1 and fn = fn - l + fn - l for n 2
By analogy with vector notation, let's write a sequence x0, x1, x2 , x3 , . . . as
x = [xo , X1, X2 , X3 , . . . )
The Fibonacci sequence then becomes
f = [f0, J1, J2 ,f3 ,
) = [O, 1, 1, 2, . . . )
We now generalize this notion.
.
2
•
.
•
•
.
2
A Fibonacci-type sequence is any sequence x [x0, x1, x 2 , x 3 ,
such that x0 and x 1 are real numbers and Xn = xn - l + Xn - z for n 2.
D e fi n it i o n
=
•
.
•
)
For example, [1, \/2, 1 + \/2, 1 + 2 \/2, 2 + 3 \/2, . . . ) is a Fibonacci-type sequence.
Write down the first five terms of three more Fibonacci-type sequences.
By analogy with vectors again, let's define the sum of two sequences x = [x0, x1,
x2 , ) and y = [y0, y 1, y2 , ) to be the sequence
Problem 1
•
.
.
•
.
•
If c is a scalar, we can likewise define the scalar multiple of a sequence by
ex
= [ cx0 , cx 1 , cx2 ,
•
•
•
)
421
428
Chapter Vector Spaces
6
(a) Using your examples from Problem 1 or other examples, compute
the sums of various pairs of Fibonacci-type sequences. Do the re­
sulting sequences appear to be Fibonacci-type?
(b) Compute various scalar multiples of your Fibonacci-type se­
quences from Problem 1. Do the resulting sequences appear to be
Fibonacci-type?
Problem 3 (a) Prove that if x and y are Fibonacci-type sequences, then so is x + y.
(b) Prove that if x is a Fibonacci-type sequence and c is a scalar, then
ex is also a Fibonacci-type sequence.
Let's denote the set of all Fibonacci-type sequences by Fib. Problem 3 shows that,
like 11;r, Fib is closed under addition and scalar multiplication. The next exercises
show that Fib has much more in common with 11;r.
Problem 4 Review the algebraic properties of vectors in Theorem 1 . 1 . Does Fib
satisfy all of these properties? What Fibonacci-type sequence plays the role of O? For a
Fibonacci-type sequence x, what is - x? Is - x also a Fibonacci-type sequence?
n
Problem 5 In !R , we have the standard basis vectors e1, e 2 ,
, e n - The Fibonacci
sequence f = [O, 1, 1, 2, . . . ) can be thought of as the analogue of e2 because its first
two terms are 0 and 1 . What sequence e in Fib plays the role of e1?
What about e3 , e4, ? Do these vectors have analogues in Fib?
Problem 6 Let x = [ x0 , x1, x2 ,
) be a Fibonacci-type sequence. Show that x is a
linear combination of e and f.
Problem 1 Show that e and f are linearly independent. (That is, show that if
ce + df = 0, then c = d = 0.)
Problem 8 Given your answers to Problems 6 and 7, what would be a sensible
value to assign to the "dimension" of Fib ? Why?
Problem 9 Are there any geometric sequences in Fib? That is, if
2
[ 1 , r, r , r 3 , . . . )
is a Fibonacci-type sequence, what are the possible values of r?
Problem 10 Find a "basis" for Fib consisting of geometric Fibonacci-type sequences.
Problem 11 Using your answer to Problem 10, give an alternative derivation of
Binet's formula [formula ( 5) in Section 4.6] :
Problem 2
•
.
•
.
•
•
.
•
.
)
(
(
)
1_ 1 Vs n
1 1 + Vs n
=
n Vs
2
Vs
2
for the terms of the Fibonacci sequence f = [f0 , f1 , f2 , ) . [Hint: Express f in terms of
j,
The
LucasLucas
sequence
is named after
Edouard
(see page
336).
_...
__
_
_
-
.
•
.
the basis from Problem 10.]
The Lucas sequence is the Fibonacci-type sequence
l = [ 10 , 11 , 12 , 13 , ) = [2, 1, 3, 4, . . . )
Problem 12 Use the basis from Problem 1 0 to find an analogue of Binet's formula
for the nth term Zn of the Lucas sequence.
Problem 13 Prove that the Fibonacci and Lucas sequences are related by the identity
fn - 1 + fn + I = Zn for n 2: 1
[Hint: The Fibonacci-type sequences C = [ 1 , 1, 2, 3, . . . ) and f + = [ 1 , 0, 1, 1, . . . )
form a basis for Fib. (Why?) ]
I n this Introduction, we have seen that the collection Fib o f all Fibonacci-type
sequences behaves in many respects like IR 2 , even though the "vectors" are actually
infinite sequences. This useful analogy leads to the general notion of a vector space
that is the subject of this chapter.
.
•
.
Section Vector Spaces and Subspaces
6. 1
429
Vecto r S p a ces a n d S u b s p aces
I n Chapters 1 and 3, we saw that the algebra o f vectors and the algebra o f matrices
are similar in many respects. In particular, we can add both vectors and matrices,
and we can multiply both by scalars. The properties that result from these two opera­
tions (Theorem 1 . 1 and Theorem 3.2) are identical in both settings. In this section,
we use these properties to define generalized "vectors" that arise in a wide variety
of examples. By proving general theorems about these "vectors:' we will therefore
simultaneously be proving results about all of these examples. This is the real power
of algebra: its ability to take properties from a concrete setting, like !R n , and abstract
them into a general setting.
Let V be a set on which two operations, called addition and scalar
multiplication, have been defined. If u and v are in V, the sum of u and v is denoted
by u + v, and if c is a scalar, the scalar multiple of u by c is denoted by cu. If the
following axioms hold for all u, v, and w in V and for all scalars c and d, then V is
D e fi n it i o n
The GermanGrassmann
mathematician
Hermann
is generallythecredited
wiath
fivector
rst introducing
i
d
ea
of
(alitnhough heUnfor­
did
nottunatelcallyspace
i, thisthat)work
wasnotveryreceidivfefi­
cult
to
read
and
did
thepersonattentiwhoondidit deserved.
Onethe
study
it
was
Italian
mathematicianInGiuseppe
Peanobook
his
Peano clarified
Grassmann'
sthe
earlier
work
and
laid
down
axioms
a vector
spacePeano'as s
webook
knowis foralsothem
today.
remarkable
forsets.
introducing
operations
on
His
notations
andand (for"is
"union;'
"intersection;'
anstilelement
of") aretheythewere
ones notwe
use,
al
t
hough
immediatel
y accepted
bys axio­
other
mathematicians.
Peano'
matic
definition
ofe ianvector
space
also
had
very
l
i
t
tl
fl
u
ence
for
many
years.
Acceptance
came
in afterrepeated
Hermannit Wey!
hisintroduction
book to Einstein'singeneralan
theory of relativity.
( 1 809-
1 877)
1 844.
1 888
called a vector space and its elements are called vectors.
Closure under addition
Commutativityvity
Associati
Closure under scalar multiplication
Distributi
Distributivviittyy
1 . u + v is in V.
2. u + v = v + u
3. (u + v) + w = u + (v + w)
4. There exists an element 0 in V, called a zero vector, such that u + 0 = u.
5. For each u in V, there is an element - u in V such that u + ( - u) = 0.
6. cu is in V.
7. C ( U + V) =
CU + CV
8. (c + d)u = cu + du
9. c (du) = (cd) u
10. lu = u
( 1 858- 1 932) .
Calcolo Geometrico,
U, n,
E
1 9 1 8,
( 1 88 5 - 1 955)
Space, Time, Matter,
R e m a rks
By "scalars" we will usually mean the real numbers. Accordingly, we should
refer to V as a real vector space (or a vector space over the real numbers). It is also pos­
sible for scalars to be complex numbers or to belong to "ll.P , where p is prime. In these
cases, V is called a complex vector space or a vector space over "ll.P , respectively. Most of
our examples will be real vector spaces, so we will usually omit the adjective "real:' If
something is referred to as a "vector space:' assume that we are working over the real
number system.
In fact, the scalars can be chosen from any number system in which, roughly
speaking, we can add, subtract, multiply, and divide according to the usual laws of
arithmetic. In abstract algebra, such a number system is called a field.
•
The definition of a vector space does not specify what the set V consists
of. Neither does it specify what the operations called "addition" and "scalar multi­
plication'' look like. Often, they will be familiar, but they need not be. See Example 6.6
and Exercises 5-7.
•
We will now look at several examples of vector spaces. In each case, we need to
specify the set V and the operations of addition and scalar multiplication and to verify
axioms 1 through 10. We need to pay particular attention to axioms 1 and 6 (closure),
430
Chapter Vector Spaces
6
axiom 4 (the existence of a zero vector in V), and axiom 5 (each vector in V must have
a negative in V).
Exa m p l e 6 . 1
For any n 2: 1 , !R n is a vector space with the usual operations of addition and scalar
multiplication. Axioms 1 and 6 follow from the definitions of these operations, and
the remaining axioms follow from Theorem 1 . 1 .
Exa m p l e 6 . 2
The set of all 2 X 3 matrices is a vector space with the usual operations of matrix
addition and matrix scalar multiplication. Here the "vectors" are actually matrices.
We know that the sum of two 2 X 3 matrices is also a 2 X 3 matrix and that multiply­
ing a 2 X 3 matrix by a scalar gives another 2 X 3 matrix; hence, we have closure.
The remaining axioms follow from Theorem 3.2. In particular, the zero vector 0 is the
2 X 3 zero matrix, and the negative of a 2 X 3 matrix A is just the 2 X 3 matrix -A.
There is nothing special about 2 X 3 matrices. For any positive integers m and n ,
the set of all m X n matrices forms a vector space with the usual operations of matrix
addition and matrix scalar multiplication. This vector space is denoted Mm n ·
Exa m p l e 6 . 3
Let <;IP 2 denote the set of all polynomials of degree 2 or less with real coefficients.
Define addition and scalar multiplication in the usual way. (See Appendix D.) If
p(x)
=
a 0 + a 1 x + a 2x 2 and q(x)
=
b0 + b 1 x + b2x 2
are in <!P 2 , then
has degree at most 2 and so is in <!P 2 • If c is a scalar, then
cp(x)
=
ca0 + ca 1 x + ca 2x 2
is also in <!P 2 . This verifies axioms 1 and 6.
The zero vector 0 is the zero polynomial-that is, the polynomial all of whose
coefficients are zero. The negative of a polynomial p(x) = a0 + a1x + a 2x 2 is the poly­
nomial -p ( x) = - a0 - a1x - a 2x 2 • It is now easy to verify the remaining axioms. We
will check axiom 2 and leave the others for Exercise 12. With p(x) and q(x) as above,
we have
p (x) + q(x)
=
=
=
=
=
(a0 + a 1 x + a 2x 2 ) + (b0 + b 1 x + b2x 2 )
(a0 + b0) + (a 1 + b 1 )x + (a 2 + b2 )x2
( b0 + a0) + (b 1 + a 1 )x + (b2 + a 2 )x2
( b0 + b 1 x + b2x 2 ) + (a0 + a 1 x + a 2x 2 )
q(x) + p(x)
where the third equality follows from the fact that addition of real numbers is
commutative.
Section Vector Spaces and Subspaces
6. 1
431
In general, for any fixed n 2: 0, the set <!J' n of all polynomials of degree less than or
equal to n is a vector space, as is the set <!J' of all polynomials.
Exa m p l e 6 . 4
Let ?Jf denote the set of all real-valued functions defined on the real line. Iff and g are
two such functions and c is a scalar, then f + g and cf are defined by
(f + g) (x)
=
f (x) + g(x) and (cj) (x)
=
cf (x)
In other words, the value off + g at x is obtained by adding together the values off
and g at x [Figure 6.1 (a) J . Similarly, the value of cf at x is just the value off at x mul­
tiplied by the scalar c [Figure 6. 1 (b) J . The zero vector in ?Jf is the constant function
f0 that is identically zero; that is, f0 (x) 0 for all x. The negative of a functionf is the
function -f defined by ( -j) (x) -f(x) [Figure 6.l (c) ] .
Axioms 1 and 6 are obviously true. Verification o f the remaining axioms is left as
Exercise 1 3 . Thus, ?Jf is a vector space.
=
=
y
y
(x, 2/(x))
\
I
(x, f(x) + g (x))
I
2f
(x, - 3/(x))
(b)
(a)
y
/-f(x))
-f
(x,
(c)
The graphs of (a) f, and f (b) f, 2f, and -3f, and (c) f and -f
Figure 6 . 1
g,
+ g,
432
Chapter Vector Spaces
6
In Example 6.4, we could also have considered only those functions defined on
some closed interval [a, b] of the real line. This approach also produces a vector space,
denoted by ?f [a, b] .
Exa m p l e 6 . 5
Exa m p l e 6 . 6
The set "1l_ of integers with the usual operations is not a vector space. To demon­
strate this, it is enough to find that one of the ten axioms fails and to give a specific
instance in which it fails (a counterexample). In this case, we find that we do not have
closure under scalar multiplication. For example, the multiple of the integer 2 by the
scalar t is (t)(2) = t , which is not an integer. Thus, it is not true that ex is in "1l_ for
every x in "1l_ and every scalar c (i.e., axiom 6 fails).
Let V = IR 2 with the usual definition of addition but the following definition of scalar
multiplication:
Then, for example,
so axiom 10 fails. [In fact, the other nine axioms are all true (check this), but we do
not need to look into them, because V has already failed to be a vector space. This
example shows the value of looking ahead, rather than working through the list of
axioms in the order in which they have been given.]
Exa m p l e 6 . 1
Let C 2 denote the set of all ordered pairs of complex numbers. Define addition and
scalar multiplication as in IR 2 , except here the scalars are complex numbers. For
example,
] [ ] [ - 2-+31� i ]
[ 2l -+ 3ii ] = [ ((ll -- i)i)((2l -+ 3i)i) ] = [ - 1 2- Si ]
( 1 - i)
[
and
- 3 + 2i
1+i
+
.
4
2 - 31
6
Using properties of the complex numbers, it is straightforward to check that all ten
axioms hold. Therefore, C 2 is a complex vector space.
In general, e n is a complex vector space for all n
Exa m p l e 6 . 8
2 1.
4
If p is prime, the set z; (with the usual definitions of addition and multiplication by
scalars from "ll_p ) is a vector space over ZP for all n 2: 1 .
4
Section Vector Spaces and Subspaces
6. 1
433
Before we consider further examples, we state a theorem that contains some use­
ful properties of vector spaces. It is important to note that, by proving this theorem
for vector spaces in general, we are actually proving it for every specific vector space.
Theorem 6 . 1
Let V be a vector space, u a vector in V, and c a scalar.
a.
b.
c.
d.
Ou = 0
co = 0
( - l)u = -u
I f cu = 0, then c = 0 o r u = 0.
Proof We prove properties (b) and ( d) and leave the proofs of the remaining proper­
ties as exercises.
(b) We have
cO = c(O + 0) = cO + cO
by vector space axioms 4 and 7. Adding the negative of cO to both sides produces
co + ( -co) = (co + co) + ( -co)
which implies
by axioms and 3
by axiom
by axiom
5
o = co + (co + ( -co))
5
= co + 0
4
= co
(d) Suppose cu = 0. To show that either c = 0 or u = 0, let's assume that c * 0. (If
c = 0, there is nothing to prove.) Then, since c * 0, its reciprocal l /c is defined, and
u = lu
1
-( cu)
c
1
-0
c
0
by axiom
10
by axiom
9
by property (b)
We will write u - v for u + ( -v), thereby defining subtraction of vectors. We will
also exploit the associativity property of addition to unambiguously write u + v + w
for the sum of three vectors and, more generally,
for a linear combination of vectors.
Subspaces
We have seen that, in !R n , it is possible for one vector space to sit inside another one,
giving rise to the notion of a subspace. For example, a plane through the origin is a
subspace of IR 3 . We now extend this concept to general vector spaces.
434
Chapter Vector Spaces
6
A subset W of a vector space V is called a subspace of V if W is
itself a vector space with the same scalars, addition, and scalar multiplication as V.
D e fi n i t i o n
As in !R n , checking to see whether a subset W of a vector space V is a subspace of
V involves testing only two of the ten vector space axioms. We prove this observation
as a theorem.
Theorem 6 . 2
Let V be a vector space and let W be a nonempty subset of V. Then W is a sub­
space of V if and only if the following conditions hold:
a. If u and v are in W, then u + v is in W.
b. If u is in W and c is a scalar, then cu is in W.
Assume that W is a subspace of V. Then W satisfies vector space axioms 1 to
10. In particular, axiom 1 is condition (a) and axiom 6 is condition (b) .
Conversely, assume that W is a subset of a vector space V, satisfying condi­
tions (a) and (b) . By hypothesis, axioms 1 and 6 hold. Axioms 2, 3, 7, 8, 9, and 10 hold
in W because they are true for all vectors in V and thus are true in particular for those
vectors in W. (We say that W inherits these properties from V.) This leaves axioms 4
and 5 to be checked.
Since W is nonempty, it contains at least one vector u. Then condition (b) and
Theorem 6. l (a) imply that Ou = 0 is also in W. This is axiom 4.
If u is in V, then, by taking c = - 1 in condition (b ), we have that -u = ( - l)u is
also in W, using Theorem 6.l (c) .
Proof
Remark
Since Theorem 6.2 generalizes the notion of a subspace from the con­
text of !R n to general vector spaces, all of the subspaces of !R n that we encountered in
Chapter 3 are subspaces of !R n in the current context. In particular, lines and planes
through the origin are subspaces of IR 3 .
Exa m p l e 6 . 9
Exa m p l e 6 . 1 0
We have already shown that the set <!/' n of all polynomials with degree at most n is a
vector space. Hence, <!/' n is a subspace of the vector space <!/' of all polynomials.
4
Let W be the set of symmetric n X n matrices. Show that W is a subspace of Mnn -
Clearly, W is nonempty, so we need only check conditions (a) and (b) in
Theorem 6.2. Let A and B be in W and let c be a scalar. Then A T = A and B T = B,
from which it follows that
Solulion
Therefore, A + B is symmetric and, hence, is in W. Similarly,
(cAf = cA T = cA
so cA is symmetric and, thus, is in W. We have shown that W is closed under addition
and scalar multiplication. Therefore, it is a subspace of Mnn ' by Theorem 6.2.
4
Section Vector Spaces and Subspaces
6. 1
Exa m p l e 6 . 1 1
435
Let C(;S be the set of all continuous real-valued functions defined on IR and let 0J be
the set of all differentiable real-valued functions defined on IR. Show that C(;S and 0J are
subspaces of �, the vector space of all real-valued functions defined on IR.
From calculus, we know that iff and g are continuous functions and c is a
scalar, then f + g and cf are also continuous. Hence, C(;S is closed under addition and
scalar multiplication and so is a subspace of �. Iff and g are differentiable, then so are
f + g and cf Indeed,
Solution
(j + g) '
=
f ' + g ' and ( cf) '
=
c (j')
So 0J is also closed under addition and scalar multiplication, making it a subspace
of �.
...+
It is a theorem of calculus that every differentiable function is continuous. Conse­
quently, 0J is contained in C(;S (denoted by 0J C C(;S), making 0J a subspace ofC(;S . It is also
the case that every polynomial function is differentiable, so rt/' C 0J , and thus rt/' is a
subspace of 0J . We therefore have a hierarchy of subspaces of �, one inside the other:
This hierarchy is depicted in Figure 6.2.
The hierarchy of subspaces of�
Figure 6 . 2
There are other subspaces of � that can be placed into this hierarchy. Some of
these are explored in the exercises.
In the preceding discussion, we could have restricted our attention to functions
defined on a closed interval [a, b] . Then the corresponding subspaces of � [a, b]
would be C(;S [a, b] , 0J [a, b] , and rt/' [a, b] .
Exa m p l e 6 . 1 2
Let S be the set of all functions that satisfy the differential equation
f" + f 0
=
That is, S is the solution set of Equation ( 1 ). Show that S is a subspace of �-
(1)
436
Chapter Vector Spaces
6
Solulion
S is nonempty, since the zero function clearly satisfies Equation ( 1 ). Let
f and g be in S and let c be a scalar. Then
(f + g) " + (f + g) = (j" + g " ) + (f + g)
= (f" + f) + (g " + g)
=O+O
=O
which shows that f + g is in S. Similarly,
(cf)" + cf = cf" + cf
= c (f " + f)
= co
=O
so cf is also in S.
Therefore, S is closed under addition and scalar multiplication and is a subspace
of ?F.
The differential Equation ( 1) is an example of a homogeneous linear differential
equation. The solution sets of such equations are always subspaces of ?F. Note that in
Example 6.12 we did not actually solve Equation ( 1) (i.e., we did not find any specific
solutions, other than the zero function) . We will discuss techniques for finding solu­
tions to this type of equation in Section 6.7.
As you gain experience working with vector spaces and subspaces, you will notice
that certain examples tend to resemble one another. For example, consider the vector
spaces IR 4 , <!/' 3 , and M22 • Typical elements of these vector spaces are, respectively,
u�
Indejathevuwords
all overof Yogi
again:'Berra, "It's
Exa m p l e 6 . 1 3
m
p (x) � a + bx + ex' + dx', and A �
[: !]
Any calculations involving the vector space operations of addition and scalar multi­
plication are essentially the same in all three settings. To highlight the similarities, in
the next example we will perform the necessary steps in the three vector spaces side
by side.
(a) Show that the set W of all vectors of the form
is a subspace of IR 4 •
(b) Show that the set W of all polynomials of the form a + bx - bx 2 + ax 3 is a
subspace of <!!' 3 .
( c) Show that the set W of all matrices of the form
[ _ : �] is a subspace of M22 .
Section Vector Spaces and Subspaces
6. 1
a=b= b=
uv
Solution
(a) W is nonempty because it con­
tains the zero vector 0. (Take
0.) Let and be in W-say,
a= b=
AA = B � �] a
[
B = [ � �]
A+B= [ -a(b++ ba ++ d]
(b) W is nonempty because it con­
tains the zero polynomial. (Take
O.) Letp (x) and q(x) be in W-say,
p ( x)
ax 3
and
q ( x)
[
ab ++ l
u+v= -ab+
[ ba ++ l
- a(b++
u+v k
Then
Then
p ( x)
c
d
-d
c
= a + bx -bx2 +
=+ +
+ = +(a(+b +
-+ ((ba ++
dx - dx2
c
q ( x)
(c) W is nonempty because it con­
tains the zero matrix 0. (Take
0. ) Let and be in W-say,
cx 3
and
c)
p ( x)
so ku is in W.
Thus, W is a nonempty subset of
IR 4 that is closed under addition and
scalar multiplication. Therefore, W is
a subspace of iR 4 , by Theorem 6.2.
d)
+ k
A+B k
= ka + kbx - kbx2 + kax3 kA = [ -kakb kakb]
kA
so p(x) q(x) is also in W (because it
has the right form).
Similarly, if is a scalar, then
k
_
c
c ) x3
so
is also in W (because it has
the right form).
Similarly, if is a scalar, then
_
Then
d) x
d ) x2
c
d
d)
c
431
c
so
is also in W (because it has
the right form).
Similarly, if is a scalar, then
so
is in W.
Thus, W is a nonempty subset of
M22 that is closed under addition and
scalar multiplication. Therefore, W is
a subspace of M22 , by Theorem 6.
so kp (x) is in W.
Thus, W is a nonempty subset of
<;IP 3 that is closed under addition and
scalar multiplication. Therefore, W is
a subspace of <;if 3 by Theorem 6.2.
4
Example 6. 1 3 shows that it is often possible to relate examples that, on the surface,
appear to have nothing in common. Consequently, we can apply our knowledge of
!R n to polynomials, matrices, and other examples. We will encounter this idea several
times in this chapter and will make it precise in Section 6.5.
Exa m p l e 6 . 1 4
+=
=
If V is a vector space, then V is clearly a subspace of itself. The set {O}, consisting of
only the zero vector, is also a subspace of V, called the zero subspace. To show this, we
simply note that the two closure conditions of Theorem 6.2 are satisfied:
0
0
0 and c 0
0 for any scalar c
The subspaces {O} and V are called the trivial subspaces of V.
438
Chapter Vector Spaces
6
An examination of the proof of Theorem 6.2 reveals the following useful fact:
If W is a subspace of a vector space V, then W contains the zero vector 0 of V.
This fact is consistent with, and analogous to, the fact that lines and planes are sub­
spaces of !R 3 if and only if they contain the origin. The requirement that every subspace
must contain 0 is sometimes useful in showing that a set is not a subspace.
Exa m p l e 6 . 1 5
Let W be the set of all 2 X 2 matrices of the form
Is W a subspace of M22 ?
Solulion
Each matrix in W has the property that its ( 1 , 2) entry is one more than its
( 1 , 1) entry. Since the zero matrix
0
=
[� �]
does not have this property, it i s not in W. Hence, W is not a subspace o f M22 •
Exa m p l e 6 . 1 6
Let W be the set of all 2 X 2 matrices with determinant equal to 0. Is W a subspace
of M22 ? (Since <let 0 0, the zero matrix is in W, so the method of Example 6. 15 is
of no use to us.)
=
Solulion
Let
A
Then <let A
=
<let B
=
=
[ � �]
and B
=
[� �]
0, so A and B are in W. But
A +B
=
[� � ]
so det (A + B) = 1 =fa 0, and therefore A + B is not in W. Thus, W is not closed under
addition and so is not a subspace of M22 .
Spanning sets
The notion of a spanning set of vectors carries over easily from !R n to general vector
spaces.
D e fi n it i o n If S {v1 , v2 , . . • , vd is a set of vectors in a vector space V, then
the set of all linear combinations of v1 , v2 , . . . , vk is called the span of v1 , v2 , . . . , vk
and is denoted by span (v1 , v2 , . . . , vk) or span ( S ) . If V = span (S ), then S is called
a spanning set for V and V is said to be spanned by S.
=
Section Vector Spaces and Subspaces
6. 1
Exa m p l e 6 . 1 1
439
Show that the polynomials 1, x, and x 2 span l!J' 2 •
By its very definition, a polynomial p (x) = a + bx + cx2 is a linear combi­
nation of 1, x, and x2 • Therefore, l!J' 2 = span ( l , x, x 2 ) .
Solution
Example 6. 1 7 can clearly b e generalized t o show that l!J' n = span ( l , x, x2 , . . . ,
n
x ) . However, no finite set of polynomials can possibly span l!J', the vector space of
all polynomials. (See Exercise 44 in Section 6.2.) But, if we allow a spanning set
to be infinite, then clearly the set of all nonnegative powers of x will do. That is,
l!J' = span ( l , x, x2 , • • • ) .
Exa m p l e 6 . 1 8
Show that M23 = span (E 1 1 , E 1 2 , E 1 3 , E2 1 , E22 , E23 ), where
El l =
E2 1 =
[�
[�
0
0
0
0
�]
�]
E1 2 =
E22 =
[�
[�
0
0
�]
�]
E1 3 =
E23 =
[� �]
[ � �]
0
0
0
0
(That is, E;j is the matrix with a 1 in row i, column j and zeros elsewhere.)
Solution
We need only observe that
Extending this example, we see that, in general, Mm n is spanned by the mn matri­
ces Eij , where i = 1 , . . . , m and j = 1, . . . , n.
Exa m p l e 6 . 1 9
In l!J' 2 , determine whether r (x) = 1 - 4x + 6x 2 is in span(p(x), q(x)), where
p ( x) = 1 - x + x 2 and q ( x) = 2 + x - 3x 2
We are looking for scalars c and d such that cp(x) + dq(x) = r (x) . This
means that
c ( l - x + x 2 ) + d ( 2 + x - 3x 2 ) = 1 - 4x + 6x 2
Solution
Regrouping according powers of x, we have
( c + 2 d) + ( - c + d) x + ( c - 3d) x 2 = 1 - 4x + 6x 2
Equating the coefficients of like powers of x gives
c + 2d = 1
- c + d = -4
c - 3d = 6
440
Chapter Vector Spaces
6
which is easily solved to give c = 3 and d =
r(x) is in span(p(x), q(x)). (Check this.)
Exa m p l e 6 . 2 0
-
1 Therefore, r(x) = 3p(x) - q(x), so
.
In ?F, determine whether sin 2x is in span(sin x, cos x) .
We set c sin x + d cos x = sin 2x and try to determine c and d so that this
equation is true. Since these are functions, the equation must be true for all values of
x. Setting x = 0, we have
S o lution
c sin O + d cos O = sin O or c ( O ) + d ( l ) = O
from which we see that d = 0. Setting x = 1T /2, we get
c sin ( 7r/2 ) + d cos ( 'TT / 2 ) = sin ( 7T ) or c ( l ) + d ( O ) = 0
giving c = 0. But this implies that sin 2x = O (sin x) + O(cos x) = 0 for all x, which
is absurd, since sin 2x is not the zero function. We conclude that sin 2x is not in
span (sin x, cos x).
Remark
It is true that sin 2x can be written in terms of sin x and cos x. For
example, we have the double angle formula sin 2x = 2 sin x cos x. However, this is not
a linear combination.
Exa m p l e 6 . 2 1
In M22 , describe the span of A =
Solution
[ � � l B = [ � � l and C = [ � �] .
Every linear combination of A, B, and C is of the form
[� �] + d [ � � ] + e[ � �]
c+d c+e
]
= [
c+ e
d
cA + dB + eC = c
This matrix is symmetric, so span (A, B, C) is contained within the subspace of sym­
metric 2 X 2 matrices. In fact, we have equality; that is, every symmetric 2 X 2 matrix is
[; �] be a symmetric 2
[yx y ] = [ cc ++ de c +d e ]
in span(A, B, C). To show this, we let
X 2 matrix. Setting
z
and solving for c and d, we find that c = x - z, d = z, and e = - x + y + z. Therefore,
�
(Check this.) It follows that span(A, B, C) is the subspace of symmetric 2 X 2 matrices .
.+
Section Vector Spaces and Subspaces
6. 1
441
As was the case in !R n , the span of a set of vectors is always a subspace of the vector
space that contains them. The next theorem makes this result precise. It generalizes
Theorem 3.19.
Theorem 6 . 3
Let V i, v2 ,
•
.
.
, vk b e vectors in a vector space V.
a. span (vi, v2 ,
b. span (vi, v2 ,
•
.
•
.
•
.
, vk) is a subspace of V.
, vk) is the smallest subspace of V that contains Vi, v2 ,
•
.
.
, vk.
(a) The proof of property (a) is identical to the proof of Theorem 3 . 1 9, with
Proof
!R n replaced by V.
(b) To establish property (b ), we need to show that any subspace of V that contains
Vi, v2 , , vk also contains span( vi, v2 , , vk) · Accordingly, let W be a subspace of V
that contains Vi, v2 , . . . , vk. Then, since W is closed under addition and scalar multi­
plication, it contains every linear combination Ci Vi + c 2v2 + + ckvk of Vi, v2 ,
,
vk. Therefore, span (vi, v2 , , vk) is contained in W.
.
•
.
•
•
•
· · ·
.
1
•
•
.
.
.
Exercises 6 . 1
In Exercises 1 - 1 1, determine whether the given set, together
with the specified operations of addition and scalar multi­
plication, is a vector space. If it is not, list all of the axioms
that fail to hold.
1 . The set of all vectors in IR 2 of the form
[: l with the
usual vector addition and scalar multiplication
2. The set of all vectors
[;] in IR2 with x 2: 0, y 2: 0 (i.e.,
the first quadrant), with the usual vector addition and
scalar multiplication
3. The set of all vectors
[�] in IR2 with xy 2: 0 (i.e., the
union of the first and third quadrants), with the usual
vector addition and scalar multiplication
4 . The set of all vectors
[; ]
in IR 2 with x 2: y, with the
usual vector addition and scalar multiplication
5. IR 2 , with the usual addition but scalar multiplication
defined by
6. IR 2 , with the usual scalar multiplication but addition
defined by
[] [] [
x,
x
+ 2
Y1
Yz
=
l]
x , + x2 +
Yi + Yz + 1
7. The set of all positive real numbers, with addition EB
defined by x EB y = xy and scalar multiplication 0
defined by c 0 x = xc
8. The set of all rational numbers, with the usual addition
and multiplication
9. The set of all upper triangular 2 X 2 matrices, with the
usual matrix addition and scalar multiplication
[: �l
10. The set of all 2 X 2 matrices of the form
where ad = 0, with the usual matrix addition and
scalar multiplication
1 1 . The set of all skew-symmetric n X n matrices, with the
usual matrix addition and scalar multiplication
(see page 162).
12. Finish verifying that <!f 2 is a vector space
(see Example 6.3).
13. Finish verifying that ge is a vector space
(see Example 6.4) .
442
E:v
Chapter Vector Spaces
6
In Exercises 1 4- 1 7, determine whether the given set, together
with the specified operations of addition and scalar multi­
plication, is a complex vector space. If it is not, list all of
the axioms that fail to hold.
14. The set of all vectors in C 2 of the form
[ �] , with the
usual vector addition and scalar multiplication
15. The set Mm n ( C) of all m X n complex matrices, with
the usual matrix addition and scalar multiplication
16. The set C 2 , with the usual vector addition but scalar
30. V = Mnn ' W = {A in Mnn : <let A = 1 }
31. V = Mnn ' W is the set of diagonal n X n matrices
32. V = Mnn > W is the set of idempotent n X n matrices
33. V = Mnn ' W = {A in Mnn : AB = BA}, where B is a
given (fixed) matrix
34. V = C!P 2 , W = {bx + cx 2 }
35. V = C!P 2 , W = {a + bx + cx 2 : a + b + c = O}
36. V = C!P 2 , W = {a + bx + cx 2 : abc = O}
z
z
multiplication defined by c 1 = � 1
37. V = C!P, W is the set of all polynomials of degree 3
CZ2
Z2
17. !R n , with the usual vector addition and scalar multiplication 38. V = ':ffe , W = {f in '2F : f ( -x) = f(x)}
39. V = ':ffe , W = {f in '2F :f( -x) = -f(x)}
In Exercises 1 8-21, determine whether the given set, together 40. V = ':ffe , w = {f in '2F :f(O) = l }
with the specified operations of addition and scalar. multipli. ';'Y .· f( ) -_ }
. is. a ector space �ver th e in. d"zca ted lL · If "t is no t, l"is t ·tt;. 41. v _- '!:Y, w _- {f m
catzon,
p
�
.
.
Jlh 42. V _- '!:Y, W is. the set of all mtegrable
funct10ns
all of the axzoms that fail to hold.
CZl! :f'(x) 2:
for all x}
18. The set of all vectors in "11._ � with an even number of Jlh 43. V = CZJJ , W = {f in
l
(2
l s, over "ll 2 with the usual vector addition and scalar � 44. V = ':ffe , W = ct; , the set of all functions with
[] [ ]
77
a;-;
u.::
1
a;-;
0 0
0
continuous second derivatives
multiplication
Jlh 45. V = ':ffe , W = {f in '2F : = lim f (x) = 00}
19. The set of all vectors in "11._ � with an odd number of
X->0
l s, over "ll 2 with the usual vector addition and scalar
46. Let V be a vector space with subspaces U and W. Prove
multiplication
that U n W is a subspace of V.
20. The set Mm n ( "llp ) of all m X n matrices with entries
47. Let V be a vector space with subspaces U and W. Give
from "11._P ' over "11._P with the usual matrix addition and
an example with v = [R 2 to show that U U W need not
scalar multiplication
be a subspace of V.
21. "ll 6 , over "ll 3 with the usual addition and multiplication
48. Let V be a vector space with subspaces U and W.
(Think this one through carefully!)
Define the sum of U and W to be
22. Prove Theorem 6.l (a) .
U + W = { u + w : u is in U, w is in W}
23. Prove Theorem 6.l (c) .
(a) If V = IR 3 , U is the x-axis, and W is the y-axis,
what is U + W ?
In Exercises 24-45, use Theorem 6.2 to determine whether
(b) If U and W are subspaces o f a vector space V,
W is a subspace of V
prove that U + W is a subspace of V.
49. If U and V are vector spaces, define the Cartesian
product of U and V to be
24. v � �' . w �
25. v � �' . w �
U X V = { (u, v) : u is in U and v is in V}
Prove that U X V is a vector space.
50. Let W be a subspace of a vector space V. Prove that
26. v = IR 3 , w =
Li = { (w, w) : w is in W } is a subspace of V X V.
a+b+l
mii
fl [ � l)
27. v � �' . w �
lUJ l
trn
1 1
In Exercises 51 and 52, let A =
and
1
1
l -1
B=
. Determine whether C is in span(A, B).
1
l 2
3 -5
51. c =
52. c =
5 -1
3 4
[
0]
[ ]
[
]
[
]
Section Linear Independence, Basis, and Dimension
6.2
i ] [o ]
In Exercises 53 and 54, let p(x) = 1 - 2x, q(x) = x - x 2,
and r(x) = - 2 + 3x + x 2 • Determine whether s(x) is in
span(p(x), q(x), r(x) ).
53. s(x) = 3 - 5x - x 2 54. s ( x) = 1 + x + x 2
In Exercises 55-58, let f(x) = sin 2x and g(x) = cos 2x.
Determine whether h(x) is in span(j(x), g(x) ).
55. h (x) = 1
56. h (x) = cos 2x
57. h (x) = sin 2x
58. h ( x) = sin x
59. Is M22 spanned by
[� � l [� �l [ � �l [� �]
-
Writi n g Project
443
1 ' 1
-1
?
0
61. Is <!!' 2 spanned by 1 + x, x + x 2 , 1 + x 2 ?
62. Is <!!' 2 spanned by 1 + x + 2 x 2 , 2 + x + 2 x 2 ,
- 1 + x + 2x 2 ?
63. Prove that every vector space has a unique zero
?
vector.
64. Prove that for every vector v in a vector space V,
there is a unique v ' in V such that v + v ' = 0.
The Rise of Vector Spaces
As noted in the sidebar on page 429, in the late 1 9th century, the mathematicians
Hermann Grassmann and Giuseppe Peano were instrumental in introducing the
idea of a vector space and the vector space axioms that we use today. Grassmann's
work had its origins in barycentric coordinates, a technique invented in 1 827 by
August Ferdinand Mobius (of Mobius strip fame). However, widespread acceptance
of the vector space concept did not come until the early 20th century.
Write a report on the history of vector spaces. Discuss the origins of the notion of
a vector space and the contributions of Grassmann and Peano. Why was the math­
ematical community slow to adopt these ideas, and how did acceptance come about?
1 . Carl B. Boyer and Uta C. Merzbach, A History of Mathematics (Third Edition)
(Hoboken, NJ: Wiley, 20 1 1 ) .
2. Jean-Luc, Dorier ( 1 995), A General Outline o f the Genesis o f Vector Space
Theory, Historia Mathematica 22 ( 1 995), pp. 227-26 1 .
3. Victor J. Katz, A History of Mathematics: A n Introduction (Third Edition) (Read­
ing, MA: Addison Wesley Longman, 2008).
f
l i n e a r I n d e p e n d e n ce . Basis , a n d D i m e n s i o n
In this section, we extend the notions of linear independence, basis, and dimension
to general vector spaces, generalizing the results of Sections 2.3 and 3.5. In most cases,
the proofs of the theorems carry over; we simply replace !R n by the vector space V.
linear Independence
D e fi n it i o n A set of vectors {v1, v2 , . . . , vd in a vector space V is linearly de­
pendent if there are scalars c1, c2 , . . . , c k, at least one of which is not zero, such that
A set of vectors that is not linearly dependent is said to be linearly independent.
444
Chapter Vector Spaces
6
As in IJ�r , {v1 , v2 ,
.
•
, vd is linearly independent in a vector space V if and only if
•
c 1 v1 + c 2v2 + · · · + ckvk = 0 implies c 1 = 0, c2 = 0, . . . , c k = 0
We also have the following useful alternative formulation of linear dependence.
Theorem 6 . 4
A set of vectors {v1 , v2 , , vk } in a vector space V is linearly dependent if and only
if at least one of the vectors can be expressed as a linear combination of the others.
.
Proof
•
.
The proof is identical to that of Theorem 2.5.
As a special case of Theorem 6.4, note that a set of two vectors is linearly depen­
dent if and only if one is a scalar multiple of the other.
Exa m p l e 6 . 2 2
In l!f 2 , the set { l + x + x 2 , 1 - x + 3x 2 , 1 + 3x - x 2 } is linearly dependent, since
2 ( 1 + x + x 2 ) - ( 1 - x + 3x 2 ) = 1 + 3x - x 2
Exa m p l e 6 . 2 3
In M22 , let
]
[ O1 J
2
, C=
1
0
Then A + B = C, so the set {A, B, C} is linearly dependent.
A=
Exa m p l e 6 . 2 4
Exa m p l e 6 . 2 5
[� � l
B=
[�
-1
In ':IF, the set {sin 2 x, cos 2 x, cos 2x} is linearly dependent, since
cos 2x = cos 2x - sin2x
Show that the set { l , x, x2 , . . . , x n } is linearly independent in <!P w
, e n are scalars such that
Co · 1 + C 1 X + Cz X 2 + · · · + C n X n = 0
Then the polynomial p (x) = c0 + c 1 x + c2 x2 + + c n x n is zero for all values of x. But
a polynomial of degree at most n cannot have more than n zeros (see Appendix D).
So p (x) must be the zero polynomial, meaning that c0 = c 1 = c2 = = en = 0.
Therefore, {l, x, x2 , . . . , x n } is linearly independent.
Solution 1
Suppose that c0 , c 1 ,
.
•
.
· · ·
· · ·
l�
We begin, as in the first solution, by assuming that
p ( x ) = Co + C 1 X + C z X 2 + · · · + C n X n = Q
Since this is true for all x, we can substitute x = 0 to obtain c0 = 0. This leaves
C 1 X + CzX 2 + · · · + C n X n = 0
Solution 2
Section Linear Independence, Basis, and Dimension
6.2
445
Taking derivatives, we obtain
C 1 + 2CzX + 3C3X2 +
+ ncnxn - I = 0
and setting x = 0, we see that c 1 = 0. Differentiating 2c 2 x + 3c 3 x 2 + + ncn x n - I = 0
and setting x = 0, we find that 2c 2 = 0, so c 2 = 0. Continuing in this fashion, we
find that k!ck = 0 for k = 0, . . . , n. Therefore, c0 = c 1 = c 2 = = e n = 0, and { 1 , x,
x2 , . . . ' x n } is linearly independent.
·
·
·
· · ·
· · ·
Exa m p l e 6 . 2 6
4
In <;!/' 2 , determine whether the set { 1 + x, x + x 2 , 1 + x 2 } is linearly independent.
Solution
Let c 1 , c 2 , and c 3 be scalars such that
Then
This implies that
the solution to which is c 1
linearly independent.
=
C1 +
c3 = 0
=0
C1 + Cz
C 2 + C3 = 0
c2 = c 3 = 0. It follows that { 1 + x, x + x 2 , 1 + x 2 } is
Remark
Compare Example 6.26 with Example 2.23(b). The system of equations
that arises is exactly the same. This is because of the correspondence between <;!/' 2 and
IR 3 that relates
I+
r n
[iJ
x + x' -
[:J
1 + x' -
m
and produces the columns of the coefficient matrix of the linear system that we have
to solve. Thus, showing that { 1 + x, x + x 2 , 1 + x 2 } is linearly independent is equiva­
lent to showing that
is linearly independent. This can be done simply by establishing that the matrix
[i : �]
has rank 3, by the Fundamental Theorem of Invertible Matrices.
446
Chapter Vector Spaces
6
Exa m p l e 6 . 2 1
In ?JP , determine whether the set {sin x, cos x} is linearly independent.
The functions f(x) = sin x and g(x) = cos x are linearly dependent if and
only if one of them is a scalar multiple of the other. But it is clear from their graphs
that this is not the case, since, for example, any nonzero multiple ofj(x) = sin x has
the same zeros, none of which are zeros ofg(x) = cos x.
This approach may not always be appropriate to use, so we offer the following
direct, more computational method. Suppose c and d are scalars such that
S o lulion
c sin x + d cos x =
0
Setting x = 0, we obtain d = 0, and setting x = n/2, we obtain c = 0. Therefore, the
set {sin x, cos x} is linearly independent.
Although the definitions of linear dependence and independence are phrased
in terms of finite sets of vectors, we can extend the concepts to infinite sets as
follows:
A set S of vectors in a vector space V is linearly dependent if it contains finitely
many linearly dependent vectors. A set of vectors that is not linearly dependent is
said to be linearly independent.
Note that for finite sets of vectors, this is just the original definition. Following is an
example of an infinite set of linearly independent vectors.
Exa m p l e 6 . 2 8
In <lP , show that S = { l , x, x2 , . . . } is linearly independent.
Suppose there is a finite subset T of S that is linearly dependent. Let x m be
n
the highest power of x in T and let x be the lowest power of x in T. Then there are
scalars en, Cn+ l ' . . . , c m , not all zero, such that
Solulion
But, by an argument similar to that used in Example 6.25, this implies that en =
Cn + i =
= cm = 0, which is a contradiction. Hence, S cannot contain finitely many
linearly dependent vectors, so it is linearly independent.
· · ·
Bases
The important concept of a basis now can be extended easily to arbitrary vector
spaces.
D e fi n it i o n
A subset l3 of a vector space V is a basis for V if
1 . l3 spans V and
2. l3 is linearly independent.
Section Linear Independence, Basis, and Dimension
6.2
441
Exa m p l e 6 . 2 9
If e; is the ith column of the n X n identity matrix, then { e1, e2 , . . . , en } is a basis for !R n ,
called the standard basis for !R n .
Exa m p l e 6 . 3 0
{ l , x, x 2 ,
Exa m p l e 6 . 3 1
The set [ = {E 11 , . • . , E 1n , E 21 , • • • , E 2n , Em1 , • • • , Emn } is a basis for Mm n ' where the
matrices Eij are as defined in Example 6. 1 8. [ is called the standard basis for Mm n ­
We have already seen that [ spans Mm n - It is easy to show that [ is linearly inde­
pendent. (Verify this! ) Hence, [ is a basis for Mm n -
Exa m p l e 6 . 3 2
Show that B
• . • , x n } is a basis for '!P n ' called the standard basis for '!P n -
=
{ l + x, x + x 2 , 1 + x 2 } is a basis for '!1' 2 •
We have already shown that B is linearly independent, in Example 6.26. To
show that B spans '!1' 2 , let a + bx + cx 2 be an arbitrary polynomial in '!1' 2 • We must
show that there are scalars c1, c2 , and c 3 such that
Solution
or, equivalently,
(c 1 + c3 ) + (c 1 + c2 )x + (c2 + c 3 )x 2 = a + bx + cx 2
Equating coefficients of like powers of x, we obtain the linear system
whi<h has a solution, since the weffideot matcix
[i : �]
has rnok 3 aod, hence,
is invertible. (We do not need to know what the solution is; we only need to know that
it exists.) Therefore, B is a basis for '!P 2 •
Ro11ark
Obmve that the matrix
O
[i �]
is the key to Example 6.32. We rnn
immediately obtain it using the correspondence between '!1' 2 and IR 3 , as indicated in
the Remark following Example 6.26.
448
Chapter Vector Spaces
6
Exa m p l e 6 . 3 3
Show that B = { l , x, x 2 , • • • } is a basis for <!P .
In Example 6.28, we saw that B is linearly independent. It also spans <!P ,
since clearly every polynomial is a linear combination of (finitely many) powers of x.
S o lulion
Exa m p l e 6 . 3 4
.+
Find bases for the three vector spaces in Example 6. 13:
Once again, we will work the three examples side by side to highlight the
similarities among them. In a strong sense, they are all the same example, but it will
take us until Section 6.5 to make this idea perfectly precise.
Solulion
(b) Since
(a) Since
we have W1 = span (u, v), where
we have W2 = span (u(x), v(x)),
where
u ( x) = 1 + x 3
and
Since {u, v} is clearly linearly in­
dependent, it is also a basis for W1 .
[ J [0 J [ 0 0J
(c) Since
a + bx - bx 2 + ax 3
= a ( l + x 3 ) + b (x - x 2 )
v ( x) = x - x 2
Since {u (x), v (x) } is clearly lin­
early independent, it is also a basis
for W2 .
1 O
1
a b
=a
+b
-b a
1
-1
[0 0J
we have W3 = span( U, V), where
U=
1
1
and V =
[ 0J
O 1
-1
Since { U, V} is clearly linearly in­
dependent, it is also a basis for W3 •
.+
Coordin a1es
Section 3.5 introduced the idea of the coordinates of a vector with respect to a basis
for subspaces of !R n . We now extend this concept to arbitrary vector spaces.
Theorem 6 . 5
Let V be a vector space and let B be a basis for V. For every vector v in V, there is
exactly one way to write v as a linear combination of the basis vectors in B.
Proof
The proofis the same as the proof ofTheorem 3.29. It works even ifthe basis B is
infinite, since linear combinations are, by definition, finite.
Section Linear Independence, Basis, and Dimension
6.2
449
The converse of Theorem 6.5 is also true. That is, if B is a set of vectors in a vector
space V with the property that every vector in V can be written uniquely as a linear
combination of the vectors in B, then B is a basis for V (see Exercise 30). In this sense,
the unique representation property characterizes a basis.
Since representation of a vector with respect to a basis is unique, the next definition
makes sense.
D e fi n it i o n Let B = {v1 , v2 ,
, vn } be a basis for a vector space V. Let v be a
vector in V, and write v = c 1 v1 + c2v2 + + cnvn - Then c 1 , c2 , , e n are called the
coordinates of v with respect to B, and the column vector
.
•
.
•
· · ·
.
•
is called the coordinate vector of v with respect to B.
Observe that if the basis B of V has n vectors, then [ v l B is a (column) vector in !R n .
Exa m p l e 6 . 3 5
Find the coordinate vector [p ( x) ] B of p (x)
dard basis B = { l , x, x 2 } of (!J> 2 .
Solution
=
2 - 3x + 5x 2 with respect to the stan­
The polynomial p(x) is already a linear combination of 1, x, and x2 , so
This is the correspondence between (!J> 2 and IR 3 that we remarked on after
Example 6.26, and it can easily be generalized to show that the coordinate vector of a
polynomial
with respect to the standard basis B
=
{ l , x, x 2 , . . . , x " } is just the vector
Remark
The order in which the basis vectors appear in B affects the order of
the entries in a coordinate vector. For example, in Example 6.35, assume that the
450
Chapter Vector Spaces
6
standard basis vectors are ordered as !3 ' = {x 2 , x, l}. Then the coordinate vector of
p (x) = 2 - 3x + 5x2 with respect to !3 ' is
[p(x) ] a �
Exa m p l e 6 . 3 6
Find the coordinate vector [A ] B of A
l3
=
=
{E 11 , E 1 2 , E2 1 > E22 } of M22 •
Solulion
[
Hl
]
2 -1
with respect to the standard basis
4
3
Since
we have
This is the correspondence between M22 and IR 4 that we noted before the intro­
duction to Example 6. 1 3 . It too can easily be generalized to give a correspondence
between Mmn and !R mn .
Exa m p l e 6 . 3 1
Find the coordinate vector [p(x) ] 8 of p (x)
C = {l + x, x + x 2 , 1 + x 2 } of \)}> 2 .
Solulion
=
1 + 2x - x 2 with respect to the basis
We need to find c1, c2 , and c3 such that
c 1 ( 1 + x) + c2 (x + x 2 ) + c 3 ( 1 + x 2 ) = 1 + 2x - x 2
or, equivalently,
(c 1 + c 3 ) + (c 1 + c2 ) x + ( c2 + c3 ) x 2
=
1 + 2x - x 2
As in Example 6.32, this means we need to solve the system
c1
+
C1 +
c3 =
c2
c2
1
2
-1
+ c3 =
whose solution is found to be c1 = 2, c2 = 0, c3 = - 1 . Therefore,
Section Linear Independence, Basis, and Dimension
6.2
[ Since this result says that p (x)
correct.]
=
2 ( 1 + x)
451
( 1 + x2 ), it is easy to check that it is
4
-
The next theorem shows that the process of forming coordinate vectors is com­
patible with the vector space operations of addition and scalar multiplication.
Theorem 6 . 6
Let B = {v1 , Vz , . . . , vn } be a basis for a vector space V. Let u and v be vectors in
V and let e be a scalar. Then
a. [ u + v l s = [ u l s + [v l s
b. [ c u ] s = e [ u ] s
Proof
We begin by writing u and v in terms of the basis vectors-say, as
Then, using vector space properties, we have
and
so
l ] [l l]
[ ] [l
d1
e1
dz
ez
.. + ..
.
dn
en
e i + d1
e z + dz
.
.
e n + dn
and
[ cu ] 8
=
ee l
eez
:
een
=
e
el
ez
:
=
e [u]8
en
An easy corollary to Theorem 6.6 states that coordinate vectors preserve linear
combinations:
( 1)
You are asked to prove this corollary in Exercise 3 1 .
The most useful aspect of coordinate vectors is that they allow us to transfer
information from a general vector space to !R n , where we have the tools of Chapters 1
to 3 at our disposal. We will explore this idea in some detail in Sections 6.3 and 6.6.
For now, we have the following useful theorem.
452
Chapter Vector Spaces
6
Theorem 6 . 1
Let B = {v1 , v2 ,
, v"} b e a basis for a vector space V and let u1, . . . , uk
be vectors in V. Then {u1, . . . , ud is linearly independent in V if and only if
{ [ u 1 ] 13, . . . , [ uk J 13} is linearly independent in IR " .
•
Proof
•
•
Assume that {u1, . . . , ud is linearly independent in V and let
in IR " . But then we have
c d u 1 ] 13
+
···
+
cd ud 13 =
[ c1u1
+
···
+
ckuk ] 13 =
0
0
using Equation ( 1 ), so the coordinates of the vector c1u1 + + ckuk with respect to
B are all zero. That is,
c 1 u 1 + · · · + ckuk = Ov 1 + Ov2 + · · · + Ov" = 0
· · ·
The linear independence of {u1 , . . . , uk} now forces c1 = c2 =
{ [ u 1 ] 13, . . . , [ uk J 13} is linearly independent.
· · ·
= ck = 0, so
The converse implication, which uses similar ideas, is left as Exercise 32.
Observe that, in the special case where U; = V;, we have
V; = 0 · V I + · · · + 1 · V; + · · · + 0 • Vn
Dimension
The definition of dimension is the same for a vector space as for a subspace of !R n -the
number of vectors in a basis for the space. Since a vector space can have more than one
basis, we need to show that this definition makes sense; that is, we need to establish
that different bases for the same vector space contain the same number of vectors.
Part (a) of the next theorem generalizes Theorem 2.8.
Theorem 6 . 8
Let B = {v 1 , v2 , . . . , vn } be a basis for a vector space V.
a. Any set of more than n vectors in V must be linearly dependent.
b. Any set of fewer than n vectors in V cannot span V.
(a) Let {u1, . . . , um} be a set of vectors in V, with m > n. Then { [ u 1 ] 13, . . . ,
[ um J 13} is a set of more than n vectors in !R n and, hence, is linearly dependent,
by Theorem 2.8. This means that {u1, . . . , um} is linearly dependent as well, by
Proof
Theorem 6.7.
(b) Let {u1, . . . , um} be a set of vectors in V, with m < n. Then S = { [ u 1 J 13, . . . , [ um J 13}
is a set of fewer than n vectors in !R n . Now span(u 1 , . . . , um) = V if and only if
span(S) = !R n (see Exercise 33). But span ( S ) is just the column space of the n X m
matrix
A = [ [ u 1 ] 13 · · · [ um J 13 ]
so dim(span(S )) = dim(col(A)) :s m < n. Hence, S cannot span !R n , so {u1, . . . , um}
does not span V.
Now we extend Theorem 3.23.
Section Linear Independence, Basis, and Dimension
6.2
Theorem 6 . 9
453
The Basis Theorem
If a vector space V has a basis with n vectors, then every basis for V has exactly n
vectors.
The proof of Theorem 3.23 also works here, virtually word for word. However, it
is easier to make use of Theorem 6.8.
Let B be a basis for V with n vectors and let B' be another basis for V with m
vectors. By Theorem 6.8, m ::::: n; otherwise, B' would be linearly dependent.
Now use Theorem 6.8 with the roles of B and B' interchanged. Since B' is a
basis of V with m vectors, Theorem 6.8 implies that any set of more than m vectors
in V is linearly dependent. Hence, n ::::: m, since B is a basis and is, therefore, linearly
independent.
Since n ::::: m and m ::::: n, we must have n = m, as required.
Proof
The following definition now makes sense, since the number of vectors in a
(finite) basis does not depend on the choice of basis.
D e fi n it i o n A vector space V is called finite-dimensional if it has a basis con­
sisting of finitely many vectors. The dimension of V, denoted by dim V, is the num ber of vectors in a basis for V. The dimension of the zero vector space { 0} is defined
to be zero. A vector space that has no finite basis is called infinite-dimensional.
Exa m p l e 6 . 3 8
Since the standard basis for !R n has n vectors, dim !R n = n. In the case of IR 3 , a one­
dimensional subspace is just the span of a single nonzero vector and thus is a line
through the origin. A two-dimensional subspace is spanned by its basis of two
linearly independent (i.e., nonparallel) vectors and therefore is a plane through the
origin. Any three linearly independent vectors must span IR 3 , by the Fundamental
Theorem. The subspaces of IR 3 are now completely classified according to dimension,
as shown in Table 6. 1 .
Ta b l e 6 . 1
dim V
3
2
0
Exa m p l e 6 . 3 9
v
IR 3
Plane through the origin
Line through the origin
{O}
The standard basis for <!J> n contains n + 1 vectors (see Example 6.30), so dim <!J> n
n + 1.
=
454
Chapter Vector Spaces
6
Exa m p l e 6 . 4 0
Exa m p l e 6 . 4 1
The standard basis for Mm n contains mn vectors (see Example 6.3 1), so
dim Mm n = mn.
Both <!P and 9F are infinite-dimensional, since they each contain the infinite linearly
independent set { l , x, x2 , } (see Exercise 44).
•
Exa m p l e 6 . 4 2
•
•
Find the dimension of the vector space W of symmetric 2 X 2 matrices (see
Example 6. 10).
Solulion
A symmetric 2 X 2 matrix is of the form
[� �] [� �] [� �] [� �]
{ [ � � l [ � �l [ � � ] }
=a
+b
+c
so W is spanned by the set
s
=
If S is linearly independent, then it will be a basis for W. Setting
we obtain
from which it immediately follows that a = b = c = 0. Hence, S is linearly indepen­
dent and is, therefore, a basis for W. We conclude that dim W = 3.
The dimension of a vector space is its "magic number:' Knowing the dimension
of a vector space V provides us with much information about V and can greatly sim­
plify the work needed in certain types of calculations, as the next few theorems and
examples illustrate.
Theorem 6 . 1 0
Let V be a vector space with dim V = n. Then:
a.
b.
c.
d.
e.
f.
Any linearly independent set in V contains at most n vectors.
Any spanning set for V contains at least n vectors.
Any linearly independent set of exactly n vectors in V is a basis for V.
Any spanning set for V consisting of exactly n vectors is a basis for V.
Any linearly independent set in V can be extended to a basis for V.
Any spanning set for V can be reduced to a basis for V.
Section Linear Independence, Basis, and Dimension
6.2
455
The proofs of properties (a) and (b) follow from parts (a) and (b) of Theo­
rem 6.8, respectively.
(c) Let S be a linearly independent set of exactly n vectors in V. If S does not span V,
then there is some vector v in V that is not a linear combination of the vectors in S.
Inserting v into S produces a set S' with n + 1 vectors that is still linearly independent
(see Exercise 54). But this is impossible, by Theorem 6.S(a). We conclude that S must
span V and therefore be a basis for V.
(d) Let S be a spanning set for V consisting of exactly n vectors. If S is linearly
dependent, then some vector v in S is a linear combination of the others. Throwing v
away leaves a set S' with n 1 vectors that still spans V (see Exercise 55). But this is
impossible, by Theorem 6.S(b ) . We conclude that S must be linearly independent and
therefore be a basis for V.
(e) Let S be a linearly independent set of vectors in V. If S spans V, it is a basis for
V and so consists of exactly n vectors, by the Basis Theorem. If S does not span V,
then, as in the proof of property (c), there is some vector v in V that is not a linear
combination of the vectors in S. Inserting v into S produces a set S' that is still linearly
independent. If S' still does not span V, we can repeat the process and expand it into
a larger, linearly independent set. Eventually, this process must stop, since no linearly
independent set in V can contain more than n vectors, by Theorem 6.S(a). When the
process stops, we have a linearly independent set S* that contains S and also spans V.
Therefore, S* is a basis for V that extends S.
(f) You are asked to prove this property in Exercise 56.
Proof
-
You should view Theorem 6. 1 0 as, in part, a labor-saving device. In many
instances, it can dramatically decrease the amount of work needed to check that a set
of vectors is linearly independent, a spanning set, or a basis.
Exa m p l e 6 . 4 3
In each case, determine whether S is a basis for V.
(a)
(b)
(c)
V
=
rzf ,
2
V
=
M ,
22
V
=
rzf ,
2
S
S
S
-
_
=
=
+ x, 2 - x + x 2, 3x - 2x2, 1 + 3x + x2}
{ [ � �], [ � - �] , [ � �] }
{l
=
{l
+ x, x + x2, 1 + x2}
(a) Since dim ( \]f 2 ) = 3 and S contains four vectors, S is linearly depen­
dent, by Theorem 6. l O(a). Hence, S is not a basis for \)f 2 .
(b) Since dim(M22 ) = 4 and S contains three vectors, S cannot span M22 , by Theo­
rem 6. l O(b). Hence, S is not a basis for M22 .
( c) Since dim ( \]f 2 ) = 3 and S contains three vectors, S will be a basis for rzf 2 if it is lin early independent or if it spans \]f 2 , by Theorem 6.10( c) or (d) . It is easier to show that
S is linearly independent; we did this in Example 6.26. Therefore, S is a basis for \]f 2 •
(This is the same problem as in Example 6.32-but see how much easier it becomes
using Theorem 6. 1 0 ! )
Solution
Exa m p l e 6 . 4 4
Extend { l + x , 1
Solution
((lf 2 ) =
-
x } t o a basis fo r \)f 2 .
First note that { l + x, 1 x} is linearly independent. (Why?) Since dim
3, we need a third vector-one that is not linearly dependent on the first two.
-
Chapter Vector Spaces
6
456
We could proceed, as in the proof of Theorem 6.I O(e), to find such a vector using trial
and error. However, it is easier in practice to proceed in a different way.
We enlarge the given set of vectors by throwing in the entire standard basis for !!f 2 .
This gives
S = { l + x, 1 - x, 1 , x, x 2 }
Now S is linearly dependent, by Theorem 6. IO(a), so we need to throw away some
vectors-in this case, two. Which ones? We use Theorem 6.l O(f), starting with the
first vector that was added, 1. Since 1 = i ( l + x) + i ( l - x), the set {l + x, 1 - x, l }
is linearly dependent, so we throw away 1 . Similarly, x = i ( l + x) - i ( l - x), so
{l + x, 1 - x, x} is linearly dependent also. Finally, we check that {l + x, 1 - x, x 2 }
is linearly independent. (Can you see a quick way to tell this?) Therefore, { l + x,
1 - x, x2 } is a basis for l!f 2 that extends { l + x, 1 - x}.
In Example 6.42, the vector space W of symmetric 2 X 2 matrices is a subspace of
the vector space M22 of all 2 X 2 matrices. As we showed, dim W = 3 ::=::: 4 = dim M22 .
This is an example of a general result, as the final theorem of this section shows.
Theorem 6 . 1 1
Let W be a subspace of a finite-dimensional vector space V. Then:
a. W is finite-dimensional and dim W ::=::: dim V.
b. dim W = dim V if and only if W = V.
Proof
(a) Let dim V = n. If W = {O}, then dim ( W ) = 0 ::=::: n = dim V. If W is
nonzero, then any basis B for V (containing n vectors) certainly spans W, since W is
contained in V. But B can be reduced to a basis B' for W (containing at most n vec­
tors), by Theorem 6. I O (f) . Hence, W is finite-dimensional and dim( W ) ::=::: n = dim V.
(b) If W = V, then certainly dim W = dim V. On the other hand, if dim W = dim
V = n, then any basis B for W consists of exactly n vectors. But these are then n lin­
early independent vectors in V and, hence, a basis for V, by Theorem 6. I O (c). There­
fore, V = span (B) = W.
I
Exercises 6 . 2
In Exercises 1 -4, test the sets of matrices for linear indepen­
dence in M22. For those that are linearly dependent, express
one of the matrices as a linear combination of the others.
4·
{ [� � l [ � �l [� � l [ � �] }
In Exercises 5-9, test the sets ofpolynomials for linear inde­
pendence. For those that are linearly dependent, express one
of the poly nomials as a linear combination of the others.
5. {x, 1 + x} in l!f
6. { l + x, 1 + x 2 , 1 - x + x 2 } in l!f 2
7. {x, 2x - x 2 , 3x + 2x 2 } in l!f 2
1
Section Linear Independence, Basis, and Dimension
6.2
8. {2x, x - x 2 , 1 + x 3 , 2 - x 2 + x 3 } in <;5} 3
9. { l - 2x, 3x + x 2 - x 3 , 1 + x 2 + 2x 3 , 3 + 2x + 3 x 3 } in <;5} 3
In Exercises 1 0-14, test the sets offunctions for linear in­
dependence in <;IF, For those that are linearly dependent,
express one of the functions as a linear combination of the
others.
1 1 . {l , sin2x, cos2x}
10. { l , sin x, cos x}
13. { l , ln (2x), ln (x 2 ) }
12. {eX, e - x }
14. {sin x, sin 2x, sin 3x}
� 15. Iff and g are in C(b (J J , the vector space of all functions
with continuous derivatives, then the determinant
I
I
f(x) g(x)
j'(x) g'(x)
is called the Wronskian off and g [named after the
W(x) =
Polish-French mathematician J 6sef Maria Hoene­
Wronski ( 1 776- 1 853), who worked on the theory of
determinants and the philosophy of mathematics] .
Show that f and g are linearly independent if their
Wronskian is not identically zero (that is, if there is
some x such that W(x) * O ).
� 16. In general, the Wronskian of f1 , • • • Jn in C(b ( n - i l is the
determinant
fz (x)
j{(x)
W(x) =
and f1 , . . . ,fn are linearly independent, provided W(x)
is not identically zero. Repeat Exercises 10-14 using
the Wronskian test.
17. Let {u, v, w} be a linearly independent set of vectors in
a vector space V.
(a) Is {u + v, v + w, u + w} linearly independent?
Either prove that it is or give a counterexample
to show that it is not.
(b) Is {u v, v w, u - w} linearly independent?
Either prove that it is or give a counterexample
to show that it is not.
- -
In Exercises 18-25, determine whether the set B is a basis
for the vector space V
18. V = M22 , B =
l9. V = M22 , B =
{ [� �l [ � �l [ _ � - � ] }
{ [ � � ] ' [ � -� ] ' [ � � ] ' [ � - � ] }
451
20. V = Mzz ,
21. V = M22 ,
B=
22. V =
23. V =
24. V =
25. V =
{ [ � � l [ � � l [ _ � �l [ � �l [� � ] }
<;IP 2 , B =
<;IP 2 , B =
<;IP 2 , B =
<;IP 2 , B =
-
{x, 1 + x, x x 2 }
{ l - x, 1 - x 2 , x x 2 }
{ l , 1 + 2x + 3x2 }
{ l , 2 - x, 3 - x2 , x + 2x 2 }
-
[� !]
[� !]
{ [� �l [� �l [ � �l [ � � ] }
26. Find the coordinate vector of A =
with
respect to the basis B = {E22 , E21 , Ew E 1 1 } of M22 •
27. Find the coordinate vector of A =
to the basis B =
of M22 .
with respect
28. Find the coordinate vector of p (x) = 1 + 2x + 3x 2
with respect to the basis B = { l + x, 1 - x, x 2 } of <;JP 2 .
29. Find the coordinate vector of p (x) = 2 - x + 3x 2 with
respect to the basis B = { l , 1 + x, - 1 + x 2 } of <;JP 2 .
30. Let B be a set of vectors in a vector space V with
the property that every vector in V can be written
uniquely as a linear combination of the vectors in B.
Prove that B is a basis for V.
31. Let B be a basis for a vector space V, let u 1 , . . . , uk
be vectors in V, and let c 1 , . . . , ck be scalars. Show that
[ c 1U1 + . . . + ckud s = C 1 [ u 1 l s + . . . + cd ud s ·
32. Finish the proof of Theorem 6.7 by showing that if
{ [ u 1 ] 8, . . . , [ uk ] 8} is linearly independent in IR " then
{ u 1 , . . . , uk } is linearly independent in V.
33. Let { u 1 , . . . , um } be a set of vectors in an
n-dimensional vector space V and let B be a basis for V.
Let S = { [ u 1 ] 8, . . . , [ um ] 8} be the set of coordinate
vectors of {u 1 , . . . , um } with respect to B. Prove that
span ( u 1 , • . • , um ) = V if and only if span(S ) = !R n .
In Exercises 34-39, find the dimension of the vector space V
and give a basis for V
34. V = {p (x) in <;JP 2 : p (O) = O}
35. V = {p ( x) in <;JP 2 : p ( l ) = O}
� 36. V = {p ( x) in <;IP 2 : xp ' ( x) = p (x) }
Chapter Vector Spaces
6
458
37. V = {A in M22 : A is upper triangular}
38. V = {A in M22 : A is skew-symmetric}
39. V = {A in M22 : AB = BA}, where B =
[ 01 ]
1
1
40. Find a formula for the dimension of the vector space
of symmetric n X n matrices.
41. Find a formula for the dimension of the vector space
of skew-symmetric n X n matrices.
42. Let U and W be subspaces of a finite-dimensional
vector space V. Prove Grassmann's Identity:
dim ( U + W) = dim U + dim W - dim( U n W)
[Hint: The subspace U + W is defined in Exercise 48
of Section 6. 1 . Let B = {v1 , . . . , vk} be a basis for
U n W. Extend B to a basis C of U and a basis D of W.
Prove that C U D is a basis for U + W.]
43. Let U and V be finite-dimensional vector spaces.
(a) Find a formula for dim( U X V) in terms of dim U
and dim V. (See Exercise 49 in Section 6. 1 . )
(b) I f W is a subspace o f V, show that dim Li =
dim W, where Li = { ( w, w) : w is in W}.
44. Prove that the vector space i!f is infinite-dimensional.
[Hint: Suppose it has a finite basis. Show that there is
some polynomial that is not a linear combination of
this basis.]
45. Extend { l + x, 1 + x + x 2 } to a basis for llf 2 .
46. Extend
47. Extend
48. Extend
{ [� � l [� � ] }
{ [ � � l [ � �l [ � � ] }
{ [ � � ], [ � � ] }
to a basis for M22 •
-
to a basis for M22 •
to a basis for the vector
space of symmetric 2 X 2 matrices.
49. Find a basis for span ( l , 1 + x, 2x) in llf 1 .
50. Find a basis for span ( l - 2x, 2x - x 2 , 1 - x 2 , 1 + x 2 )
in i!f 2 .
5 1 . Find a basis for span ( l - x, x - x 2 , 1 - x 2 , 1 - 2x +
x 2 ) in i!f 2 •
O O l
52. Find a basis for span
,
,
1 1 0
1
[ � -�])
_
in M22 .
( [�
J [ ] [-1
53. Find a basis for span(sin2x, cos 2x, cos 2x) in '!fa.
54. Let S = {v1 , . . . , vn } be a linearly independent set in
a vector space V. Show that if v is a vector in V that is
not in span(S ), then S' = {v1 , • • • , vn , v} is still linearly
independent.
55. Let S = {v 1 , • • • , vn } be a spanning set for a vector
space V. Show that if vn is in span (v1 , . . . , vn _ 1 ), then
S' = {v1 , • . • , vn _ 1 } is still a spanning set for V.
56. Prove Theorem 6.l O(f) .
57. Let {v1 , • • • , vJ be a basis fo r a vector space V
and let c 1 , • . . , cn be nonzero scalars. Prove that
{c 1 V1 , . . . , C nvn } is also a basis for V.
58. Let {v1 , • • • , vJ be a basis for a vector space V. Prove
that
{vi , V1 + V2 , V1 + Vz + V3 , . . . , V1 + . . . + vJ
is also a basis for V.
Let a0, a 1 , • • • , a n be n + 1 distinct real numbers. Define
polynomials p0(x), p 1 ( x) , . . . , Pn ( x) by
( x - a0) · • · ( x - a ; _ 1 ) (x - a; 1 ) · • · ( x - a n )
P; ( x) = (a - a0) • • · (a ; - a ; _ ) (a ; - a ;+ ) • • · (a ; - a n )
;
1
+1
These are called the Lagrange polynomials associated
with a0, a 1 , . . . , aw [Joseph-Louis Lagrange (1 736- 1 8 1 3)
was born in Italy but spent most of his life in Germany and
France. He made important contributions to such fields as
number theory, algebra, astronomy, mechanics, and the
calculus of variations. In 1 773, Lagrange was the first to give
the volume interpretation of a determinant (see Chapter 4).]
59. ( a) Compute the Lagrange polynomials associated
with a0 = 1, a 1 = 2, a 2 = 3.
(b) Show, in general, that
p ; (a) =
{o
if i * j
1 if i = j
60. (a) Prove that the set B = {p0(x), p 1 ( x) , . . . , P n ( x) }
of Lagrange polynomials is linearly independent
in i!f w [Hint: Set c0p0(x) +
+ c np n (x) = 0 and
use Exercise 59(b) .]
(b) Deduce that B is a basis for i!f w
61. If q(x) is an arbitrary polynomial in i!f n ' it follows from
Exercise 60(b) that
q ( x) = CoPo ( x) + . . . + cnp n (x)
( 1)
for some scalars c0, . . . , cw
( a) Show that C; = q (a ; ) for i = 0, . . . , n, and deduce
that q ( x) = q(a0)p0(x) +
+ q(a n )P n (x) is the
unique representation of q(x) with respect to the
basis B.
·
·
·
·
·
·
Section Linear Independence, Basis, and Dimension
6.2
,
(b) Show that for any n + 1 points ( a0, c0 ) , ( a1, c1 ) ,
( a n , en ) with distinct first components, the func­
.
.
•
tion q(x) defined by Equation ( 1 ) is the unique
polynomial of degree at most n that passes
through all of the points. This formula is known
as the Lagrange interpolation formula. (Com­
pare this formula with Problem 19 in Explora­
tion: Geometric Applications of Determinants in
Chapter 4.)
(c) Use the Lagrange interpolation formula to find
the polynomial of degree at most 2 that passes
through the points
459
(i) ( 1 , 6), (2, - 1 ), and (3, - 2)
(ii) ( - 1 , 10), (0, 5), and (3, 2)
62. Use the Lagrange interpolation formula to show that
if a polynomial in i!f n has n + 1 zeros, then it must be
the zero polynomial.
63. Find a formula for the number of invertible matrices
in Mnn ( !f_p ) · [Hint: This is the same as determining the
number of different bases for z;. (Why?) Count the
number of ways to construct a basis for z;, one vector
at a time.]
Exp loration
Magic S qu a r e s
The engraving shown o n page 46 1 i s Albrecht Durer's Melancholia I ( 1 5 14). Among
the many mathematical artifacts in this engraving is the chart of numbers that hangs
on the wall in the upper right-hand corner. (It is enlarged in the detail shown.) Such
an array of numbers is known as a magic square. We can think of it as a 4 X 4 matrix
3 2
10 1 1
6 7
1 5 14
Observe that the numbers in each row, in each column, and in both diagonals have
the same sum: 34. Observe further that the entries are the integers 1 , 2, . . . , 16. (Note
that Durer cleverly placed the 1 5 and 14 adjacent to each other in the last row, giving
the date of the engraving.) These observations lead to the following definition.
An n x n matrix M is called a magic square if the sum of the
entries is the same in each row, each column, and both diagonals. This common
sum is called the weight of M, denoted wt(M). If M is an n X n magic square that
contains each of the entries 1 , 2, . . . , n 2 exactly once, then M is called a classical
D e fi n it i o n
magic square.
1 . If M is a classical n X n magic square, show that
wt ( M)
[Hint: Use Exercise 5 1 in Section 2.4.]
=
n(_n_2_+_l_)
_
2
2. Find a classical 3 X 3 magic square. Find a different one. Are your two ex­
amples related in any way?
460
3. Clearly, the 3 X 3 matrix with all entries equal to t is a magic square with
weight 1 . Using your answer to Problem 2, find a 3 X 3 magic square with weight 1 ,
all of whose entries are different. Describe a method fo r constructing a 3 X 3 magic
square with distinct entries and weight w for any real number w.
Let Magn denote the set of all n X n magic squares, and let Mag� denote the set of all
n X n magic squares of weight 0.
4. (a) Prove that Mag3 is a subspace of M33 .
(b) Prove that Mag� is a subspace of Mag3 .
5. Use Problems 3 and 4 to show that if M is a 3 X 3 magic square with weight
w, then we can write M as
M = M0 + kJ
where M0 is a 3 X 3 magic square of weight 0, J is the 3 X 3 matrix consisting entirely
of ones, and k is a scalar. What must k be? [Hint: Show that M - kJ is in Maij for an
appropriate value of k.]
Let's try to find a way of describing all 3 X 3 magic squares. Let
be a magic square with weight 0. The conditions on the rows, columns, and diag­
onals give rise to a system of eight homogeneous linear equations in the variables a,
b, . . . , i.
6. Write out this system of equations and solve it. [Note: Using a CAS will
facilitate the calculations.]
461
7. Find the dimension of Mag�. Hint: By doing a substitution, if necessary, use
your solution to Problem 6 to show that M can be written in the form
M
=
[
s
-s - t
t
s- t
0
-s + t
-t
s+ t
-s
l
8. Find the dimension of Mag3 • [Hint: Combine the results of Problems 5 and 7.]
9. Can you find a direct way of showing that the ( 2, 2) entry of a 3 X 3 magic
square with weight w must be w/3? [Hint: Add and subtract certain rows, columns,
and diagonals to leave a multiple of the central entry.]
10. Let M be a 3 X 3 magic square of weight 0, obtained from a classical 3 X 3
magic square as in Problem 5. If M has the form given in Problem 7, write out an
equation for the sum of the squares of the entries of M. Show that this is the equation
of a circle in the variables s and t, and carefully plot it. Show that there are exactly
eight points (s, t) on this circle with both s and t integers. Using Problem 8, show that
these eight points give rise to eight classical 3 X 3 magic squares. How are these magic
squares related to one another?
462
Section 6.3 Change of Basis
463
C h a n g e o f Basis
In many applications, a problem described using one coordinate system may be
solved more easily by switching to a new coordinate system. This switch is usually
accomplished by performing a change of variables, a process that you have prob­
ably encountered in other mathematics courses. In linear algebra, a basis provides
us with a coordinate system for a vector space, via the notion of coordinate vectors.
Choosing the right basis will often greatly simplify a particular problem. For example,
consider the molecular structure of zinc, shown in Figure 6.3(a). A scientist studying
zinc might wish to measure the lengths of the bonds between the atoms, the angles
between these bonds, and so on. Such an analysis will be greatly facilitated by intro­
ducing coordinates and making use of the tools of linear algebra. The standard basis
and the associated standard xyz coordinate axes are not always the best choice. As
Figure 6.3(b) shows, in this case {u, v, w} is probably a better choice of basis for IR 3
than the standard basis, since these vectors align nicely with the bonds between the
atoms of zinc.
v
(a)
(b)
figure 6 . 3
Chan ge-of-Basis Matrices
Figure 6.4 shows two different coordinate systems for IR 2 , each arising from a different
basis. Figure 6.4(a) shows the coordinate system related to the basis !3 = { u , , u2 } ,
while Figure 6.4(b) arises from the basis C = {v1 , vz } , where
The same vector x is shown relative to each coordinate system. It is clear from the
diagrams that the coordinate vectors of x with respect to !3 and C are
[x]8
=
[�]
and [ x ] c
=
[ _�]
respectively. It turns out that there is a direct connection between the two coordinate
vectors. One way to find the relationship is to use [ x] 8 to calculate
464
Chapter 6 Vector Spaces
y
y
2
-2
x
-2
-4
-4
(a)
(b)
Figure 6 . 4
Then we can find [ x ] c by writing x as a linear combination of v1 and v2 • However,
there is a better way to proceed-one that will provide us with a general mechanism
for such problems. We illustrate this approach in the next example.
Exa m p l e 6 . 4 5
Using the bases B and C above, find [ x ] c, given that [ x ] 8 =
1
.
3
Solulion
Since x = u 1 + 3u2 , writing u 1 and u2 in terms ofv 1 and v2 will give us the
required coordinates of x with respect to C. We find that
u1 =
and
so
[ -�] -3 [�] 2 [ � ]
[ �] [�] [ � ]
u2 =
+
_
=3
This gives
[x] c =
in agreement with Figure 6.4(b) .
2
= - 3v1 + v2
= 3v1 - V2
[ _�]
This method may not look any easier than the one suggested prior to Example 6.45,
but it has one big advantage: We can now find [ y ] c from [ y ] 8 for any vector y in IR 2
Section 6.3 Change of Basis
465
with very little additional work. Let's look at the calculations in Example 6.45 from a
different point of view. From x = u1 + 3u2, we have
by Theorem 6.6. Thus,
[� ]
[ -� -�] [�]
[ x ] c = [ [ u 1 J d u2 l c l
=
where P is the matrix whose columns are [ u 1 ] c and [ u2 ] c· This procedure generalizes
very nicely.
Let B = { u 1 , . . . , u"} and C = {v1 , . . . , v"} be bases for a vector
space V. The n X n matrix whose columns are the coordinate vectors [u1 ] C• . . . ,
[un l c of the vectors in B with respect to C is denoted by Pc<- B and is called the
change-of-basis matrix from B to C. That is,
D e fi n it i o n
Pc<- B = [ [ u 1 J d u2 l c
···
[ un l c l
Think of B as the "old" basis and C as the "new" basis. Then the columns of Pc<- B
are just the coordinate vectors obtained by writing the old basis vectors in terms of
the new ones. Theorem 6.12 shows that Example 6.45 is a special case of a general
result.
Theorem 6 . 1 2
and C = {v1 , . . . , vn } be bases for a vector space V and let
Pc<- B be the change-of-basis matrix from B to C. Then
Let B
= {u 1 , . . . , uJ
a. Pc,_ s [ x] 8 = [x] c for all x in V.
b. PC <- 8 is the unique matrix P with the property that P [ x ] 8
c. Pc<- B is invertible and (Pc ,_ 8) - 1 = Ps <- c·
Proof
(a) Let x be in V and let
+ . . . + C n un J c
C1 [ u 1 J c + . . . + Cn [ un J c
[ x ] c = [ C 1 U1
=
�
I [ u. J c
= Pc<-s [ x ] s
[ u" ] '
tl
= [x] c
for all x in V.
466
Chapter 6 Vector Spaces
(b) Suppose that P is an n X n matrix with the property that P [ x] 8 = [ x] c fo r all x
in V. Taking x = U;, the ith basis vector in B, we see that [x] 8 = [ u;] 8 = e;, so the ith
column of P is
P; = Pe ; = P [ u; ] B = [ U; ] c
which is the ith column of Pc ,_ 8, by definition. It follows that P = Pc.-B·
( c) Since {u1, . . . , un } is linearly independent in V, the set { [u1 J c, . . . , [ un l d is linearly
independent in ll�r, by Theorem 6.7. Hence, Pc .-8 = [ [ u 1 J c
[ u ] c is invert­
ible, by the Fundamental Theorem.
For all x in V, we have PC<- B [ x ] 8 = [ x ] c· Solving for [ x l t» we find that
·
·
·
[ x ] 8 = (Pc <-B ) - 1 [ x ] c
for all x in V. Therefore, (Pc .-8) - 1 is a matrix that changes bases from C to B. Thus, by
the uniqueness property (b), we must have (Pc .-8) - 1 = P8.- c .
R e m a rks
You may find it helpful to think of change of basis as a transformation (indeed,
it is a linear transformation) from !R n to itself that simply switches from one coordi­
nate system to another. The transformation corresponding to Pc<-B accepts [ x ] 8 as
input and returns [x]c as output; (Pc ,_ 8) - 1 = p8 ,_ c does just the opposite. Figure 6.5
gives a schematic representation of the process.
•
/
� [ ]B
/ Multiplication �
[ Jc
[x]• c
Change of basis
Figure 6 . 5
v
x•
by Pc� B
Multiplication
[x]• B
by PB� c = (Pc� B ) - 1
The columns of Pc.-8 are the coordinate vectors of one basis with respect to
the other basis. To remember which basis is which, think of the notation C +--- B as
saying "B in terms of C:' It is also helpful to remember that Pc <-t3 [ x] 8 is a linear com­
bination of the columns of Pc <-B· But since the result of this combination is [ xJ c, the
columns of Pc .-8 must themselves be coordinate vectors with respect to C.
•
Exa m p l e 6 . 4 6
Find the change-of-basis matrices Pc.-8 and P8.- c for the bases B = { 1 , x, x 2 } and C =
{ I + x, x + x 2 , 1 + x 2 } of \]]> 2 . Then find the coordinate vector of p(x) = 1 + 2x - x 2
with respect to C.
Changing to a standard basis is easy, so we find PB <- c first. Observe that the
coordinate vectors for C in terms of B are
Solution
Section 6.3 Change of Basis
461
(Look back at the Remark following Example 6.26.) It follows that
To find PC<- l3• we could express each vector in l3 as a linear combination of the vec­
tors in C (do this), but it is much easier to use the fact that Pc,_13 = (P13,_c) - 1 , by
Theorem 6.1 2(c). We find that
Pc<-- 13
=
(P13,_c) - 1
It now follows that
[p(x) ] c
=
=
[ -�: -�: -�::l
2
1
:-
2
2
1
1
:-
-
Pc,_13 [p(x) ] 13
which agrees with Example 6.37.
Ifwe do not need Pc,_13 explicitly, we can find [p(x) l e from [p(x) ] 13 and
P13,_c using Gaussian elimination. Row reduction produces
Remark
[I I [p(x) l c l
(See the next section on using Gauss-Jordan elimination.)
It is worth repeating the observation in Example 6.46: Changing to a standard
basis is easy. If £ is the standard basis for a vector space V and l3 is any other basis,
then the columns of PE<- l3 are the coordinate vectors of l3 with respect to £, and these
are usually "visible:' We make use of this observation again in the next example.
Exa m p l e 6 . 4 1
In M22 , let l3 be the basis {E 11, E 2 1, E1 2 , E 22 } and let C be the basis {A, B, C, D}, where
A-
[ � ! ].
[ J [
1 O
]
�]
[
1 1
1
CB'
'
O O
1
o o
Find the change-of-basis matrix Pc,_13 and verify that [XJ c
=
Pc,_13 [X] 13 for X
=
468
Chapter 6 Vector Spaces
To solve this problem directly, we must find the coordinate vectors of l3
with respect to C. This involves solving four linear combination problems of the form
X = aA bB cC dD, where X is in l3 and we must find a, b, c, and d. However,
here we are lucky, since we can find the required coefficients by inspection.
Clearly, E11 = A, E2 1 = - B C, E1 2 = -A B, and E22 = - C D. Thus,
+ + +
Solulion 1
[E u l c �
+
+
+
m [ I} [ � l [ -!]
:
�
[ � -!]
-
[ £,, ] , �
-
[E , , ] , �
[E,, J c �
Pc�,; � [ [E u l c [E,. ] , [E ,, J c [E,, J c ]
rn
If X
=
[� !l
then
[X] s �
Pc +- 13 [ X ] 13 =
and
[i
0
-1
1
0
rn
-1
1
0
0
-!][ ;] [ =]
This is the coordinate vector with respect to C of the matrix
- A - B - C + 4D
=
[� �] - [� �] - [ � �] + 4 [ � � ]
[� !]
-
=
x
as it should be.
We can compute Pc..-13 in a different way, as follows. As you will be
asked to prove in Exercise 2 1 , if £ is another basis for M22 > then Pc ..-13 = Pc ..-EPE..-13 =
(PE..-c) - 1 pE+- l3· If £ is the standard basis, then PE+-l3 and PE+-c can be found by inspec­
tion. We have
Solulion 2
Section 6.3 Change of Basis
�
469
(Do you see why?) Therefore,
Pc +- B = (Pt:+- c) - 1 Pt:+- B
[j J[j �]
[j �m �]
� [j
-� ]
1
0 1
0 0
0
-1
1 -1
1
0
0
0
0 -1
1
-1
1
0
0
0
0
0
1
0
-
0
1
0
0
0 0
0 1
1 0
0 0
which agrees with the first solution.
.+
The second method has the advantage of not requiring the computa­
tion of any linear combinations. It has the disadvantage of requiring that we find a
matrix inverse. However, using a CAS will facilitate finding a matrix inverse, so in
general the second method is preferable to the first. For certain problems, though,
the first method may be just as easy to use. In any event, we are about to describe yet
a third approach, which you may find best of all.
Remark
The Gauss-Jordan Melhod for Compuling a Change-of-Basis Malrix
Finding the change-of-basis matrix to a standard basis is easy and can be done
by inspection. Finding the change-of-basis matrix from a standard basis is almost
as easy, but requires the calculation of a matrix inverse, as in Example 6.46. If we do
it by hand, then (except for the 2 X 2 case) we will usually find the necessary inverse
by Gauss-Jordan elimination. We now look at a modification of the Gauss-Jordan
method that can be used to find the change-of-basis matrix between two nonstandard
bases, as in Example 6.47.
Suppose B = {u 1 , . . . , u"} and C = {v1 , , vn } are bases for a vector space V
and Pc +-B is the change-of-basis matrix from B to C. The ith column of P is
.
•
.
so U; = p 1;v1 + · · · + P n ivn - If £ is any basis for V, then
[ u; ]t; = [ P1 ;V 1
+ · · · + Pnivn l t: = P1d v1 J t: + · · · + Pnd vn l t:
This can be rewritten in matrix form as
410
Chapter 6 Vector Spaces
which we can solve by applying Gauss-Jordan elimination to the augmented matrix
[vn l t: I [ u; ] EJ
There are n such systems of equations to be solved, one for each column of Pc,_ 8,
but the coefficient matrix [ [ v1 ] E • • • [ vn ] E ] is the same in each case. Hence, we can
solve all the systems simultaneously by row reducing the n X 2n augmented matrix
[ [v 1 L�·
[ [v1 J t:
···
···
[ vn l t: l [ u 1 J t:
···
[ un l t: l = [ C I B J
Since {V1 ' . . . ' vn } is linearly independent, so is { [ v I l E> ' [ vn l E} ' by Theorem 6. 7.
Therefore, the matrix C whose columns are [ v1 ] E> , [ vn ] E has the n X n identity
matrix I for its reduced row echelon form, by the Fundamental Theorem. It follows
that Gauss-Jordan elimination will necessarily produce
•
•
•
•
•
•
[ C I B J ---+ [ I I P J
where P = Pc<-- 8.
We have proved the following theorem.
Theorem 6 . 1 3
Let l3 = {u 1 , , un } and C = {v1 , . . . , vn } be bases for a vector space V. Let
B = [ [ u 1 l t: . . . [ un l t: l and C = [ [v 1 ] ,<: . . . [vn l t: L where £ is any basis for V.
Then row reduction applied to the n X 2n augmented matrix [ C I B ] produces
[ C I B J ---+ [ I I Pc.-al
.
•
.
If £ i s a standard basis, this method i s particularly easy t o use, since i n that case
B = PE<-- B and C = PE<-- C · We illustrate this method by reworking the problem in
Example 6.47.
Exa m p l e 6 . 4 8
Rework Example 6.47 using the Gauss-Jordan method.
Taking £ to be the standard basis for M22 , we see that
Solution
B = PE<-- B =
[l f l
0 0
0
1 0
0 0
Row reduction produces
IC I BJ
�
�
[l
1
1
1
1
0 1
0 0 1
1
0
0
0
0
0
1
0
and C = PE.-c =
[
f: l
� [l
-�]
0
1
0
0
�
(Verify this row reduction.) It follows that
as we found before.
P
,�,
0 -1
1
-1
1 0
0 0
[l : :
0 1
0 0
0 0 0 1 0 -1
0 0 0 -1
1
0 1 0 0 1 0
0 0
0 0 0
-�]
I
Section Change of Basis
6.3
Exercises 6 . 3
[x] 8 [x]c x
[x]
In Exercises 1 -4:
(a) Find the coordinate vectors
and
of with
respect to the bases B and C, respectively.
(b) Find the change-of-basis matrix Pc,_8from B to C.
(c) Use your answer to part (b) to compute c, and
compare your answer with the one found in part (a).
(d) Find the change-of-basis matrix P8,_ c from C to B.
(e) Use your answers to parts (c) and (d) to compute
and compare your answer with the one found in part (a).
x = [ � ], = { [ � ], [ � ] } ,
= { [�l [ �] }
x = [ �l = { [�l [ � ] }
= { [�l [�] }
x � [ H � un m. [�J l.
c � m H:WJ l
4. x � [; J s � m J m . [ m
c � m rn rn J l
.
x
== = = = == - = =
1.
[ x ] 8,
B
C
2.
in � 2
_
_
B
,
in � 2
C
3.
B
_
in � '
in � '
In Exercises 5-8, follow the instructions for Exercises 1 -4
using p ( x) instead of
5. p ( x) 2 - x, B { l , x}, C {x, 1 + x} in 0\
6. p(x) 1 + 3x, B = { l + x, 1 x},
C = {2x, 4} in 0 \
7. p(x) 1 + x 2, B { l + x + x 2, x + x 2, x 2},
C
411
{ l , x, x 2} in rzf 2
8. p(x) 4 2x - x 2, B
C = { l , 1 + x, x 2} in rzf 2
{x, 1 + x2 , x + x 2},
x.
= [ � � ], =
c = { [ � - � l [ � � ], [ � � ], [ � � ] }
= [l 1]
= { [ � � l [ � �l [ � �l [ � � ] } ,
= { [ � � l [ � �l [ � � l [ � � ] }
.
x
== - =
=
=
In Exercises 9 and 1 0, follow the instructions for
Exercises 1 -4 using A instead of
9. A
B
_
10. A
1 1
the standard basis,
in M22
,
B
C
in M22
In Exercises 1 1 and 12, follow the instructions for
Exercises 1 -4 usingf(x) instead of
1 1 . f(x) 2 sin x 3 cos x, B {sin x + cos x, cos x},
C {sin x + cos x, sin x - cos x} in span (sin x, cos x)
12. f(x) = sin x, B = {sin x + cos x, cos x},
C {cos x - sin x, sin x + cos x} in span (sin x, cos x)
13. Rotate the xy-axes in the plane counterclockwise
through an angle () 60° to obtain new x' y' -axes.
Use the methods of this section to find (a) the
x'y'-coordinates of the point whose xy-coordinates
are (3, 2) and (b) the xy-coordinates of the point
whose x'y' -coordinates are (4, - 4).
14. Repeat Exercise 1 3 with ()
=
1 35°.
= { [ � ] [�] }
[ � -�]
15. Let B and C be bases for � 2 . If C
,
the change-of-basis matrix from B to C is
Pc<-B =
_
find B.
16. Let B and C be bases for rzf 2 . If B = {x, 1 + x,
1 x + x2 } and the change-of-basis matrix
from B to C is
-
find C.
and
412
Chapter 6 Vector Spaces
In calculus, you learn that a Taylor polynomial of degree n
about a is a poly nomial of the form
p(x) = a0 + a 1 (x - a) + a 2 (x - a) 2 + · · · + a " (x - a) "
where a n * 0. In other words, it is a polynomial that has
been expanded in terms ofpowers of x - a instead ofpow­
ers of x. Taylor poly nomials are very useful for approximat­
ingfunctions that are "well behaved" near x = a.
The set B = { l , x - a, (x - a) 2 , , (x - a) " } is a basis
for rtl'n for any real number a. (Do you see a quick way to
show this? Try using Theorem 6. 7.) This fact allows us to use
the techniques of this section to rewrite a poly nomial as a
Taylor poly nomial about a given a.
17. Express p (x) = 1 + 2x - 5x2 as a Taylor polynomial
about a = 1 .
•
•
•
18. Express p ( x) = 1 + 2 x - sx2 as a Taylor polynomial
about a = - 2 .
19. Express p (x) = x3 as a Taylor polynomial about a = - 1 .
20. Express p ( x) = x3 as a Taylor polynomial about a = t .
21. Let B, C, and V be bases for a finite-dimensional vec­
tor space V. Prove that
Pv +-cPc+-B = Pv +- B
22. Let V be an n-dimensional vector space with basis
B = {v 1 ,
, vn }. Let P be an invertible n X n matrix
.
•
.
and set
fo r i = 1 , . . . , n. Prove that C = {u 1 ,
for V and show that P = Ps+-c·
•
.
•
, uJ is a basis
linear Tra n st o r m a l i o n s
We encountered linear transformations in Section 3 . 6 in the context o f matrix trans­
formations from IR " to !R m . In this section, we extend this concept to linear transfor­
mations between arbitrary vector spaces.
A linear transformation from a vector space V to a vector space
W is a mapping T : V ---+ W such that, for all u and v in V and for all scalars c,
D e fi n it i o n
1. T(u + v) = T(u) + T(v)
2. T(cu) = cT(u)
It is straightforward to show that this definition is equivalent to the requirement
that T preserve all linear combinations. That is,
T : V ---+ W is a linear transformation if and only if
T(c 1 v1 + c2v2 + · · · + ckvk ) = c 1 T(v1 ) + c2 T(v2 ) + · · · + ck T(vk )
for all v1 , . . . , vk in V and scalars c 1 , . . . , ck .
Exa m p l e 6 . 4 9
Every matrix transformation is a linear transformation. That is, if A is an m X n
matrix, then the transformation TA : IR " ---+ !R m defined by
TA ( x) = Ax for x in IR "
is a linear transformation. This is a restatement of Theorem 3.30.
Section 6.4 Linear Transformations
Exa m p l e 6 . 5 0
Define T : Mnn ---+ Mnn by T(A)
Solution
=
413
A T. Show that T is a linear transformation.
We check that, for A and B in Mnn and scalars c,
T(A + B)
and
=
(A + Bf = A T + B T = T(A) + T(B)
T(cA)
=
(cAf = cA T = cT(A)
Therefore, T is a linear transformation.
Exa m p l e 6 . 5 1
Let D be the differential operator D : 0J ---+ '2F defined by D(j)
linear transformation.
= f' .
Show that D is a
Let f and g be differentiable functions and let c be a scalar. Then, from
calculus, we know that
Solution
D(f + g)
and
=
(j + g) '
D(cj)
=
=
(cf) '
f' + g'
=
cf'
=
=
D (j) + D(g)
cD(j)
Hence, D is a linear transformation.
In calculus, you learn that every continuous function on [a, b] is integrable. The
next example shows that integration is a linear transformation.
Exa m p l e 6 . 5 2
Define S : � [ a, b ] ---+
Solution
IR:
by S(j)
=
J: f (x) dx. Show that S is a linear transformation.
Let f and g be in � [a, b] . Then
S(f + g)
=
r(j + g) (x) dx
r(j(x) + g(x)) dx
rf(x) dx + rg(x) dx
a
=
a
=
=
and
S(cf)
a
S(f) + S(g)
=
r(cj)(x) dx
fcf(x) dx
cff(x) dx
a
=
a
=
=
It follows that S is linear.
a
a
cS(f)
414
Chapter 6 Vector Spaces
Exa m p l e 6 . 5 3
Show that none of the following transformations is linear:
(a) T : M22 ---+ IR defined by T(A) = <let A
(b) T : IR ---+ IR defined by T(x) = 2x
(c) T : IR ---+ IR defined by T(x) = x + 1
In each case, we give a specific counterexample to show that one of the
properties of a linear transformation fails to hold.
Solution
(a) Let A =
[� �]
and B =
[ � � ].
[� �l
I � �I
I� � I I � � I
Then A + B =
T(A + B ) = <let (A + B ) =
But
T(A ) + T( B ) = detA + detB =
so
= 1
+
=0+0=0
so T(A + B) * T(A) + T(B) and T is not linear.
(b) Let x = 1 and y = 2. Then
T ( x + y) = T ( 3 ) = 2 3 = 8 * 6 = 2 1 + 2 2 = T ( x) + T (y)
so T is not linear.
(c) Let x = 1 and y = 2. Then
T( x + y) = T( 3 ) = 3 + 1 = 4 * 5 = ( 1 + 1) + (2 + 1) = T ( x) + T(y)
Therefore, T is not linear.
Example 6.53(c) shows that you need to be careful when you encounter
the word "linear:' As a function, T(x) = x + 1 is linear, since its graph is a straight
line. However, it is not a linear transformation from the vector space IR to itself, since
it fails to satisfy the definition. (Which linear functions from IR to IR will also be linear
transformations?)
Remark
�
4
There are two special linear transformations that deserve to b e singled out.
Exa m p l e 6 . 5 4
(a) For any vector spaces V and W, the transformation T0 : V ---+ W that maps every
vector in V to the zero vector in W is called the zero transformation. That is,
T0 ( v) = 0 for all v in V
(b) For any vector space V, the transformation I : V ---+ V that maps every vector in V
to itself is called the identity transformation. That is,
J ( v) = v for all v in V
(If it is important to identify the vector space V, we may write Iv for clarity.) The
proofs that the zero and identity transformations are linear are left as easy exercises.
4
Section 6.4 Linear Transformations
415
Properties of linear Tra nsformations
_.,
Theorem 6 . 1 4
In Chapter 3, all linear transformations were matrix transformations, and their
properties were directly related to properties of the matrices involved. The fol­
lowing theorem is easy to prove for matrix transformations. (Do it! ) The full
proof for linear transformations in general takes a bit more care, but it is still
straightforward.
Let T : V ---+ W b e a linear transformation. Then:
a. T (O) = 0
b. T ( -v) = - T(v) for all v in V.
c. T (u - v) = T (u) - T (v) for all u and v in V.
We prove properties (a) and (c) and leave the proof of property (b) for
Exercise 2 1 .
Proof
_.,
(a) Let v be any vector in V. Then T (O) = T (Ov) = O T (v) = 0, as required. (Can you
give a reason for each step?)
(c) T(u - v) = T(u + ( - l )v) = T(u) + ( - l ) T(v) = T(u) - T(v)
Remark
Property (a) can be useful in showing that certain transformations are
not linear. As an illustration, consider Example 6.53(b ). If T(x) = 2x, then T(O) = 2 ° =
1 * 0, so T is not linear, by Theorem 6. 14(a) . Be warned, however, that there are lots
of transformations that do map the zero vector to the zero vector but that are still not
linear. Example 6.53(a) is a case in point: The zero vector is the 2 X 2 zero matrix 0,
so T( 0) = det 0 = 0, but we have seen that T(A) = det A is not linear.
The most important property of a linear transformation T : V ---+ W is that T is
completely determined by its effect on a basis for V. The next example shows what
this means.
Exa m p l e 6 . 5 5
Suppose T is a linear transformation from IR 2 to <!J' 2 such that
[�]
[ � J [�J
r
Find r
_.,
Solution
-
span (B) . Solving
- 3x + x2 and r
[�]
=
1 - x2
.
and r
Since B
= 2
=
{ [ � ] , [ � ] } is a basis for IR2 (why?), every vector in IR2 is in
416
Chapter 6 Vector Spaces
we find that c 1 = - 7 and c2 = 3. Therefore,
= - 7r
[ � ] [�]
+ 3T
= - 7 ( 2 - 3x + x 2 ) + 3 ( 1 - x 2 )
= - 1 1 + 21x - 1 0x 2
Similarly, we discover that
so
r
[:J
[ :]
= ( 3a - 2 b)
(
[�J
[ �J
[�]
+ ( b - a)
= r ( 3 a - 2b )
+ ( b - a)
= ( 3 a - 2b ) r
+ ( b - a) r
[�]
[�] )
[�J
= ( 3 a - 2b ) ( 2 - 3x + x 2 ) + ( b - a ) ( l - x 2 )
= ( Sa - 3b) + ( - 9a + 6b ) x + ( 4a - 3b ) x 2
II-
(Note tha\by setting a = - 1 and b = 2, we recover the solution r
2lx - lOx
.)
[ -�]
= -11 +
4
The proof of the general theorem is quite straightforward.
Theorem 6 . 1 5
Let T : V -+ W be a linear transformation and let l3 = {v 1 , , vn } be a spanning
set for V. Then T ( B ) = { T ( v1 ) , , T ( vn ) } spans the range of T.
•
•
•
•
•
•
The range of T is the set of all vectors in W that are of the form T(v), where
v is in V. Let T(v) be in the range of T. Since l3 spans V, there are scalars c 1 ,
, en
such that
Proof
.
.
•
Applying T and using the fact that it is a linear transformation, we see that
T ( v) = T ( c 1 v 1 +
· · · + cnvn )
=
c 1 T ( v1 ) +
In other words, T(v) is in span ( T(B)), as required.
· · · + cn T (vn )
Theorem 6. 1 5 applies, in particular, when l3 is a basis for V. You might guess that,
in this case, T(B) would then be a basis for the range of T. Unfortunately, this is not
always the case. We will address this issue in Section 6.5.
Composition of linear Tra nsformations
In Section 3.6, we defined the composition of matrix transformations. The definition
extends to general linear transformations in an obvious way.
Section 6.4 Linear Transformations
S T
0
is read of
"S
T:'
411
D e fi n it i o n If T : U ---+ V and S : V ---+ W are linear transformations, then the
composition of S with T is the mapping S 0 T, defined by
(S 0 T) (u)
=
S( T(u))
where u is in U.
Observe that S 0 T is a mapping from U to W (see Figure 6.6) . Notice also that for
the definition to make sense, the range of T must be contained in the domain of S.
u
U•
T
•
T(u)
v
w
s
--+
S(T(u))
•
=
(S T)(u)
0
Composition of linear transformations
Figure 6 . 6
Exa m p l e 6 . 5 6
Let T : IR 2 ---+ <!/' 1 and S : <!/' 1 ---+ <!/' 2 be the linear transformations defined by
Find (S 0 T)
Solution
[ �]
_
r[ :]
=
a + (a + b )x and S(p(x))
and (S 0 T)
=
xp (x)
[ :].
We compute
(S 0 T)
[ _� ] s( r[ _� ] )
=
=
and
(S 0
=
S(3 + (3 - 2)x)
=
S(3 + x)
=
S(a + (a + b)x)
=
x(a + (a + b)x)
=
x(3 + x)
3x + x 2
T) [ :J ( r [ :] )
=
s
=
ax + (a + b)x 2
Chapter 3 showed that the composition of two matrix transformations was
another matrix transformation. In general, we have the following theorem.
Theorem 6 . 1 6
If T : U ---+ V and S : V ---+ W are linear transformations, then S 0 T : U ---+ W is a
linear transformation.
418
Chapter 6 Vector Spaces
Proof
Let u and v be in U and let c be a scalar. Then
(S T) (u + v)
0
=
=
=
=
and
(S T)(cu)
0
S( T(u + v))
S( T(u) + T(v))
S( T(u)) + S(T(v))
(S T)(u) + (S T) (v)
=
=
=
=
0
0
since Tis linear
since is linear
S
since Tis linear
since is linear
S( T(cu))
S(cT(u))
cS( T(u))
c(S T) (u)
S
0
Therefore, S T is a linear transformation.
0
The algebraic properties of linear transformations mirror those of matrix trans­
formations, which, in turn, are related to the algebraic properties of matrices. For
example, composition of linear transformations is associative. That is, if R, S, and T
are linear transformations, then
provided these compositions make sense. The proof of this property is identical to
that given in Section 3.6.
The next example gives another useful (but not surprising) property of linear
transformations.
Exa m p l e 6 . 5 1
Let S : U � V and T : V � W be linear transformations and let I : V � V be the iden­
tity transformation. Then for every v in V, we have
( T J)(v)
0
=
T(I(v))
=
T(v)
Since T I and T have the same value at every v in their domain, it follows that
T I = T. Similarly, I S = S.
0
o
o
Remark
The method of Example 6.57 is worth noting. Suppose we want to show
that two linear transformations Ti and T2 (both from V to W ) are equal. It suffices to
show that Ti ( v) = T2 (v) for every v in V.
Further properties of linear transformations are explored in the exercises.
Inverses or Linear Transrormalions
D e fi n il i O D A linear transformation T : V � W is invertible if there is a linear
transformation T' : W � V such that
T' T = Iv and T T ' = Iw
In this case, T' is called an inverse for T.
o
0
Section 6.4 Linear Transformations
419
R e m a rks
The domain V and codomain W of T do not have to be the same, as they do in
the case of invertible matrix transformations. However, we will see in the next section
that V and W must be very closely related.
•
The requirement that T' be linear could have been omitted from this definition.
For, as we will see in Theorem 6.24, if T' is
mapping from W to V such that
T' 0 T = Iv and T 0 T' = Iw, then T' is forced to be linear as well.
•
If T' is an inverse for T, then the definition implies that T is an inverse for T'.
Hence, T' is invertible too.
•
any
Exa m p l e 6 . 5 8
Verify that the mappings T : IR 2 ---+ r!J> 1 and T'
r [ : ] a + (a +
=
:
r!J> 1
b ) x and T ' ( c
are inverses.
and
IR 2 defined by
+ [d � J
dx)
=
We compute
Solution
(T'
---+
o
r) [ :J r' (r [ :] ) T' (a + (a + [ (a + :) _ a ] [ :]
+
+ r[ ] + +
+
=
( T 0 T ') (c
dx)
Hence, T ' T
other.
o
=
=
b ) x)
=
T( T ' ( c
dx))
IIR2 and T 0 T '
=
=
c
d-c
=
=
c
(c
( d - c )) x
=
c
dx
I21' 1 • Therefore, T and T' are inverses of each
4
As was the case for invertible matrices, inverses of linear transformations are
unique if they exist. The following theorem is the analogue of Theorem 3.6.
Theorem 6 . 1 1
If T is an invertible linear transformation, then its inverse is unique.
The proof is the same as that of Theorem 3.6, with products of matrices re­
placed by compositions of linear transformations. (You are asked to complete this
proof in Exercise 3 1 .)
Proof
Thanks to Theorem 6. 1 7, if T is invertible, we can refer to the inverse of T. It will
be denoted by r - 1 (pronounced " T inverse"). In the next two sections, we will ad­
dress the issue of determining when a given linear transformation is invertible and
finding its inverse when it exists.
480
..
I
Chapter 6 Vector Spaces
Exercises 6 . 4
In Exercises 1 - 12, determine whether T is a linear
transformation.
1. T : M22 M22 defined by
a b
a+b
=
r
c d
0
c+d
2. T : M22 M22 defined by
1
w-z
w x
r
1
x-y
y z
3. T : Mnn Mnn defined by T (A ) = AB, where B is a
fixed n X n matrix
4. T : Mnn Mnn defined by T (A ) = AB - BA, where B
is a fixed n X n matrix
5. T : Mnn IR defined by T (A ) = tr (A )
6. T : Mnn IR defined by T(A) = a ll a 22 a nn
7. T : Mnn IR defined by T (A ) = rank (A )
8. T : <JP2 <!P2 defined by T(a + bx + cx 2 ) = (a + 1 ) +
(b + l) x + (c + l ) x 2
9. T : <!P2 <!P2 defined by T(a + bx + cx 2 ) = a +
b(x + 1 ) + b(x + 1 ) 2
10. T : gjf ---,l> gji defined by T (j) = f (x 2 )
1 1 . T : gji gji defined by T (j) = (j(x) ) 2
12. T : gji IR defined by T (j) = j (c) , where c is a fixed
OJ
--,lo
[ ] [
[ ] [
]
--,lo
--,lo
16. Let T : <JP2 <JP2 be a linear transformation for which
T( l) = 3 - 2x, T(x) = 4x - x2 , and T(x2 ) = 2 + 2x2
Find T(6 + x - 4x2 ) and T(a + bx + cx 2 ).
17. Let T : <!P2 <!P2 be a linear transformation for which
T( l + x) = 1 + x 2 , T(x + x 2 ) = x - x 2 ,
T( 1 + x 2) = 1 + x + x 2
--,lo
--,lo
Find T ( 4 - x + 3x 2 ) and T(a + bx + cx 2 ).
18. Let T : M22 --,lo IR be a linear transformation for which
--,lo
--,lo
--,lo
•
•
•
--,lo
--,lo
--,lo
--,lo
--,lo
scalar
13. Show that the transformations S and T in Exam­
ple 6.56 are both linear.
14. Let T : IR 2 --,lo IR 3 be a linear transformation for which
[� J [�J
[ �]
[�]
[ - �] [ �J.
Find r
15. Let T : IR
T
Find r
and r
2
--,lo
=
.
<JP2 be a linear transformation for which
1 - 2x and T
and r
_
=
x + 2x 2
Find r
[� OJ [� ]
[� ] [� �]
[� � ] [ : �].
r
0
r
1
0
=
1' r
=
3, r
1
0
=
2'
=
4
and r
19. Let T : M22
--,lo IR be a linear transformation. Show that
there are scalars a, b, c, and d such that
[ ; :J
[ ; :J
r
for all
=
aw + bx + cy + dz
in M22 .
20. Show that there is no linear transformation T : IR 3
such that
{]
� 1
+ x, r
t! J
�
[�]
�
--,lo
2 - x + x' .
- 2 + 2x'
21. Prove Theorem 6. 14(b) .
22. Let {v1 , , vn } be a basis for a vector space V and
•
.
•
<JP 2
let T : V --,lo V be a linear transformation. Prove that if
T (v1) = V1, T (v2) = V2 . . . , T ( vn ) = vn , then T is the
identity transformation on V.
� 23. Let T : <JP n --,lo <JP n be a linear transformation such that
T(x k ) = kx k - I for k = 0, 1 , . . . , n. Show that T must
be the differential operator D.
Section 6.5 The Kernel and Range of a Linear Transformation
24. Let v1 , • • • , v" be vectors in a vector space V and let
T : V ---+ W be a linear transformation.
( a ) If { T(v1 ), . . . , T(vn )} is linearly independent in W,
show that {v1 , • • • , vJ is linearly independent in V.
(b) Show that the converse of part (a) is false.
That is, it is not necessarily true that if
{v1 , . • . , vn } is linearly independent in V, then
{T(v1 ), • • • , T(v)} is linearly independent in W.
Illustrate this with an example T : IR 2 ---+ IR 2 •
25. Define linear transformations S : IR 2 ---+ M22 and
T : IR 2 ---+ IR 2 by
s[ � ] [ :
a
=
b
Compute (S T)
0
compute ( T S)
0
�b
[�]
[;]?
a
]
[ �] [ : ]
[;].
2c d
_
and T
and (S T)
0
Can you
If so, compute it.
0
0
0
---+
Find (S T) (p(x)) and ( T S) (p(x)). [Hint: Remember
the Chain Rule.]
� 28. Define linear transformations S : <;if n ---+ <;if n and
T : <;if n -+ <;if n by
0
S(p (x)) = p (x + 1 ) and T(p(x))
Find (S T)(p(x)) and ( T S)(p(x)).
0
0
II
s[x] [ ]
[] [ ]
In Exercises 29 and 30, verify that S and T are inverses.
4x + y
=
29. S : IR 2 -+ IR 2 defined by
and T: IR 2 -+ IR 2
3x + y
y
x = x-y
defined by T
y
- 3x + 4y
30. S : <!P 1 ---+ <!P 1 defined by S(a + bx) =
( - 4a + b) + 2ax and T : <!P 1 ---+ <!P 1 defined by
T(a + bx) = b/2 + (a + 2b)x
31. Prove Theorem 6. 17.
32. Let T : V ---+ V be a linear transformation such that
T o T = I.
(a) Show that {v, T(v)} is linearly dependent if and
only if T(v) = ±v.
(b) Give an example of such a linear transformation
with V = IR 2 •
33. Let T : V ---+ V be a linear transformation such that
T T = T.
( a) Show that {v, T(v)} is linearly dependent if and
only if T(v) = v or T(v) = 0.
0
26. Define linear transformations S : <!P 1 ---+ <!P 2 and
T : <!P 2 ---+ <!P 1 by
S(a + bx) = a + (a + b )x + 2bx 2
and
T(a + bx + cx 2 ) = b + 2cx
Compute (S T)(3 + 2x - x 2 ) and
(S T)(a + bx + cx 2 ) . Can you compute
( T S) (a + bx)? If so, compute it.
� 27. Define linear transformations S : <;if " <;if " and
T : <!P n -+ <!P n by
S(p(x)) = p(x + 1 ) and T(p(x)) = p ' (x)
0
481
=
xp ' (x)
(b) Give an example of such a linear transformation
with V = IR 2 .
The set of all linear transformations from a vector space V
to a vector space W is denoted by ;£', ( V, W ) . If S and T are
in ;£', ( V, W ) , we can define the sum S + T of S and T by
(S + T) (v) = S(v) + T(v)
for all v in V If c is a scalar, we define the scalar multiple
cT of T by c to be
(cT)(v) = cT(v)
for all v in V Then S + T and cT are both transformations
from V to W
34. Prove that S + T and cT are linear transformations.
35. Prove that H', ( V, W ) is a vector space with this addi­
tion and scalar multiplication.
36. Let R, S, and T be linear transformations such that the
following operations make sense. Prove that:
( a)
(b)
R o (S + T) = R o S + R o T
c(R S) = (cR) S = R (cS) for any scalar c
0
0
0
T h e Kernel a n d R a n g e o f a l i n e a r Transfo r m a t i o n
The null space and column space are two o f the fundamental subspaces associated
with a matrix. In this section, we extend these notions to the kernel and range of a
linear transformation.
482
Chapter 6 Vector Spaces
wordEnglish word
is derived froma
theThe
Old
form oforthe"graiwordn:' Like ameaning
"seed"
kerneltrans­of
corn,
the
kernel
of
a
linear
formation
is itist"core"
orinforma­
"seed" in
thetion
sense
that
carries
about ofmanythe transformation.
of the important
properties
kernel
cyrnel,
corn,
Exa m p l e 6 . 5 9
D e fi n it i o n Let T: V---+ W be a linear transformation. The kernel of T, denoted
ker ( T), is the set of all vectors in V that are mapped by T to 0 in W. That is,
ker ( T) = {v in V: T ( v) = O}
The range of T, denoted range(T), is the set of all vectors in W that are images of
vectors in V under T. That is,
range ( T ) = { T ( v) : v in V}
= {w in W : w = T ( v) for some v in V}
Let A be an m X n matrix and let T = TA be the corresponding matrix transformation
from !R n to !R m defined by T(v) = Av. Then, as we saw in Chapter 3, the range of T is
the column space of A.
The kernel of T is
ker ( T) = {v in !R n : T (v) = O}
= {v in !R n : Av = O}
= null (A )
In words, the kernel of a matrix transformation is just the null space of the corre­
sponding matrix.
Exa m p l e 6 . 6 0
Find the kernel and range of the differential operator D : <;if 3 ---+ <!P 2 defined by
D (p(x)) = p'(x).
Solulion
Since D (a + bx + cx 2 + dx 3 )
=
b + 2cx + 3dx 2 , we have
ker ( D ) = {a + bx + cx 2 + dx 3 : D(a + bx + cx 2 + dx 3 ) = O}
= {a + bx + cx 2 + dx 3 : b + 2cx + 3dx 2 = O}
But b + 2cx + 3dx 2
d = 0. Therefore,
=
0 if and only if b
=
2c
=
3d = 0, which implies that b = c =
ker ( D ) = {a + bx + cx 2 + dx 3 : b = c = d = O}
= {a : a in IR}
In other words, the kernel of D is the set of constant polynomials.
The range of D is all of<!P 2 , since every polynomial in <!P 2 is the image under D (i.e.,
the derivative) of some polynomial in <!P 3 . To be specific, if a + bx + cx 2 is in <!P 2 , then
Section 6.5 The Kernel and Range of a Linear Transformation
Exa m p l e 6 . 6 1
Let S : <!/' 1
�
IR be the linear transformation defined by
p x rp (x)dx
S ( ( ))
Find the kernel and range of S.
Solution
483
=
0
In detail, we have
S(a + bx) r (a + bx)dx
= [ ax + !x I
= ( a + %) - o = a + %
a + bx: S(a + bx)
{ a + bx : a + % 0 }
{ a + bx : a - !}
{ - !+ bx }
y
=
0
2
b
2
ker ( S )
Therefore,
1
2
=
{
=
O}
=
b
2
=
Ify b + bx
then rydx
Figure 6 . 1
--
=
2
0
=
'
0
Geometrically, ker(S ) consists of all those linear polynomials whose graphs have the
property that the area between the line and the x-axis is equally distributed above and
below the axis on the interval [O, l ] (see Figure 6.7).
The range of S is IR, since every real number can be obtained as the image under
S of some polynomial in <!/' 1 . For example, if is an arbitrary real number, then
so
Exa m p l e 6 . 6 2
a a
=
{ dx
0
S( ) .
a
=
a
x -o a
[a n
=
a
=
Let T : M22 � M22 b e the linear transformation defined by taking transposes:
T(A) = A T. Find the kernel and range of T.
Solution
We see that
ker ( T)
{A in M22 : T (A) = O}
= {A in M22 : AT = O}
But if A T = 0, then A = (A T ) T = o T = 0. It follows that ker (T) = { O}.
Since, for any matrix A in M22 , we have A = (A T ) T = T(A T ) (and A T is in M22 ),
we deduce that range(T) = M22 .
=
In all of these examples, the kernel and range of a linear transformation are sub­
spaces of the domain and co domain, respectively, of the transformation. Since we are
generalizing the null space and column space of a matrix, this is perhaps not surpris­
ing. Nevertheless, we should not take anything for granted, so we need to prove that
it is not a coincidence.
484
Chapter 6 Vector Spaces
Theorem 6 . 1 8
Let T : V ---+ W b e a linear transformation. Then:
a. The kernel of T is a subspace of V.
b. The range of T is a subspace of W.
(a) Since T (O) = 0, the zero vector of V is in ker(T), so ker(T) is nonempty.
Let u and v be in ker ( T ) and let c be a scalar. Then T (u) = T (v) = 0, so
Proof
T(u + v) = T(u) + T(v) = 0 + 0 = 0
and
T(cu) = cT(u) = c O = 0
Therefore, u + v and c u are in ker( T), and ker(T) is a subspace of V.
(b) Since 0 = T(O), the zero vector of W is in range(T), so range(T) is nonempty.
Let T (u) and T (v) be in the range of T and let c be a scalar. Then T(u) + T(v) =
T(u + v) is the image of the vector u + v. Since u and v are in V, so is u + v, and
hence T(u) + T(v) is in range (T). Similarly, cT(u) = T(c u) . Since u is in V, so is cu,
and hence cT(u) is in range(T). Therefore, range(T) is a nonempty subset of W that is
closed under addition and scalar multiplication, and thus it is a subspace of W.
Figure 6.8 gives a schematic representation of the kernel and range of a linear
transformation.
rne
v
(T)
�
a
0
T
The kernel and range of
•o
g
W
Figure 6 . 8
T : V --+ W
In Chapter 3, we defined the rank of a matrix to be the dimension of its column
space and the nullity of a matrix to be the dimension of its null space. We now extend
these definitions to linear transformations.
Let T : V ---+ W be a linear transformation. The rank of T is the
dimension of the range of T and is denoted by rank( T). The nullity of T is
the dimension of the kernel of T and is denoted by nullity( T).
Definition
Exa m p l e 6 . 6 3
If A is a matrix and T = TA is the matrix transformation defined by T (v) = Av, then
the range and kernel of T are the column space and the null space of A, respectively,
by Example 6.59. Hence, from Section 3.5, we have
rank ( T) = rank (A ) and nullity ( T) = nullity (A )
Section 6.5 The Kernel and Range of a Linear Transformation
Exa m p l e 6 . 6 4
485
Find the rank and the nullity of the linear transformation D : l!P 3 � <!J 2 defined by
D(p(x)) = p'(x).
Solution
In Example 6.60, we computed range ( D ) = <!J 2 , so
rank ( D ) = dim <!f 2 = 3
The kernel of D is the set of all constant polynomials: ker (D) = {a : a in IR} = {a
in IR}. Hence, { 1 } is a basis for ker (D), so
·
1:a
nullity ( D ) = dim ( ker ( D )) = 1
Exa m p l e 6 . 6 5
Find the rank and the nullity of the linear transformation S : <!J 1
S(p(x)) =
Solution
r
0
�
IR defined by
p (x) dx
From Example 6.6 1 , range(S ) = IR and rank(S ) = dim IR = 1. Also,
ker ( S ) =
{�
-
+ bx : b in IR
}
= {b ( - t + x) : b in IR}
= span ( - t + x)
so { -t + x} is a basis for ker(S ). Therefore, nullity(S ) = dim(ker (S )) = 1 .
Exa m p l e 6 . 6 6
Find the rank and the nullity of the linear transformation T : M22 � M22 defined by
T (A) = A T.
Solution
In Example 6.62, we found that range( T ) = M22 and ker( T) = {O}. Hence,
rank ( T )
=
dim M22
=
4 and nullity ( T ) = dim{O} = 0
In Chapter 3, we saw that the rank and nullity of an m X n matrix A are related
by the formula rank(A) + nullity(A) = n. This is the Rank Theorem (Theorem 3.26).
Since the matrix transformation T = TA has !R n as its domain, we could rewrite the
relationship as
rank (A ) + nullity (A ) = dim !R n
This version of the Rank Theorem extends very nicely to general linear transforma­
tions, as you can see from the last three examples:
rank ( D ) + nullity ( D ) = 3 + 1 = 4 = dim l!P3
rank ( S ) + nullity ( S) = 1 + 1 = 2 = dim <!J 1
rank ( T ) + nullity ( T ) = 4 + 0 = 4 = dim M22
Example 6.64
Example 6.65
Example 6.66
486
Chapter 6 Vector Spaces
Theorem 6 . 1 9
The Rank Theorem
Let T : V ---+ W be a linear transformation from a finite-dimensional vector space
V into a vector space W. Then
rank ( T ) + nullity ( T )
=
dim V
In the next section, you will see how to adapt the proof of Theorem 3.26 to prove
this version of the result. For now, we give an alternative proof that does not use
matrices.
Let dim V = n and let {v1, . . . , vk } be a basis for ker (T) [so that nullity( T) =
dim(ker( T)) = k] . Since {v1, . . . , vd is a linearly independent set, it can be extended
to a basis for V, by Theorem 6.28. Let B = {v1 , . . . , vk > vk+ 1 , . . . , vn } be such a basis.
If we can show that the set C = { T (vk + 1), . . . , T (vn ) } is a basis for range( T), then we
will have rank( T) = dim(range( T)) = n - k and thus
Proof
rank ( T ) + nullity ( T )
=
k + (n
-
k)
=
n
=
dim V
as required.
Certainly C is contained in the range of T. To show that C spans the range of T, let
T (v) be a vector in the range of T. Then v is in V, and since B is a basis for V, we can
find scalars c1, . . . , en such that
Since v1,
.
•
.
, vk are in the kernel of T, we have T (v1 )
T ( v)
=
=
=
= · · · =
T( vk )
=
0, so
T(c 1 V1 + . . . + ckvk + Ck+1 Vk+1 + . . . + cnvn )
c 1 T(v1 ) + · · · + ck T(vk ) + ck+1 T(vk+1 ) + · · · + cn T(vn )
Ck+ 1 T(vk+1 ) + . . . + c n T(vn )
This shows that the range of T is spanned by C.
To show that C is linearly independent, suppose that there are scalars ck + 1, . . . , en
such that
Then T ( ck + I Vk + I + + cnvn ) = 0, which means that Ck + ! Vk + l + + cnvn is in the
kernel of T and is, hence, expressible as a linear combination of the basis vectors
v1, . . . , vk of ker ( T)-say,
· · ·
· · ·
But now
and the linear independence of B forces c1 = = en = 0. In particular, ck + 1 = =
en = 0, which means C is linearly independent.
We have shown that C is a basis for the range of T, so, by our comments above, the
proof is complete.
· · ·
· · ·
We have verified the Rank Theorem for Examples 6.64, 6.65, and 6.66. In practice,
this theorem allows us to find the rank and nullity of a linear transformation with
only half the work. The following examples illustrate the process.
Section 6.5 The Kernel and Range of a Linear Transformation
Exa m p l e 6 . 6 1
�
Find the rank and nullity of the linear transformation T : <!/' 2
T(p(x)) = xp(x) . (Check that T really is linear.)
�
481
<!/' 3 defined by
In detail, we have
T ( a + bx + cx 2 ) = ax + bx 2 + cx 3
It follows that
ker ( T ) = {a + bx + cx 2 : T ( a + bx + cx 2 ) = O}
= {a + bx + cx 2 : ax + bx 2 + cx 3 = O}
= {a + bx + cx 2 : a = b = c = o}
= {o}
so we have nullity(T) = dim(ker( T)) = The Rank Theorem implies that
Solution
0.
rank ( T ) = dim <!/' 2 - nullity ( T ) = 3 -
0
= 3
Remarll In Example 6.67, it would be just as easy to find the rank of T first, since
{x, x2 , x3 } is easily seen to be a basis for the range of T. Usually, though, one of the two
(the rank or the nullity of a linear transformation) will be easier to compute; the Rank
Theorem can then be used to find the other. With practice, you will become better at
knowing which way to proceed.
Exa m p l e 6 . 6 8
Let W be the vector space of all symmetric 2 X 2 matrices. Define a linear transfor­
mation T : W � <!/' 2 by
r [ : �]
iJ)-'"--
= ( a - b) + ( b - c)x + ( c - a )x 2
(Check that T is linear.) Find the rank and nullity of T.
Solution
as follows:
The nullity of T is easier to compute directly than the rank, so we proceed
{ [ : �] [ : �] 0 }
{ [ : �]
{ [ : �] :(a
{ [ : �] }
{ [ � �] } ( [ � � ] )
{ [� �] }
ker ( T) =
:T
=
: ( a - b) + ( b - c ) x + ( c - a ) x 2 =
- b) = ( b - c) = ( c - a ) =
0}
O}
:a = b = c
= span
Therefore,
is a basis for the kernel of T, so nullity( T) = dim (ker ( T)) = 1 .
The Rank Theorem and Example 6.42 tell us that rank( T) = dim W - nullity( T) =
3 - 1 = 2.
488
Chapter 6 Vector Spaces
one-to-one and Onto Linear Tra nsformations
We now investigate criteria for a linear transformation to be invertible. The keys to
the discussion are the very important properties one-to-one and onto.
A linear transformation T : V ---+ W is called one-to-one if T maps
distinct vectors in V to distinct vectors in W. If range( T) = W, then T is called onto.
Definition
R e m a rks
•
The definition of one-to-one may be written more formally as follows:
T : V ---+ W is one-to-one if, for all u and v in V,
u * v implies that T ( u ) * T ( v)
The above statement is equivalent to the following:
T : V ---+ W is one-to-one if, for all u and v in V,
T ( u)
=
T ( v) implies that u
=
v
Figure 6.9 illustrates these two statements.
v
(a) T is
Fioure 6 . 9
•
one-to-one
w
v
(b) T is
not one-to-one
w
Another way to write the definition of onto is as follows:
T : V ---+ W is onto if, for all w in W, there is at least one v in V such that
w
=
T ( v)
In other words, given w in W, does there exist some v in V such that w = T(v) ? If,
for an arbitrary w, we can solve this equation for v, then T is onto (see Figure 6. 10).
Section 6.5 The Kernel and Range of a Linear Transformation
v
(a) T is
Figure 6 . 1 0
Exa m p l e 6 . 6 9
onto
w
v
(b) T is
not onto
489
w
r [;] [ � ]
Which of the following linear transformations are one-to-one? onto?
(') T
!I ' -> !I '
defined by
�
x
y
(b) D : <!/' 3 ---+ <!1' 2 defined by D(p(x)) = p' (x)
(c) T : M22 ---+ M22 defined by T (A) = A T
Solution
(a) Let r
[;: J [;: J
= r
[;: ] [;: ],
[; l m
. Then
so 2x 1 = 2x2 and x 1 - y 1 = x2 - y2 • Solving these equations, we see that x 1 = x2 and
so T is one-to-one.
y 1 = Yz . Hence,
=
T is not onto, since its range is not all of IR 3 . To be specific, there is no vector
It-"-
[; l
in !I ' 'uch th" r
�
(Why not?)
(b) In Example 6.60, we showed that range(D) = <!/' 2 , s o D is onto. D is not one­
to-one, since distinct polynomials in <!/' 3 can have the same derivative. For example,
x 3 -=F x 3 + 1 , but D(x3 ) = 3x2 = D(x3 + 1 ) .
(c) Let A and B b e in M22 , with T (A) = T(B). Then A T = B T, s o A = (A T ) T =
(B T l = B. Hence, T is one-to-one. In Example 6.62, we showed that range(T) = M22 •
Hence, T is onto.
4
It turns out that there is a very simple criterion for determining whether a linear
transformation is one-to-one.
Theorem 6 . 2 0
A linear transformation T : V ---+ W is one-to-one if and only if ker (T) = {O}.
490
Chapter 6 Vector Spaces
Proof
Assume that T is one-to-one. If v is in the kernel of T, then T(v) = 0. But
we also know that T (O) = 0, so T (v) = T (O). Since T is one-to-one, this implies that
v = 0, so the only vector in the kernel of T is the zero vector.
Conversely, assume that ker (T) = {O}. To show that T is one-to-one, let u and v be
in V with T (u) = T(v) . Then T (u - v) = T(u) - T(v) = 0, which implies that u - v
is in the kernel of T. But ker ( T ) = {O}, so we must have u - v = 0 or, equivalently,
u = v. This proves that T is one-to-one.
Exa m p l e 6 . 10
Show that the linear transformation T : IR 2 ---+ rtP 1 defined by
is one-to-one and onto.
Solulion
If
[�]
r[�]
= a + ( a + b)x
is in the kernel of T, then
0=
r[�]
= a + ( a + b)x
It follows that a = 0 and a + b = 0. Hence, b = 0, and therefore
quently, ker ( T) =
{ [�] }
[ � ] [ � ].
Conse-
, and T is one-to-one, by Theorem 6.20.
By the Rank Theorem,
rank ( T ) = dim IR 2 - nullity ( T ) = 2 - 0 = 2
Therefore, the range of T is a two-dimensional subspace of IR 2 , and hence
range(T) = IR 2 . It follows that T is onto.
�
Theorem 6 . 2 1
For linear transformations between two n-dimensional vector spaces, the proper­
ties of one-to-one and onto are closely related. Observe first that for a linear trans­
formation T : V ---+ W, ker (T) = {O} if and only if nullity(T) = 0, and T is onto if and
only if rank( T) = dim W. (Why?) The proof of the next theorem essentially uses the
method of Example 6.70.
Let dim V = dim W = n . Then a linear transformation T : V ---+ W is one-to-one
if and only if it is onto.
Proof
Assume that T is one-to-one. Then nullity( T) = 0 by Theorem 6.20 and the
remark preceding Theorem 6.2 1 . The Rank Theorem implies that
rank ( T ) = dim V - nullity ( T ) = n - 0 = n
Therefore, T is onto.
Conversely, assume that T is onto. Then rank(T) = dim W = n. By the Rank
Theorem,
nullity ( T ) = dim V - rank ( T ) = n - n = 0
Hence, ker( T ) = {O}, and T is one-to-one.
Section 6.5 The Kernel and Range of a Linear Transformation
491
In Section 6.4, we pointed out that if T : V ---+ W is a linear transformation, then
the image of a basis for V under T need not be a basis for the range of T. We can now
give a condition that ensures that a basis for V will be mapped by T to a basis for W.
Theorem 6 . 2 2
Let T : V ---+ W be a one-to-one linear transformation. If S = {v1 , . . . , vk} is a lin­
early independent set in V, then T (S ) = { T (v1 ), . . . , T(vk) } is a linearly indepen­
dent set in W.
Let c 1 , . . . , ck be scalars such that
Proof
Then T(c 1 v1 + · · · + ckvk) = 0, which implies that c 1 v1 + · · · + ckvk is in the kernel of
T. But, since T is one-to-one, ker ( T ) = {O}, by Theorem 6.20. Hence,
c 1 v1 +
· · ·
+ ckvk
=
0
But, since {v1 , • • • , vd is linearly independent, all of the scalars C; must be 0. Therefore,
{ T (v1 ), . . . , T (vk) } is linearly independent.
c o r o n a rv 6 . 2 3
Let dim V = dim W = n. Then a one-to-one linear transformation T : V ---+ W
maps a basis for V to a basis for W.
Let B = {v1 , . • . , vn } be a basis for V. By Theorem 6.22, T ( B) = { T (v 1 ), • • • ,
T ( vn ) } is a linearly independent set in W, so we need only show that T ( B ) spans
W. But, by Theorem 6. 15, T ( B ) spans the range of T. Moreover, T is onto, by Theo­
rem 6.2 1, so range(T) = W. Therefore, T ( B ) spans W, which completes the proof.
Proof
Exa m p l e 6 . 1 1
Let T : IR 2 ---+ \fP 1 be the linear transformation from Example 6.70, defined by
r
=
[ : J = a + (a + b)x
Then, by Corollary 6.23, the standard basis [
T ( [) { T ( e 1 ) , T ( e2 ) } of \fP 1 • We find that
T( e1 ) = r
[�]
=
= { e1 , e2 } for IR 2 is mapped to a basis
1 + x and T ( e2 ) = r
[ �] = x
It follows that { 1 + x, x} is a basis for \fP 1 .
We can now determine which linear transformations T : V ---+ W are invertible.
Theorem 6 . 2 4
A linear transformation T : V ---+ W is invertible if and only if it is one-to-one
and onto.
492
Chapter 6 Vector Spaces
Assume that T is invertible. Then there exists a linear transformation T - 1 :
W � V such that
Proof
T - 1 0 T = Iv and T 0 T - 1 = Iw
To show that T is one-to-one, let v be in the kernel of T. Then T (v) = 0. Therefore,
Y - 1 ( T (v)) = Y - 1 ( 0 ) =:> ( T - 1 0 T ) ( v) = 0
=:> I(v) = 0
=:> v = O
which establishes that ker ( T ) = { O}. Therefore, T is one-to-one, by Theorem 6.20.
To show that T is onto, let w be in W and let v = T - 1 (w) . Then
T ( v) =
=
=
=
T ( T - 1 ( w))
( T o T - 1 ) ( w)
I ( w)
w
which shows that w is the image of v under T. Since v is in V, this shows that T is onto.
Conversely, assume that T is one-to-one and onto. This means that nullity(T) = 0
and rank( T) = dim W. We need to show that there exists a linear transformation
T ' : W � V such that T' 0 T = Iv and T 0 T' = Iw.
Let w be in W. Since T is onto, there exists some vector v in V such that T(v) = w.
There is only one such vector v, since, if v' is another vector in V such that T ( v' ) = w,
then T(v) = T (v' ); the fact that T is one-to-one then implies that v = v' . It therefore
makes sense to define a mapping T' : W � V by setting T'(w) = v.
It follows that
( T ' o T ) ( v) = T ' ( T ( v)) = T ' ( w) = v
and
( T 0 T ' ) (w) = T ( T ' ( w)) = T ( v) = w
It then follows that T' 0 T = Iv and T 0 T' = Iw. Now we must show that T' is a linear
transformation.
To this end, let w 1 and w2 be in W and let c 1 and c2 be scalars. As above, let
T (v1 ) = w 1 and T(v2 ) = w2 • Then v1 = T ' (w1 ) and v2 = T' (w2 ) and
T ' ( c 1 w 1 + c2w2 ) = T ' ( c 1 T ( v1 ) + c2 T ( v2 ))
= T ' ( T ( c 1 v 1 + c2v2 ))
= I ( c 1 V 1 + CzVz )
Consequently, T' is linear, so, by Theorem 6. 17, T' = T - 1 •
Section 6. 5 The Kernel and Range of a Linear Transformation
The words are derived from
and
the"equalGreek
words meaning
meaning
;
'
and
"shape:'isomorphic
Thus, figurati
velspaces
y speak­have
ing,
vector
"equal shapes:'
isomorphism
isomorphic
isos,
morph,
Exa m p l e 6 . 12
493
Isomorphisms of Vector Spaces
We now are in a position to describe, in concrete terms, what it means for two vector
spaces to be "essentially the same:'
A linear transformation T : V � W is called an isomorphism if it
is one-to-one and onto. If V and W are two vector spaces such that there is an iso­
morphism from V to W, then we say that V is isomorphic to W and write V = W.
Definition
Show that <!P n - I and !R n are isomorphic.
Solution The process of forming the coordinate vector of a polynomial provides us
with one possible isomorphism (as we observed already in Section 6.2, although we did
not use the term isomorphism there). Specifically, define T : <;JP n - i � !R n by T(p(x) ) =
[p(x) ], 0, where £ = { 1 , x, . . . , x" - 1 } is the standard basis for <;JP n - i That is,
·
Theorem 6.6 shows that T is a linear transformation. If p(x)
a n - 1X n- l is in the kernel of T, then
=
a0 + a1x + · · · +
Hence, a0 = a1 = = a n - i = 0, so p(x) = 0. Therefore, ker ( T ) = {O}, and T is one­
to-one. Since dim <!P n- i = dim IR " = n, T is also onto, by Theorem 6.2 1 . Thus, T is
an isomorphism, and <!P n - i = IR " .
· · ·
Exa m p l e 6 . 13
Show that Mm n and !R m " are isomorphic.
Once again, the coordinate mapping from Mm n to !R m " (as in Example 6.36 )
is an isomorphism. The details of the proof are left as an exercise.
Solution
In fact, the easiest way to tell if two vector spaces are isomorphic is simply to
check their dimensions, as the next theorem shows.
494
Chapter 6 Vector Spaces
Theorem 6 . 2 5
Let V and W b e two finite-dimensional vector spaces (over the same field of scalars).
Then V is isomorphic to W if and only if dim V = dim W.
Let n = dim V. If V is isomorphic to W, then there is an isomorphism
T : V ---+ W. Since T is one-to-one, nullity( T) = 0. The Rank Theorem then implies
that
Proof
rank ( T )
=
dim V - nullity ( T )
=
n-0
=
n
Therefore, the range of T is an n-dimensional subspace of W. But, since T is onto,
W = range(T), so dim W = n, as we wished to show.
Conversely, assume that Vand Whave the same dimension, n. Let B = {v1 , , vn }
be a basis for V and let C = {w1, . . . , wn } be a basis for W. We will define a linear
transformation T : V ---+ W and then show that T is one-to-one and onto. An arbitrary
vector v in V can be written uniquely as a linear combination of the vectors in the
basis B-say,
.
•
.
We define T by
.-..
It is straightforward to check that T is linear. (Do so.) To see that T is one-to-one,
suppose v is in the kernel of T. Then
and the linear independence of C forces c1
= · · · =
en = 0. But then
so ker( T) = {O}, meaning that T is one-to-one. Since dim V = dim W, T is also onto,
by Theorem 6.21 . Therefore, T is an isomorphism, and V = W.
Exa m p l e 6 . 14
Show that u;g n and <lP n are not isomorphic.
Solution
Since dim u;g n
Theorem 6.25.
Exa m p l e 6 . 15
=
n* n+ 1
=
dim <lP n' u;g n and <lP n are not isomorphic, by
Let W be the vector space of all symmetric 2 X 2 matrices. Show that W is isomorphic
to IFR 3 .
In Example 6.42, we showed that dim W = 3. Hence, dim W = dim IFR 3 ,
so W = IFR 3 , by Theorem 6.25. (There is an obvious candidate for an isomorphism
T : W ---+ IFR 3 . What is it?)
Solution
Section 6.5 The Kernel and Range of a Linear Transformation
..
I
495
Remark
Our examples have all been real vector spaces, but the theorems we
have proved are true for vector spaces over the complex numbers C or ZP ' where p is
prime. For example, the vector space M22 (Z 2 ) of all 2 X 2 matrices with entries from
2 2 has dimension 4 as a vector space over 2 2 , and hence M22 (Z 2 ) � Zi.
Exercises 6 . 5
j]h 4. Let T : '!!' 2 ---+ '!!' 2 be the linear transformation defined by
T (p ( x)) = xp ' (x).
( a) Which, if any, of the following polynomials are in
1. Let T : M22 ---+ M22 be the linear transformation
defined by
1
(a) Which, if any, of the following matrices are in
ker( T)?
(ii)
[� �]
(iii)
[3 J
O
0 -3
(b) Which, if any, of the matrices in part (a) are in
range(T)?
(c) Describe ker( T) and range(T).
2. Let T : M22 ---+ IR be the linear transformation defined
by T (A) = tr (A ) .
( a) Which, if any, of the following matrices are in
ker ( T ) ?
(i)
[ - 11 ]
2
3
(ii)
[� �]
(iii)
[1 -1]
0
3
(b) Which, if any, of the following scalars are in
range(T)?
(i) 0
(ii) 2
(iii) Vl/2
( c) Describe ker( T ) and range(T).
3. Let T : '1P 2 ---+ IR 2 be the linear transformation
defined by
T ( a + bx + cx 2 )
=
[ ab -+ bc ]
1
ker (T)?
(i) + x
(ii) x - x 2
(iii) + x - x 2
(b) Which, if any, of the following vectors are in
range(T)?
(i)
(ii)
[�]
(iii)
( c) Describe ker ( T ) and range(T).
[�]
In Exercises 5-8, find bases for the kernel and range of the
linear transformations T in the indicated exercises. In each
case, state the nullity and rank of T and verify the Rank
Theorem.
6. Exercise 2
5. Exercise
8. Exercise 4
7. Exercise 3
1
In Exercises 9- 1 4, find either the nullity or the rank of T
and then use the Rank Theorem to find the other.
a b = a-b
9. T : M22 ---+ IR 2 defined by T
c d
c d
[ ] [ ]
_
10. T : r;; 2 ---+ 1R 2 defined by TCpCx))
[ - 11 - 11 ]
[� � ]
1 1 . T : M22 ---+ M22 defined by T (A )
B
=
12. T : M22 ---+ M22 defined by T(A)
(a) Which, if any, of the following polynomials are in
1
[�]
ker( T)?
(i)
(ii) x
(iii) x 2
(b) Which, if any, of the polynomials in part (a) are in
range(T)?
( c ) Describe ker( T ) and range(T).
B=
J]h 13. T : '1P 2 ---+ IR defined by T(p(x))
14. T : M33 ---+ M33 defined by T (A )
[� � � n
=
=
AB, where
=
AB - BA, where
=
p '(O)
=
A - Ar
In Exercises 1 5-20, determine whether the linear transfor­
mation T is (a) one-to-one and (b) onto.
x
2x - y
15. T : IR 2 ---+ IR 2 defined by T =
y
x + 2y
[] [
]
Chapter 6 Vector Spaces
496
16. T : IR 2 ---+ <.5P 2 defined by
r[ :J
3 1 . Show that 'ii; [ O, l ] = � [ 0, 2 ] .
32. Show that 'ii; [ a, b ] = 'ii; [ c, d ] for all a < b and c < d.
33. Let S : V---+ W and T : U---+ V be linear transformations.
( a ) Prove that if S and T are both one-to-one, so is
S T.
(b) Prove that if S and T are both onto, so is S T.
= (a - 2b ) + ( 3a + b ) x + ( a + b ) x2
[� ]
17. T : <.5P 2 ---+ IR 3 defined by
0
-b
2
T ( a + bx + cx ) = a b - 3c
c-a
0
18. T : <.5P 2 ---+ IR 2 defined by T (p ( x)) =
[]
r[�l
19. T : IR 3 ---+ M 22 defined by r
20.
1" '
[
nl' --+
w defined by
]
:
c
=
[� � � ; ]
[ aa + b
34. Let S : V---+ W and T : U---+ V be linear
transformations.
( a ) Prove that if S T is one-to-one, so is T.
(b) Prove that if S T is onto, so is S.
0
]
b+c
- b b- c
�
a + b + c b - 2c
, where W is the vector space of
b - 2c
a-c
0
35. Let T : V---+ W be a linear transformation between two
finite-dimensional vector spaces.
( a) Prove that if dim V < dim W, then T cannot be
onto.
(b) Prove that if dim V > dim W, then T cannot be
one-to-one.
36. Let a 0, a 1 , , a n be n + 1 distinct real numbers.
Define T : <.5P n ---+ !R n+ I by
•
•
•
all symmetric 2 X 2 matrices
In Exercises 21 -26, determine whether V and W are
isomorphic. If they are, give an explicit isomorphism
T : V ---+ W.
2 1 . V = D 3 (diagonal 3 X 3 matrices), W = IR 3
22. V = S 3 (symmetric 3 X 3 matrices), W = U3 (upper
triangular 3 X 3 matrices)
23. V = S 3 (symmetric 3 x 3 matrices), W = S� (skew­
symmetric 3 X 3 matrices)
24. V = <.5P 2 , W = {p ( x) in <.5P 3 : p ( O ) = O}
F+"S7 25. V = C, W = IR 2
26. V = {A in M22 : tr (A ) = O}, W = IR 2
� 27. Show that T : <.5Pn ---+ <.5Pn defined by T(p(x) ) = p(x) +
p' (x) is an isomorphism.
28. Show that T : <.5P n ---+ <.5P n defined by T(p(x) ) = p(x - 2)
is an isomorphism.
(�)
29. � how_ that T : <.IP_n ---+ <.5P n defined by T (p ( x)) = x np
1s an 1Somorph1sm.
30. (a) Show that 'ii; [O, l] = 'ii; [2, 3 ] . [Hint: Define T :
� [O, l ] ---+ � [2, 3] by letting T (j) be the function
whose value at x is ( T(j)) ( x) = j ( x - 2 ) for x in
[2, 3 ] . ]
(b) Show that � [ O , l ] = <'.{;; [ a , a + l ] for all a.
Prove that T is an isomorphism.
37. If V is a finite-dimensional vector space and T : V ---+ V is
a linear transformation such that rank( T) = rank( T 2 ),
prove that range(T) n ker( T) = {O}. [Hint: T 2 denotes
T T. Use the Rank Theorem to help show that the
kernels of T and T 2 are the same.]
38. Let U and W be subspaces of a finite-dimensional
vector space V. Define T : U X W ---+ V by T(u, w) =
u - w.
( a) Prove that T is a linear transformation.
(b) Show that range(T) = U + W.
(c) Show that ker(T) = U n W. [Hint: See Exercise 50
in Section 6. 1 . ]
(d) Prove Grassmann's Identity:
0
dim ( U + W) = dim U + dim W - dim( U n W)
[Hint: Apply the Rank Theorem, using results
(a) and (b) and Exercise 43(b) in Section 6.2.]
Section 6.6 The Matrix of a Linear Transformation
491
T h e Matrix 01 a l i n e a r Transto r m a t i o n
Theorem 6. 15 showed that a linear transformation T : V � W is completely deter­
mined by its effect on a spanning set for V. In particular, if we know how T acts on a
basis for V, then we can compute T (v) for any vector v in V. Example 6.55 illustrated
the process. We implicitly used this important property of linear transformations
in Theorem 3.31 to help us compute the standard matrix of a linear transformation
T : !Rn � !Rm. In this section, we will show that every linear transformation between
finite-dimensional vector spaces can be represented as a matrix transformation.
Suppose that V is an n-dimensional vector space, W is an m-dimensional vector
space, and T : V � W is a linear transformation. Let B and C be bases for V and W,
respectively. Then the coordinate vector mapping R (v) = [ v ] 6 defines an isomor­
phism R : V � !Rn. At the same time, we have an isomorphism S : W � !Rm given by
S(w) = [wJ c, which allows us to associate the image T (v) with the vector [ T (v) J c in
!Rm. Figure 6. 1 1 illustrates the relationships.
v
v
•
R- 1
U
T (v)
T
•
ls
R
.
[Rn
- - - - - - -
[v] s S 0 T 0 R - l
w
[Rnl
•
�
[T(v) J c
figure 6 . 1 1
Since R is an isomorphism, it is invertible, so we may form the composite mapping
S o T o R- 1 : [Rn � [R m
which maps [v] 6 to [ T (v) J c. Since this mapping goes from !Rn to !Rm, we know
from Chapter 3 that it is a matrix transformation. What, then, is the standard
matrix of S 0 T 0 R - 1 ? We would like to find the m X n matrix A such that A [ v J 6
( S 0 T 0 R- 1 ) ( [v] 6 ) . 0r, since (S 0 T 0 R- 1 ) ( [v] 6) = [ T(v) ] c, we require
A [v ] s
[ T ( v) ] c
=
It turns out to be surprisingly easy to find. The basic idea is that of Theorem 3.3 1 . The
columns of A are the images of the standard basis vectors for !Rn under S 0 T 0 R - 1 •
But, if B = {v 1 , • • • , vn } is a basis for V, then
R (v; )
=
[v; ] 6
0
+--
0
ith entry
498
Chapter 6 Vector Spaces
s o R - 1 (e;) = v;. Therefore, the ith column of the matrix A we seek is given by
(S T R - 1 ) ( e; ) = S ( T ( R - 1 ( e; )))
= S ( T ( v; ))
= [ T ( v; ) l e
which is the coordinate vector of T (v;) with respect to the basis C of W.
We summarize this discussion as a theorem.
0
Theorem 6 . 2 6
0
Let V and W be two finite-dimensional vector spaces with bases B and C, respec­
tively, where B = {v1 , , vJ. If T : V ---+ W is a linear transformation, then the
m X n matrix A defined by
A = [ [ T (v 1 ) l c ! [ T (v2 ) l c ! · · · ! [ T(vn ) l c l
satisfies
A [v] 8 = [ T ( v) J c
•
•
•
for every vector v in V.
The matrix A in Theorem 6.26 is called the matrix of T with respect to the bases B
and C. The relationship is illustrated below. (Recall that TA denotes multiplication
by A.)
T
v -----+
J,
TA
T ( v)
J,
[v] 8 -----+ A [v] 8 = [ T ( v) J c
The matrix of a linear transformation T with respect to bases B and C is some­
times denoted by [ T] c<-- B · Note the direction of the arrow: right-to-left (not left-to­
right, as for T: V---+ W ) . With this notation, the final equation in Theorem 6.26 becomes
R e m a rks
•
[ T J c <-- B [v] s = [ T ( v) J c
Observe that the Bs in the subscripts appear side by side and appear to "cancel" each
other. In words, this equation says, "The matrix for T times the coordinate vector for
v gives the coordinate vector for T (v) :'
In the special case where V = W and B = C, we write [ T] 8 (instead of [ T] B <-- B) .
Theorem 6.26 then states that
[ T J s [vJ s = [ T ( v) J s
•
The matrix of a linear transformation with respect to given bases is unique.
That is, for every vector v in V, there is only one matrix A with the property specified
by Theorem 6.26-namely,
A [v ] 8 = [ T ( v) J c
(You are asked to prove this in Exercise 39. )
Section 6.6 The Matrix of a Linear Transformation
499
The diagram that follows Theorem 6.26 is sometimes called a commutative
diagram because we can start in the upper left-hand corner with the vector v and get
•
to [T(v) J c in the lower right-hand corner in two different, but equivalent, ways. If, as
before, we denote the coordinate mappings that map v to [ v J 8 and w to [w]c by R and
S, respectively, then we can summarize this "commutativity" by
s
0
T
=
TA R
0
The reason for the term commutative becomes clearer when V = W and B
then R = S too, and we have
R T
0
=
= C,
for
TA R
0
suggesting that the coordinate mapping R commutes with the linear transformation T
(provided we use the matrix version of T-namely, TA = Tl Tl 8-where it is required).
•
The matrix [ TJ c<-- B depends on the order of the vectors in the bases B and
C. Rearranging the vectors within either basis will affect the matrix [ TJ c ._ 8. [See
Example 6.77(b) .]
Exa m p l e 6 . 16
Let T : IR 3 ---+ IR be the linear transformation defined by
2
and let B
=
{ e 1 , e2 , e J and C = { e2 , ei } be bases for IR 3 and !R 2 , respectively. Find the
mote ix of T whh '"P"' to B ond C and vecify Themem 6.26 foe v �
Solution
First, we compute
Next, we need their coordinate vectors with respect to C. Since
we have
Therefore, the matrix of T with respect to B and C is
=
[�
-2
-
�]
[ _� l
·
500
Chapter 6 Vector Spaces
To verify Theorem 6.26 for v, we first compute
T ( v) �
Then
[ T ( v) l e =
=
(Check these.)
Using all of these facts, we confirm that
A [v] 8 �
Exa m p l e 6 . 11
�
[v i s �
and
_.,..
T [ J [ �� ]
u[ �� L [ � � ]
[: - � -�1 [ J [ � � ]
�
� [ T ( v) J ,
Let D : l!P3 � <!P2 be the differential operator D(p(x)) = p' (x). Let B = { l , x, x 2 , x 3 }
and C = { l , x, x 2 } be bases for l!P3 and l!P2 , respectively.
(a) Find the matrix A of D with respect to B and C.
(b) Find the matrix A ' of D with respect to B ' and C, where B ' = {x 3 , x 2 , x, l}.
(c) Using part (a), compute D(S
Theorem 6.26.
Solulion
-
x + 2x3 ) and D(a + bx + ex 2 + dx 3 ) to verify
First note that D(a + bx + cx 2 + dx 3 ) = b + 2cx + 3dx 2 • (See Example 6.60.)
(a) Since the images of the basis B under D are D( l ) = 0, D(x) = 1, D(x 2 ) = 2x, and
D (x 3 ) = 3x 2 , their coordinate vectors with respect to C are
[D(l) J, �
[H
[ D (x) J , �
Consequently,
[O
[H
[ D (x ' ) J , �
OJ
[H
[ D (x ' ) J , �
A = [ D Jc.- s = [ [ D ( l ) l e ! [ D (x) l e ! [ D (x2 ) l e ! [ D (x3 ) l e l
1 0
= 0 0 2 0
0 0 0 3
(b) Since the basis B ' is just B in the reverse order, we see that
A ' = [ D b- s · = [ [ D (x3 ) l e ! [ D (x2 ) l e ! [ D (x) l e ! [ D ( l ) l e l
=
[� � � �i
3 0 0 0
m
Section 6.6 The Matrix of a Linear Transformation
501
(This shows that the order of the vectors in the bases l3 and C affects the matrix of a
transformation with respect to these bases.)
( c) First we compute D(S x + 2x 3 ) = - 1 + 6x 2 directly, getting the coordinate
vector
-
[D (S
-
-
x + 2x') ] , � [ 1 + 6x' ] , �
[ -�]
On the other hand,
so
[D (S
-
x + 2x 3 ) l e
which agrees with Theorem 6.26. We leave proof of the general case as an exercise .
.+
Since the linear transformation in Example 6. 77 is easy to use directly, there is re­
ally no advantage to using the matrix of this transformation to do calculations. How­
ever, in other examples-especially large ones-the matrix approach may be simpler,
as it is very well-suited to computer implementation. Example 6.78 illustrates the
basic idea behind this indirect approach.
Exa m p l e 6 . 18
Let T : <!J' 2 � <!J' 2 be the linear transformation defined by
T (p (x)) = p(2x - 1 )
(a) Find the matrix of T with respect to £ = { l , x, x2 }.
(b) Compute T(3 + 2x - x 2 ) indirectly, using part (a).
(a) We see that
Solution
T ( l ) = 1 , T ( x) = 2x
so the coordinate vectors are
Therefore,
-
1, T(x 2 ) = (2x
-
1 )2 = 1
-
4x + 4x 2
502
Chapter 6 Vector Spaces
(b) We apply Theorem 6.26 as follows: The coordinate vector of p(x)
with respect to E is
[p (x) ] , �
Therefore, by Theorem 6.26,
[ T(3 + 2x - x 2 ) ] £
=
[J
=
3 + 2x - x 2
[ T(p(x)) ] £
[ T ] £ l p(x) ] £
[ � � -: J [ J U l
It follows that T(3 + 2x - x 2 ) = 0 · 1 + 8 · x - 4 · x 2 = 8x - 4x 2 . [Verify this by
computing T(3 + 2x - x 2 ) = 3 + 2(2x - 1 ) - (2x - 1 ) 2 directly.]
The matrix of a linear transformation can sometimes be used in surprising ways.
Example 6.79 shows its application to a traditional calculus problem.
Exa m p l e 6 . 19
Let <f/J be the vector space of all differentiable functions. Consider the subspace W of
given by W = span(e 3x, xe 3X, x 2 e 3x ) . Since the set B = {e 3X, xe 3X , x 2 e 3x} is linearly
independent (why?), it is a basis for W.
<f/J
(a) Show that the differential operator D maps W into itself.
(b) Find the matrix of D with respect to B.
(c) Compute the derivative of 5e 3x + 2xe 3x - x 2 e 3x indirectly, using Theorem 6.26,
and verify it using part (a) .
(a) Applying D to a general element of W, we see that
D ( ae 3x + bxe 3x + cx 2 e 3x) = ( 3a + b ) e 3x + ( 3b + 2c ) xe 3x + 3cx 2 e 3x
Solulion
�
(check this), which is again in W.
(b) Using the formula in part (a), we see that
D (e 3x ) = 3e 3X , D ( xe 3x ) = e 3x + 3xe 3X, D ( x 2 e 3x)
so
It follows that
=
2xe 3x + 3x 2 e 3x
Section 6.6 The Matrix of a Linear Transformation
503
(c) Forf(x) = 5e 3x + 2xe 3x - x 2 e 3X, we see by inspection that
V (x) ] a �
Hence, by Theorem 6.26, we have
[ D (j (x)) l s = [ D l s [f(x) l s =
[J
[� � �i [ �i [ :]
=
1
0 0 3 -1
-3
which, in turn, implies thatf'(x) = D(j(x)) = 1 7e 3x + 4xe 3x - 3x 2 e 3x, in agreement
4
with the formula in part (a) .
Remark The point of Example 6. 79 is not that this method is easier than direct
differentiation. Indeed, once the formula in part (a) has been established, there is
little to do. What is significant is that matrix methods can be used at all in what
appears, on the surface, to be a calculus problem. We will explore this idea further in
Example 6.83.
Exa m p l e 6 . 8 0
Let V be an n-dimensional vector space and let I be the identity transformation on
V. What is the matrix of I with respect to bases B and C of V if B = C (including the
order of the basis vectors) ? What if B * C?
Solution
Let B = {v 1 , . . . , vJ. Then J(v1 ) = v1 , . . . , J(vn ) = vn , so
and, if B = C,
[ I l s = [ [ J(v1 ) l s ! [ J(vz ) l s ! · · · ! [ J(vn ) l s l
= [ e 1 ! ez ! · · · ! e n l
= In
the n X n identity matrix. (This is what you expected, isn't it?)
In the case B * C, we have
[ J ( v1 ) l e = [vi l e ,
, [ J ( vn ) l e = [vn l e
so
[ IJ e ..- s = [ [vi l e ! · · · ! [vn l e l
= Pe ..- s
the change-of-basis matrix from B to C.
·
·
.
Malrices of Composite and Inverse linear Tra nsformations
We now generalize Theorems 3.32 and 3.33 to get a theorem that will allow us to
easily find the inverse of a linear transformation between finite-dimensional vector
spaces (if it exists) .
504
Chapter 6 Vector Spaces
Theorem 6 . 2 1
Let U, V, and W b e finite-dimensional vector spaces with bases B, C, and
respectively. Let T : U � V and S : V � W be linear transformations. Then
D,
R e m a rks
In words, this theorem says, "The matrix of the composite is the product of
the matrices:'
•
Notice how the "inner subscripts" C must match and appear to cancel each
other out, leaving the "outer subscripts" in the form D +--- B.
•
We will show that corresponding columns of the matrices [ S T] v<---- 6 and
[ S J v<-- c l T] c<---- 6 are the same. Let V; be the ith basis vector in B. Then the ith column
of [ S T ] v<---- 6 is
Proof
0
0
[ ( S T) (v; ) ] v
0
=
=
[ S ( T( v; ) J v
[ S l v<-- c l T(v;) ] c
[ s l v<-- c l T l C<-- 6 [ v; ] 6
by two applications of Theorem 6.26. But [ v; ] 6
= e;
(why?), so
is the ith column of the matrix [ S l v<-- c l T l c<---- 6 . Therefore, the ith columns of
[ S T l v<---- 6 and [ S l v<-- c [ T] C<-- 6 are the same, as we wished to prove.
0
Exa m p l e 6 . 8 1
Use matrix methods to compute ( S T)
0
Example 6.56.
Solution
[:]
for the linear transformations S and T of
Recall that T : IR 2 � <J/' 1 and S : <JP , � <!P 2 are defined by
r[:]
= a
+ ( a + b ) x and S ( a + bx)
= ax
+ bx 2
Choosing the standard bases £, [', and £" for IR 2 , <JP 1 , and <JP 2 , respectively, we see that
�
(Verify these.) By Theorem 6.27, the matrix of S T with respect to [ and £" is
0
Section 6.6 The Matrix of a Linear Transformation
505
Thus, by Theorem 6.26,
Consequently, ( S T)
0
Example 6.56.
[ �]
= ax + (a + b )x 2 , which agrees with the solution to
In Theorem 6.24, we proved that a linear transformation is invertible if and only
if it is one-to-one and onto (i.e., if it is an isomorphism) . When the vector spaces in­
volved are finite-dimensional, we can use the matrix methods we have developed to
find the inverse of such a linear transformation.
Theorem 6 . 2 8
Let T : V ---+ W be a linear transformation between n-dimensional vector spaces V
and W and let B and C be bases for V and W, respectively. Then T is invertible if
and only if the matrix [ T] C +- B is invertible. In this case,
Observe that the matrices of T and r - 1 (if T is invertible) are n X n. If T is
invertible, then r - 1 T = Iv. Applying Theorem 6.27, we have
In = [ Iv l s = [ Y- 1 TJ s
[ Y - 1 J B+- c [ T] C+- B
This shows that [ T ] c.- B is invertible and that ( [ T] c..-8 ) - 1 = [ r- 1 ] B+- C·
Conversely, assume that A = [ T J c..- B is invertible. To show that T is invertible, it
is enough to show that ker (T) = {O}. (Why?) To this end, let v be in the kernel of T.
Then T(v) = 0 , so
Proof
0
0
�
A [v ] a = [ T l c..- a [v] a = [ T(v) J c = [ O l e = 0
which means that [ v] 8 is in the null space of the invertible matrix A. By the Fundamen­
tal Theorem, this implies that [ v J 8 = 0, which, in turn, implies that v = 0, as required.
Exa m p l e 6 . 8 2
In Example 6.70, the linear transformation T : IR 2 ---+ !Jl 1 defined by
r [�]
= a + (a + b)x
was shown to be one-to-one and onto and hence invertible. Find r - 1 .
506
Chapter 6 Vector Spaces
Solulion
In Example 6.8 1 , we found the matrix of T with respect to the standard
bases £ and £' for IR 2 and <!P 1 , respectively, to be
[ T] £' +- £ =
[ � �]
By Theorem 6.28, it follows that the matrix of r - 1 with respect to £' and [ is
[ Y - 1 1 £+- £' = ( [ T] £' +- £) - 1 =
By Theorem 6.26,
[ l 0] -1 = [ OJ
1
1
-1
1
1
[ _� � ][�]
[b : a]
This means that
(Note that the choice of the standard basis makes this last calculation virtually
irrelevant.)
4
The next example, a continuation of Example 6.79, shows that matrices can be
used in certain integration problems in calculus. The specific integral we consider is
usually evaluated in a calculus course by means of two applications of integration by
parts. Contrast this approach with our method.
Exa m p l e 6 . 8 3
Show that the differential operator, restricted to the subspace W = span (e 3x, xe3x,
x 2 e 3x) of '20 , is invertible, and use this fact to find the integral
I x 2 e 3x dx
In Example 6.79, we found the matrix of D with respect to the basis
B = { e 3X, xe 3X , x 2 e 3x} of W to be
Solulion
l [ t �:i
By Theorem 6.28, therefore, D is invertible on W, and the matrix of D - 1 is
o -1
2
3
= 0
0
Section 6.6 The Matrix of a Linear Transformation
501
Since integration is antidifferentiation, this is the matrix corresponding to integration
on W. We want to integrate the function x 2 e 3x whose coordinate vector is
Consequently, by Theorem 6.26,
[�
It follows that
(To be fully correct, we need to add a constant of integration. It does not show up here
because we are working with linear transformations, which must send zero vectors to
zero vectors, forcing the constant of integration to be zero as well.)
Warning
In general, differentiation is not an invertible transformation. (See
Exercise 22.) What the preceding example shows is that, suitably restricted, it some­
times is. Exercises 27-30 explore this idea further.
Change of Basis and Simila rilv
Suppose T : V --+ V is a linear transformation and 13 and C are two different bases for
V. It is natural to wonder how, if at all, the matrices [ T ] 8 and [ T ] c are related. It turns
out that the answer to this question is quite satisfying and relates to some questions
we first considered in Chapter 4.
Figure 6. 1 2 suggests one way to address this problem. Chasing the arrows around
the diagram from the upper left-hand corner to the lower right-hand corner in two
different, but equivalent, ways shows that I T = T I, something we already knew,
since both are equal to T. However, if the "upper" version of T is with respect to the
0
v
•V
i
I
•V
Figure 6 . 1 2
Jo T = ToJ
v
0
v
T
-
T
-
• T(v)
i1
• T(v)
v
}
basis C
}
basis 13
508
Chapter 6 Vector Spaces
basis C and the "lower" version is with respect to B, then T = I T = T I is with
respect to C in its domain and with respect to B in its codomain. Thus, the matrix of
T in this case is [ T] B<- C · But
0
and
[ T] a.-- c
=
[I T] a.-- c
=
[IJ a.-- d T l c.-- c
[ T] a.-c
=
[ T I] a.-c
=
[ T] a.-- a [I] a.-- c
0
0
0
Therefore, [I l a.-- d T] c.-c = [ T] a.-- a [I] a.-c·
From Example 6.80, we know that [I] 8 .-c = P8 .-c, the (invertible) change-of­
basis matrix from C to B. If we denote this matrix by P, then we also have
p- 1
=
With this notation,
(P5 .-- c ) - 1
=
Pc.-- a
so
Thus, the matrices [ T] 8 and [ T] c are similar, in the terminology of Section 4.4.
We summarize the foregoing discussion as a theorem.
Theorem 6 . 2 9
Let V be a finite-dimensional vector space with bases B and C and let T : V � V
be a linear transformation. Then
where P is the change-of-basis matrix from C to B.
Remark As an aid in remembering that P must be the change-of-basis matrix
from C to B, and not B to C, it is instructive to look at what Theorem 6.29 says when
written in full detail. As shown below, the "inner subscripts" must be the same (all Bs)
and must appear to cancel, leaving the "outer subscripts;' which are both Cs.
S ame
Theorem 6.29 is often used when we are trying to find a basis with respect to
which the matrix of a linear transformation is particularly simple. For example, we
can ask whether there is a basis C of V such that the matrix [ T J c of T : V � V is a
diagonal matrix. Example 6.84 illustrates this application.
Exa m p l e 6 . 8 4
Let T :
IR:2 IR:2
�
be defined by
If possible, find a basis C for
IR:2
T
x
x
[] [ ]
y
=
+ 3y
2x + 2y
such that the matrix of T with respect to C is diagonal.
Section 6.6 The Matrix of a Linear Transformation
Solution
509
The matrix of T with respect to the standard basis E is
[ T] E
=
[� �]
This matrix is diagonalizable, as we saw in Example 4.24. Indeed, if
p
=
[ � -�J
and D
=
[0 0]
4
-1
then P - 1 [ T] Ep = D. If we let C be the basis of IR 2 consisting of the columns of P, then
P is the change-of-basis matrix PE.- c from C to E. By Theorem 6.29,
[ T] c = P - 1 [ T] EP = D
so the matrix of T with respect to the basis C
=
{ [ � ] , [ _�] }
is diagonal.
R e m a rks
It is easy to check that the solution above is correct by computing [ T ] c directly.
We find that
•
r[ � ] [! ] 4 [ � ] o [ _ � J r[ _ � J [ - � ] o [ � ] [ _ � J
[ r[ � J L [�] [ r[ _ � J L [ _ �J
=
+
and
=
Thus, the coordinate vectors that form the columns of [ T ] c are
and
in agreement with our solution above.
•
The general procedure for a problem like Example 6.84 is to take the stan­
dard matrix [ T] E and determine whether it is diagonalizable by finding bases for its
eigenspaces, as in Chapter 4. The solution then proceeds exactly as in the preceding
example.
Example 6.84 motivates the following definition.
D e f i n i t i o n Let V be a finite-dimensional vector space and let T : V ---+ V be a
linear transformation. Then T is called diagonalizable if there is a basis C for V
such that the matrix [ T ] c is a diagonal matrix.
It is not hard to show that if B is any basis for V, then T is diagonalizable if and
only if the matrix [ T] 8 is diagonalizable. This is essentially what we did, for a special
case, in the last example. You are asked to prove this result in general in Exercise 42.
Sometimes it is easiest to write down the matrix of a linear transformation with
respect to a "nonstandard" basis. We can then reverse the process of Example 6.84
to find the standard matrix. We illustrate this idea by revisiting Example 3.59.
Exa m p l e 6 . 8 5
Let e be the line through the origin in IR 2 with direction vector d
standard matrix of the projection onto e.
=
[ �: ] .
Find the
510
Chapter 6 Vector Spaces
df d
Let T denote the projection. There is no harm in assuming that d is
a unit vector (i.e.,
+ � = 1), since any nonzero multiple of d can serve as a
Solulion
direction vector for
-€. Let d' [ -d�2 ] so that d and d' are orthogonal. Since
=
d' is also a unit vector, the set D = {d, d'} is an orthonormal basis for IR 2 .
As Figure 6. 1 3 shows, T(d) = d and T(d') = 0. Therefore,
[ T(d) J v
=
[�]
and [ T(d') J v
=
[�]
y
Projection onto -€
Figure 6 . 1 3
so
[ T] v
=
[� �]
The change-of-basis matrix from D to the standard basis [ is
so the change-of-basis matrix from E to D is
By Theorem 6.29, then, the standard matrix of T is
which agrees with part (b) of Example 3.59.
Section 6.6 The Matrix of a Linear Transformation
Exa m p l e 6 . 8 6
- 1)
Let T : <!P 2 ---+ <!P 2 be the linear transformation defined by
T(p(x))
=
p ( 2x
511
1
(a) Find the matrix of T with respect to the basis B = { l + x, - x, x 2 } of <!f 2 .
(b) Show that T is diagonalizable and find a basis C for <!P 2 such that [ T ] c is a diago­
nal matrix.
(a) In Example 6.78, we found that the matrix of T with respect to the
standard basis £ = { l , x, x 2 } is
Solution
[ T] ,
[ � � -:l
11 -� o�
[0 l
�
The change-of-basis matrix from B to £ is
p = PE<-B =
It follows that the matrix of T with respect to B is
[ T] B
=
-1 - [
u -! � m 0 : J i i � ]
[ - 0� �0 -:i
1,
p- 1 [ T] EP
2
-
4
�
�
(Check this.)
(b) The eigenvalues of [ T] E are 2, and 4 (why?), so we know from Theorem 4.25
that [ T] E is diagonalizable. Eigenvectors corresponding to these eigenvalues are
[ � JTJ H l
[� � :l [� � �l
respectively. Therefore, setting
P�
-
-
and D �
we have p- i [ T] Ep = D. Furthermore, P is the change-of-basis matrix from a basis
C to £, and the columns of P are thus the coordinate vectors of C in terms of £. It
follows that
C = {l,
+ x, - 2x + x 2 }
and [ T l c = D.
-1 1
Chapter 6 Vector Spaces
512
The preceding ideas can be generalized to relate the matrices [ T ] C<-- B and
[ T] c '<-- B ' of a linear transformation T : V ---+ W, where 13 and 13 ' are bases for V and C
and C' are bases for W. (See Exercise 44.)
We conclude this section by revisiting the Fundamental Theorem of Invertible
Matrices and incorporating some results from this chapter.
Theorem 6 . 3 0
The Fundamental Theorem of Invertible Matrices: Version 4
Let A be an n X n matrix and let T : V ---+ W be a linear transformation whose
matrix [ T] C<-- B with respect to bases 13 and C of V and W, respectively, is A. The
following statements are equivalent:
a.
b.
c.
d.
e.
f.
g.
h.
i.
j.
k.
A is invertible.
Ax = b has a unique solution for every b in !R n .
Ax = 0 has only the trivial solution.
The reduced row echelon form of A is Iw
A is a product of elementary matrices.
rank(A) = n
nullity(A) = 0
The column vectors of A are linearly independent.
The column vectors of A span !R n .
The column vectors of A form a basis for !R n .
The row vectors of A are linearly independent.
1. The row vectors of A span !R n .
m. The row vectors of A form a basis for !R n .
n. det A * 0
o. 0 is not an eigenvalue of A.
p. T is invertible.
q. T is one-to-one.
r. T is onto.
s. ker ( T) = {O}
t. range(T) = W
The equivalence (q) -¢:? (s) is Theorem 6.20, and (r) -¢:? (t) is the definition of
onto. Since A is n X n, we must have dim V = dim W = n. From Theorems 6.2 1 and
6.24, we get (p) -¢:? ( q) -¢:? ( r) . Finally, we connect the last five statements to the others
by Theorem 6.28, which implies that (a) -¢:? (p ).
Proof
..
I
Exercises 6 . 6
In Exercises 1 - 12, find the matrix [ T] C<-- B of the linear
transformation T : V ---+ W with respect to the bases 13 and C
of V and W, respectively. Verify Theorem 6.26 for the vector v
by computing T(v) directly and using the theorem.
1 . T : <!P 1 ---+ <!P 1 defined by T ( a + bx) = b - ax,
13 = C = { l , x}, v = p(x) = 4 + 2x
..
2. T : <!P 1 ---+ <!P 1 defined by T(a + bx) = b - ax,
13 = { l + x, 1 - x}, C = {l, x}, v = p(x) = 4 + 2x
3. T : <!P 2 ---+ <!P 2 defined by T(p(x) ) = p(x + 2),
13 = { l , x, x 2 }, C = { l , x + 2, (x + 2 ) 2 },
v = p(x) = a + bx + cx 2
Section 6.6 The Matrix of a Linear Transformation
4. T : <J/l 2 ---+ <J/l 2 defined by T(p(x) ) = p(x + 2),
B = { l , x + 2, (x + 2 ) 2 }, C = { l , x, x 2 },
v = p(x) = a + bx + cx2
5. T : <J/l 2 ---+ IR 2 defined by T(p(x) ) =
B = { l , x, x 2 }, C = {e 1 , ez },
[� � � � l
v=
p(x) = a + bx + cx 2
6. T : <J/l 2 ---+ IR 2 defined by T(p(x) ) =
B= {x2, x, l },C = { [ � ] , [ � ] } ,
v=
[� � � � l
p(x) = a + bx + cx 2
7. T : IR 2 ---+ IR 3 defined by
a 2b
B
a
r[:J [ � l � WH - � ] } .
c � m H J [: l l · v � n
[ �] .
[ : �]
8. Repeat Exercise 7 with v =
9. T : M22 ---+ M22 defined by T(A) = A r, B = C =
{E 1 1 , E 1 2 , E 2 1 , E 22 }, v = A =
10. Repeat Exercise 9 with B = {E w E 2 1 , E 1 2 > E 11 } and
C = {E 1 2 , E 2 1 > E w E 11 }.
11. T: M22 ---+ M22 defined by T (A) = AB - BA, where
B= [
l
-1
v=A =
]
[ : �]
-l
, B = C = {E 11 , E 1 2 > E2 1 , E22 },
1
12. T : M22 ---+ M22 defined by T(A) = A - A r, B =
C = {E 11 , E 1 2 > E2 1 , E22 }, v = A =
[ : �]
� 13. Consider the subspace W of '.2ll , given by
W = span (sin x, cos x).
(a) Show that the differential operator D maps W into
513
ilili_ 14. Consider the subspace W of '.2ll , given by
W = span ( e 2x, e-2x) .
(a) Show that the differential operator D maps W into
itself.
(b) Find the matrix of D with respect to B = { e 2 , e - zx}.
(c) Compute the derivative of f ( x) = e 2x - 3e - zx
indirectly, using Theorem 6.26, and verify that it
agrees with j'( x) as computed directly.
ilili_ 15. Consider the subspace W of '.2ll , given by W = span ( e 2 ,
e 2x cos x, e 2x sin x).
( a) Find the matrix of D with respect to B = {e 2x,
e 2x cos x, e 2x sin x}.
(b) Compute the derivative off(x) = 3 e 2x - e 2x cos x +
2e 2x sin x indirectly, using Theorem 6.26, and
verify that it agrees with f' (x) as computed directly.
ilili_ 16. Consider the subspace W of '.2ll , given by
W = span (cos x, sin x, x cos x, x sin x).
( a) Find the matrix of D with respect to B = {cos x,
sin x, x cos x, x sin x}.
(b) Compute the derivative off( x) = cos x + 2x cos x
indirectly, using Theorem 6.26, and verify that it
agrees with f' (x) as computed directly.
X
X
In Exercises 1 7 and 1 8, T : U ---+ V and S : V ---+ W are linear
transformations and B, C, and D are bases for U, V, and
W, respectively. Compute [S T l v +- B in two ways: (a) by
finding S T directly and then computing its matrix and
(b) by finding the matrices of S and T separately and using
Theorem 6.27.
p(O)
17. T : <J/l 1 ---+ IR 2 defined by T(p(x)) =
, S : IR 2 ---+ IR 2
p( l)
2
B = { l , x},
=
defined by s
a
C = D = {e 1 , ez }
18. T : <J/l 1 ---+ <J/l 2 defined by T(p(x) ) = p(x + 1 ) ,
S : <JP 2 ---+ <JP 2 defined by S(p(x) ) = p(x + 1 ) ,
B = { l , x}, C = D = { l , x, x2 }
0
0
[ �] [ ; -- � l
[ ]
In Exercises 1 9-26, determine whether the linear transfor­
mation T is invertible by considering its matrix with respect
to the standard bases. If T is invertible, use Theorem 6.28
itself.
and the method of Example 6.82 to find T - 1 •
(b) Find the matrix of D with respect to
20. T in Exercise 5
19. T in Exercise 1
B = {sin x, cos x}.
21. T in Exercise 3
(c) Compute the derivative off( x) = 3 sin x - 5 cos x
22. T : <Jfl 2 -+ <Jfl 2 defined by T(p (x) ) = p' (x)
indirectly, using Theorem 6.26, and verify that it
agrees withj' (x) as computed directly.
� 23. T : <J/l 2 ---+ <JP 2 defined by T(p(x) ) = p(x) + p' ( x)
514
Chapter 6 Vector Spaces
24. T : M22 � M22 defined by T (A ) = AB, where
[� �]
B=
25. T in Exercise 1 1
26. T in Exercise 1 2
� In Exercises 27-30, use the method of Example 6.83
to evaluate the given integral.
27. f (sin x - 3 cos x) dx. (See Exercise 13.)
28. f 5e - 2x dx. (See Exercise 14.)
29. f (e2x cos x - 2e2x sin x) dx. (See Exercise 15.)
30. f (x cos x + x sin x) dx. (See Exercise 16.)
In Exercises 31 -36, a linear transformation T : V � V is
given. Ifpossible, find a basis C for V such that the matrix
[ T] c of T with respect to C is diagonal.
[�] [ ]
r[ �] [ : : �]
- 4b
a + Sb
31. T : IR 2 � IR 2 defined by r
32. T : IR 2 � IR 2 defined by
=
33. T : <!/' 1 � <!/' 1 defined by T(a + bx) = (4a + 2b) +
(a + 3b)x
34. T : <!1' 2 � <!1' 2 defined by T(p(x)) = p(x + 1 )
� 35. T : <!/' 1 � <!/' 1 defined by T(p( x)) = p(x) + xp' (x)
36. T : <!1' 2 � <!1' 2 defined by T(p( x)) = p(3x + 2)
37. Let € be the line through the origin in IR 2 with direction
vector d =
[ �: ]
it to compute the orthogonal projection of v onto W,
where
. Use the method of Example 6.85 to
find the standard matrix of a reflection in e.
38. Let W be the plane in IR 3 with equation x - y +
2z = 0. Use the method of Example 6.85 to find
the standard matrix of an orthogonal projection
onto W. Verify that your answer is correct by using
Compare your answer with Example 5. 1 1 .
[Hint: Find an orthogonal decomposition of IR 3 as
IR 3 = W + WJ_ using an orthogonal basis for W. See
Example 5.3.]
39. Let T : V � W be a linear transformation between
finite-dimensional vector spaces and let l3 and C be
bases for V and W, respectively. Show that the matrix
of T with respect to l3 and C is unique. That is, if A is a
matrix such that A [ v ] 8 = [ T ( v) l e for all v in V, then
A = [ T l c +- B· [Hint: Find values of v that will show
this, one column at a time.]
In Exercises 40-45, let T : V � W be a linear transforma­
tion between finite-dimensional vector spaces V and W
Let l3 and C be bases for V and W, respectively, and let
A = [ T] c +- B·
40. Show that nullity( T) = nullity(A).
41. Show that rank( T) = rank(A) .
42. I f V = W and l3 = C , show that T is diagonalizable if
and only if A is diagonalizable.
43. Use the results of this section to give a matrix­
based proof of the Rank Theorem (Theorem 6. 19).
44. If !3 ' and C' are also bases for V and W, respectively,
what is the relationship between [ T] c +- B and [ T] c · +- B '?
Prove your assertion.
45. If dim V = n and dim W = m, prove that � ( V, W ) =
Mm n - (See the exercises for Section 6.4.) [Hint: Let l3
and C be bases for V and W, respectively. Show that the
mapping cp ( T) = [ T ] c._8, for T in � ( V, W ) , defines
a linear transformation <p : � ( V, W) � Mmn that is an
isomorphism.]
46. If V is a vector space, then the dual space of V is
the vector space V* = � ( V, IR). Prove that if V is
finite-dimensional, then V* = V.
Exp loration
Tilings , L att i c e s , and t h e
C rystall o gr ap h i c Restriction
Repeating patterns are frequently found in nature and i n art. The molecular struc­
ture of crystals often exhibits repetition, as do the tilings and mosaics found in the
artwork of many cultures. Tiling (or tessellation) is covering of a plane by shapes that
do not overlap and leave no gaps. The Dutch artist M. C. Escher ( 1 898- 1 972) pro­
duced many works in which he explored the possibility of tiling a plane using fanciful
shapes (Figure 6. 14).
M. C. Escher's "Symmetry Drawing El03"
Figure 6 . 1 4
515
translationDrawing El03"
M.Invariance
C. Escher'under
s "Symmetry
figure 6 . 1 5
•
•
•
•
•
•
"
•
•
A lattice
l
•
•
•
figure 6 . 1 6
•
•
"
•
u
•
•
•
•
•
•
•
M.Rotational
C. Escher'symmetry
s "Symmetry Drawing El03"
figure 6 . 1 1
In this exploration, we will be interested in patterns such as those in Figure 6. 14,
which we assume to be infinite and repeating in all directions of the plane. Such a
pattern has the property that it can be shifted (or translated) in at least two directions
(corresponding to two linearly independent vectors) so that it appears not to have
been moved at all. We say that the pattern is invariant under translations and has
translational symmetry in these directions. For example, the pattern in Figure 6. 14
has translational symmetry in the directions shown in Figure 6. 1 5 .
If a pattern has translational symmetry in two directions, it has translational sym­
metry in infinitely many directions.
1 . Let the two vectors shown in Figure 6. 1 5 be denoted by u and v. Show that the
pattern in Figure 6.14 is invariant under translation by any integer linear combination
of u and v-that is, by any vector of the form au + bv, where a and b are integers.
For any two linearly independent vectors u and v in IR 2 , the set of points deter­
mined by all integer linear combinations of u and v is called a lattice. Figure 6. 1 6
shows an example o f a lattice.
2. Draw the lattice corresponding to the vectors u and v of Figure 6.15.
Figure 6. 14 also exhibits rotational symmetry. That is, it is possible to rotate
the entire pattern about some point and have it appear unchanged. We say that it is
invariant under such a rotation. For example, the pattern of Figure 6. 14 is invariant
under a rotation of 1 20° about the point 0, as shown in Figure 6. 17. We call 0 a center
of rotational symmetry (or a rotation center) .
Note that if a pattern is based on an underlying lattice, then any symmetries of the
pattern must also be possessed by the lattice.
516
3. Explain why, if a point 0 is a rotation center through an angle e, then it is
a rotation center through every integer multiple of e. Deduce that if 0 < e :s 360°,
then 360 / e must be an integer. (If 360 / e = n, we say the pattern or lattice has n-fold
rotational symmetry.)
4. What is the smallest positive angle of rotational symmetry for the lattice in
Problem 2? Does the pattern in Figure 6.14 also have rotational symmetry through
this angle?
5. Take various values of e such that 0 < e :s 360° and 360/e is an integer. Try
to draw a lattice that has rotational symmetry through the angle e. In particular, can
you draw a lattice with eight-fold rotational symmetry?
We will show that values of e that are possible angles of rotational symmetry for a
lattice are severely restricted. The technique we will use is to consider rotation trans­
formations in terms of different bases. Accordingly, let Re denote a rotation about the
origin through an angle e and let £ be the standard basis for IR 2 . Then the standard
matrix of Re is
[
cos e - sin e
[ Re l E -
sm e
.
cos e
]
6. Referring to Problems 2 and 4, take the origin to be at the tails of u and v.
(a) What is the actual (i.e., numerical) value of [ R e ] E in this case?
(b) Let B be the basis { u, v}. Compute the matrix [ R 0 ] a7. In general, let u and v be any two linearly independent vectors in IR 2 and
suppose that the lattice determined by u and v is invariant under a rotation through
an angle e. If B = {u, v}, show that the matrix of R11 with respect to B must have the
form
where a, b, c, and d are integers.
8. In the terminology and notation of Problem 7, show that 2 cos e must be an
integer. [Hint: Use Exercise 35 in Section 4.4 and Theorem 6.29.]
9. Using Problem 8, make a list of all possible values of e, with 0 < e :s 360°,
that can be angles of rotational symmetry of a lattice. Record the corresponding val­
ues of n, where n = 360/e, to show that a lattice can have n-fold rotational symmetry
if and only if n = 1 , 2, 3, 4, or 6. This result, known as the crystallographic restriction,
was first proved by W Barlow in 1 894.
10. In the library or on the Internet, see whether you can find an Escher tiling
for each of the five possible types of rotational symmetry-that is, where the smallest
angle of rotational symmetry of the pattern is one of those specified by the crystal­
lographic restriction.
511
518
Chapter 6 Vector Spaces
A p p l icati o n s
Jlh
Homogeneous Linear Diffe rential Equalions
In Exercises 69-72 in Section 4.6, we showed that if y = y(t) is a twice-differentiable
function that satisfies the differential equation
y " + ay ' + by = 0
(1)
then y is of the form
if ,\ 1 and ,\ 2 are distinct roots of the associated characteristic equation ,\ 2 + a,\ +
b = 0. (The case where ,\ 1 = ,\ 2 was left unresolved.) Example 6. 1 2 and Exercise 20
in this section show that the set of solutions to Equation ( 1 ) forms a subspace of 9F,
the vector space of functions. In this section, we pursue these ideas further, paying
particular attention to the role played by vector spaces, bases, and dimension.
To set the stage, we consider a simpler class of examples. A differential equation
of the form
y' + ay = 0
�
Theorem 6 . 3 1
(2)
is called a first-order, homogeneous, linear differential equation. ("First-order"
refers to the fact that the highest derivative that is involved is a first derivative, and
"homogeneous" means that the right-hand side is zero. Do you see why the equa­
tion is "linear"?) A solution to Equation (2) is a differentiable function y = y (t) that
satisfies Equation (2) for all values of t.
It is easy to check that one solution to Equation (2) is y = e at (Do it.) However,
we would like to describe all solutions-and this is where vector spaces come in. We
have the following theorem.
-
.
The set S of all solutions to y' + ay = 0 is a subspace of ?F.
Since the zero function certainly satisfies Equation (2), S is nonempty. Let x
and y be two differentiable functions of t that are in S and let c be a scalar. Then
Proof
x' + ax = 0 and y' + ay = 0
so, using rules for differentiation, we have
( x + y) ' +
a (x + y) = x' + y' + ax + ay = (x' + ax) + (y ' + ay) = 0 + 0 = 0
and
(cy) ' + a (cy) = cy ' + c (ay) = c (y' + ay) = c · O = 0
Hence, x + y and cy are also in S, so S is a subspace of ?F.
Now we will show that S is a one-dimensional subspace of 9F and that { e - at } is a
basis. To this end, let x = x ( t) be in S. Then, for all t,
x' ( t) + ax (t) = 0 or x' ( t) = - ax ( t)
Section 6.7 Applications
519
Define a new function z (t) = x ( t) eat . Then, by the Chain Rule for differentiation,
z' ( t) = x ( t) ae at + x'(t)e at
= a x (t)eat - ax (t)ea t
=O
Since z ' is identically zero, z must be a constant function-say, z ( t) = k. But this
means that
x (t)e at = z ( t) = k for all t
so x ( t) = ke - a1 • Therefore, all solutions to Equation (2) are scalar multiples of the
single solution y = e - at . We have proved the following theorem.
Theorem 6 . 3 2
If S is the solution space of y' + ay = 0, then dim S = 1 and {e - a1 } is a basis for S.
One model for population growth assumes that the growth rate of the popula­
tion is proportional to the size of the population. This model works well if there are
few restrictions (such as limited space, food, or the like) on growth. If the size of the
population at time t is p(t), then the growth rate, or rate of change of the population,
is its derivative p' (t) . Our assumption that the growth rate of the population is pro­
portional to its size can be written as
p ' (t) = kp (t)
where k is the proportionality constant. Thus, p satisfies the differential equation
p' - kp = 0, so, by Theorem 6.32,
p (t) = ce k t
for some scalar c. The constants c and k are determined using experimental data.
Exa m p l e 6 . 8 1
in Michael
Crichton'is(Nmentioned
sewnovel
York:
Deln" inl, 1969),
alwasthough
the
"vi
l
a
i
that
novel
supposedl
y
an
alien
vi
r
us.
In
real
life,
contaminated
the
town
waterin 2000,
supplyresulting
of Walkerton,
Ontario,
in seven
deaths
and
causing
hundreds
people to become seriously ill.of
E. coli
The Andromeda
Strain
E. coli
The bacterium Escherichia coli (or E. coli, for short) is commonly found in the
intestines of humans and other mammals. It poses severe health risks if it escapes
into the environment. Under laboratory conditions, each cell of the bacterium divides
into two every 20 minutes. If we start with a single E. coli cell, how many will there
be after 1 day?
We do not need to use differential equations to solve this problem, but we
will, in order to illustrate the basic method.
To determine c and k, we use the data given in the statement of the problem. If
we take 1 unit of time to be 20 minutes, then we are given that p(O) = 1 and p(l) = 2.
Therefore,
Solution
c = c · 1 = ce k · o
=
1 and 2 = ee k · !
=
i
It follows that k = ln 2, so
p (t) = e t ln 2 = e ln 2 ' = 2 t
After 1 day, t = 72, so the number of bacteria cells will be p(72) = 2 7 2
(see Figure 6. 18).
=
4.72 X 1 0 2 1
520
Chapter 6 Vector Spaces
p (t)
5 x 1021
4.72 x 1021
4 x 1021
3 x 1021
I
I
I
I
I
I
I
I
I
I
2 x 1021
1
x
1021
0
0 10 20 30 40 50 60 70"' 72
'---+---+--I--+--.+---+�-+-'--
t
Exponential growth
Figure 6 . 1 8
Radioactive substances decay by emitting radiation. If m(t) denotes the mass of
the substance at time t, then the rate of decay is m '(t). Physicists have found that the
rate of decay of a substance is proportional to its mass; that is,
m '(t) = km (t) or m ' - km = 0
where k is a negative constant. Applying Theorem 6.32, we have
m (t) = ce k t
for some constant c. The time required for half of a radioactive substance to decay is
called its half-life.
Exa m p l e 6 . 8 8
After 5.5 days, a 1 00 mg sample of radon-222 decayed to 37 mg.
(a) Find a formula for m(t), the mass remaining after t days.
(b) What is the half-life of radon-222?
(c) When will only 1 0 mg remain?
Solution
(a) From m(t)
=
c/1, we have
100 = m (O) = ce k · o = c • 1 = c
so
m ( t) = l OOe k t
With time measured in days, we are given that m ( 5.5)
100 e 5 · 5 k = 37
so
e 55k = 0.37
Solving for k, we find
5.5k = ln ( 0.37 )
so
1
Therefore, m ( t) = l OOe- 0 · 81.
k=
ln ( 0.37 )
5.5
=
-0.18
=
37. Therefore,
Section 6.7 Applications
521
m (t)
1 00
80
60
40
"so
I
I
I
I
I
I
I
I
20
2
3.85 :
""
4
6
8
10
Radioactive decay
Figure 6 . 1 9
(b) To find the half-life of radon-222, we need the value of t for which m(t)
Solving this equation, we find
=
50.
l00e - 0 · 1 8 t = 50
e - o . 1 s t = 0.50
so
Hence,
- 0. 1 8t
and
t
=
ln ( t)
ln 2
0. 1 8
=
= -=
- ln 2
3.85
Thus, radon-222 has a half-life of approximately 3.85 days. (See Figure 6. 1 9.)
(c) We need to determine the value of t such that m(t) = 10. That is, we must solve
the equation
IOOe - o . 1 s t = 1 0 or e - o . 1 s t
=
0.1
Taking the natural logarithm o f both sides yields - 0 . 1 8t = ln 0. 1 . Thus,
ln 0. 1
12.79
-0.18
s o 1 0 mg o f the sample will remain after approximately 12.79 days.
t
SeeFriedberg, A. J. Insel,by andS. H.L. E.
Spence
(Englewood
Prentice-Hall,
1979).Cliffs, NJ:
Linear Algebra
= -- =
The solution set S of the second-order differential equation y " + ay' + by = 0
is also a subspace of ?F (Exercise 20), and it turns out that the dimension of S is 2.
Part (a) of Theorem 6.33, which extends Theorem 6.32, is implied by Theorem 4.40.
Our approach here is to use the power of vector spaces; doing so allows us to obtain
part (b) of Theorem 6.33 as well, a result that we could not obtain with our previous
methods.
Chapter 6 Vector Spaces
522
Theorem 6 . 3 3
Let S be the solution space of
y " + ay' + by = 0
and let A 1 and A 2 be the roots of the characteristic equation A 2 + a,\ + b = 0.
a. If A 1 * A 2 , then {e ;\ ' t, e ;\21} is a basis for S.
b. If A 1 = A 2 , then {e A1 1, te ;\ 1 1} is a basis for S.
R e m a rks
•
Observe that what the theorem says, in other words, is that the solutions of
y " + ay' + by = 0 are of the form
in the first case and
in the second case.
•
Compare Theorem 6.33 with Theorem 4.38. Linear differential equations
and linear recurrence relations have much in common. Although the former belong to
continuous mathematics and the latter to discrete mathematics, there are many parallels.
(a) We first show that {e A1 1, e A21} is contained in S. Let A be any root of the
characteristic equation and let f( t) = eA t . Then
Proof
J'(t) = Ae A t and J"(t) = A 2 e At
from which it follows that
+ af' + bf = A 2 e A t + aAe A t + be A t
= (A2 + a,\ + b)eA t
= 0 · e At = 0
Therefore, f is in S. But, since A 1 and A 2 are roots of the characteristic equation, this
means that e A, t and e A, t are in S.
The set {e A1 1, e A21} is also linearly independent, since if
f"
then, setting t = 0, we have
y
r
Y = e
e A2
Next, we set t = 1 to obtain
c 1 + c2 = 0 or c2 = - c 1
------------
But e A , - e A2 * 0, since e A , - e A2 = 0 implies that e A , = e\ which is clearly im­
possible if A 1 * A 2 . (See Figure 6.20.) We deduce that c 1 = 0 and, hence, c2 = 0, so
{e A1 1, e A2t} is linearly independent.
Since dim S = 2, {e A , t, e A21} must be a basis for S.
Figure 6 . 2 0
(b) You are asked to prove this property in Exercise 2 1 .
Section 6.7 Applications
Exa m p l e 6 . 8 9
523
Find all solutions of y 11 - Sy' + 6y = 0.
The characteristic equation is A 2 - SA + 6 = (A - 2) (,\ - 3) = 0. Thus,
the roots are 2 and 3, so {e 2 1, e 3 1} is a basis for the solution space. It follows that the
solutions to the given equation are of the form
Solution
The constants c 1 and c2 can be determined if additional equations, called bound­
ary conditions, are specified.
Exa m p l e 6 . 9 0
Find the solution of y11 + 6y' + 9y = 0 that satisfies y(O) = 1, y'(O) = 0.
The characteristic equation is A 2 + 6A + 9 = (A + 3) 2 = 0, so - 3 is a
repeated root. Therefore, {e - 3 1, te - 3 1} 
Download