Linear Independence of Vector Functions

advertisement
Chapter 3 Outline continued
Section 3.4: Determinants
Determinant of a Matrix
The determinant of square matrix A is a number |A| associated with the matrix.
The determinant of a two-by-two matrix is the product of the diagonal elements minus
the product of the off-diagonal elements:
a11 a12
 a11a22  a12 a21 .
a21 a22
Before calculating the determinant of an arbitrary n  n matrix, we need some
preliminaries. To calculate the determinant of an n  n matrix, we use a recursive
approach. A determinant of n  n matrix is related the determinant of matrices of
order (n  1)  (n  1) . Below outlines important concepts needed to calculate the
determinants.
Associated with the entry aij of the n  n matrix A =  aij  is a matrix of size
(n  1)  (n  1) called the minor of aij and denoted M ij ; The minor M ij is obtained
from A by deleting the ith row and the jth column. For each minor we calculate a
number called the cofactor of aij , Cij  (1)i  j M ij . The determinant of an n  n matrix
can be expressed in terms of the cofactors of any particular row or column. (It doesn’t
matter which one because they all give the same answer.)
Example to find the minor of a32 , cross out row 3 and column 2 of the matrix






 
  
(note the unfilled circles denote the entries crossed out) to obtain


 
   
M 32     .
  
The cofactor of an element of a matrix attaches alternate signs to the determinants of the
minors, according to checkerboard pattern:
    
    

 . For a the sign would be negative.
32
    


    
Determinant of an n  n Matrix A: Choose any row or column and expand by the
appropriate cofactor formula using either expansion by the ith row
n
A   aij Cij
j 1
or, equivalently, expansion by the jth column:
n
A   aij Cij .
i 1
Repeat this process, obtaining smaller matrices at each step. The definition is completed
with the 2  2 case.
NOTE: To shorten the amount of calculation, the strategy is to choose the row or column
for expansion that has the most zero entries!
Row Operations and Determinants:
Let A be a square matrix.
 If two rows of A are interchanged to produce matrix B, then
B  A .

If one row of A is multiplied by a constant k and then added to another row to
produce matrix B, then B  A .

If one row of A is multiplied by a constant k to produce matrix B, then B  k A .
Invertible Matrix Characterization (Using Determinants):
Let A be an n  n matrix. The following statements are equivalent:
 A is an invertible matrix
 A 0
Properties of Determinants
 For n  n matrices A and B, the determinant of AB is given by
AB  A B

AT  A
1
.
A

If A  0 , then A1 

If A is an upper triangular matrix or lower triangular matrix, then
m
A   aii , i.e. the determinant is the product of the diagonal elements.
i 1

If one row or column of A consists entirely of zeros, then A  0 .

If two rows or two column vectors of A are equal, then A  0 .
Section 3.5 Vector Spaces and Subspaces
Vector Space: A vector space V is a nonempty collection of objects called vectors for
which there are defined two operations, vector addition, denoted by x + y , and scalar
multiplication (multiplication by a real constant), denoted by cx , that satisfy the
following 10 properties. ( denotes the set of real numbers.)
Closure Properties:
C1. x + y  V whenever x, y  V
C2. cx  V whenever x  V and c 
Addition Properties:
A1. There is a zero vector, denoted by 0, in V such tat for all x  V ,
x + 0 = x . (Additive Identity)
A2. For every vector x  V , there is a vector in V called its negative, denoted -x , such
that x + (-x) = 0 . (Additive Inverse)
A3. (x + y )  z = x + (y  z ) , for all x, y, z  V . (Associativity)
A4. x + y = y  x , for all x, y  V . (Commutativity)
Scalar Multiplication Properties:
S1. 1x = x , for all x  V . (Scalar Multiplicative Identity)
S2. c(x + y ) = cx + cy , for all x, y  V and c  . (First Distributive Property)
S3. (c  d )x = cx + dx , for all x  V and c, d  . (Second Distributive Property)
S4. c(dx) = (cd )x , for all x  V and c, d  (Associativity)
Note that one check the two closure requirements at once by verifying the following
property, called the closure under linear combination:
C0. cx + dy  V whenever x, y  V and c, d  .
Definition
Vector spaces that are important in DEs (as well as other branches of mathematics) are
function spaces. Here the “vectors” are functions defined on an interval I, and they are
added and multiplied in the usual way:
( f  g )(t )  f (t )  g (t ) and (cf )(t )  cf (t ) , for all t  I .
Review Examples 1-6 on pp.167-168 and see how one would verify if something is a
vector space. Any of these problems would make up a good true/false question. I.E.
True/false The set V of all solutions of the first-order linear homogeneous DE,
y  p(t ) y  0 , defined on some interval I, where p(t) is a continuous function on I, is a
vector space.
Note that it is necessary that DEs be linear for the solutions to form a vector space and it
is also necessary for the DEs to be homogeneous!!
Prominent Vector Spaces
2
: space of all ordered pairs (or 2-vectors)
3
: space of all ordered triples (or 3-vectors)
n
: space of all ordered n-tuples (or n-vectors)
P : space of all polynomials
Pn : space of all polynomials of degree less than or equal to n
M mn : space of all m n matrices
C ( I ) : space of all continuous functions on the interval I (I may be an open or closed
interval, finite or infinite).
C n ( I ) : space of all functions on interval I having n continuous derivatives (I may
be an open or closed interval, finite or infinite).
C n : space of all ordered n-tuples of complex numbers (a  bi, c  di, )
Vector Subspaces
Often a vector space V has a subset W , that itself is a vector space. Most of the
addition and scalar multiplication are “inherited” from V . Checking whether a subset is
a subspace boils down to checking for closure.
Vector Subspace Theorem: A nonempty subset W of a vector space V is a subspace
of V if it is closed under addition and scalar multiplication, that is,
i)
If u, v  W , u + v  W
ii)
If u  W and c  , then cu  W .
An easy check for W being nonempty is to see if 0 is in W . If it is not, then W is not a
subspace.
As noted earlier, it is often efficient to check both closure properties at once by verifying
closure under “linear combinations”
If u, v  W , and a, b  , then au + bv  W .
Note: the only subspaces of 2 are the following:
 The zero subspace {(0,0)}, (the zero vector alone)
 Lines passing through the origin and
2

itself
Since the set consisting of the zero vector alone and the set V itself are always
subspaces, we call them trivial subspaces.
Note: the only subspaces of 3 are the following:
 The zero subspace {(0,0)}, (the zero vector alone)
 Lines and planes passing that contain the origin and
3

itself
Section 3.6 Basis and Dimension
Given vectors v1 , v 2 , , v n , we can form a new vector v = c1v1  c2 v 2   cn v n , where
c1 , c2 ,, cn are scalars, called a linear combination of the vectors. The set of all of
vectors we can make from a given set of vectors is called its span.
Definition
Span: The span of a set {v1 , v 2 , , v n } of vectors in a vector space V , denoted
Span {v1 , v 2 , , v n } , is the set of all linear combinations of these vectors.
1 
0 
0 
The vectors e1   0  , e2  1  , and e3  0  , which are the columns of the identity
 
 
 
 0 
0
1 
matrix I 3 , are called the standard basis vectors of 3 .
Definition Spanning sets in n
A vector b in n is in Span {v1 , v 2 , , v n } , where v1 , v 2 , , v n are vectors in n ,
provided that there is at least one solution of the matrix-vector equation Ax  b , where A
is the matrix whose column vectors are v1 , v 2 , , v n .
Span Theorem: For a set of vectors {v1 , v 2 ,
Span {v1 , v 2 , , v n } is a subspace of V .
, v n } in vector space V ,
Definition Column Space The number of elements m in the column vector of an m n
matrix to see that ColA is a subspace of m .
Linear Dependence and Independence
Definition
Linear Independence: A set {v1 , v 2 , , v n } in vector space V is linearly independent
if no vector of the set can be written as a linear combination of the others. Otherwise it is
linearly dependent.
A convenient criterion for checking if a set of vectors is linearly independent is to show
that if a linear combination of the vi is the zero vector, then all the coefficients must be
zero: c1v1  c2 v 2   cn v n  0  c1  c2   cn  0 .
Note: If any set of vectors includes 0, the zero vector, then the set is linearly dependent.
Linear Independence of Column Vectors:
The column vectors of the coefficient matrix A for Ax  0 if and only if the solution
x  0 is unique.
Linear Independence of Vector Functions
Linear Independence and Dependence of Vectors: The set of vectors
{v1 (t ), v 2 (t ), , v n (t )} on a vector space V is linearly dependent on an interval I if
there are scalars c1 , c2 ,, cn not all zero such that c1v1 (t )  c2 v 2 (t )   cn v n (t )  0 for all
t in I. Otherwise the set is linearly independent.
The Wronskian
Wronskian of Functions f1 , f 2 ,
W [ f1 , f 2 ,
, f n ](t ) 
, f n on I:
f1 (t )
f1(t )
f 2 (t )
f 2(t )
f n (t )
f n(t )
f1n 1 (t )
f 2n 1 (t )
f nn 1 (t )
defined on I.
The Wronskian and Linear Independence:
If W [ f1 , f 2 , , f n ](t )  0 , the zero function on I, where f1 , f 2 ,
{ f1 , f 2 ,
, f n are defined on I, then
, f n } is linearly independent set of functions on I.
Basis of a Vector Space
Definition
Basis of a Vector Space: The set {v1 , v 2 ,
i) {v1 , v 2 , , v n } is linearly independent
ii) Span {v1 , v 2 , , v n } = V .
The Standard Basis for
n
, v n } is a basis for vector space V provided
: The column vectors of the identity matrix I n are denoted
1 
0
0
0
1 
0
 
 
 
by e1 , e2 , , en , where e1  0  , e 2  0  , , e n   0  .
 
 
 
 
 
 
0 
0 
1 
The standard basis for n is { e1 , e2 , , en }.
Dimension of a Vector Space
Note the number of vectors in a basis is always the same for a particular vector space.
This fact allows us to define the dimension of a vector space.
The dimension of a (finite-dimensional) vector space is the number of vectors in any
basis.
A vector space so large that no finite set of vectors spans it is called infinitedimensional.
The Dimension of the Column Space of a Matrix
Column Space of a Matrix:
The pivot columns of a matrix A form a basis for ColA. Then the dimension of the
column space, denoted dim(ColA), is the number of pivot columns in A and is called the
rank of A, denoted rank A.
Invertible Matrix Characterization (Basis for Col A):
Let A be an n  n matrix. The following statements are equivalent:
 A is invertible
 The column vectors of A are linearly independent
 Every column of A is a pivot column
 The column vectors of A form a basis for ColA
 RankA=n.
Download