Module in Linear Algebra and Matrix Theory

advertisement
1
MODULE 1
MATRICES OVER A FIELD ๐‘ญ
Introduction
In this lesson, we will discuss the different operations on matrices and its properties,
the transpose of a matrix, the different types of matrices and the special types of square
matrices.
Objectives
1.
2.
3.
4.
5.
After going through this chapter, you are expected to be able to do the following:
Define matrix.
Give the different types of matrices.
Perform fundamental operations on matrices.
Find the transpose of a given matrix.
Identify special types of square matrices.
1.1 Definition of a Field
Definition 1.1.1: By a field ๐น we mean a nonempty set of elements with two laws of
combination, which we call addition and multiplication, satisfying the following conditions:
๐น1 Closure Properties
To every pair of elements ๐‘Ž, ๐‘ ∈ ๐น, ๐‘Ž + ๐‘ ∈ ๐น
To every pair of elements ๐‘Ž, ๐‘ ∈ ๐น, ๐‘Ž๐‘ ∈ ๐น
๐น2 Commutative Laws
๐‘Ž+๐‘ =๐‘+๐‘Ž
๐‘Ž๐‘ = ๐‘๐‘Ž ∀๐‘Ž, ๐‘ ∈ ๐น
๐น3 Associative Laws
(๐‘Ž + ๐‘) + ๐‘ = ๐‘Ž + (๐‘ + ๐‘)
(๐‘Ž๐‘)๐‘ = ๐‘Ž(๐‘๐‘) ∀๐‘Ž, ๐‘, ๐‘ ∈ ๐น
๐น4 Distributive Laws
(๐‘Ž + ๐‘)๐‘ = ๐‘Ž๐‘ + ๐‘๐‘
๐‘(๐‘Ž + ๐‘) = ๐‘๐‘Ž + ๐‘๐‘ ∀๐‘Ž, ๐‘, ๐‘ ∈ ๐น
๐น5 Identity Elements
∃0 ∈ ๐น such that ๐‘Ž + 0 = ๐‘Ž ∀๐‘Ž ∈ ๐น
∃1 ∈ ๐น, 1 ≠ 0 such that ๐‘Ž โˆ™ 1 = ๐‘Ž ∀๐‘Ž ∈ ๐น
2
๐น6 Inverse Elements
∀๐‘Ž ∈ ๐น, ∃ − ๐‘Ž ∈ ๐น such that ๐‘Ž + (−๐‘Ž) = 0
∀๐‘Ž ∈ ๐น, ๐‘Ž ≠ 0, ∃๐‘Ž−1 ∈ ๐น such that ๐‘Ž๐‘Ž −1 = 1
The elements of ๐น are called scalars. The set of real numbers โ„ and complex numbers โ„‚ are
examples of fields under the usual addition and multiplication in these sets
1.2 Definition of a Matrix
Definition 1.2.1: A matrix over a field ๐น is a rectangular array of elements from ๐น arranged
in ๐‘š horizontal rows and ๐‘› vertical columns:
๐‘Ž11 ๐‘Ž12 … ๐‘Ž1๐‘›
๐‘Ž21 ๐‘Ž22 … ๐‘Ž2๐‘›
๐ด=[ โ‹ฎ
โ‹ฎ
โ‹ฎ ]
๐‘Ž๐‘š1 ๐‘Ž๐‘š2 … ๐‘Ž๐‘š๐‘›
(1)
The ith row of A is
[๐‘Ž๐‘–1 ๐‘Ž๐‘–2 … ๐‘Ž1๐‘› ] (1 ≤ ๐‘– ≤ ๐‘š);
the jth column of A is
๐‘Ž1๐‘—
๐‘Ž2๐‘—
[ โ‹ฎ ] (1 ≤ ๐‘— ≤ ๐‘›).
๐‘Ž๐‘š๐‘—
The symbol ๐‘€๐‘š,๐‘› (๐น) denotes the collection of all ๐‘š x ๐‘› matrices over ๐น. We usually denote
matrices by capital letters. If ๐‘š is equal to ๐‘›, then the matrix is called a square matrix of
order ๐’.
Example 1: Below are examples of matrices of different sizes.
2 −5
i) ๐ด = [
] is a 2 x 2 matrix
3
4
3
ii) ๐ต = [−1
0
(A is a square matrix of order 2)
4 −8 −5
5
3
1 ] is a 3 x 4 matrix
7
2 −4
−8
The second row is [−1 5 3 1] and the third column is [ 3 ]
2
3
iii) ๐ถ = [
1
2
1/2
0
] is a 2 x 3 matrix
3
4/3
We will use the notation ๐ด = [๐‘Ž๐‘–๐‘— ] to denote a matrix A with entries ๐‘Ž๐‘–๐‘— . The subscripts ๐‘–
and ๐‘— will be used to denote the position of an entry ๐‘Ž๐‘–๐‘— in a matrix. Thus, the ๐‘–๐‘—th element
or entry of a matrix A is the number appearing in the ๐‘–th row and ๐‘—th column of A.
1 0
1 0 4
Example 2: Let ๐ด = [
] and ๐ต = [5 1
−2 2 3
3 2
−1
4 ].
−2
Then A is a 2 x 3 matrix with ๐‘Ž12 = 0 (the element appearing in the first row and the second
column), ๐‘Ž13 = 4, and ๐‘Ž23 = 3; B is a 3 x 3 matrix with ๐‘12 = 0, ๐‘13 = −1 , ๐‘21 = 5 , ๐‘23 =
4, ๐‘31 = 3 , and ๐‘32 = 2 .
If A is a square matrix of order ๐‘›, then the numbers ๐‘Ž11 , ๐‘Ž22 , … , ๐‘Ž๐‘›๐‘› form the main
diagonal. Thus in matrix B above, the elements ๐‘11 = 1, ๐‘22 = 1, and ๐‘33 = −2 form the
main diagonal.
Definition 1.2.2: (Equality of Matrices) Two ๐‘š x ๐‘› matrix ๐ด = [๐‘Ž๐‘–๐‘— ] and ๐ต = [๐‘๐‘–๐‘— ] are said
to be equal if ๐‘Ž๐‘–๐‘— = ๐‘๐‘–๐‘— , 1 ≤ ๐‘– ≤ ๐‘š , 1 ≤ ๐‘— ≤ ๐‘› , that is, if corresponding elements agree.
Example 3:
1 2
๐ด=[
2 −5
−3
1 2
] and ๐ต = [
4
2 ๐‘ฅ
๐‘ค
] are equal if ๐‘ฅ = −5 and ๐‘ค = −3.
4
1.3 Matrix Operations and Its Properties
Definition 1.3.1: (Addition of Matrices) If ๐ด = [๐‘Ž๐‘–๐‘— ] and ๐ต = [๐‘๐‘–๐‘— ] are ๐‘š x ๐‘› matrices, then
the sum of A and B is the ๐‘š x ๐‘› matrix ๐ถ = [๐‘๐‘–๐‘— ] defined by
[๐‘๐‘–๐‘— ] = [๐‘Ž๐‘–๐‘— ] + [๐‘๐‘–๐‘— ]
(1 ๏‚ฃ ๐‘– ≤ ๐‘š, 1 ๏‚ฃ ๐‘— ≤ ๐‘›)
That is, C is obtained by adding the corresponding elements of A and B.
4
๏ƒฉ1 ๏€ญ2 4 ๏ƒน
๏ƒฉ0 ๏€ญ2 4 ๏ƒน
Example 1: Let A ๏€ฝ ๏ƒช
and B ๏€ฝ ๏ƒช
๏ƒบ
๏ƒบ
๏ƒซ3 ๏€ญ1 5 ๏ƒป
๏ƒซ1 ๏€ญ3 1 ๏ƒป
Then ๐ด + ๐ต = [
1+0
3+1
(−2) + (−2) 4 + 4
1
]= [
(−1) + (−3) 5 + 1
4
−4 8
]
−4 6
Note that only matrices with the same size can be added.
Definition 1.3.2: (Scalar Multiplication) If ๐ด = [๐‘Ž๐‘–๐‘— ] is an ๐‘š x ๐‘› matrix and ๐‘Ÿ is a real
number, then the scalar multiple of A by ๐‘Ÿ, ๐‘ŸA, is the ๐‘š x ๐‘› matrix ๐ต = [๐‘๐‘–๐‘— ] , where
bij ๏€ฝ ra ij
(1 ๏‚ฃ ๐‘– ≤ ๐‘š, 1 ๏‚ฃ ๐‘— ≤ ๐‘›)
That is, B is obtained by multiplying each element of A by ๐‘Ÿ.
๏€ญ 2๏ƒน
๏ƒฉ4 3
Example 2: Let ๐‘Ÿ = −3 and A ๏€ฝ ๏ƒช
๏ƒบ . Then
๏ƒซ2 ๏€ญ 5 0 ๏ƒป
4
๐‘Ÿ๐ด = −3 [
2
−3(4) −3(3)
3 −2
] =[
−3(2) −3(−5)
−5
0
−3(−2)
−12 −9 6
]= [
]
−3(0)
−6 15 0
Definition 1.3.3: (Matrix Multiplication) If ๐ด = [๐‘Ž๐‘–๐‘— ] is an ๐‘š x ๐‘› matrix and ๐ต = [๐‘๐‘–๐‘— ] is an
๐‘› x ๐‘ matrix, then the product of A and B is the ๐‘š x ๐‘ matrix ๐ถ = [๐‘๐‘–๐‘— ], defined by
๐‘๐‘–๐‘— = ∑๐‘›๐‘˜=1 ๐‘Ž๐‘–๐‘˜ ๐‘๐‘˜๐‘—
To illustrate this, let us consider an ๐‘š x ๐‘› matrix ๐ด and an ๐‘› x ๐‘ matrix ๐ต.
๐‘—th column of ๐ต
๐‘–th row of ๐ด
๐‘Ž11
๐‘Ž21
โ‹ฎ
๐‘Ž๐‘–1
โ‹ฎ
[ ๐‘Ž๐‘š1
๐‘Ž12
๐‘Ž22
โ‹ฎ
๐‘Ž๐‘–2
โ‹ฎ
๐‘Ž๐‘š2
… ๐‘Ž1๐‘›
๐‘11 ๐‘12 … ๐‘1๐‘— … ๐‘1๐‘
… ๐‘Ž2๐‘›
๐‘21 ๐‘22 … ๐‘2๐‘— … ๐‘2๐‘
โ‹ฎ
โ‹ฎ
โ‹ฎ
โ‹ฎ
โ‹ฎ
… ๐‘Ž๐‘–๐‘›
๐‘๐‘›1 ๐‘๐‘›2 … ๐‘๐‘›๐‘— … ๐‘๐‘›๐‘
โ‹ฎ
[
]
… ๐‘Ž๐‘š๐‘› ]
Note that the ๐‘–th row of ๐ด and the ๐‘—th column of ๐ต must have the same number of
components for product AB to be defined.
5
If we let AB = C then the ๐‘–๐‘—th element of AB is
๐‘๐‘–๐‘— = [๐‘Ž๐‘–1 ๐‘Ž๐‘–2
Example 3: If ๐ด = [
๐‘1๐‘—
๐‘
… ๐‘Ž๐‘–๐‘› ] 2๐‘— = ๐‘Ž๐‘–1 ๐‘๐‘–๐‘— + ๐‘Ž๐‘–2 ๐‘2๐‘— + … + ๐‘Ž๐‘–๐‘› ๐‘๐‘›๐‘—
โ‹ฎ
[๐‘๐‘›๐‘— ]
1 3
−2 6
] and ๐ต = [
] then
−2 5
4 7
−2
๐‘11 = [1 3] [ ] = 1(−2) + 3(4) = 10
4
−2
๐‘21 = [−2 5] [ ] = (−2)(−2) + 5(4) = 24
4
6
๐‘12 = [1 3] [ ] = 1(6) + 3(7) = 27
7
6
๐‘22 = [−2 5] [ ] = (−2)(6) + 5(7) = 23
7
10
Thus ๐ด๐ต = [
24
Similarly, ๐ต๐ด = [
27
]
23
(−2)(1) + 6(−2) (−2)(3) + 6(5)
−2 6
1 3
][
]= [
]
4(1) + 7(−2)
4(3) + 7(5)
4 7 −2 5
−14
= [
−10
24
]
47
Example 3 shows that matrix multiplication, in general, is not commutative.
7 −1
2
0 5
Example 4: Let ๐ด = [
] and ๐ต = [−2
5
4 −3 1
3
2
Find a) AB and b) BA.
1
6
4 −4].
−3
2
6
Solution:
a. Note that the number of columns of A is equal to the number of rows of B hence the
product AB is defined. By Definition 1.3.3, AB is a 2 x 4 matrix.
2
๐ด๐ต = [
4
= [
= [
7 −1
0 5
] [−2
5
−3 1
3
2
14 + 0 + 15
28 + 6 + 3
29
8
37 −17
1
6
4 −4]
−3
2
(−2) + 0 + 10
(−4) + (−15) + 2
2 + 0 + (−15)
4 + (−12) + (−3)
12 + 0 + 10
]
24 + 12 + 2
−13 22
]
−11 38
b. The product BA is not defined because the number of columns of B is not equal to
number of rows of A.
the
SAQ 1-1
−2 5
3 −1
] and ๐ต = [−4 3]. Find
1 −4
3 1
a. ๐ด๐ต and ๐ต๐ด
1
Let ๐ด = [
2
Solve SAQ1-1 in your notebook and compare your answer with ASAQ1-1.
ASAQ 1-1
The number of columns of ๐ด is equal to the number of rows of ๐ต hence ๐ด๐ต is defined.
Similarly, the number of columns of ๐ต is equal to the number of rows of ๐ด thus ๐ต๐ด is also
defined. Multiplying we have
๏ƒฉ1(๏€ญ2) ๏€ซ 3(๏€ญ4) ๏€ซ (๏€ญ1)(3) 1(5) ๏€ซ 3(3) ๏€ซ (๏€ญ1)(1) ๏ƒน
AB ๏€ฝ ๏ƒช
๏ƒบ
๏ƒซ 2(๏€ญ2) ๏€ซ 1(๏€ญ4) ๏€ซ (๏€ญ4)(3) 2(5) ๏€ซ 1(3) ๏€ซ (๏€ญ4)(1) ๏ƒป
๏ƒฉ ๏€ญ17 13๏ƒน
๏€ฝ๏ƒช
๏ƒบ
๏ƒซ ๏€ญ20 9 ๏ƒป
7
and
๏ƒฉ ๏€ญ2 5๏ƒน
๏ƒฉ8 ๏€ญ1 ๏€ญ18๏ƒน
๏ƒฉ1 3 ๏€ญ1๏ƒน ๏ƒช
๏ƒช
๏ƒบ
BA ๏€ฝ ๏ƒช ๏€ญ4 3๏ƒบ ๏ƒช
๏€ฝ ๏ƒช 2 ๏€ญ9 ๏€ญ8 ๏ƒบ๏ƒบ
๏ƒบ
2 1 ๏€ญ4๏ƒป
๏ƒช๏ƒซ3 1 ๏ƒบ๏ƒป ๏ƒซ
๏ƒช๏ƒซ5 10 ๏€ญ7 ๏ƒบ๏ƒป
Remarks: If ๐ด = [๐‘Ž๐‘–๐‘— ] is an ๐‘š x ๐‘› matrix and ๐ต = [๐‘๐‘–๐‘— ] is an ๐‘› x ๐‘ matrix then,
1. BA may not be defined as in Example 4.b; this will take place if ๐‘ ≠ ๐‘š.
2. If BA is defined, which means ๐‘ = ๐‘š, then BA is ๐‘› x ๐‘› while AB is ๐‘š x ๐‘; thus if ๐‘› ≠
๐‘, AB and BA are of different size (see ASAQ1-1).
3. If AB and BA are both of the same size, they may be equal.
4. If AB and BA are both of the same size, they may be unequal (see Example 3).
Definition 1.3.4: (Additive Inverse of a Matrix) Let ๐ด = [๐‘Ž๐‘–๐‘— ]. Then – ๐ด is the matrix
obtained by replacing the elements of ๐ด by their additive inverses; that is
−๐ด = −[๐‘Ž๐‘–๐‘— ] = [−๐‘Ž๐‘–๐‘— ]
Example 5. Let ๐ด = [
−2
5
1/2
] then
3 2/3 −17
–๐ด = [
2
−5
−1/2
]
−3 −2/3
17
Definition 1.3.5: (The Zero Matrix) An ๐‘š x ๐‘› matrix ๐ด = [๐‘Ž๐‘–๐‘— ] is called a zero matrix and is
denoted by 0 if ๐‘Ž๐‘–๐‘— = 0 , 1 ๏‚ฃ ๐‘– ≤ ๐‘š, 1 ๏‚ฃ ๐‘— ≤ ๐‘›; that is all entries are equal to zero.
Properties of Matrix Operations
Theorem 1.3.1: (Properties of Matrix Addition) Let A, B, and C be ๐‘š x ๐‘› matrices then:
(a) A + B = B + A
(commutative law for matrix addition)
(b) A + ( B + C ) = ( A + B ) + C
(associative law for matrix addition)
(c) A + O = A
(The matrix O is the ๐’Ž x ๐’ zero matrix)
(d) there is a unique ๐‘š x ๐‘› matrix (−๐ด) such that
๐ด + (−๐ด) = ๐‘‚
We will prove part (a) and leave the proof of the remaining parts as an exercise.
8
Proof of part (a):
Let ๐ด = [๐‘Ž๐‘–๐‘— ] and ๐ต = [๐‘๐‘–๐‘— ] be ๐‘š x ๐‘› matrices. By Definition 1.3.1
(1 ๏‚ฃ ๐‘– ≤ ๐‘š, 1 ๏‚ฃ ๐‘— ≤ ๐‘›)
๐ด + ๐ต = ๐‘Ž๐‘–๐‘— + ๐‘๐‘–๐‘—
= ๐‘๐‘–๐‘— + ๐‘Ž๐‘–๐‘—
since a, b ∈ ๐‘… then a + b = b + a for any real
numbers a and b.
= ๐ต + ๐ด
Therefore, matrix addition is commutative.
Example 6: To illustrate part (a) of Theorem 1.3.1, let
๏ƒฉ 4 ๏€ญ1๏ƒน
๏ƒฉ 4 6๏ƒน
A๏€ฝ๏ƒช
and B ๏€ฝ ๏ƒช
๏ƒบ
๏ƒบ
๏ƒซ3 2 ๏ƒป
๏ƒซ10 2 ๏ƒป
Then,
๏ƒฉ 4 ๏€ญ1๏ƒน ๏ƒฉ 4 6 ๏ƒน ๏ƒฉ 8 5 ๏ƒน
A๏€ซ B ๏€ฝ ๏ƒช
๏ƒบ๏€ซ๏ƒช
๏ƒบ๏€ฝ๏ƒช
๏ƒบ
๏ƒซ 3 2 ๏ƒป ๏ƒซ10 2๏ƒป ๏ƒซ13 4๏ƒป
and
๏ƒฉ 4 6 ๏ƒน ๏ƒฉ 4 ๏€ญ1๏ƒน ๏ƒฉ 8 5 ๏ƒน
B๏€ซ A๏€ฝ ๏ƒช
๏ƒบ๏€ซ๏ƒช
๏ƒบ๏€ฝ๏ƒช
๏ƒบ
๏ƒซ10 2 ๏ƒป ๏ƒซ 3 2 ๏ƒป ๏ƒซ13 4๏ƒป
Example 7: To illustrate part (c) of Theorem 1.3.1, let
๏ƒฉ1 2 ๏€ญ1๏ƒน
๏ƒฉ0 0 0 ๏ƒน
๏ƒช
๏ƒบ
A ๏€ฝ ๏ƒช 3 2 ๏€ญ2๏ƒบ and O ๏€ฝ ๏ƒช๏ƒช0 0 0๏ƒบ๏ƒบ . Note that O is the 3 x 3 zero matrix.
๏ƒช๏ƒซ 4 5 3 ๏ƒบ๏ƒป
๏ƒช๏ƒซ0 0 0๏ƒบ๏ƒป
Then,
1
๐ด + ๐‘‚ = [3
4
2 −1
0 0
2 −2] + [0 0
5
3
0 0
0
1
0] = [3
0
4
Example 8: To illustrate part (d) of Theorem 1.3.1, let
๏ƒฉ2 2 3 ๏ƒน
๏ƒฉ ๏€ญ2 ๏€ญ2 ๏€ญ3๏ƒน
A๏€ฝ๏ƒช
and ๏€ญ A ๏€ฝ ๏ƒช
๏ƒบ
๏€ญ2 ๏ƒบ๏ƒป
๏ƒซ3 ๏€ญ1 2 ๏ƒป
๏ƒซ ๏€ญ3 1
2 −1
2 −2] = ๐ด
5
3
9
Then,
๏ƒฉ 2 2 3 ๏ƒน ๏ƒฉ ๏€ญ2 ๏€ญ2 ๏€ญ3๏ƒน ๏ƒฉ0 0 0 ๏ƒน
A ๏€ซ (๏€ญ A) ๏€ฝ ๏ƒช
๏€ฝ
๏ƒบ๏€ซ๏ƒช
๏€ญ2๏ƒบ๏ƒป ๏ƒช๏ƒซ0 0 0 ๏ƒบ๏ƒป
๏ƒซ3 ๏€ญ1 2๏ƒป ๏ƒซ ๏€ญ3 1
Theorem 1.3.2: (Properties of Matrix Multiplication) Let ๐ด = [๐‘Ž๐‘–๐‘— ] be an ๐‘š x ๐‘› matrix, ๐ต =
๐‘๐‘–๐‘— be an ๐‘› x ๐‘ matrix, and ๐ถ = ๐‘๐‘–๐‘— be a ๐‘ x ๐‘ž matrix. Then
(a) (๐ด๐ต)๐ถ = ๐ด(๐ต๐ถ)
(associative law for matrix multiplication)
(b) ๐ด(๐ต + ๐ถ) = ๐ด๐ต + ๐ด๐ถ
(left distributive law for matrix multiplication)
(c) (๐ด + ๐ต)๐ถ = ๐ด๐ถ + ๐ต๐ถ
(right distributive law for matrix multiplication)
Proof of part (a):
Note that AB is an ๐‘š x ๐‘ matrix hence (AB)C is an ๐‘š x ๐‘ž matrix. Similarly, BC is an ๐‘› x ๐‘ž
matrix hence A(BC) is an ๐‘š x ๐‘ž matrix. We see that (AB)C and A(BC) have the same size.
Now we must show that the corresponding parts of (AB)C and A(BC) are equal. Let AB = D =
[๐‘‘๐‘–๐‘— ]. Then
๐‘‘๐‘–๐‘— = ∑๐‘›๐‘˜=1 ๐‘Ž๐‘–๐‘˜ ๐‘๐‘˜๐‘—
The ๐‘–๐‘—th component of (AB)C = DC is
∑๐‘๐‘ =1 ๐‘‘๐‘–๐‘  ๐‘๐‘ ๐‘— = ∑๐‘๐‘ =1(∑๐‘›๐‘˜=1 ๐‘Ž๐‘–๐‘˜ ๐‘๐‘˜๐‘  )๐‘๐‘ ๐‘—
๐‘
= ∑๐‘ =1 ∑๐‘›๐‘˜=1 ๐‘Ž๐‘–๐‘˜ ๐‘๐‘˜๐‘  ๐‘๐‘ ๐‘—
Now we let BC = E = [๐‘’๐‘–๐‘— ] then
๐‘
๐‘’๐‘˜๐‘— = ∑๐‘ =1 ๐‘๐‘˜๐‘  ๐‘๐‘ ๐‘—
Thus the ๐‘–๐‘—th component of A(BC) = AE is
∑๐‘›๐‘˜=1 ๐‘Ž๐‘–๐‘˜ ๐‘’๐‘˜๐‘— = ∑๐‘›๐‘˜=1 ∑๐‘๐‘ =1 ๐‘Ž๐‘–๐‘˜ ๐‘๐‘˜๐‘  ๐‘๐‘ ๐‘— = ๐‘–๐‘—th component of (AB)C.
This shows that (๐ด๐ต)๐ถ = ๐ด(๐ต๐ถ).
10
Example 9: To illustrate part (a) of Theorem 1.3.2 we let
๏ƒฉ1
๏ƒฉ 2 3 ๏€ญ1 0 ๏ƒน
๏ƒช2
๏ƒฉ3 ๏€ญ2 4 ๏ƒน
๏ƒช
๏ƒบ
๏ƒช
A๏€ฝ๏ƒช
๏ƒบ , B ๏€ฝ ๏ƒช0 2 ๏€ญ2 2๏ƒบ , and C ๏€ฝ ๏ƒช0
๏€ญ
1
๏€ญ
2
3
๏ƒซ
๏ƒป
๏ƒช๏ƒซ1 0 ๏€ญ1 3 ๏ƒบ๏ƒป
๏ƒช
๏ƒซ ๏€ญ2
0 ๏€ญ2 ๏ƒน
๏€ญ3 0 ๏ƒบ๏ƒบ
1
๏€ญ3 ๏ƒบ
๏ƒบ
0 0 ๏ƒป
Then,
๏ƒฉ 8 ๏€ญ10 ๏€ญ1๏ƒน
๏€ญ18 ๏€ญ11๏ƒน
๏ƒฉ3 ๏€ญ2 4๏ƒน ๏ƒช
๏ƒบ ๏€ฝ ๏ƒฉ4
A( BC ) ๏€ฝ ๏ƒช
0
๏€ญ
8
6
๏ƒบ
๏ƒช
๏ƒบ
๏ƒช
๏ƒบ
๏ƒซ ๏€ญ1 ๏€ญ2 3 ๏ƒป ๏ƒช ๏€ญ5 ๏€ญ1 1 ๏ƒบ ๏ƒซ ๏€ญ23 23 ๏€ญ8 ๏ƒป
๏ƒซ
๏ƒป
and
๏ƒฉ1
๏€ญ3 8 ๏ƒน ๏ƒช๏ƒช 2
๏ƒฉ10 5
( AB)C ๏€ฝ ๏ƒช
๏ƒบ
๏ƒซ1 ๏€ญ7 2 5๏ƒป ๏ƒช0
๏ƒช
๏ƒซ ๏€ญ2
0 ๏€ญ2 ๏ƒน
๏€ญ3 0 ๏ƒบ๏ƒบ ๏ƒฉ 4
๏€ญ18 ๏€ญ11๏ƒน
๏€ฝ๏ƒช
1
๏€ญ3 ๏ƒบ ๏ƒซ ๏€ญ23 23 ๏€ญ8 ๏ƒบ๏ƒป
๏ƒบ
0 0 ๏ƒป
This shows that matrix multiplication is associative.
Proof of part (b):
Let ๐ด = [๐‘Ž๐‘–๐‘— ] be an ๐‘š x ๐‘› matrix and let ๐ต = [๐‘๐‘–๐‘— ] and ๐ถ = [๐‘๐‘–๐‘— ] be ๐‘› x ๐‘ matrices. Then
the ๐‘˜๐‘—th element of B + C is
๐‘๐‘˜๐‘— + ๐‘๐‘˜๐‘—
and the ๐‘–๐‘—th element of A(B + C) is
∑๐‘›๐‘˜=1 ๐‘Ž๐‘–๐‘˜ (๐‘๐‘˜๐‘— + ๐‘๐‘˜๐‘— ) = ∑๐‘›๐‘˜=1 ๐‘Ž๐‘–๐‘˜ ๐‘๐‘˜๐‘— + ∑๐‘›๐‘˜=1 ๐‘Ž๐‘–๐‘˜ ๐‘๐‘˜๐‘—
= ๐‘–๐‘—th component of AB + ๐‘–๐‘—th component of AC
Thus, A(B + C) = AB + AC. The proof of part (c) is similar to the proof of part (b) and
is left as an exercise.
0๏ƒน
๏ƒฉ1
๏ƒฉ0 2 ๏ƒน
๏ƒฉ2 2 4๏ƒน
๏ƒช
๏ƒบ
Example 10: Let A ๏€ฝ ๏ƒช
, B ๏€ฝ ๏ƒช ๏€ญ2 3 ๏ƒบ , and C ๏€ฝ ๏ƒช๏ƒช ๏€ญ1 1 ๏ƒบ๏ƒบ
๏ƒบ
๏ƒซ ๏€ญ4 ๏€ญ1 3 ๏ƒป
๏ƒช๏ƒซ5 1 ๏ƒบ๏ƒป
๏ƒช๏ƒซ 2 0 ๏ƒบ๏ƒป
11
Then
2๏ƒน
๏ƒฉ1
๏ƒฉ2 2 4๏ƒน ๏ƒช
๏ƒฉ 24 16 ๏ƒน
A( B ๏€ซ C ) ๏€ฝ ๏ƒช
๏€ญ3 4๏ƒบ๏ƒบ ๏€ฝ ๏ƒช
๏ƒบ
๏ƒบ
๏ƒช
๏ƒซ ๏€ญ4 ๏€ญ1 3 ๏ƒป ๏ƒช7 1 ๏ƒบ ๏ƒซ 20 ๏€ญ9๏ƒป
๏ƒซ
๏ƒป
and
๏ƒฉ18 10๏ƒน ๏ƒฉ6 6 ๏ƒน ๏ƒฉ 24 16 ๏ƒน
AB ๏€ซ AC ๏€ฝ ๏ƒช
๏ƒบ๏€ซ๏ƒช
๏ƒบ๏€ฝ๏ƒช
๏ƒบ
๏ƒซ13 0 ๏ƒป ๏ƒซ7 ๏€ญ9๏ƒป ๏ƒซ 20 ๏€ญ9๏ƒป
Theorem 1.3.3: (Properties of Scalar Multiplication) If r and s are real numbers and A and B
are matrices then
(a) r(sA) = (rs)A
(b) (r + s)A = rA + sA
(c) r(A + B) = rA + rB
(d) A(rB) = r(AB) = (rA)B
We prove part (c) of the theorem and leave the proof of the other parts as an exercise.
Proof:
Let ๐ด = [๐‘Ž๐‘–๐‘— ] and ๐ต = [๐‘๐‘–๐‘— ] be ๐‘š x ๐‘› matrices. Then the
๐‘–๐‘—th entry of ๐‘Ÿ(๐ด + ๐ต) = ๐‘Ÿ(๐‘Ž๐‘–๐‘— + ๐‘๐‘–๐‘— )
definition of scalar multiplication
= ๐‘Ÿ๐‘Ž๐‘–๐‘— + ๐‘Ÿ๐‘๐‘–๐‘—
distributive property of real numbers
= ๐‘–๐‘—th entry of ๐‘Ÿ๐ด + ๐‘–๐‘—th entry of ๐‘Ÿ๐ต
definition of scalar
multiplication
Hence ๐‘Ÿ(๐ด + ๐ต) = ๐‘Ÿ๐ด + ๐‘Ÿ๐ต.
Example 11: To illustrate part (d) of Theorem 1.3.3 we let
Then,
3
1 3 2
๐‘Ÿ = −3, ๐ด = [
], and ๐ต = [2
2 −1 3
0
๐ด(๐‘Ÿ๐ต) = [
3 −4
1
3 2
] ((−3) [2
1 ])
2 −1 3
0
5
−9
12
1
3 2
= [
] [−6 −3 ]
2 −1 3
0 −15
−4
1]
5
12
−27 −27
= [
]
−12 −18
and
3 −4
1
3 2
๐‘Ÿ(๐ด๐ต) = −3 ([
] [2
1 ])
2 −1 3
0
5
= −3 [
9 9
]
4 6
−27 −27
= [
]
−12 −18
Hence ๐ด(๐‘Ÿ๐ต) = ๐‘Ÿ(๐ด๐ต).
1.4 Transpose of a Matrix
Definition 1.4.1: (The Transpose of a Matrix) If ๐ด = [๐‘Ž๐‘–๐‘— ] is an ๐‘š x ๐‘› matrix, then the ๐‘› x ๐‘š
matrix AT ๏€ฝ [aij ] , where
T
aij ๏€ฝ a ji
T
(1 ๏‚ฃ ๐‘– ≤ ๐‘š, 1 ๏‚ฃ ๐‘— ≤ ๐‘›)
is called the transpose of A. Thus the transpose of A is obtained by interchanging the rows
and columns of A.
4 −2 −3
Example 1: Let ๐ด = [
], ๐ต = [3
0 −5 −2
4
0
๐ด๐‘‡ = [−2 −5]
−3 −2
2
],
and
๐ถ
=
[
−1]. Then
5 −1
4
Note that the first row of A became the first column of ๐ด๐‘‡ and the second row of A became
the second column of ๐ด๐‘‡ . Similarly
3
๐ต ๐‘‡ = [ 5 ] and ๐ถ ๐‘‡ = [2
−1
−1 4]
13
Theorem 1.4.1: (Properties of Transpose) If r is a scalar and A and B are matrices, then
(a) (AT)T = A
(b) (A + B)T = AT + BT
(c) (AB)T = BTAT
(d) (rA)T = rAT
Proof of part (a):
Let ๐ด = [๐‘Ž๐‘–๐‘— ]. Then the ๐‘–๐‘—th entry of A is
๐‘Ž๐‘–๐‘— = ๐‘Ž๐‘—๐‘–๐‘‡
definition of transpose of a matrix
= (๐‘Ž๐‘‡ )๐‘‡๐‘–๐‘—
definition of a transpose of a matrix
๐‘‡ ๐‘‡
Hence (๐ด ) = ๐ด.
Proof of part (b):
Let [(๐ด + ๐ต)๐‘‡ ]๐‘–๐‘— be the ๐‘–๐‘—th component of (๐ด + ๐ต)๐‘‡ and (๐ด + ๐ต)๐‘—๐‘– denote the ๐‘—๐‘–th
component of ๐ด + ๐ต. Then
[(๐ด + ๐ต)๐‘‡ ]๐‘–๐‘— = (๐ด + ๐ต)๐‘—๐‘–
definition of a transpose
= ๐‘Ž๐‘—๐‘– + ๐‘๐‘—๐‘–
definition of addition of matrices
= (๐ด๐‘‡ )๐‘–๐‘— + (๐ต ๐‘‡ )๐‘–๐‘—
definition of a transpose
= ๐‘–๐‘—th entry of ๐ด๐‘‡ + ๐‘–๐‘—th entry of ๐ต ๐‘‡
Thus (๐ด + ๐ต)๐‘‡ = ๐ด๐‘‡ + ๐ต ๐‘‡ .
The proof of parts (c) and (d) is left as an exercise.
Example 2: To illustrate part (c) of Theorem 1.4.1 we let
๏ƒฉ0 ๏€ญ1๏ƒน
B ๏€ฝ ๏ƒช๏ƒช 2 ๏€ญ2๏ƒบ๏ƒบ
๏ƒช๏ƒซ3 ๏€ญ1๏ƒบ๏ƒป
2 ๏ƒน
๏ƒฉ1
๏ƒฉ0 2 3 ๏ƒน ๏ƒช
๏ƒฉ0 7 ๏ƒน
๏ƒบ ๏€ฝ ๏ƒฉ0 7 ๏ƒน
T T
T
๏€ญ
3
๏€ญ
1
Then ( AB) ๏€ฝ ๏ƒช
and B A ๏€ฝ ๏ƒช
๏ƒบ
๏ƒบ
๏ƒบ
๏ƒช
๏ƒบ ๏ƒช
๏ƒซ 3 ๏€ญ3๏ƒป
๏ƒซ ๏€ญ1 ๏€ญ2 ๏€ญ1๏ƒป ๏ƒช 2 3 ๏ƒบ ๏ƒซ3 ๏€ญ3๏ƒป
๏ƒซ
๏ƒป
๏ƒฉ1 ๏€ญ3 2๏ƒน
A๏€ฝ๏ƒช
๏ƒบ and
๏ƒซ 2 ๏€ญ1 3 ๏ƒป
14
1.5 Special Types of Square Matrices
Definition 1.5.1: A square matrix ๐ด = [๐‘Ž๐‘–๐‘— ] for which every term off the main diagonal is
zero, that is, ๐‘Ž๐‘–๐‘— = 0 for ๐‘– ≠ ๐‘—, is called a diagonal matrix.
Example 1: The following are diagonal matrices.
๏ƒฉ ๏€ญ5 0 0๏ƒน
๏ƒฉ3 0 ๏ƒน
and H ๏€ฝ ๏ƒช๏ƒช 0 ๏€ญ8 0 ๏ƒบ๏ƒบ
G๏€ฝ๏ƒช
๏ƒบ
๏ƒซ0 ๏€ญ1๏ƒป
๏ƒช๏ƒซ 0 0 3๏ƒบ๏ƒป
Definition 1.5.2: A diagonal matrix ๐ด = [๐‘Ž๐‘–๐‘— ], for which all terms on the main diagonal are
equal, that is, ๐‘Ž๐‘–๐‘— = ๐‘ for ๐‘– = ๐‘— and ๐‘Ž๐‘–๐‘— = 0 for ๐‘– ≠ ๐‘—, is called a scalar matrix.
Example 2: The following are examples of scalar matrices.
3
๐ต = [0
0
0 0
−7
0
]
3 0] and ๐ถ = [
0 −7
0 3
Definition 1.5.3: A matrix ๐ด = [๐‘Ž๐‘–๐‘— ] is called upper triangular if a ij ๏€ฝ 0 for i > j. It is called
lower triangular if a ij ๏€ฝ 0 for i < j.
Definition 1.5.4: A matrix ๐ด = [๐‘Ž๐‘–๐‘— ] is called strictly upper triangular if a ij ๏€ฝ 0 for i
is called strictly lower triangular if a ij ๏€ฝ 0 for i
Example 3:
0
๐ท = [0
0
8
0
0
j. It
j.
๏ƒฉ๏€ญ 2 0 0๏ƒน
๏ƒฉ1 3 5 ๏ƒน
0 0 0
๏ƒช
๏ƒบ
๏ƒช
๏ƒบ
Let A = ๏ƒช0 2 ๏€ญ 1๏ƒบ , B = ๏ƒช 5 ๏€ญ 1 0๏ƒบ , ๐ถ = [2 0 0] and
3 7 0
๏ƒช๏ƒซ 6 ๏€ญ 3 4๏ƒบ๏ƒป
๏ƒช๏ƒซ0 0 ๏€ญ 4๏ƒบ๏ƒป
4
−9]
0
Matrix A is upper triangular since all entries below the main diagonal are zero. Matrix B is
lower triangular since all entries above the main diagonal are zero while C and D are strictly
lower triangular and strictly upper triangular, respectively. Strictly upper/lower triangular
matrices are upper/ lower triangular matrices with all entries on the main diagonal equal to
zero.
15
There are two important types of square matrices that can be defined in terms of
the transpose operation. These are symmetric and skew symmetric matrices which we
define as follows:
Definition 1.5.5: A matrix ๐ด = [๐‘Ž๐‘–๐‘— ] is called symmetric if AT = A.
๏ƒฉ 1 ๏€ญ2 ๏€ญ3๏ƒน
1 −2 −3
๏ƒช
๏ƒบ
๐‘‡
Example 4: Let A ๏€ฝ ๏ƒช ๏€ญ2 5 0 ๏ƒบ then ๐ด = [−2
5
0 ]. Since ๐ด๐‘‡ = ๐ด then A is
−3
0
4
๏ƒช๏ƒซ ๏€ญ3 0 4 ๏ƒบ๏ƒป
symmetric.
SAQ 1-2
Let ๐ด and ๐ต be symmetric matrices. Show that ๐ด + ๐ต is symmetric.
ASAQ 1-2
(๐ด + ๐ต)๐‘‡ = ๐ด๐‘‡ + ๐ต ๐‘‡
=๐ด+๐ต
Part (b) of Theorem 1.4.1
Definition of symmetric matrix
Hence ๐ด + ๐ต is symmetric.
Definition 1.5.6: A matrix ๐ด = [๐‘Ž๐‘–๐‘— ] is called skew symmetric if ๐ด๐‘‡ = −๐ด.
๏ƒฉ 0 ๏€ญ3 2 ๏ƒน
Example 5: The matrix B ๏€ฝ ๏ƒช๏ƒช 3 0 4 ๏ƒบ๏ƒบ is skew symmetric because
๏ƒช๏ƒซ ๏€ญ2 ๏€ญ4 0 ๏ƒบ๏ƒป
0 3
๐ต ๐‘‡ = [−3 0
2 4
−2
−4] = −๐ต.
0
16
SAQ 1-3
Let ๐ด and ๐ต be skew-symmetric matrices. Show that ๐ด + ๐ต is skew-symmetric.
ASAQ 1-3
(๐ด + ๐ต)๐‘‡ = ๐ด๐‘‡ + ๐ต ๐‘‡
= (−๐ด) + (−๐ต)
= −(๐ด + ๐ต)
Part (b) of Theorem 1.4.1
Definition of skew-symmetric matrix
Thus ๐ด + ๐ต is skew-symmetric.
Definition 1.5.7: (The Identity Matrix) The ๐‘› x ๐‘› matrix ๐ผ๐‘› = [๐‘Ž๐‘–๐‘— ] defined by ๐‘Ž๐‘–๐‘— = 1 if ๐‘– =
๐‘—, ๐‘Ž๐‘–๐‘— = 0 if ๐‘– ≠ ๐‘—, is called the ๐‘› x ๐‘› identity matrix of order ๐’.
Example 6: Below are examples of identity matrices.
๏ƒฉ1 0 0 ๏ƒน
I 3 ๏€ฝ ๏ƒช๏ƒช0 1 0๏ƒบ๏ƒบ
๏ƒซ๏ƒช0 0 1 ๏ƒบ๏ƒป
๏ƒฉ1
๏ƒช0
I4 ๏€ฝ ๏ƒช
๏ƒช0
๏ƒช
๏ƒซ0
is a 3 x 3 identity matrix of order 3
0 0 0๏ƒน
1 0 0 ๏ƒบ๏ƒบ
is a 4 x 4 identity matrix of order 4
0 1 0๏ƒบ
๏ƒบ
0 0 1๏ƒป
Note that an identity matrix is a scalar matrix where all the elements on the main diagonal
are 1.
17
Remarks:
The identity matrix ๐ผ๐‘› functions for ๐‘› x ๐‘› matrices the way the number 1 functions
for real numbers. In other words, the identity matrix ๐ผ๐‘› is actually a multiplicative identity
for ๐‘› x ๐‘› matrices. That is
๐ด๐ผ๐‘› = ๐ผ๐‘› ๐ด = ๐ด
for every ๐‘› x ๐‘› matrix A.
๏ƒฉ4 0 1๏ƒน
Example 7: Let A ๏€ฝ ๏ƒช๏ƒช 5 ๏€ญ3 2 ๏ƒบ๏ƒบ . Then
๏ƒช๏ƒซ 2 ๏€ญ1 4 ๏ƒบ๏ƒป
๏ƒฉ 4 0 1 ๏ƒน ๏ƒฉ1 0 0 ๏ƒน ๏ƒฉ 4 0 1 ๏ƒน
AI3 ๏€ฝ ๏ƒช๏ƒช5 ๏€ญ3 2๏ƒบ๏ƒบ ๏ƒช๏ƒช0 1 0๏ƒบ๏ƒบ ๏€ฝ ๏ƒช๏ƒช5 ๏€ญ3 2 ๏ƒบ๏ƒบ ๏€ฝ A
๏ƒช๏ƒซ 2 ๏€ญ1 4๏ƒบ๏ƒป ๏ƒช๏ƒซ0 0 1 ๏ƒบ๏ƒป ๏ƒช๏ƒซ 2 ๏€ญ1 4๏ƒบ๏ƒป
and
๏ƒฉ1 0 0 ๏ƒน ๏ƒฉ 4 0 1 ๏ƒน ๏ƒฉ 4 0 1 ๏ƒน
I 3 A ๏€ฝ ๏ƒช๏ƒช0 1 0๏ƒบ๏ƒบ ๏ƒช๏ƒช5 ๏€ญ3 2๏ƒบ๏ƒบ ๏€ฝ ๏ƒช๏ƒช5 ๏€ญ3 2๏ƒบ๏ƒบ ๏€ฝ A
๏ƒช๏ƒซ0 0 1 ๏ƒบ๏ƒป ๏ƒช๏ƒซ 2 ๏€ญ1 4๏ƒบ๏ƒป ๏ƒช๏ƒซ 2 ๏€ญ1 4๏ƒบ๏ƒป
Hence, AI3 = I3A = A
Suppose that A is a square matrix. If p is a positive integer, then we define
1. Ap = A.A…A
p factors
If A is n x n, we also define
2. A0 = In
3. ApAq = Ap+q
4. (Ap)q = Apq
NOTE:
1. The rule (AB)P = APBP holds only if AB = BA.
2. The rule “if AB = 0 then either A = 0 or B = 0” is not true for matrices.
18
๏ƒฉ2
Example 8: Let A ๏€ฝ ๏ƒช
๏ƒซ ๏€ญ8
Then,
๏ƒฉ2
AB ๏€ฝ ๏ƒช
๏ƒซ ๏€ญ8
๏€ญ3๏ƒน
๏ƒฉ3 6๏ƒน
and B ๏€ฝ ๏ƒช
๏ƒบ
๏ƒบ
12 ๏ƒป
๏ƒซ 2 4๏ƒป
๏€ญ3๏ƒน ๏ƒฉ 3 6 ๏ƒน ๏ƒฉ0 0๏ƒน
๏€ฝ
12 ๏ƒบ๏ƒป ๏ƒช๏ƒซ 2 4๏ƒบ๏ƒป ๏ƒช๏ƒซ0 0๏ƒบ๏ƒป
ACTIVITY
1
1. Let ๐ด = [
3
2 0
3 5 1
2 4
3 5
], ๐ต = [3 4 ], ๐ถ = [−2 7 2 ], ๐ท = [
],
−1 5
2 −4
1 −5
6 4 −3
2 4 5
−5 4
๐ธ = [0 1 4], and ๐น = [
].
2 3
3 −2 1
If possible, compute
a. ( BT + A )C
b. AB
c. 2D – 3F
d) ( C + E )T
e) AB + DF
f) ( 3C – 2E )TB
2. Let ๐ด and ๐ต be skew-symmetric matrices, show ๐ด๐ต is symmetric if and only if
๐ด๐ต = ๐ต๐ด.
3. Let ๐ด be an ๐‘› x ๐‘› matrix. Show that
a. ๐ด + ๐ด๐‘‡ is symmetric.
b. ๐ด − ๐ด๐‘‡ is skew-symmetric.
c. ๐ด๐ด๐‘‡ and ๐ด๐‘‡ ๐ด are symmetric.
๏ƒฉ 4 2๏ƒน
4. Let A ๏€ฝ ๏ƒช
๏ƒบ . Find
๏ƒซ1 3 ๏ƒป
a) A2 + 3A
b) 2A3 + 3A2 + 4A + 5I2
19
MODULE 2
LINEAR EQUATIONS AND MATRICES
Introduction
In this lesson we will discuss the Gauss-Jordan reduction method and the Gaussian
Elimination method and their application to the solution of linear systems, inverse of a
matrix and the practical method for finding the inverse of a matrix, determinants and its
properties, cofactor expansion and Cramer’s rule.
Objectives
After going through the lessons in this chapter, you are expected to be able to do the
following:
1. Explain solution of system of linear equations by substitution and by elimination through
simultaneous equations.
2. Transform a given matrix into a row-echelon and reduced row-echelon form.
3. Solve systems of linear equations using the Gauss-Jordan reduction method and the
Gaussian Elimination method.
4. Define homogenous systems.
5. Differentiate between singular and non-singular matrices.
6. Enumerate the properties of the Inverse.
7. Find the inverse of a given matrix.
8. Solve systems of linear equations using the inverse of a matrix.
9. Define permutation.
10. Evaluate the determinant of a matrix using permutation.
11. Reduce the problem of evaluating determinants by co-factor expansion.
12. Find the inverse of a matrix using determinants and co-factor expansion.
13. Solve systems of linear equations by using the Cramer’s Rule.
2.1 Solutions of Systems of Linear Equations
In this section we introduce a systematic technique of eliminating unknowns in
solving systems of linear equations. To illustrate this technique, let us consider the following
system:
Example 1:
3๐‘ฅ1 + 2๐‘ฅ2 − ๐‘ฅ3 = −2
2๐‘ฅ1 − 3๐‘ฅ2 + 2๐‘ฅ3 = 14
๐‘ฅ1 + 2๐‘ฅ2 + 3๐‘ฅ3 = 6
(1)
20
Interchanging the first and third equations gives us
๐‘ฅ1 + 2๐‘ฅ2 + 3๐‘ฅ3 = 6
2๐‘ฅ1 − 3๐‘ฅ2 + 2๐‘ฅ3 = 14
3๐‘ฅ1 + 2๐‘ฅ2 − ๐‘ฅ3 = −2
(2)
If we multiply the first equation in (2) by – 2 and add the result to the second equation we
get
−2๐‘ฅ1 − 4๐‘ฅ2 − 6๐‘ฅ3 = −12
2๐‘ฅ1 − 3๐‘ฅ2 + 2๐‘ฅ3 = 14
−7๐‘ฅ2 − 4๐‘ฅ3 = 2
The new equation obtained may replace either the first or second equation in system (2),
(the two equations used to obtain it). Next let us multiply the first equation in (2) by – 3
and add the result to the third equation. This gives us
−3๐‘ฅ1 − 6๐‘ฅ2 − 9๐‘ฅ3 = −18
3๐‘ฅ1 + 2๐‘ฅ2 − ๐‘ฅ3 = −2
or
−5๐‘ฅ2 − 10๐‘ฅ3 = −20
๐‘ฅ2 + 2๐‘ฅ3 = 4
(dividing both sides of the
equation by –5 )
The new equation obtained may replace either the first or the third equation in system (2).
Thus if we replace the second equation in (2) by −7๐‘ฅ2 − 4๐‘ฅ3 = 2 and the third equation
by ๐‘ฅ2 + 2๐‘ฅ3 = 4, we obtain the new system:
๐‘ฅ1 + 2๐‘ฅ2 + 3๐‘ฅ3 = 6
−7๐‘ฅ2 − 4๐‘ฅ3 = 2
๐‘ฅ2 + 2๐‘ฅ3 = 4
(3)
Note that the variable ๐‘ฅ1 has been eliminated from the second and third equations
of system (3). This new system is equivalent to the original system (1).
Next we multiply the third equation of (3) by 7 and add the result to the second
equation. This gives us
7 ๐‘ฅ2 + 14๐‘ฅ3 = 28
−7๐‘ฅ2 − 4๐‘ฅ3 = 2
or
10๐‘ฅ3 = 30
๐‘ฅ3 = 3
(dividing both sides of the
equation by 10)
21
The new equation obtained may replace either the second or third equation in (3).
Replacing the third equation we get an equivalent system
๐‘ฅ1 + 2๐‘ฅ2 + 3๐‘ฅ3 = 6
−7๐‘ฅ2 − 4๐‘ฅ3 = 2
๐‘ฅ3 = 3
(4)
To solve for the variable ๐‘ฅ2 we substitute ๐‘ฅ3 = 3 to the second equation in (4):
−7๐‘ฅ2 – 4(3) = 2
๐‘ฅ2 = −2
To solve for the variable ๐‘ฅ1 , we substitute ๐‘ฅ2 = −2 and ๐‘ฅ3 = 3 to the first equation in
(4):
๐‘ฅ1 + 2(−2) + 3(3) = 6
๐‘ฅ1 = 1
Thus the solution to the system (1) is the ordered triple (1, 2, 3).
The method used here is called the Gaussian elimination. The objective of Gaussian
elimination is to reduce a given system to triangular or echelon (staircase pattern) form
and then use back substitution to find the solution of the system. The Gaussian elimination
was modified by Carl Friedrich Gauss (1777 – 1855) and Wilhelm Jordan (1842 – 1899). They
called it the Gauss-Jordan elimination. To illustrate it, let us start from system (3) of
Example 1.
Example 2:
๐‘ฅ1 + 2๐‘ฅ2 + 3๐‘ฅ3 = 6
−7๐‘ฅ2 − 4๐‘ฅ3 = 2
๐‘ฅ2 + 2๐‘ฅ3 = 4
(3)
Interchange the second and third equations in (3):
๐‘ฅ1 + 2๐‘ฅ2 + 3๐‘ฅ3 = 6
๐‘ฅ2 + 2๐‘ฅ3 = 4
−7๐‘ฅ2 − 4๐‘ฅ3 = 2
(4)
Now let us multiply the second equation in (4) by – 2 and add the result to the first
equation. This gives us
๐‘ฅ1 + 2๐‘ฅ2 + 3๐‘ฅ3 = 6
−2 ๐‘ฅ2 − 4๐‘ฅ3 = −8
๐‘ฅ1
− ๐‘ฅ3 = −2
22
Next we multiply the second equation in (4) by 7 and add the result to the third equation.
This gives us
7 ๐‘ฅ2 + 14๐‘ฅ3
−7๐‘ฅ2 − 4๐‘ฅ3
10๐‘ฅ3
or
๐‘ฅ3
=
=
=
=
28
2
30
3
Replacing the first equation in (4) by ๐‘ฅ1 − ๐‘ฅ3 = −2 and the third equation by ๐‘ฅ3 = 3, we
obtain the new system
๐‘ฅ1
− ๐‘ฅ3 = −2
๐‘ฅ2 + 2๐‘ฅ3 = 4
๐‘ฅ3 = 3
(5)
Note that the variable ๐‘ฅ2 has been eliminated from the first and third equations of (5). This
time let us add the first equation in (5) to the third equation. This gives us
๐‘ฅ1
๐‘ฅ1
− ๐‘ฅ3 = −2
๐‘ฅ3 = 3
= 1
Finally we multiply the third equation in (5) by – 2 and add the result to the second
equation.
๐‘ฅ2 + 2๐‘ฅ3 = 4
−2๐‘ฅ3 = 3
๐‘ฅ2
= −2
System (5) now becomes
๐‘ฅ1
๐‘ฅ2
= 1
= −2
๐‘ฅ3 = 3
Thus the solution to the system is the ordered triple (1, -2, 3). The Gauss-Jordan elimination
is an algorithm that reduces a given system to reduced row echelon form or row canonical
form. This method is less efficient than the Gaussian elimination in solving systems of
equations however, it is well suited for calculating the inverse of a matrix which we will
discuss in a later section.
23
2.2 Elementary Row Operations
A system of linear equations can be placed in matrix form. This notation is very
efficient especially when we are dealing with large systems. To begin, let us consider an ๐‘š x
๐‘› system of linear equations
๐‘Ž11 ๐‘ฅ1 + ๐‘Ž12 ๐‘ฅ2 + … + ๐‘Ž1๐‘› ๐‘ฅ๐‘› = ๐‘1
๐‘Ž21 ๐‘ฅ1 + ๐‘Ž22 ๐‘ฅ2 + … + ๐‘Ž2๐‘› ๐‘ฅ๐‘› = ๐‘2
โ‹ฎ
๐‘Ž๐‘š1 ๐‘ฅ1 + ๐‘Ž๐‘š2 ๐‘ฅ2 + … + ๐‘Ž๐‘š๐‘› ๐‘ฅ๐‘› = ๐‘๐‘š
(1)
The coefficients of this linear system can be written in a rectangular array having ๐‘š rows
and ๐‘› columns, and we designate this array as ๐ด:
๐‘Ž11 ๐‘Ž12 … ๐‘Ž1๐‘›
๐‘Ž21 ๐‘Ž22 … ๐‘Ž2๐‘›
โ‹ฎ
โ‹ฎ
โ‹ฎ
๐ด=
๐‘Ž๐‘š1 ๐‘Ž๐‘š2 … ๐‘Ž๐‘š๐‘›
[
]
This ๐‘š x ๐‘› matrix ๐ด is called the coefficient matrix for the given system (1) above. If we add
another column in ๐ด to include the constants b1, b2, …, bm , we will have a matrix that
expresses compactly all the relevant information contained in (1). Such a matrix is called the
augmented matrix for (1), and is usually denoted as [Aโ”‚B]. If we let C = [Aโ”‚B] be the
augmented matrix for the system (1), then C is the ๐‘š x ( ๐‘›+1 ) matrix given by
๐‘Ž11 ๐‘Ž12 … ๐‘Ž1๐‘› ๐‘1
๐‘Ž21 ๐‘Ž22 … ๐‘Ž2๐‘› ๐‘2
C=
โ‹ฎ
โ‹ฎ
โ‹ฎ
โ‹ฎ
๐‘Ž๐‘š1 ๐‘Ž๐‘š2 … ๐‘Ž๐‘š๐‘› ๐‘๐‘š
[
]
We also write ๐ด๐‘‹ = ๐ต, where
๐‘Ž11 ๐‘Ž12 … ๐‘Ž1๐‘›
๐‘Ž21 ๐‘Ž22 … ๐‘Ž2๐‘›
โ‹ฎ
โ‹ฎ
โ‹ฎ
๐ด=
and
๐‘Ž๐‘š1 ๐‘Ž๐‘š2 … ๐‘Ž๐‘š๐‘›
[
]
represent the coefficients and
๐‘ฅ1
๐‘ฅ2
๐‘‹=[ โ‹ฎ ]
๐‘ฅ๐‘›
๐‘1
๐‘2
๐ต=[ ]
โ‹ฎ
๐‘๐‘š
24
represents the variables. For example, the array
1 −3 0 2
[0
4 5 −1
2 −1 3 0
2
−3]
4
represents the system of 3 linear equations
๐‘ฅ1 − 3๐‘ฅ2
+ 2๐‘ฅ4 = 2
4๐‘ฅ2 + 5๐‘ฅ3 − ๐‘ฅ4 = −3
2๐‘ฅ1 − ๐‘ฅ2 + 3๐‘ฅ3
= 4
The task of this section is to manipulate the augmented matrix representing a given
linear system into reduced row echelon form. But before we continue, let us introduce
some terminology. We have seen from Examples 1 and 2 that three important operations
were applied to solve the given system. These are: multiplying or dividing both sides of an
equation by a nonzero number, adding a multiple of one equation to another equation, and
interchanging two equations. These three operations when applied to the rows of an
augmented matrix are called elementary row operations and will result in the matrix of an
equivalent system.
Definition 2.2.1: An elementary row operation on an ๐‘š x ๐‘› matrix ๐ด = [๐‘Ž๐‘–๐‘— ] is any one of
the following operations:
1. Interchange rows ๐‘– and ๐‘— of A.
2. Multiply row ๐‘– of ๐ด by any nonzero real number.
3. Add ๐‘ times row ๐‘– of ๐ด to row ๐‘— of ๐ด, ๐‘– ≠ ๐‘—.
The process of applying elementary row operations to simplify an augmented matrix
is called row reduction. The following notations will be used to denote the elementary row
operation used:
1. ๐‘…๐‘–
๐‘…๐‘— means interchanging the ๐‘–th row and the ๐‘—th row
2. ๐‘๐‘…๐‘–
๐‘…๐‘– means that the ๐‘–th row is replace with ๐‘ times the ๐‘–th row.
3. ๐‘๐‘…๐‘– + ๐‘…๐‘—
๐‘…๐‘— means that the ๐‘—th row is replaced with the sum of the ๐‘—th row
and the ๐‘–th row multiplied by ๐‘.
Before we illustrate the three elementary operations, let us define first a matrix in
reduced row echelon form.
25
Definition 2.2.2: An ๐‘š x ๐‘› matrix is said to be in reduced row echelon form when it satisfies
the following properties:
a) All rows consisting entirely of zeros, if any, are at the bottom of the matrix.
b) The first nonzero entry in each row that does not consist entirely of zeros is a 1,
called the leading entry(pivot) of its row.
c) If rows ๐‘– and ๐‘– + 1 are two successive rows that do not consist entirely of zeros, then
the leading entry of row ๐‘– + 1 is to the right of the leading entry of row ๐‘–.
d) If a column contains a leading entry of some row, then all other entries in that
column are zero.
Note that a matrix in reduced row echelon form might not have any rows that consist
entirely of zeros.
Example 1: Below are matrices in reduced row echelon form:
๏ƒฉ1 0 0 ๏€ญ4 ๏ƒน
A ๏€ฝ ๏ƒช๏ƒช0 1 0 3 ๏ƒบ๏ƒบ
๏ƒช๏ƒซ0 0 1 2 ๏ƒบ๏ƒป
๏‚ท The first nonzero entry in each row is 1 called the
pivot.
๏‚ท The leading entry (pivot) for row 2 is to the right of
the leading entry for row 1 and the leading entry for
row 3 is to the right of the leading entry for row 2.
๏‚ท All other entries for the columns containing a leading
entry are zero.
๏‚ท There are entries other than zero in column 4 since
there is no pivot in that column.
๏ƒฉ1 2 0 0๏ƒน
B ๏€ฝ ๏ƒช๏ƒช0 0 1 0๏ƒบ๏ƒบ
๏ƒช๏ƒซ0 0 0 1 ๏ƒบ๏ƒป
๏‚ท The leading entry for row 2 appears in column 3 and
the leading entry for row 3 appears in column 4.
๏‚ท Column 2 does not contain a pivot.
๏ƒฉ1
๏ƒช0
C๏€ฝ๏ƒช
๏ƒช0
๏ƒช
๏ƒซ0
๏‚ท All rows consisting of zeros are at the bottom of the
matrix.
๏‚ท Columns 1 and 3 contain a pivot hence all other
entries in those columns are zero.
0 0 1/ 3๏ƒน
0 1 0 ๏ƒบ๏ƒบ
0 0 0 ๏ƒบ
๏ƒบ
0 0 0 ๏ƒป
26
Example 2: Below are matrices not in reduced row echelon form:
๏ƒฉ1 2 0 4๏ƒน
D ๏€ฝ ๏ƒช๏ƒช0 0 0 0๏ƒบ๏ƒบ
๏ƒช๏ƒซ0 0 1 3 ๏ƒบ๏ƒป
๏‚ท Fails property (a), the row consisting entirely of zeros
must be at the bottom of the matrix.
๏‚ท Fails property (b), the first nonzero entry in row 2 is
not equal to 1.
๏‚ท Fails property (d), column 3 contains a leading entry
but the other entries in that column are not equal to
zero.
4๏ƒน
๏ƒฉ1 0 3
E ๏€ฝ ๏ƒช๏ƒช0 2 ๏€ญ 2 5 ๏ƒบ๏ƒบ
๏ƒช๏ƒซ0 0 1
2๏ƒบ๏ƒป
๏ƒฉ1
๏ƒช0
F๏€ฝ๏ƒช
๏ƒช0
๏ƒช
๏ƒซ0
0 3 4๏ƒน
1 ๏€ญ 2 5๏ƒบ๏ƒบ
1 2 2๏ƒบ
๏ƒบ
0 0 0๏ƒป
๏‚ท Fails property (c), the leading entry of the third row
must be to the right of the leading entry of row 2.
Example 3: The three elementary operations are illustrated below:
๏ƒฉ0 1 0 2 ๏ƒน
Let A ๏€ฝ ๏ƒช๏ƒช2 3 0 ๏€ญ 4๏ƒบ๏ƒบ
๏ƒช๏ƒซ2 1 6 9 ๏ƒบ๏ƒป
(a) Interchanging rows 1 and 3 of A, (๐‘…1
๐‘…3 ) , we obtain:
๏ƒฉ2 1 6 9 ๏ƒน
B ๏€ฝ ๏ƒช๏ƒช2 3 0 ๏€ญ 4๏ƒบ๏ƒบ
๏ƒช๏ƒซ0 1 0 2 ๏ƒบ๏ƒป
(b) Replacing row 1 of B by ½ times row 1 of B, ( 12๐‘…1
1
9
1 2 3 2
๐ถ = [ 2 3 0 − 4]
0 1 0 2
๐‘…1 ) , we obtain:
27
(c) Replacing row 2 by the sum of -2 times row 1 of ๐ถ plus row 2 of ๐ถ,
(−2๐‘…1 + ๐‘…2
๐‘…2 ) we obtain:
1
9
1 2
3
2
๐ท = [ 0 2 − 6 − 13]
0 1
0 2
Definition 2.2.3: An ๐‘š x ๐‘› matrix ๐ด is said to be row equivalent to an ๐‘š x ๐‘› matrix ๐ต if ๐ต
can be obtained by applying a finite sequence of elementary row operations to ๐ด.
Hence in Example 3, ๐ด is row equivalent to ๐ท because ๐ท is obtained from ๐ด after
applying a series of elementary row operations on A.
NOTE:
1. Every matrix is row equivalent to itself.
2. If A is row equivalent to B, then B is row equivalent to A.
3. If A is row equivalent to B and B is row equivalent to C, then A is row equivalent to C.
Theorem 2.2.1: Every nonzero ๐‘š x ๐‘› matrix is row equivalent to a unique matrix in reduced
row echelon form.
2.3 The Gauss-Jordan Method
The Gauss-Jordan method is a systematic technique for applying matrix row
transformations in an attempt to reduce a matrix to diagonal form also called reduced row
echelon form.
The following are steps in using the Gauss-Jordan method to put a matrix into reduced row
echelon form:
STEP 1: Form the augmented matrix [ Aโ”‚B ].
STEP 2: Obtain 1 as the first element of the first column.
STEP 3: Use the first row to transform the remaining entries in the first column to 0.
STEP 4: Obtain 1 as the second entry in the second column.
STEP 5: Use the second row to transform the remaining entries in the second column to 0.
STEP 6: Continue in this manner as far as possible.
28
๏ƒฉ0 0 1 2 ๏ƒน
Example 1: Transform A ๏€ฝ ๏ƒช๏ƒช2 3 0 ๏€ญ 2๏ƒบ๏ƒบ to reduced row echelon form.
๏ƒช๏ƒซ3 3 6 ๏€ญ 9 ๏ƒบ๏ƒป
To transform A to reduced row echelon form, we perform Gauss-Jordan elimination
method. The sequence of elementary row operations, beginning with A, follows.
Interchange rows 1 and 3 of ๐ด to obtain ๐ด1
0 0 1 2
๐ด = [ 2 3 0 −2 ]
3 3 6 −9
๐‘…1
๐‘…3
3 3 6 −9
[2 3 0 −2] = ๐ด1
0 0 1 2
(Multiply row 1 of ๐ด1 by 1/3 to obtain ๐ด2 )
1
3
๐‘…1
๐‘…1
1
[2
0
1 2 −3
3 0 −2] = ๐ด2
0 1 2
(Multiply row 1 of ๐ด2 by -2 and add to row 2 to obtain ๐ด3 )
−2๐‘…1 + ๐‘…2
๐‘…2
1
[0
0
1 2 −3
1 −4 4 ] = ๐ด3
0 1
2
(Multiply row 2 of ๐ด3 by -1 and add to row 1 to obtain ๐ด4 )
1๐‘…2 + ๐‘…1
๐‘…1
1 0
[0 1
0 0
6 −7
−4 4 ] = ๐ด4
1
2
Multiply row 3 of ๐ด4 by −6 and add to row 1; multiply row 3 of ๐ด4 by 4 and add to row 2 to
obtain ๐ด5 )
−6๐‘…3 + ๐‘…1
4๐‘…3 + ๐‘…2
๐‘…1
๐‘…3
1 0 0 −19
[0 1 0 12 ] = ๐ด5
0 0 1
2
๐ด5 is already in reduced row echelon form.
We now apply these results to the solution of linear system.
29
Theorem 2.3.1. Let AX = B and CX = D be two linear systems each of ๐‘š equations in ๐‘›
unknowns. If the augmented matrices [ Aโ”‚B ] and [ Cโ”‚D ] of these systems are row
equivalent, then both linear systems have exactly the same solutions.
Example 2. Solve the linear system
x + y + 2z = -1
x – 2y + z = -5
3x + y + z = 3
by Gauss-Jordan reduction.
Solution: Let the augmented matrix be
1
๐ด = [1
3
1 2
−2 1
1 1
−1
−5]
3
Applying the Gauss-Jordan elimination we have
1
1 2
[1 −2 1
3
1 1
−1 ๐‘… + ๐‘…
1
2
−5] 3๐‘… + ๐‘…
1
3
3
1
− 3 ๐‘…2
−๐‘…2 + ๐‘…1
2๐‘…2 + ๐‘…3
3
− 13 ๐‘…3
5
− 3 ๐‘…3 + ๐‘…1
1
− 3 ๐‘…3 + ๐‘…2
๐‘…2
๐‘…3
1
[0
0
1
2
−3 −1
−2 −5
๐‘…2
1
[0
0
1
2
1 1/3
−2 −5
1
[0
0
0
5/3
−7/3
1
1/3
4/3 ]
0 −13/3 26/3
๐‘…1
๐‘…3
๐‘…3
๐‘…1
๐‘…2
1
[0
0
1
[0
0
0
1
0
0
1
0
−1
−4]
6
5/3
1/3
1
0
0
1
−1
4/3]
6
−7/3
4/3 ]
−2
1
2]
−2
The last matrix is in reduced row echelon form and is row equivalent to the augmented
matrix representing the given system. As an augmented matrix, it represents the system
x
y
z
= 1
= 2
= -2
30
Thus the solution set is {(1, 2, - 2)}. The given system is consistent.
Example 3: Solve the linear system
๐‘ฅ1 + ๐‘ฅ2 − ๐‘ฅ3 = 7
4๐‘ฅ1 − ๐‘ฅ2 + 5๐‘ฅ3 = 4
6๐‘ฅ1 + ๐‘ฅ2 + 3๐‘ฅ3 = 0
Solution: Let the augmented matrix be
1
[4
6
1 −1
−1
5
1
3
7
4]
0
Applying the Gauss-Jordan elimination we have
1
1 −1
[4 −1
5
6
1
3
7 −4๐‘… + ๐‘…
4] −6๐‘…1 + ๐‘…2
1
3
0
1
๐‘…2 + ๐‘…1
−๐‘…2 + ๐‘…3
5
1
− 5 ๐‘…2
๐‘…2 1
[
๐‘…3 0
0
1 −1
−5
9
−5
9
๐‘…1
๐‘…3
1
0 4/5
[0 −5
9
0
0
0
๐‘…2
1
[0
0
0
1
0
7
−24]
−42
11/5
−24 ]
−18
4/5
−9/5
0
11/5
24/5 ]
−18
The last equation reads 0๐‘ฅ1 + 0๐‘ฅ2 + 0๐‘ฅ3 = −18, which is impossible since 0 ≠ −18.
Hence the given system has no solution. In this case the system is said to be inconsistent.
Example 4: Solve the system
2๐‘ฅ1 + 6๐‘ฅ2 − 4๐‘ฅ3 + 2๐‘ฅ4 = 4
๐‘ฅ1
− ๐‘ฅ3 + ๐‘ฅ4 = 5
−3๐‘ฅ1 + 2๐‘ฅ2 − 2๐‘ฅ3
= −2
Solution: Applying the Gauss-Jordan elimination to the augmented matrix, we have
2 6 −4
[ 1 0 −1
−3 2 −2
2
1
0
4
5]
−2
๐‘…1
๐‘…2
1 0 −1
[ 2 6 −4
−3 2 −2
1
2
0
5
4]
−2
31
๐‘…2
๐‘…3
1 0
[ 0 6
0 2
๐‘…2
๐‘…2
1
[ 0
0
−2๐‘…2 + ๐‘…3
๐‘…3
1 0
[ 0 1
0 0
๐‘…3
1
1 0
−1
0
[ 0 1 −1/3
0 0
1 −9/13
−2๐‘…1 + ๐‘…2
3๐‘…1 + ๐‘…3
1
6
3
− 13 ๐‘…3
๐‘…3 + ๐‘…1
1
3
๐‘…3 + ๐‘…2
๐‘…1
๐‘…2
−1
−2
−5
1
0
3
0
−1
1 −1/3
2
−5
5
−6]
13
1
0
3
5
−1]
13
−1
1
−1/3 0
−13/3 3
5
−1]
15
1 0 0 4/13
[ 0 1 0 3/13
0 0 1 −9/13
5
−1 ]
−45/13
20/13
−28/13]
−45/13
Column 4 has no pivot hence ๐‘ฅ4 is a free variable. If we let ๐‘ฅ4 = ๐‘Ÿ, ๐‘Ÿ ∈ ๐‘…, then we obtain:
๐‘ฅ1 =
20
13
−
4
13
๐‘Ÿ , ๐‘ฅ2 = −
28
13
20
Thus the solution set is in the form of (
13
−
−
3
13
4
13
๐‘Ÿ , and ๐‘ฅ3 = −
๐‘Ÿ, −
28
13
−
3
13
45
13
๐‘Ÿ, −
+
45
13
9
13
+
๐‘Ÿ.
9
13
๐‘Ÿ, ๐‘Ÿ), ๐‘Ÿ ∈
๐‘…. The given system has infinitely many solution.
Corollary 2.3.1. If A and C are row equivalent ๐‘š x ๐‘› matrices, then the linear systems
AX = 0 and CX = 0 have exactly the same solutions.
SAQ 2-1
Consider the system
2๐‘ฅ1 − ๐‘ฅ2 + 3๐‘ฅ3 = ๐‘Ž
3๐‘ฅ1 + ๐‘ฅ2 − 5๐‘ฅ3 = ๐‘
−5๐‘ฅ1 − 5๐‘ฅ2 + 21๐‘ฅ3 = ๐‘
Show that the system is inconsistent if ๐‘ ≠ 2๐‘Ž − 3๐‘.
32
ASAQ 2-1
Reducing the augmented matrix of the given system to row echelon form, we have
2 −1 3
[3
1 −5
−5 −5 21
๐‘Ž
๐‘]
๐‘
1 −1/2
๐‘…1 [ 3
1
−5
−5
1
๐‘…
2 1
−3๐‘…1 + ๐‘…2
5๐‘…1 + ๐‘…3
๐‘…2
๐‘…3
1
[0
0
3/2
−5
21
−1/2
3/2
5/2
−19/2
−15/2 57/2
๐‘Ž/2
๐‘ ]
๐‘
๐‘Ž/2
−3๐‘Ž+2๐‘
2
5๐‘Ž+2๐‘
]
2
3๐‘…2 + ๐‘…3
๐‘…3
1
[0
0
−1/2
3/2
5/2 −19/2
0
0
๐‘Ž/2
−3๐‘Ž+2๐‘
2
]
−2๐‘Ž + 3๐‘ + ๐‘
The last equation reads 0๐‘ฅ1 + 0๐‘ฅ2 + 0๐‘ฅ3 = −2๐‘Ž + 3๐‘ + ๐‘. The given system has no
solution if −2๐‘Ž + 3๐‘ + ๐‘ ≠ 0. Thus the system is inconsistent if ๐‘ ≠ 2๐‘Ž − 3๐‘.
2.4 Homogeneous Systems
A linear system of the form
a11 x1 ๏€ซ a12 x2 ๏€ซ ... ๏€ซ a1n xn ๏€ฝ 0
a21 x1 ๏€ซ a22 x2 ๏€ซ ... ๏€ซ a2 n xn ๏€ฝ 0
(2)
am1 x1 ๏€ซ am 2 x2 ๏€ซ ... ๏€ซ amn xn ๏€ฝ 0
is called a homogeneous system. We can write it in matrix form as
๐ด๐‘‹ = 0
The solution to the homogeneous system (2) is called the trivial solution if
x1 ๏€ฝ x2 ๏€ฝ ... ๏€ฝ xn ๏€ฝ 0 and a solution x1 , x2 ,..., x n in which not all the xi are zero is called a
nontrivial solution.
33
Example 1. Consider the homogeneous system
๐‘ฅ + 2๐‘ฆ + ๐‘ง = 0
2๐‘ฅ + 5๐‘ฆ + 3๐‘ง = 0
3๐‘ฅ + 7๐‘ฆ + 5๐‘ง = 0
The augmented matrix of this system is
1 2
[2 5
3 7
1
3
5
0
0]
0
which is row equivalent to (you must do the calculation here)
1 0
[0 1
0 0
0
0
1
0
0]
0
Hence the solution to the given homogeneous system is ๐‘ฅ = ๐‘ฆ = ๐‘ง = 0 (a trivial solution).
SAQ 2-2
Find the solution of the homogeneous system
๐‘ฅ + ๐‘ฆ + ๐‘ง + ๐‘ค = 0
2๐‘ฅ + ๐‘ฆ
๐‘ง
= 0
3๐‘ฅ + ๐‘ฆ + ๐‘ง + 2๐‘ค = 0
ASAQ 2-2
The augmented matrix of this system is
1 1 1 1
[2 1 −1 0
3 1 1 2
0
0]
0
34
Transforming it to reduced row echelon form we have
1 1 1 1
[2 1 −1 0
3 1 1 2
0 −2๐‘… + ๐‘…
0] −3๐‘…1 + ๐‘…2
1
3
0
−๐‘…2
1
๐‘…2 1 1
[
๐‘…3 0 −1 −3
0 −2 −2
๐‘…2
1
[0
0
1
1
1
3
−2 −2
1 0
−2 0]
−1 0
1 0
2 0]
−1 0
๐‘…1 1
[
๐‘…3 0
0
0 −2
1 3
0 4
−1 0
2 0]
3 0
๐‘…
4 3
1
๐‘…3 [0
0
0 −2
1 3
0 1
−1 0
2
0]
3/4 0
2๐‘…3 + ๐‘…1
−3๐‘…3 + ๐‘…2
1
๐‘…1
[
๐‘…2 0
0
−๐‘…2 + ๐‘…1
2๐‘…2 + ๐‘…3
1
0 0
1 0
0 1
1/2 0
−1/4 0]
3/4 0
The fourth column has no pivot hence ๐‘ค is a free variable. If we let ๐‘ค = ๐‘ , ๐‘  ∈ ๐‘…, then the
solution of the given homogeneous system is
1
๐‘ฅ = −2๐‘ 
๐‘ฆ =
1
4
๐‘ 
3
๐‘ง = −4๐‘ 
๐‘ค = ๐‘ 
1
1
3
where ๐‘  is any real number. Thus the solution set is in the form of {(− 2 ๐‘ , 4 ๐‘ , − 4 ๐‘ , ๐‘ )},
๐‘  ∈ ๐‘…. This shows that the system has a nontrivial solution. This result is generalized in the
next theorem.
Theorem 2.4.1: A homogeneous system of ๐‘š equations in ๐‘› unknowns always has a
nontrivial solution if ๐‘š < ๐‘›, that is , if the number of unknowns exceeds the number of
equations.
35
ACTIVITY
1. Find all solutions to the given linear systems.
a. ๐‘ฅ1 − ๐‘ฅ2 − ๐‘ฅ3 = 0
b. 2๐‘ฅ1 + 3๐‘ฅ2 − ๐‘ฅ3 = 0
2๐‘ฅ1 + ๐‘ฅ2 + 2๐‘ฅ3 = 0
−4๐‘ฅ1 + 2๐‘ฅ2 + ๐‘ฅ3 = 0
๐‘ฅ1 − 4๐‘ฅ2 − 5๐‘ฅ3 = 0
7๐‘ฅ1 + 3๐‘ฅ2 − 9๐‘ฅ3 = 0
c. ๐‘ฅ1 + ๐‘ฅ2 + ๐‘ฅ3 + ๐‘ฅ4 = 6
2๐‘ฅ1
− ๐‘ฅ3 − ๐‘ฅ4 = 4
3 ๐‘ฅ3 + 6๐‘ฅ4 = 3
๐‘ฅ1
− ๐‘ฅ4 = 5
2. Find all values of ๐‘Ž for which the resulting system has (a) no solution, (b) a unique
solution, and (c) infinitely many solutions.
๐‘ฅ1 + ๐‘ฅ2 − ๐‘ฅ3 = 3
๐‘ฅ1 − ๐‘ฅ2 + 3๐‘ฅ3 = 4
๐‘ฅ1 + ๐‘ฅ2 + (๐‘Ž2 − 10)๐‘ฅ3 = ๐‘Ž
2.5 The Inverse of a Matrix
When working with real numbers, we know that a number ๐‘Ž times its inverse is one
provided that ๐‘Ž ≠ 0. Thus the equation ๐‘Ž๐‘ฅ = ๐‘ could be solved for ๐‘ฅ by dividing both sides
(or multiplying both sides by 1/๐‘Ž) of the equation by ๐‘Ž to get ๐‘ฅ = ๐‘/๐‘Ž provided that ๐‘Ž ≠ 0.
However the matrix equation ๐ด๐‘‹ = ๐ต cannot be divided by the matrix ๐ด on both sides
because there is no matrix division.
This section is concerned on finding a matrix whose function is similar to the inverse
of a real number ๐‘Ž. We like to find a matrix inverse such that ๐ด times its inverse is equal to
“one”. The matrix equivalent of “one” is called the “identity matrix”.
Definition 2.5.1: An ๐‘› x ๐‘› matrix ๐ด is called nonsingular (or invertible) if there exists an ๐‘› x
๐‘› matrix ๐ต such that
๐ด๐ต = ๐ต๐ด = ๐ผ๐‘›
The matrix ๐ต is called an inverse of ๐ด. If there exists no such matrix ๐ต, then ๐ด is called
singular (or non-invertible).
36
−2
Example 1: Let ๐ด = [
3
−4/11
1
] and ๐ต = [
3/11
4
−2 1 −4/11
๐ด๐ต = [
][
3 4 3/11
−4/11
๐ต๐ด = [
3/11
1/11
] then,
2/11
1/11
1 0
] = [
] = ๐ผ2
2/11
0 1
1/11 −2
][
2/11
3
1
1
]= [
4
0
and
0
] = ๐ผ2
1
Since ๐ด๐ต = ๐ต๐ด = ๐ผ2 , we conclude that ๐ต is an inverse of ๐ด and that ๐ด is nonsingular.
4
Example 2: Let ๐ด = [
−3
Solution: Let ๐ด−1 = [
๐‘Ž
๐‘
2
], find ๐ด−1 .
1
๐‘
] then
๐‘‘
๐ด๐ด−1 = ๐ผ2
[
[
4 2 ๐‘Ž
][
−3 1 ๐‘
4๐‘Ž + 2๐‘
−3๐‘Ž + ๐‘
1 0
๐‘
]= [
]
0 1
๐‘‘
4๐‘ + 2๐‘‘
1
]= [
0
−3๐‘ + ๐‘‘
0
]
1
Two matrices are equal if and only if their corresponding parts are equal. Hence if we
equate the corresponding parts, we get
4๐‘Ž + 2๐‘ = 1
−3๐‘Ž + ๐‘ = 0
and
4๐‘ + 2๐‘‘ = 0
−3๐‘ + ๐‘‘ = 1
Solving the system we have a = 1/10, b = -1/5, c = 3/10, and d = 2/5. Thus
๐ด−1 = [
1/10
3/10
−1/5
]
2/5
You can always check your answer by taking the product ๐ด๐ด−1 and making sure that the
answer is the identity matrix ๐ผ2 .
There is a simple procedure for finding the inverse of a 2 x 2 matrix. It can be done
easily as follows:
Let ๐ด = [
๐‘Ž
๐‘
๐‘ค
๐‘
]. We are looking for a matrix [ ๐‘ฆ
๐‘‘
๐‘ฅ
๐‘ง ] such that
37
[
[
๐‘Ž
๐‘
๐‘Ž๐‘ค + ๐‘๐‘ฆ
๐‘๐‘ค + ๐‘‘๐‘ฆ
๐‘ ๐‘ค
][
๐‘‘ ๐‘ฆ
๐‘ฅ
1 0
๐‘ง ] = [0 1]
๐‘Ž๐‘ฅ + ๐‘๐‘ง
1 0
] = [
]
๐‘๐‘ฅ + ๐‘‘๐‘ง
0 1
Equating the corresponding parts we have
๐‘Ž๐‘ค + ๐‘๐‘ฆ = 1
(1)
๐‘๐‘ค + ๐‘‘๐‘ฆ = 0
(2)
๐‘Ž๐‘ฅ + ๐‘๐‘ง = 0
(3)
๐‘๐‘ฅ + ๐‘‘๐‘ง = 1
(4)
Multiplying equation (1) by ๐‘ and equation (2) by ๐‘Ž we get:
๐‘Ž๐‘๐‘ค + ๐‘๐‘๐‘ฆ = ๐‘
๐‘Ž๐‘๐‘ค + ๐‘Ž๐‘‘๐‘ฆ = 0
(1a)
(2a)
Subtracting equation (1a) from equation (2a) and solving for ๐‘ฆ gives us:
๐‘ฆ=
−๐‘
๐‘Ž๐‘‘−๐‘๐‘
Multiplying equation (3) by ๐‘ and equation (4) by ๐‘Ž we obtain:
๐‘Ž๐‘๐‘ฅ + ๐‘๐‘๐‘ง = 0
๐‘Ž๐‘๐‘ฅ + ๐‘Ž๐‘‘๐‘ง = ๐‘Ž
(3a)
(4a)
Subtracting equation (3a) from equation (4a) and solving for ๐‘ง gives us:
๐‘ง=
๐‘Ž
๐‘Ž๐‘‘−๐‘๐‘
In a similar manner we can solve for ๐‘ค by multiplying equation (1) by ๐‘‘ and equation (2) by
๐‘:
๐‘‘
๐‘ค=
๐‘Ž๐‘‘−๐‘๐‘
Finally we multiply equation (3) by ๐‘‘ and equation (4) by ๐‘ to solve for ๐‘ฅ:
๐‘ฅ=
−๐‘
๐‘Ž๐‘‘−๐‘๐‘
38
Thus, the inverse of matrix ๐ด = [
๐‘‘
−๐‘
[๐‘Ž๐‘‘−๐‘๐‘
−๐‘
๐‘Ž๐‘‘−๐‘๐‘
๐‘Ž ]
๐‘Ž๐‘‘−๐‘๐‘
๐‘Ž๐‘‘−๐‘๐‘
=
๐‘Ž
๐‘
๐‘
] is the matrix
๐‘‘
1
๐‘‘
[
๐‘Ž๐‘‘−๐‘๐‘ −๐‘
−๐‘
] provided that ๐‘Ž๐‘‘ − ๐‘๐‘ ≠ 0.
๐‘Ž
4 2
Example 3: Solve the inverse of matrix ๐ด = [
] using the formula.
−3 1
Solution: Let ๐ด = [
4 2
๐‘Ž
]=[
−3 1
๐‘
๐ด−1 =
1
1
[
10 3
๐‘
] then ๐‘Ž๐‘‘ − ๐‘๐‘ = 4(1) − 2(−3) = 10. Thus
๐‘‘
1/10
−2
] = [
3/10
4
−1/5
]
2/5
Note that this procedure only works for 2 x 2 matrices. In general, the procedure in finding
the inverse of any ๐‘› x ๐‘› matrix is as follows:
1. Form the ๐‘› x 2๐‘› partitioned matrix [ ๐ดโ”‚๐ผ๐‘› ] obtained by adjoining the identity matrix
๐ผ๐‘› to the given matrix ๐ด.
2. Use elementary row operations to transform the matrix obtained in Step 1 to
reduced row echelon form. Remember that any elementary row operation we do to
a row of ๐ด we also do to the corresponding row of ๐ผ๐‘› .
3. The series of elementary row operations which reduces ๐ด to ๐ผ๐‘› will reduce ๐ผ๐‘› to
๐ด−1 . If ๐ด cannot be reduced to ๐ผ๐‘› then ๐ด is singular and ๐ด−1 does not exist.
1 2
Example 4. Find the inverse of the matrix ๐ด = [2 5
1 0
3
3].
8
Solution:
We form the 3 x 6 partitioned matrix [A โ”‚I3 ] by adjoining the 3 x 3 identity matrix to ๐ด and
transform it to reduced row echelon form by applying the Gauss-Jordan elimination.
๐ด
1
[2
1
2
5
0
๐ผ3
3
3
8
1
0
0
0
1
0
0 −2๐‘… + ๐‘…
0] −๐‘… 1+ ๐‘… 2
1
3
1
๐‘…2
๐‘…3
39
1
2
[0
1
0 −2
3
−3
5
9
−3
−1
1 0 0
−2 1 0]
−1 0 1
−2๐‘…2 + ๐‘…1
2๐‘…2 + ๐‘…3
1
[0
0
0
1
0
1
[0
0
0
9
1 −3
0
1
5 −2
0
−2
1
0]
5 −2 −1
1
[0
0
0
1
0
−40
13
5
0
0
1
5 −2
−2 1
−5 2
0
0] −๐‘…3
1
๐‘…1
๐‘…3
๐‘…3
−9๐‘…3 + ๐‘…1
3๐‘…3 + ๐‘…2
๐‘…1
๐‘…2
16
9
−5 −3].
−2 −1
−40
Since ๐ด has been transformed to ๐ผ3 then ๐ด−1 = [ 13
5
16
−5
−2
9
−3].
−1
You can check your answer by taking the product ๐ด๐ด−1 and making sure that the answer is
the identity matrix ๐ผ3 .
2
1
Example 5. Find the inverse of the matrix ๐ด = [ 1 −2
−3 −1
−1
−3].
2
Solution: We form the 3 x 6 matrix [A โ”‚I3 ] and transform it to reduced row echelon form
(you must do the calculation here):
2
1 −1 1
[ 1 −2 −3 0
−3 −1
2 0
0 0
1 0
]
is
row
equivalent
to
[
1 0
0 1
0 1
0 0
2/5
1/5
0
−1
1/5
−2/5
0 ]
1
0 −1/5 −1/35 −1/7
Since ๐ด is row equivalent to a matrix which has a row consisting of zeros, then ๐ด cannot be
reduced to ๐ผ3 . Thus ๐ด has no inverse and we say that ๐ด is a singular matrix.
Example 4 shows that an inverse exists if the partitioned matrix [๐ด ๐ผ๐‘› ] can be
reduced to [๐ผ๐‘› ๐ต] where ๐ต = ๐ด−1 . That is, A is row equivalent to ๐ผ๐‘› . This is stated in the
following theorem.
Theorem 2.5.1: An ๐‘› x ๐‘› matrix is nonsingular if and only if it is row equivalent to ๐ผ๐‘› .
Theorem 2.5.2: If a matrix has an inverse, then the inverse is unique.
40
Proof:
Suppose ๐ด has two inverses, say ๐ต and C. Then by Definition 2.5.1, ๐ด๐ต = ๐ต๐ด = ๐ผ๐‘› and
๐ด๐ถ = ๐ถ๐ด = ๐ผ๐‘› . We now have
๐ต(๐ด๐ถ) = (๐ต๐ด)๐ถ
Matrix multiplication is associative
Then
๐ต = ๐ต๐ผ๐‘› = ๐ต(๐ด๐ถ) = (๐ต๐ด)๐ถ = ๐ผ๐‘› ๐ถ = ๐ถ
By transitivity, we conclude that ๐ต = ๐ถ and the theorem is proved.
Theorem 2.5.3: ( Properties of the Inverse )
a) If ๐ด is a nonsingular matrix, then ๐ด−1 is nonsingular and
(๐ด−1 )−1 = ๐ด
b) If ๐ด and ๐ต are nonsingular matrices, then ๐ด๐ต is nonsingular and
(๐ด๐ต)−1 = ๐ต −1 ๐ด−1
c) If A is a nonsingular matrix, then
(๐ด๐‘‡ )−1 = (๐ด−1 )๐‘‡
We prove part (b) of the theorem and leave the proof of parts (a) and (c) as an exercise.
Proof of part (b): We have to show that
(๐ด๐ต)(๐ต−1 ๐ด−1 ) = (๐ต −1 ๐ด−1 )(๐ด๐ต) = ๐ผ๐‘› , that is the inverse of ๐ด๐ต = ๐ต −1 ๐ด−1 .
(๐ด๐ต)(๐ต−1 ๐ด−1 ) = ๐ด(๐ต๐ต −1 )๐ด−1
Matrix multiplication is associative
= ๐ด(๐ผ๐‘› )๐ด−1
Definition of an inverse of a matrix
= ๐ด๐ด−1
๐ผ๐‘› is a multiplicative identity for an ๐‘› x ๐‘› matrix
= ๐ผ๐‘›
Definition of an inverse of a matrix
Similarly
(๐ต −1 ๐ด−1 )(๐ด๐ต) = ๐ต −1 (๐ด−1 ๐ด)๐ต
Matrix multiplication is associative
= ๐ต −1 (๐ผ๐‘› )๐ต
Definition of an inverse of a matrix
= ๐ต −1 ๐ต
๐ผ๐‘› is a multiplicative identity for an ๐‘› x ๐‘› matrix
= ๐ผ๐‘›
Definition of an inverse of a matrix
41
Hence (๐ด๐ต)−1 = ๐ต −1 ๐ด−1.
Corollary 2.5.1: If ๐ด1 , ๐ด2 , โ‹ฏ , ๐ด๐‘Ÿ are ๐‘› x ๐‘› nonsingular matrices, then ๐ด1 ๐ด2 โ‹ฏ ๐ด๐‘Ÿ
is nonsingular and
(๐ด1 ๐ด2 โ‹ฏ ๐ด๐‘Ÿ )−1 = ๐ด๐‘Ÿ −1 ๐ด๐‘Ÿ−1 −1 โ‹ฏ ๐ด1 −1
Theorem 2.5.4: Suppose that ๐ด and ๐ต are ๐‘› x ๐‘› matrices
a) If ๐ด๐ต = ๐ผ๐‘› then ๐ต๐ด = ๐ผ๐‘› .
b) If ๐ต๐ด = ๐ผ๐‘› then ๐ด๐ต = ๐ผ๐‘› .
2.6 Linear Systems and Inverses
Suppose that matrix ๐ด is invertible. Consider the system ๐ด๐‘‹ = ๐ต where
๐‘ฅ1
๐‘1
๐‘ฅ2
๐‘
๐‘‹ = [ โ‹ฎ ] and ๐ต = [ 2 ] are ๐‘› x 1 matrices where ๐‘ฅ1 , ๐‘ฅ2 , โ‹ฏ , ๐‘ฅ๐‘› are variables and
โ‹ฎ
๐‘ฅ๐‘›
๐‘๐‘›
๐‘1 , ๐‘2 , โ‹ฏ , ๐‘๐‘› are real numbers. Since ๐ด is invertible then ๐ด−1 exists and
๐ด๐‘‹ = ๐ต
๐ด−1 ๐ด๐‘‹ = ๐ด−1 ๐ต
Multiplying both sides by ๐ด−1
๐ผ๐‘› ๐‘‹ = ๐ด−1 ๐ต
Definition of an inverse of a matrix
๐‘‹ = ๐ด−1 ๐ต
๐ผ๐‘› is the multiplicative identity for
an ๐‘› x ๐‘› matrix
Hence ๐‘‹ = ๐ด−1 ๐ต is the unique solution of the system.
Example 1. Let ๐ด = [
1 1
5
], ๐ต = [ ]. Solve ๐ด๐‘‹ = ๐ต by using the inverse of A.
1 2
7
Solution:
๐ด−1 = [
2 −1
] (verify)
−1 1
42
Hence
๐‘‹ = ๐ด−1 ๐ต
2 −1 5
3
๐‘‹=[
][ ]= [ ]
−1 1 7
2
Hence the unique solution is ๐‘‹ = (3, 2).
Theorem 2.6.1: If ๐ด is an ๐‘› x ๐‘› matrix, the homogeneous system ๐ด๐‘‹ = 0 has a nontrivial
solution if and only if A is singular.
Example 2: Consider the homogeneous system
2๐‘ฅ1 + ๐‘ฅ2 − ๐‘ฅ3 = 0
๐‘ฅ1 − 2๐‘ฅ2 − 3๐‘ฅ3 = 0
−3๐‘ฅ1 − ๐‘ฅ2 + 2๐‘ฅ3 = 0
2
1 −1
where the coefficient matrix ๐ด = [ 1 −2 −3] is the singular matrix of Example 5 of
−3 −1
2
the previous section. The augmented matrix
2
1 −1
[ 1 −2 −3
−3 −1
2
0
1 0
0] is row equivalent to [0 1
0
0 0
−1
1
0
0
0]
0
which implies that ๐‘ฅ1 = ๐‘Ÿ, ๐‘ฅ2 = −๐‘Ÿ, and ๐‘ฅ3 = ๐‘Ÿ, where ๐‘Ÿ ∈ ๐‘…. Thus the system has a
nontrivial solution.
Theorem 2.6.2: If ๐ด is an ๐‘› x ๐‘› matrix, then ๐ด is nonsingular if and only if the linear system
๐ด๐‘‹ = ๐ต has a unique solution for every ๐‘› x 1 matrix ๐ต.
43
SAQ 2-3
1
Let ๐ด = [1
2
๐ต
1 −1
−1 2 ]. Find the inverse of ๐ด then solve ๐ด๐‘‹ = ๐ต where
1 −1
1
is the 3 x 1 matrix (a) [−2],
3
4
(b) [−3], and
5
5
(c) [ 7 ]
−4
ASAQ 2-3
To solve for ๐ด−1 let us form the 3 x 6 matrix [๐ด ๐ผ3 ] and transform it to reduced row echelon form.
1 1 −1
[1 −1 2
2 1 −1
1 0 0 −๐‘… + ๐‘…
1
2
0 1 0] −2๐‘… + ๐‘…
1
3
0 0 1
๐‘…2
−๐‘…2
๐‘…2 1
[
๐‘…3 0
0
1 −1
−2 3
−1 1
1 0
−1 1
−2 0
0
0]
1
1 1 −1
๐‘…3 [0 −1 1
0 −2 3
1 0 0
−2 0 1]
−1 1 0
1 1 −1
๐‘…2 [0 1 −1
0 −2 3
1
2
−1
0 0
0 −1]
1 0
−1 0
2 0
3 1
1
−1]
−2
−๐‘…2 + ๐‘…1
2๐‘…2 + ๐‘…3
๐‘…1
๐‘…3
1 0
[0 1
0 0
0
−1
1
๐‘…3 + ๐‘…2
๐‘…2
1 0 0
[0 1 0
0 0 1
−1 0 1
5 1 −3]
3 1 −2
−1 0 1
Thus ๐ด−1 = [ 5 1 −3]. Using this to solve the linear system ๐ด๐‘‹ = ๐ต for ๐‘‹ where
3 1 −2
44
1
(a) ๐ต = [−2] we have
3
๐‘‹ = ๐ด−1 ๐ต
−1 0
= [5 1
3 1
1
1
−3] [−2]
−2 3
2
= [−6]
−5
4
(b) ๐ต = [−3] we have
5
๐‘‹ = ๐ด−1 ๐ต
4
−1 0 1
= [ 5 1 −3] [−3]
3 1 −2 5
1
= [2]
−1
5
(c) ๐ต = [ 7 ] we have
−4
๐‘‹ = ๐ด−1 ๐ต
−1 0 1
5
= [ 5 1 −3] [ 7 ]
3 1 −2 −4
−9
= [ 44 ]
30
This only shows that when a matrix is nonsingular, then the system ๐ด๐‘‹ = ๐ต has a unique
solution for every ๐‘› x 1 matrix ๐ต.
45
Based on the preceding theorems and examples, it can be noted that the following
statements are equivalent:
1. The matrix ๐ด is invertible (nonsingular).
2. The system ๐ด๐‘‹ = 0 has only the trivial solution.
3. The matrices ๐ด and ๐ผ๐‘› are row equivalent.
4. The system ๐ด๐‘‹ = ๐ต has a unique solution for every ๐‘› x 1 matrix ๐ต.
ACTIVITY
1. Find the inverse of the following matrices:
๏ƒฉ 1 2 ๏€ญ3 4 ๏ƒน
๏ƒฉ1 1 1๏ƒน
๏ƒช๏€ญ 4 2 1
3 ๏ƒบ๏ƒบ
a. ๏ƒช๏ƒช0 2 3๏ƒบ๏ƒบ
b. ๏ƒช
๏ƒช 3 0 0 ๏€ญ 3๏ƒบ
๏ƒช๏ƒซ5 5 1๏ƒบ๏ƒป
๏ƒช
๏ƒบ
๏ƒซ 2 0 ๏€ญ2 3 ๏ƒป
1 4
1 0 5
2. Let ๐ต = [2 1 4] and ๐ท = [0 2
3 5
3 4 7
2
1]. Calculate ๐ต๐ท and ๐ท−1 ๐ต −1 . Verify
3
that (๐ต๐ท)(๐ท−1 ๐ต−1 ) = ๐ผ3 .
3. Find all values of a
๏ƒฉ1
A ๏€ฝ ๏ƒช๏ƒช1
๏ƒช๏ƒซ1
-1
exists. What is A ?
for which the inverse of
1 0๏ƒน
0 0 ๏ƒบ๏ƒบ
2 a ๏ƒบ๏ƒป
4. Show that if ๐ด, ๐ต, and ๐ถ are invertible matrices, then ๐ด๐ต๐ถ is invertible and
(๐ด๐ต๐ถ)−1 = ๐ถ −1 ๐ต −1 ๐ด−1 .
5. Prove parts (a) and (c) of Theorem 2.5.4.
46
2.7 Determinants
Definition 2.7.1: Let S = {1, 2, … , n } be the set of integers from 1 to n, arranged in ascending
order. A rearrangement j1 j 2 ... j n of the elements of S is called a permutation of S.
The set of all permutations of S is denoted by ๐‘†๐‘› . For example, {1, 2} has two
permutations namely ๐‘†2 = { (12), (21) } and {1, 2, 3} has six permutations namely
๐‘†3 = { (123), (132), (213), (231), (312), (321) }.
The number of permutations of the set {1, 2, …, ๐‘›} can be determined without
writing a list as in the example above. Notice that there are ๐‘› possible positions to be filled.
There are ๐‘› choices for the first position, ๐‘› − 1 for the second, ๐‘› − 2 for the third, and only
one element for the ๐‘›th position. Thus the total number of permutations of ๐‘› elements is
๐‘›(๐‘› − 1)(๐‘› − 2) … 2.1 = ๐‘›!
A permutation j1 j 2 ... j n of S = { 1, 2, … , n } is said to have an inversion if a larger
integer jr precedes a smaller one j s . For example, consider the permutation (23541). The
following pairs of numbers form an inversion: 21, 31, 54, 51.
A permutation is called even if the total number of inversions in it is even and odd if
the total number of inversions in it is odd.
Example 1: The even permutations in S3 are:
123 ( There is no inversion )
231 ( There are two inversions; 21 and 31 )
312 ( There are two inversions; 31 and 32 )
The odd permutations in S3 are:
132 ( There is one inversion; 32 )
213 ( There is one inversion; 21 )
321 ( There are three inversions; 32, 31, and 21 )
Note that the number of odd permutations is equal to the number of even permutations.
Definition 2.7.2: Let A ๏€ฝ [ a ij ] be an ๐‘› x ๐‘› matrix. We define the determinant of ๐ด (written
โ”‚๐ดโ”‚) by
A ๏€ฝ ๏ƒฅ ๏‚ฑ a1 j1a2 j2 ...an jn ,
where the summation ranges over all permutations j1 j 2 ... j n of the set S = {1, 2, …, n}.
47
Note that each term ±๐‘Ž1๐‘—1 ๐‘Ž2๐‘—2 … ๐‘Ž๐‘›๐‘—๐‘› of |๐ด| is a product of ๐‘› elements of ๐ด
such that one and only one element comes from each row and one and only one element
comes from each column. Thus if the factors come from successive rows then the first
number in the subscripts are in the natural order 1, 2, …, ๐‘›. Since all rows are different then
the subscripts ๐‘—1 , ๐‘—2 , … , ๐‘—๐‘› are distinct. This means that {๐‘—1 , ๐‘—2 , … , ๐‘—๐‘› } is a permutation of {1,
2, … , ๐‘› }.
The sign of the term ๐‘Ž1๐‘—1 ๐‘Ž2๐‘—2 … ๐‘Ž๐‘›๐‘—๐‘› is if the permutation ๐‘—1 , ๐‘—2 , … , ๐‘—๐‘› is even
otherwise, the sign is .
๐‘Ž11
Example 2: Let ๐ด = [๐‘Ž
21
๐‘Ž12
๐‘Ž22 ].
To illustrate Definition 2.7.2, let us consider exactly one element from each row and each
column. There are two possibilities:
๐‘Ž11
[ ∗
∗
๐‘Ž22 ]
(1)
∗
[๐‘Ž
๐‘Ž12
∗ ]
(2)
21
In (1), the second subscripts form the even permutation 12 hence the sign of the
product ๐‘Ž11 ๐‘Ž22 is positive. In (2), the second subscripts form the odd permutation 21 hence
the sign of the product ๐‘Ž12 ๐‘Ž21 is negative. Thus
โ”‚Aโ”‚= a11a22 ๏€ญ a12 a21 .
The determinant in Example 2 can be solved by writing the terms
๐‘Ž1___ ๐‘Ž2___
and
๐‘Ž1___ ๐‘Ž2___
(there are only two terms because
๐‘†2 has only two permutations)
Next we fill in the blanks with the two permutations of ๐‘†2 .
๐‘Ž11 ๐‘Ž22
and
๐‘Ž12 ๐‘Ž21
We assign a sign for the even permutation 12 and a
Taking the sum of the two terms we have
โ”‚Aโ”‚= a11a22 ๏€ญ a12 a21 .
sign for the odd permutation 21.
48
Example 3: Let ๐ด = [
2 −4
] then
3 5
|๐ด| = (2)(5) − (−4)(3) = 22
๏ƒฉ a11
Example 4: Let A ๏€ฝ ๏ƒช๏ƒชa 21
๏ƒช๏ƒซa31
terms
a12
a 22
a32
a13 ๏ƒน
a 23 ๏ƒบ๏ƒบ . Since ๐‘†3 has 6 permutations then we write the 6
a33 ๏ƒบ๏ƒป
๐‘Ž1___ ๐‘Ž2___ ๐‘Ž3___ ,
๐‘Ž1___ ๐‘Ž2___ ๐‘Ž3___,
๐‘Ž1___ ๐‘Ž2___ ๐‘Ž3___,
๐‘Ž1___ ๐‘Ž2___ ๐‘Ž3___,
๐‘Ž1___ ๐‘Ž2___ ๐‘Ž3___,
and
๐‘Ž1___ ๐‘Ž2___ ๐‘Ž3___
Next we fill in the blanks with the six permutations of ๐‘†3 . The even permutations are 123,
231, and 312 and the odd permutations are 213, 132, and 321. Thus
|๐ด| = ๐‘Ž11 ๐‘Ž22 ๐‘Ž33 + ๐‘Ž12 ๐‘Ž23 ๐‘Ž31 + ๐‘Ž13 ๐‘Ž21 ๐‘Ž32 − ๐‘Ž12 ๐‘Ž21 ๐‘Ž33 − ๐‘Ž11 ๐‘Ž23 ๐‘Ž32 − ๐‘Ž13 ๐‘Ž22 ๐‘Ž31
We can also obtain the determinant of ๐ด by augmenting the first two columns as
shown below:
๐‘Ž11 ๐‘Ž12 ๐‘Ž13 ๐‘Ž11 ๐‘Ž12
๐‘Ž21 ๐‘Ž22 ๐‘Ž23 ๐‘Ž21 ๐‘Ž22
๐‘Ž31 ๐‘Ž32 ๐‘Ž33 ๐‘Ž31 ๐‘Ž32
− −
− +
+ +
We form the product of each of the three entries joined by the line from left to right and
precede each product by a plus sign. Next we form the product of each of the three entries
joined by the line from right to left and precede each product by a minus sign.
1
2
Example 5: Let ๐ด = [−2 1
3 −1
1
2 3
|๐ด| = |−2 1 4
3 −1 2
1
−2
3
3
4]. Then
2
2
1|
−1
= (1)(1)(2) + (2)(4)(3) + (3)(-2)(-1) – (3)(1)(3) – (1)(4)(-1) – (2)(-2)(2)
= 35
49
2.8 Properties of Determinants
Theorem 2.8.1: The determinant of a matrix ๐ด and its transpose are equal.
Example 1. Let A be the matrix of Example 5. Then
1
๐ด = [2
3
−2 3
1 −1]
4
2
๐‘‡
The determinant of AT is
1 −2 3
๐ด = |2 1 −1
3 4
2
๐‘‡
1 −2
2 1|
3 4
= (1)(1)(2) + (-2)(-1)(3) + (3)(2)(4) – (3)(1)(3) – (1)(-1)(4) – (-2)(2)(2)
= 35
Because of this property we can now replace “row” by “column” in the succeeding
theorems about the determinants of a matrix ๐ด.
Theorem 2.8.2: Let ๐ต be the matrix obtained from a matrix ๐ด by
(i)
multiplying a row (column) of ๐ด by a scalar ๐‘; then |๐ต| = ๐‘|๐ด|.
(ii) interchanging two rows(columns) of |๐ด|; then |๐ต| = −|๐ด|.
(iii) adding a multiple of a row(column) of ๐ด to another; then |๐ต| = |๐ด|.
2
Example 2. To illustrate part (i) of Theorem 2.8.2, let ๐ต = [
4
factor of the entries in column 1, then
5
]. Since 2 is a common
−7
2 5
1 5
๏€ฝ2
๏€ฝ 2(๏€ญ7 ๏€ญ 10) ๏€ฝ ๏€ญ34
4 ๏€ญ7
2 ๏€ญ7
๏ƒฉ 2 ๏€ญ1๏ƒน
๏ƒฉ5 ๏€ญ2๏ƒน
Example 3: To illustrate part (ii) of Theorem 2.8.2, let A ๏€ฝ ๏ƒช
and B ๏€ฝ ๏ƒช
๏ƒบ
๏ƒบ . Then
๏ƒซ5 ๏€ญ2๏ƒป
๏ƒซ 2 ๏€ญ1๏ƒป
A ๏€ฝ 4 + 5 = 1 and B = 5 + 4 = 1. This shows that |๐ต| = −|๐ด|.
50
6
Example 4: To illustrate part (iii) of theorem 2.8.2, let ๐ด = [−1
3
A
๏ƒฉ 6 9 ๏€ญ12๏ƒน
๏ƒช ๏€ญ1 0 2 ๏ƒบ 3๐‘… + ๐‘…
2
3
๏ƒช
๏ƒบ
๏ƒช๏ƒซ 3 0 ๏€ญ8 ๏ƒบ๏ƒป
๐‘…3
9 −12
0
2 ].
0 −8
B
๏ƒฉ 6 9 ๏€ญ12๏ƒน
๏ƒช ๏€ญ1 0 2 ๏ƒบ
๏ƒช
๏ƒบ
๏ƒช๏ƒซ 0 0 ๏€ญ2 ๏ƒบ๏ƒป
|๐ด| = 54 − 72 = −18 and |๐ต| = 0 − 18 = −18. Thus |๐ด| = |๐ต|.
Theorem 2.8.3: Let ๐ด be a square matrix.
(i) If two rows (columns) of ๐ด are equal, then โ”‚๐ดโ”‚ = 0.
(ii) If a row (column) of ๐ด consists entirely of zeros, then โ”‚๐ดโ”‚ = 0.
Proof of part (i):
Suppose the rows ๐‘š and ๐‘› of ๐ด are equal. Interchange rows ๐‘š and ๐‘› of ๐ด to obtain a matrix
๐ต. By theorem 2.8.2 (ii), |๐ต| = −|๐ด|. Since rows ๐‘š and ๐‘› are equal then ๐ต = ๐ด, so |๐ต| =
|๐ด|. Thus
|๐ด| = −|๐ด|
(By substitution)
Hence |๐ด| = 0.
๏ƒฉ 3 2 1๏ƒน
Example 5: Let A ๏€ฝ ๏ƒช๏ƒช ๏€ญ1 0 4 ๏ƒบ๏ƒบ (row 1 is equal to row 3).The determinant of A is
๏ƒช๏ƒซ 3 2 1 ๏ƒบ๏ƒป
3 2 1
A ๏€ฝ ๏€ญ1 0 4 = 3(0)(4) + 2(4)(3) +(-1)(2)(1)] – 3(2)(4) - 2(-1)(1) - 1(0)(3)
3 2 1
= 0 + 24 – 2 – 24 + 2 – 0
= 0
51
Proof of part (ii):
Let the ๐‘›th row(column) of ๐ด consist entirely of zero. Since each term in the |๐ด| contains a
factor from each row(column) then each term in the |๐ด| is equal to zero. Hence |๐ด| = 0.
Theorem 2.8.4: If a matrix A ๏€ฝ [ a ij ] is upper (lower) triangular, then
A ๏€ฝ a11a22 ...ann
that is, the determinant of a triangular matrix is the product of the elements on the main
diagonal.
Example 6: Evaluate the determinant of each matrix by applying theorem 2.8.4.
2
(a) ๐ด = [0
1
4
2
(b) ๐ต = [ 3 −2
−2
0
8 −2
3 −4
−4 2 ]
−1 5
3 −4
1
5]
1 −3
6
4
Solution:
(a) We transform matrix ๐ด into triangular form by applying elementary operations and
taking note of the corresponding changes in the determinant.
2
|0
1
3 −4
−4 2 |
−1 5
๐‘…1
๐‘…3
−2๐‘…1 + ๐‘…3
1
๐‘…
4 2
5๐‘…2 + ๐‘…3
1
= − |0
2
(By theorem 2.8.2.ii)
๐‘…3
1 −1
5
= − |0 −4
2 |
0 5 −14
(By theorem 2.8.2.iii)
๐‘…2
1
= (−4) |0
0
(By theorem 2.8.2.i)
๐‘…3
1 −1
5
1/2 |
= ( −4) |0 −1
0 0 −23/2
The last matrix is in triangular form thus
23
−1 5
−4 2 |
3 −4
|๐ด| = (−4)(1)(−1) (− ) = −46
2
−1
5
−1 1/2 |
5 −14
(By theorem 2.8.2.iii)
52
4
2
3
−2
(b) |
−2
0
8 −2
3 −4
1
5| ๐‘…
1
1 −3
6
4
๐‘…3
2๐‘…1 + ๐‘…3
4๐‘…1 + ๐‘…4
1
− ๐‘…1
2
1
0
= −(−2) | 3 −2
0
2
0 −2
๐‘…1
−๐‘…3 + ๐‘…4
๐‘…2
๐‘…3
๐‘…4
๐‘…4
1 −3
1
5 | (By theorem 2.8.2.ii)
3 −4
6
4
−2
0 1 −3
3
−2
1
5 | (By theorem 2.8.2.iii)
= −|
0
2
5 −10
0 −2 10 −8
๐‘…3
๐‘…4
−3๐‘…1 + ๐‘…2
๐‘…2 + ๐‘…3
−๐‘…2 + ๐‘…4
−2
0
3
−2
= −|
4
2
8 −2
1
0
= 2 | 0 −2
0
2
0 −2
−1/2 3/2
1
5 | (By theorem 2.8.2.i)
5 −10
10
−8
−1/2
5/2
5
10
1 0 −1/2
0 −2 5/2
= 2|
15/2
0 0
15/2
0 0
3/2
1/2
|(By theorem 2.8.2.iii)
−10
−8
3/2
1/2
|(By theorem 2.8.2.iii)
−19/2
−17/2
3/2
1 0 −1/2
5/2
1/2
= 2 | 0 −2
|(By theorem 2.8.2.iii)
0 0 15/2 −19/2
0 0
0
1
15
Thus |๐ต| = 2(1)(−2) ( 2 ) (1) = −30.
Theorem 2.8.5: The determinant of a product of two matrices is the product of their
determinants; that is
AB ๏€ฝ A B
Corollary 2.8.1: If A is nonsingular, then A ๏‚น 0 and
A ๏€ญ1 ๏€ฝ
1
.
A
53
๏ƒฉ 2 ๏€ญ4๏ƒน
๏ƒฉ 5 / 22 2 /11๏ƒน
Example 7. Let A ๏€ฝ ๏ƒช
then A๏€ญ1 ๏€ฝ ๏ƒช
๏ƒบ.
๏ƒบ
๏ƒซ ๏€ญ3 / 22 1/11 ๏ƒป
๏ƒซ3 5 ๏ƒป
The determinant of A is โ”‚Aโ”‚= 10 – (-12) = 22 and the determinant of A-1 is
1
A-1โ”‚= (5/22)(1/11) – (2/11)(-3/22) = 1/22. Hence |๐ด−1 | =
.
|๐ด|
ACTIVITY
1. Evaluate the determinants of the following matrices.
๏ƒฉ 4 3 2๏ƒน
a. A ๏€ฝ ๏ƒช๏ƒช3 ๏€ญ 2 5๏ƒบ๏ƒบ
๏ƒช๏ƒซ2 4 6๏ƒบ๏ƒป
๏ƒฉ 3 ๏€ญ 1 2๏ƒน
b. B ๏€ฝ ๏ƒช๏ƒช4 5 6๏ƒบ๏ƒบ
๏ƒช๏ƒซ7 1 2๏ƒบ๏ƒป
4 ๏ƒน
๏ƒฉ ๏€ญ1 2
๏ƒช
c. ๏ƒช 4 ๏€ญ8 ๏€ญ16 ๏ƒบ๏ƒบ
๏ƒช๏ƒซ 3 0
5 ๏ƒบ๏ƒป
2. Find all values of ๏ฌ for which
a.
๏ฌ ๏€ญ2
3
2
๏€ฝ0
๏ฌ ๏€ญ3
1 0 ๏€ญ1
b. ๏ฌ I3 ๏€ญ A ๏€ฝ 0 where A ๏€ฝ 2 0 1
0 0 ๏€ญ1
3. The matrix ๐ด is called idempotent if ๐ด2 = ๐ด. What are the possible values for the |๐ด| if ๐ด
is idempotent?
4. Prove Corollary 2.8.1.
2.9 Minors and Cofactor Expansion
For large matrices such as ๐‘› ≥ 4, evaluating the determinants using the permutation
formula could be very tedious. We see from the preceding section that transforming a
matrix into a triangular matrix could make the computation easier. Another method that is
also efficient in the computation of determinants of large matrices is by means of cofactor
expansion.
54
Definition 2.9.1: Let A ๏€ฝ [ a ij ] be an ๐‘› x ๐‘› matrix. Let M ij be the (๐‘› -1) x (๐‘› – 1) submatrix
of ๐ด obtained by deleting the ๐‘–th row and the ๐‘—th column of ๐ด. The determinant M ij is
called the minor of ๐‘Ž๐‘–๐‘— . The cofactor Aij of ๐‘Ž๐‘–๐‘— is defined as
Aij ๏€ฝ (๏€ญ1) i ๏€ซ j M ij
๏ƒฉ 5 ๏€ญ1 6 ๏ƒน
Example 1: Let A ๏€ฝ ๏ƒช๏ƒช 3 4 2 ๏ƒบ๏ƒบ . If we delete the first row and second column of ๐ด we
๏ƒช๏ƒซ7 2 1 ๏ƒบ๏ƒป
3 2
obtain the 2 x 2 submatrix ๐‘€12 = [
]. The determinant of ๐‘€12 is the minor of ๐‘Ž12 =
7 1
−1. Thus
M 12 ๏€ฝ
3 2
๏€ฝ 3 ๏€ญ 14 ๏€ฝ ๏€ญ11 .
7 1
Likewise, if we delete the second row and the third column of ๐ด we obtain the 2 x 2
5 −1
submatrix ๐‘€23 = [
]. The determinant |๐‘€23 | is called the minor of ๐‘Ž23 = 2. Thus
7 2
M 23 ๏€ฝ
5 ๏€ญ1
๏€ฝ 10 ๏€ญ (๏€ญ7) ๏€ฝ 17 .
7 2
By Definition 2.9.1, the cofactor of ๐‘Ž12 is
๐ด12 = (−1)1+2 |๐‘€12 |
= −(−11)
= 11
Similarly the cofactor of ๐‘Ž23 is
๐ด23 = (−1)2+3 |๐‘€23 |
= (−1)(17)
= −17
55
Theorem 2.9.1: Let A ๏€ฝ [ a ij ] be an ๐‘› x ๐‘› matrix. Then for each 1 ≤ ๐‘– ≤ ๐‘›,
A ๏€ฝ ai1 Ai1 ๏€ซ ai 2 Ai 2 ๏€ซ ... ๏€ซ ain Ain
(expansion of โ”‚๐ดโ”‚ about the ๐‘–th row)
and for each 1 ≤ ๐‘— ≤ ๐‘›,
A ๏€ฝ a1 j A1 j ๏€ซ a2 j A2 j ๏€ซ ... ๏€ซ anj Anj
(expansion of โ”‚๐ดโ”‚ about the ๐‘—th column)
2 2 −3 1
0 1
2 −1
Example 2: Evaluate the determinant of ๐ด = [
] by cofactor expansion.
3 −1
4 1
2 3
0 0
Solution: You can expand about any row or column of your choice, however it is best to
expand about the fourth row because it has the most number of zeros. Thus if we expand
about the fourth row we obtain
|๐ด| = ๐‘Ž41 ๐ด41 + ๐‘Ž42 ๐ด42 + ๐‘Ž43 ๐ด43 + ๐‘Ž44 ๐ด44
= 2๐ด41 + 3๐ด42 + 0๐ด43 + 0๐ด44
Notice that the cofactor of a zero entry need not be calculated, so if we can get another
zero on the fourth row then the computation would be a lot easier. Let us apply
elementary operation on the determinant of ๐ด and take note of the changes in the
determinant.
2 2 −3 1
0 1
2 −1
|
|
3 −1
4 1
2 3
0 0
1
๐ถ
2 1
−3๐ถ1 + ๐ถ2
๐ถ1
1
2 −3 1
0
1
2 −1
= 2|
|
3/2 − 1
4 1
1
3
0 0
By theorem 2.8.2.i
1
−1 −3
1
0
1
2 −1
๐ถ2 = 2 |
| By theorem 2.8.2.iii
3/2 − 11/2 4
1
1
0
0
0
1
−1 −3
1
0
1
2 −1
Now let us evaluate the determinant |
| . Expanding about the fourth
3/2 − 11/2 4
1
1
0
0
0
row we have
56
1
−1 −3
1
0
1
2 −1
|
| = (1)๐ด41 + 0๐ด42 + 0๐ด43 + 0๐ด44
3/2 − 11/2 4
1
1
0
0
0
=
(1)(−1)4+1
−1 −3 1
| 111 2 −1|
−2
4
1
−1 −3 1
= − | 111 2 −1|
−2
4
1
−1
Next we evaluate | 111
−
2
−3
2
4
1
−1| by expanding about the third row. (Note, you can
1
expand about any row or column of your choice.)
−1
| 111
−
2
−3
2
4
1
−1| = (− 11) (−1)3+1 |−3
2
2
1
1
−1 1
| + (4)(−1)3+2 |
|
−1
1 −1
−1 −3
+ (1)(−1)3+3 |
|
1
2
11
= (− 2 ) (1) + (−4)(0) + (1)(1)
= −9/2
Substituting, we have
1
−1 −3
1
9
0
1
2 −1
|
| = − (− 2) = 9/2
3/2 − 11/2 4
1
1
0
0
0
57
Hence
2 2 −3 1
9
0 1
2 −1
|
| = 2( ) = 9
3 −1
4 1
2
2 3
0 0
SAQ 2-4
๏ƒฉ4 ๏€ญ 4 2 1 ๏ƒน
๏ƒช1 2
0 3 ๏ƒบ๏ƒบ
๏ƒช
Let A ๏€ฝ
. Evaluate the determinant by
๏ƒช2 0
3 4๏ƒบ
๏ƒช
๏ƒบ
๏ƒซ0 ๏€ญ 3 2 1 ๏ƒป
(a) reducing ๐ด into a triangular matrix, and
(b) expanding about the second row.
ASAQ 2-4
(a) Reducing ๐ด into a triangular matrix using elementary row operation we have
4 −4
1 2
|
2 0
0 −3
2
0
3
2
1
3
| ๐‘…1
4
1
−4๐‘…1 + ๐‘…2
−2๐‘…1 + ๐‘…3
3๐ถ4 + ๐ถ2
−2๐ถ4 + ๐ถ3
1 2
4 −4
๐‘…2 = − |
2 0
0 −3
0
2
3
2
3
1
|
4
1
By theorem 2.8.2.ii
0
3
2 − 11
|
3 −2
2 1
By theorem 2.8.2.iii
1 11 − 6 3
๐ถ2
0 − 45 24 − 11
= −|
|
๐ถ3
0 − 10 7 − 2
0 0 0
1
By theorem 2.8.2.iii
1
2
๐‘…2
0 − 12
= −|
๐‘…3
0 −4
0 −3
58
1
9
๐‘…2
−2๐‘…2 + ๐‘…3
−6
1 11
24/9
๐‘…2 = −9 |0 −5
0 −10
7
0
0
0
1 11
๐‘…3 = −9 | 0 −5
0 −10
0
0
3
−11/9
|
−2
1
By theorem 2.8.2.i
−6
3
24/9 −11/9
| By theorem 2.8.2.iii
5/3
4/9
0
1
5
Hence |๐ด| = −9(1)(−5) (3) (1) = 75
(b) Before we expand about the second row, let us introduce more zeros on the second row
by applying elementary column operations on the determinant.
4 −4
1 2
|
2 0
0 −3
2
0
3
2
1
3 −2๐ถ1 + ๐ถ2
|
4 −3๐ถ1 + ๐ถ4
1
4 − 12
๐ถ2
1
0
= |
๐ถ4
2 −4
0 −3
2 − 11
0
0
|
3 −2
2
1
By theorem 2.8.2.iii
Expanding about the 2nd row we have
4 − 12
1
0
|
2 −4
0 −3
2 − 11
0
0
| = 1๐ด21
3 −2
2
1
−12 2 −11
= 1(−1)2+1 | −4 3 −2 |
−3 2
1
Next we introduce zeros on the third row by using column operations.
−12 2 −11 3๐ถ + ๐ถ
3
1
− | −4 3 −2 |
−2๐ถ3 + ๐ถ2
−3 2
1
−45 24
๐ถ1
= − |−10 7
๐ถ2
0
0
−11
−2 |
1
−45 24
= −(−1)3+3 |
| expanding about the
−10 7
3rd row
= −(−315 + 240)
= 75
59
SAQ 2-5
Prove: If the ๐‘–th row of A is multiplied by a scalar ๐‘, then the
determinant of ๐ด is multiplied by ๐‘.
ASAQ 2-5
๐‘Ž11
๐‘Ž21
| โ‹ฎ
Let |๐ต| = ๐‘๐‘Ž
| ๐‘–1
โ‹ฎ
๐‘Ž๐‘›1
๐‘Ž12
๐‘Ž22
โ‹ฎ
๐‘๐‘Ž๐‘–2
โ‹ฎ
๐‘Ž๐‘›2
โ‹ฏ ๐‘Ž1๐‘›
โ‹ฏ ๐‘Ž2๐‘›
โ‹ฎ |
โ‹ฏ ๐‘๐‘Ž๐‘–๐‘› |
โ‹ฎ
โ‹ฏ ๐‘Ž๐‘›๐‘›
If we expand about the ๐‘–th row, we have
|๐ต| = ๐‘๐‘Ž๐‘–1 ๐ด๐‘–1 + ๐‘๐‘Ž๐‘–2 ๐ด๐‘–2 + โ‹ฏ + ๐‘๐‘Ž๐‘–๐‘› ๐ด๐‘–๐‘›
= ๐‘(๐‘Ž๐‘–1 ๐ด๐‘–1 + ๐‘Ž๐‘–2 ๐ด๐‘–2 + โ‹ฏ + ๐‘Ž๐‘–๐‘› ๐ด๐‘–๐‘› )
= ๐‘|๐ด|
Definition 2.9.2: If A ๏€ฝ [ a ij ] is an ๐‘› x ๐‘› matrix, the adjoint of ๐ด denoted by adj๐ด, is the
transpose of the matrix of cofactors. Thus,
๏ƒฉ A11
๏ƒชA
adjA ๏€ฝ ๏ƒช 12
๏ƒช
๏ƒช
๏ƒซ๏ƒช A1n
A21 ... An1 ๏ƒน
A22 ... An 2 ๏ƒบ๏ƒบ
๏ƒบ
๏ƒบ
A2 n ... Ann ๏ƒป๏ƒบ
60
๏ƒฉ 2 1 3๏ƒน
Example 3: Let A ๏€ฝ ๏ƒช๏ƒช ๏€ญ1 2 0 ๏ƒบ๏ƒบ . Compute adj ๐ด.
๏ƒช๏ƒซ 3 ๏€ญ2 1 ๏ƒบ๏ƒป
Solution:
2 0
1 3
A11 ๏€ฝ (๏€ญ1)1๏€ซ1
๏€ฝ2
A21 ๏€ฝ (๏€ญ1) 2๏€ซ1
๏€ฝ ๏€ญ7
๏€ญ2 1
๏€ญ2 1
A31 ๏€ฝ (๏€ญ1)3๏€ซ1
1 3
๏€ฝ ๏€ญ6
2 0
A12 ๏€ฝ (๏€ญ1)1๏€ซ 2
๏€ญ1 0
๏€ฝ1
3 1
A22 ๏€ฝ (๏€ญ1) 2๏€ซ 2
2 3
๏€ฝ ๏€ญ7
3 1
A32 ๏€ฝ (๏€ญ1)3๏€ซ 2
2 3
๏€ฝ ๏€ญ3
๏€ญ1 0
A13 ๏€ฝ (๏€ญ1)1๏€ซ3
๏€ญ1 2
๏€ฝ ๏€ญ4
3 ๏€ญ2
A23 ๏€ฝ (๏€ญ1) 2๏€ซ3
2 1
๏€ฝ7
3 ๏€ญ2
A33 ๏€ฝ (๏€ญ1)3๏€ซ3
2 1
๏€ฝ5
๏€ญ1 2
Hence,
๏ƒฉ 2 ๏€ญ7 ๏€ญ6 ๏ƒน
adj ๐ด = ๏ƒช๏ƒช 1 ๏€ญ7 ๏€ญ3๏ƒบ๏ƒบ
๏ƒช๏ƒซ ๏€ญ4 7 5 ๏ƒบ๏ƒป
Theorem 2.9.2: If A ๏€ฝ [ a ij ] is an ๐‘› x ๐‘› matrix, then
๐ด(๐‘Ž๐‘‘๐‘—๐ด) = (๐‘Ž๐‘‘๐‘—๐ด)๐ด = |๐ด|๐ผ๐‘›
๏ƒฉ 2 1 3๏ƒน
Example 4: Let A ๏€ฝ ๏ƒช๏ƒช ๏€ญ1 2 0 ๏ƒบ๏ƒบ be the matrix of Example 3. Compute |๐ด| using Theorem
๏ƒช๏ƒซ 3 ๏€ญ2 1 ๏ƒบ๏ƒป
2.9.2.
Solution:
๏ƒฉ 2 1 3๏ƒน ๏ƒฉ 2 ๏€ญ7 ๏€ญ6๏ƒน
๏ƒฉ ๏€ญ7 0 0 ๏ƒน
๏ƒฉ1 0 0 ๏ƒน
๏ƒช
๏ƒบ
๏ƒช
๏ƒบ
๏ƒช
๏ƒบ
๐ด(adj ๐ด) = ๏ƒช ๏€ญ1 2 0๏ƒบ ๏ƒช 1 ๏€ญ7 ๏€ญ3๏ƒบ = ๏ƒช 0 ๏€ญ7 0 ๏ƒบ ๏€ฝ ๏€ญ7 ๏ƒช๏ƒช0 1 0 ๏ƒบ๏ƒบ
๏ƒซ๏ƒช 3 ๏€ญ2 1 ๏ƒบ๏ƒป ๏ƒช๏ƒซ ๏€ญ4 7 5 ๏ƒป๏ƒบ
๏ƒซ๏ƒช 0 0 ๏€ญ7 ๏ƒป๏ƒบ
๏ƒซ๏ƒช0 0 1 ๏ƒป๏ƒบ
Thus, A ๏€ฝ ๏€ญ7
61
Corollary 2.9.1: If A is ๐‘› x ๐‘› matrix and A ๏‚น 0 , then
A ๏€ญ1 ๏€ฝ
1
(adjA)
A
๏ƒฉ 2 1 3๏ƒน
Example 5: Let A ๏€ฝ ๏ƒช๏ƒช ๏€ญ1 2 0 ๏ƒบ๏ƒบ be the matrix of Example 3, find ๐ด−1 .
๏ƒช๏ƒซ 3 ๏€ญ2 1 ๏ƒบ๏ƒป
Solution:
From previous examples we have
๏ƒฉ 2 ๏€ญ7 ๏€ญ6 ๏ƒน
|๐ด| = −7 and adj ๐ด = ๏ƒช 1 ๏€ญ7 ๏€ญ3๏ƒบ
๏ƒช
๏ƒบ
๏ƒช๏ƒซ ๏€ญ4 7 5 ๏ƒบ๏ƒป
Hence,
−1
๐ด
๏ƒฉ 2 ๏€ญ7 ๏€ญ6 ๏ƒน
= −7 ๏ƒช๏ƒช 1 ๏€ญ7 ๏€ญ3๏ƒบ๏ƒบ
๏ƒช๏ƒซ ๏€ญ4 7 5 ๏ƒบ๏ƒป
1
−2/7
= [−1/7
4/7
1
6/7
1
3/7 ]
−1 −5/7
You can check your answer by taking the product ๐ด๐ด−1 and making sure that the answer is
the identity matrix ๐ผ3 .
Theorem 2.9.3: A matrix A is nonsingular if and only if A ๏‚น 0 .
Corollary 2.9.2: If A is an ๐‘› x ๐‘› matrix, then the homogeneous system AX = 0 has a nontrivial
solution if and only if A ๏€ฝ 0 .
62
Theorem 2.9.4: (Cramer’s Rule)
Let
a11 x1 ๏€ซ a12 x2 ๏€ซ ... ๏€ซ a1n xn ๏€ฝ b1
a21 x1 ๏€ซ a22 x2 ๏€ซ ... ๏€ซ a2n xn ๏€ฝ b2
an1 x1 ๏€ซ an 2 x2 ๏€ซ ... ๏€ซ ann xn ๏€ฝ bn
be a linear system of ๐‘› equations in ๐‘› unknowns and let A ๏€ฝ [ a ij ] be the coefficient matrix
so that we can write the given system as AX = B, where
๏ƒฉb1 ๏ƒน
๏ƒชb ๏ƒบ
B ๏€ฝ ๏ƒช 2๏ƒบ
๏ƒช ๏ƒบ
๏ƒช ๏ƒบ
๏ƒซ๏ƒชbn ๏ƒป๏ƒบ
If A ๏‚น 0 , then the system has the unique solution
x1 ๏€ฝ
A1
A
, x2 ๏€ฝ
A2
A
, … , xn ๏€ฝ
An
A
,
where ๐ด๐‘– is the matrix obtained by replacing the ๐‘–th column of ๐ด by ๐ต.
Example 6: Using Cramer’s Rule, determine the solution of the linear system
3x + y – z = 4
- x + y + 3z = 0
x + 2y + z = 1
Solution:
The determinant of the coefficient matrix is (you must do the calculation here)
3 1 ๏€ญ1
A ๏€ฝ ๏€ญ 1 1 3 = −8 and
1 2 1
4 1 ๏€ญ1
4
A1 ๏€ฝ 0 1 3 = −16 (๐ด1 is obtained by replacing the first column of ๐ด by ๐ต = [0])
1
1 2 1
63
3
4 ๏€ญ1
A2 ๏€ฝ ๏€ญ 1 0
1 1
3
3 =8,
1
1 4
A3 ๏€ฝ ๏€ญ 1 1 0 = −8
1 2 1
4
(๐ด2 is obtained by replacing the second column of ๐ด by ๐ต = [0])
1
4
(๐ด3 is obtained by replacing the third column of ๐ด by ๐ต = [0])
1
By Cramer’s Rule,
x๏€ฝ
A1
A
๏€ฝ
A2
A3 ๏€ญ 8
8
๏€ญ 16
๏€ฝ
๏€ฝ ๏€ญ1, and z ๏€ฝ
๏€ฝ 2, y ๏€ฝ
๏€ฝ
๏€ฝ1
A
๏€ญ8
๏€ญ8
A
๏€ญ8
Thus the solution set is {(2, -1, 1)}.
NOTE: Cramer’s rule is only applicable to the case where we have ๐‘› equations in ๐‘›
unknowns and where the coefficient matrix ๐ด is nonsingular. Cramer’s rule becomes
computationally inefficient for ๐‘› > 4, and it is better to use the Gauss – Jordan method.
ACTIVITY
I. Solve the following system using Cramer’s Rule.
1. 2๐‘ฅ1 + 5๐‘ฅ2 − ๐‘ฅ3 = −1
4๐‘ฅ1 + ๐‘ฅ2 + 3๐‘ฅ3 = 3
−2๐‘ฅ1 + 2๐‘ฅ2
= 0
2. ๐‘ฅ1 + ๐‘ฅ2 + ๐‘ฅ3 + ๐‘ฅ4 = 6
2๐‘ฅ1
− ๐‘ฅ3 − ๐‘ฅ4 = 4
3 ๐‘ฅ3 + 6๐‘ฅ4 = 3
๐‘ฅ1
− ๐‘ฅ4 = 5
1 a a2
3. Show that 1 b b 2 = (b-a)(c-a)(c-b). Also, find the inverse of ๐ด by using the formula
1 c c2
1
๐ด−1 = |๐ด| adj ๐ด.
2
2
4. Find the determinant of ๐ด = 4
1
[2
1
3
5
0
1
0
1
2
1
0
−1 3
2
5
3 −3 .
2
4
1
0]
64
MODULE 3
VECTOR SPACES OVER A FIELD
Introduction
In this chapter we will discuss vectors in โ„๐‘› and its basic properties, properties and
structure of a vector space and subspace, linear combination and spanning sets.
Objectives
At the end of this chapter, you are expected to be able to do the following:
1. Define vector, vector spaces and subspaces.
2. Enumerate and explain the properties of vector spaces and subspaces.
3. Determine whether a set with given operations is a vector space.
4. Determine whether the given subsets of โ„๐‘› are subspaces.
5. Define linear combination and spanning.
6. Check whether a given set of vectors spans a vector space V.
3.1 Vectors in the Plane
Vectors are measurable quantities which has magnitude and direction. Example of
vectors are velocity, force, and acceleration.
Vectors in โ„๐Ÿ
A vector on the plane โ„๐Ÿ can be described as an ordered pair ๐— = (๐‘ฅ1 , ๐‘ฅ2 ) where
๐‘ฅ1 , ๐‘ฅ2 ∈ โ„. It can also be denoted by a 2 x 1 matrix
๐‘ฅ1
๐— = [๐‘ฅ ]
2
With X we associate the directed line segment with initial point at the origin O and the
โƒ—โƒ—โƒ—โƒ—โƒ— .
terminal point at ๐‘ƒ(๐‘ฅ1 , ๐‘ฅ2 ), denoted by ๐‘‚๐‘ƒ
The direction of a directed line segment is the angle made with the positive X-axis
and the magnitude of a directed line segment is its length.
65
Vector Operations
Definition 3.1.1: Let ๐— = (๐‘ฅ1 , ๐‘ฅ2 ) and ๐’€ = (๐‘ฆ1 , ๐‘ฆ2 ) be two vectors in the plane. The sum of
the vectors X and Y, denoted by ๐‘ฟ + ๐’€, is the vector
(๐‘ฅ1 + ๐‘ฆ1 , ๐‘ฅ2 + ๐‘ฆ2 ).
Example 1: Let X = (2, 3) and Y = (-4, 1). Then
X + Y = (-2, 4)
Definition 3.1.2: If X = (๐‘ฅ, ๐‘ฆ) and ๐‘ is a scalar (a real number), then the scalar multiple ๐‘๐‘ฟ of
๐‘ฟ by ๐‘ is the vector (๐‘๐‘ฅ, ๐‘๐‘ฆ).
If ๐‘ > 0, then ๐‘๐‘ฅ is in the same direction as ๐‘ฟ, whereas if ๐‘‘ < 0, then ๐‘‘๐‘ฅ is in the
opposite direction.
Example 2: Let c = 3, d = -2, and X = (3, -2). Then
cX = (9, -6) and dX = (-6, 4)
NOTE:
1. ๐‘‚ = (0, 0) is called the zero vector.
2. X + ๐‘‚ = X
3. X + (-1)X = ๐‘‚
4. (-1)X = X; the negative of X
5. X + (-1)Y = X – Y; the difference between X and Y.
3.2. ๐’-Vectors
Definition 3.2.1: An ๐‘›-vector is an ๐‘› x 1 matrix
๐‘ฅ1
๐‘ฅ2
๐‘ฟ=[โ‹ฎ ]
๐‘ฅ๐‘›
where ๐‘ฅ1 , ๐‘ฅ2 , … , ๐‘ฅ๐‘› are real numbers, which are called the components of X. The set of all
๐‘›-vectors is denoted by โ„๐‘› and is called ๐‘›-space.
66
Definition 3.2.2: Let ๐‘ฟ = (๐‘ฅ1 , ๐‘ฅ2 , … , ๐‘ฅ๐‘› ) and ๐’€ = (๐‘ฆ1 , ๐‘ฆ2 , … , ๐‘ฆ๐‘› ) be two vectors in โ„๐‘› .
The sum of the vectors X and Y is the vector
(๐‘ฅ1 + ๐‘ฆ1 , ๐‘ฅ2 + ๐‘ฆ2 , … , ๐‘ฅ๐‘› + ๐‘ฆ๐‘› )
and is denoted by X + Y.
Example 1: If X = (1, 2, -3) and Y = (0, 1, -2) are vectors in R3, then
X + Y = (1+0, 2+1, -3 + (-2) ) = ( 1, 3, -5 ).
Definition 3.2.3: If ๐‘ฟ = (๐‘ฅ1 , ๐‘ฅ2 , … , ๐‘ฅ๐‘› ) is a vector in โ„๐‘› and ๐‘ is a scalar, then the scalar
multiple ๐‘๐‘ฟ of ๐‘ฟ by ๐‘ is the vector
(๐‘๐‘ฅ1 , ๐‘๐‘ฅ2 , … , ๐‘๐‘ฅ๐‘› )
Example 2: If X = ( 2, 1, 5, -2 ) is a vector in โ„4 and c = -3, then
cX = (-3) ( 2, 1, 5, -2 ) = (-6, -3, -15, 6 )
The operations of vector addition and scalar multiplication satisfy the following properties:
Theorem 3.2.1: Let X, Y, and Z be any vectors in โ„๐‘› ; let ๐‘ and d be any scalars. Then
I. X + Y is a vector in โ„๐‘› (that is, โ„๐‘› is closed under the operation of vector addition).
a. X + Y = Y + X
b. X + ( Y + Z ) = ( X + Y ) +Z
c. There is a unique vector in โ„๐‘› , 0 = (0, 0, …, 0) such that X + 0 = 0 + X = X.
d. There is a unique vector – X,
-X = (- x1 , - x2 , … , - xn ) such that X + (- X) = 0
67
II. cX is a vector in โ„๐‘›
a. ( X + Y ) = cX + cY
b. ( c + d )X = cX + dX
c. c(dX) = (cd)X
d. 1X = X
3.3 Vector Spaces and Subspaces
Definition 3.3.1: A vector space V over a field ๐น is a nonempty set of elements, called
vectors, together with two operations called vector addition and scalar multiplication
satisfying the following properties:
[A1] (Closure under addition)
To every pair of vectors X, Y ∈ V, then X + Y ∈ V.
[A2] (Commutative law of vector addition)
X+ Y= Y+X
∀ X , Y ๏ƒŽ V.
[A3] (Associative law of vector addition)
X + ( Y + Z ) = ( X + Y ) + Z ∀ X, Y, and Z ๏ƒŽ V.
[A4] (Zero Element)
∃๐‘ถ ∈ V such that
X + ๐‘ถ = ๐‘ถ + X = X ∀ X ๏ƒŽ V.
[A5] (Negative Elements)
∀ X ๏ƒŽ V, ∃ X ∈ V such that
X + ( X) = ๐‘ถ
[M1] (Closure under scalar multiplication)
To every X ๏ƒŽ V and ๐‘ ∈ ๐น, then ๐‘ X ๏ƒŽ V.
[M2] (Distributive Law)
c ( X + Y ) = c X + c Y, ∀๐‘ ∈ ๐น , ∀ X, Y ๏ƒŽ V.
68
[M3] (Distributive Law)
(๐‘ + ๐‘‘ ) X = ๐‘ X + ๐‘‘ X, ∀๐‘, ๐‘‘ ∈ ๐น, ∀ X ๏ƒŽ V.
[M4] (Associative Law of Scalar Multiplication)
(๐‘๐‘‘ )X = ๐‘(๐‘‘ X ), ∀๐‘, ๐‘‘ ∈ ๐น, ∀ X ๏ƒŽ V.
[M5] (Identity Element)
1 X = X, ∀ X ๏ƒŽ V.
Example 1. Consider the set ๐‘ƒ๐‘› of all polynomials of degree ≤ ๐‘› together with the zero
polynomial. Show that ๐‘ƒ๐‘› is a vector space.
Solution: We have to show that all properties are satisfied. A polynomial in ๐‘ƒ๐‘› is expressible
as
๐‘(๐‘ฅ) = ๐‘0 ๐‘ฅ ๐‘› + ๐‘1 ๐‘ฅ ๐‘›−1 + … + ๐‘๐‘›−1 ๐‘ฅ + ๐‘๐‘›
where ๐‘0 , ๐‘1 , … , ๐‘๐‘› are real numbers. Let ๐‘(๐‘ฅ), ๐‘ž(๐‘ฅ), and ๐‘Ÿ(๐‘ฅ) ∈ ๐‘ƒ๐‘› and ๐‘, ๐‘‘ ∈ โ„. Then
[A1] ๐‘(๐‘ฅ) + ๐‘ž(๐‘ฅ) = (๐‘0 + ๐‘ž0 )๐‘ฅ ๐‘› + (๐‘1 + ๐‘ž1 )๐‘ฅ ๐‘›−1 + โ‹ฏ + (๐‘๐‘›−1 + ๐‘ž๐‘›−1 )๐‘ฅ + (๐‘๐‘› + ๐‘ž๐‘› )
Clearly, the sum of two polynomials of degree ≤ ๐‘› is another polynomial with degree ≤ ๐‘›.
Hence ๐‘ƒ๐‘› is closed under addition.
[A2] ๐‘(๐‘ฅ) + ๐‘ž(๐‘ฅ) = (๐‘0 + ๐‘ž0 )๐‘ฅ ๐‘› + (๐‘1 + ๐‘ž1 )๐‘ฅ ๐‘›−1 + โ‹ฏ + (๐‘๐‘›−1 + ๐‘ž๐‘›−1 )๐‘ฅ + (๐‘๐‘› + ๐‘ž๐‘› )
= (๐‘ž0 + ๐‘0 )๐‘ฅ ๐‘› + (๐‘ž1 + ๐‘1 )๐‘ฅ ๐‘›−1 + โ‹ฏ + (๐‘ž๐‘›−1 + ๐‘๐‘›−1 )๐‘ฅ + (๐‘ž๐‘› + ๐‘๐‘› )
= ๐‘ž(๐‘ฅ) + ๐‘(๐‘ฅ)
Thus addition is commutative.
[A3] [๐‘(๐‘ฅ) + ๐‘ž(๐‘ฅ)] + ๐‘Ÿ(๐‘ฅ) = [(๐‘0 + ๐‘ž0 )๐‘ฅ ๐‘› + (๐‘1 + ๐‘ž1 )๐‘ฅ ๐‘›−1 + โ‹ฏ + (๐‘๐‘›−1 + ๐‘ž๐‘›−1 )๐‘ฅ
+(๐‘๐‘› + ๐‘ž๐‘› )] + ๐‘Ÿ0 ๐‘ฅ ๐‘› + ๐‘Ÿ1 ๐‘ฅ ๐‘›−1 + … + ๐‘Ÿ๐‘›−1 ๐‘ฅ + ๐‘Ÿ๐‘›
= [(๐‘0 + ๐‘ž0 ) + ๐‘Ÿ0 ]๐‘ฅ ๐‘› + [(๐‘1 + ๐‘ž1 )+๐‘Ÿ1 ]๐‘ฅ ๐‘›−1 + โ‹ฏ +
[(๐‘๐‘›−1 + ๐‘ž๐‘›−1 ) + ๐‘Ÿ๐‘›−1 ]๐‘ฅ +[(๐‘๐‘› + ๐‘ž๐‘› ) + ๐‘Ÿ๐‘› ]
= [๐‘0 + (๐‘ž0 + ๐‘Ÿ0 )]๐‘ฅ ๐‘› + [๐‘1 + (๐‘ž1 + ๐‘Ÿ1 )]๐‘ฅ ๐‘›−1 + โ‹ฏ +
[๐‘๐‘›−1 + (๐‘ž๐‘›−1 + ๐‘Ÿ๐‘›−1 )]๐‘ฅ + [๐‘๐‘› + (๐‘ž๐‘› + ๐‘Ÿ๐‘› )]
69
= (๐‘0 ๐‘ฅ ๐‘› + ๐‘1 ๐‘ฅ ๐‘›−1 + … + ๐‘๐‘›−1 ๐‘ฅ + ๐‘๐‘› ) +
[(๐‘ž0 + ๐‘Ÿ0 )๐‘ฅ ๐‘› + (๐‘ž1 + ๐‘Ÿ1 )๐‘ฅ ๐‘›−1 + โ‹ฏ + (๐‘ž๐‘›−1 + ๐‘Ÿ๐‘›−1 )๐‘ฅ +
(๐‘ž๐‘› + ๐‘Ÿ๐‘› )]
= ๐‘(๐‘ฅ) + [๐‘ž(๐‘ฅ) + ๐‘Ÿ(๐‘ฅ)]
Thus vector addition is associative.
[A4] Let ๐‘‚ = 0๐‘ฅ ๐‘› + 0๐‘ฅ ๐‘›−1 + … + 0๐‘ฅ + 0 be the zero polynomial then
๐‘(๐‘ฅ) + ๐‘‚ = ๐‘‚ + ๐‘(๐‘ฅ) = ๐‘(๐‘ฅ)
[A5] Let – ๐‘(๐‘ฅ) = −๐‘0 ๐‘ฅ ๐‘› − ๐‘1 ๐‘ฅ ๐‘›−1 − โ‹ฏ − ๐‘๐‘›−1 ๐‘ฅ − ๐‘๐‘› , then
๐‘(๐‘ฅ) + (−๐‘(๐‘ฅ)) = ๐‘‚
[M1]
๐‘๐‘(๐‘ฅ) = ๐‘๐‘0 ๐‘ฅ ๐‘› + ๐‘๐‘1 ๐‘ฅ ๐‘›−1 + … + ๐‘๐‘๐‘›−1 ๐‘ฅ + ๐‘๐‘๐‘›
where ๐‘๐‘0 , ๐‘๐‘1 , … , ๐‘๐‘๐‘› ∈ โ„ . Clearly , the product of a real number and a polynomial is also
a polynomial. Hence ๐‘ƒ๐‘› is closed under scalar multiplication.
[M2] ๐‘[๐‘(๐‘ฅ) + ๐‘ž(๐‘ฅ)] = ๐‘[(๐‘0 + ๐‘ž0 )๐‘ฅ ๐‘› + (๐‘1 + ๐‘ž1 )๐‘ฅ ๐‘›−1 + โ‹ฏ + (๐‘๐‘›−1 + ๐‘ž๐‘›−1 )๐‘ฅ
+(๐‘๐‘› + ๐‘ž๐‘› )]
= ๐‘(๐‘0 + ๐‘ž0 )๐‘ฅ ๐‘› + ๐‘(๐‘1 + ๐‘ž1 )๐‘ฅ ๐‘›−1 + โ‹ฏ + ๐‘(๐‘๐‘›−1 + ๐‘ž๐‘›−1 )๐‘ฅ
+ ๐‘(๐‘๐‘› + ๐‘ž๐‘› )
= (๐‘๐‘0 + ๐‘๐‘ž0 )๐‘ฅ ๐‘› + (๐‘๐‘1 + ๐‘๐‘ž1 )๐‘ฅ ๐‘›−1 + โ‹ฏ + (๐‘๐‘๐‘›−1 + ๐‘๐‘ž๐‘›−1 )๐‘ฅ
+ (๐‘๐‘๐‘› + ๐‘๐‘ž๐‘› )
= (๐‘๐‘0 ๐‘ฅ ๐‘› + ๐‘๐‘1 ๐‘ฅ ๐‘›−1 + … + ๐‘๐‘๐‘›−1 ๐‘ฅ + ๐‘๐‘๐‘› ) +
(๐‘๐‘ž0 ๐‘ฅ ๐‘› + ๐‘๐‘ž1 ๐‘ฅ ๐‘›−1 + … + ๐‘๐‘ž๐‘›−1 ๐‘ฅ + ๐‘๐‘ž๐‘› )
= ๐‘๐‘(๐‘ฅ) + ๐‘๐‘ž(๐‘ฅ)
[M3] (๐‘ + ๐‘‘)๐‘(๐‘ฅ) = (๐‘ + ๐‘‘)(๐‘0 ๐‘ฅ ๐‘› + ๐‘1 ๐‘ฅ ๐‘›−1 + … + ๐‘๐‘›−1 ๐‘ฅ + ๐‘๐‘› )
= (๐‘ + ๐‘‘)๐‘0 ๐‘ฅ ๐‘› + (๐‘ + ๐‘‘)๐‘1 ๐‘ฅ ๐‘›−1 + … + (๐‘ + ๐‘‘)๐‘๐‘›−1 ๐‘ฅ + (๐‘ + ๐‘‘)๐‘๐‘›
70
= (๐‘๐‘0 + ๐‘‘๐‘0 )๐‘ฅ ๐‘› + (๐‘๐‘1 + ๐‘‘๐‘1 )๐‘ฅ ๐‘›−1 + โ‹ฏ + (๐‘๐‘๐‘›−1 + ๐‘‘๐‘๐‘›−1 )๐‘ฅ
+ (๐‘๐‘๐‘› + ๐‘‘๐‘๐‘› )
= (๐‘๐‘0 ๐‘ฅ ๐‘› + ๐‘๐‘1 ๐‘ฅ ๐‘›−1 + … + ๐‘๐‘๐‘›−1 ๐‘ฅ + ๐‘๐‘๐‘› ) +
๐‘‘๐‘0 ๐‘ฅ ๐‘› + ๐‘‘๐‘1 ๐‘ฅ ๐‘›−1 + … + ๐‘‘๐‘๐‘›−1 ๐‘ฅ + ๐‘‘๐‘๐‘›
= ๐‘๐‘(๐‘ฅ) + ๐‘‘๐‘(๐‘ฅ)
[M4] (๐‘๐‘‘)๐‘(๐‘ฅ) = (๐‘๐‘‘)๐‘0 ๐‘ฅ ๐‘› + (๐‘๐‘‘)๐‘1 ๐‘ฅ ๐‘›−1 + … + (๐‘๐‘‘)๐‘๐‘›−1 ๐‘ฅ + (๐‘๐‘‘)๐‘๐‘›
= ๐‘(๐‘‘๐‘0 )๐‘ฅ ๐‘› + ๐‘(๐‘‘๐‘1 )๐‘ฅ ๐‘›−1 + … + ๐‘(๐‘‘๐‘๐‘›−1 )๐‘ฅ + ๐‘(๐‘‘๐‘๐‘› )
= ๐‘(๐‘‘๐‘0 ๐‘ฅ ๐‘› + ๐‘‘๐‘1 ๐‘ฅ ๐‘›−1 + … + ๐‘‘๐‘๐‘›−1 ๐‘ฅ + ๐‘‘๐‘๐‘› )
= ๐‘(๐‘‘๐‘(๐‘ฅ))
[M5] 1๐‘(๐‘ฅ) = 1๐‘0 ๐‘ฅ ๐‘› + 1๐‘1 ๐‘ฅ ๐‘›−1 + … + 1๐‘๐‘›−1 ๐‘ฅ + 1๐‘๐‘›
= ๐‘0 ๐‘ฅ ๐‘› + ๐‘1 ๐‘ฅ ๐‘›−1 + … + ๐‘๐‘›−1 ๐‘ฅ + ๐‘๐‘›
= ๐‘(๐‘ฅ)
Since all ten properties are satisfied, then ๐‘ƒ๐‘› is a vector space.
Example 2. Let V be the set of all ordered triples of real numbers (๐‘ฅ, ๐‘ฆ, ๐‘ง) with the
operations
(๐‘ฅ, ๐‘ฆ, ๐‘ง) + (๐‘ฅ ′ , ๐‘ฆ ′ , ๐‘ง ′ ) = (๐‘ฅ + ๐‘ฅ ′ , ๐‘ฆ + ๐‘ฆ ′ , ๐‘ง + ๐‘ง ′ )
๐‘(๐‘ฅ, ๐‘ฆ, ๐‘ง) = (๐‘ฅ, 1, ๐‘ง)
Determine if V is a vector space.
Solution: To prove that V is a vector space, we have to show that all ten properties are
satisfied.
Let ๐‘ฟ = (๐‘ฅ, ๐‘ฆ, ๐‘ง), ๐’€ = (๐‘ฅ ′ , ๐‘ฆ ′ , ๐‘ง ′ ) and ๐’ = (๐‘ฅ ′′ , ๐‘ฆ ′′ , ๐‘ง ′′ )
[A1]
V and ๐‘, ๐‘‘ ∈ โ„. Then
๐‘ฟ + ๐’€ = (๐‘ฅ + ๐‘ฅ ′ , ๐‘ฆ + ๐‘ฆ ′ , ๐‘ง + ๐‘ง ′ )
Since ๐‘ฅ + ๐‘ฅ ′ , ๐‘ฆ + ๐‘ฆ ′ , and ๐‘ง + ๐‘ง ′
addition.
∈ โ„, then ๐‘ฟ + ๐’€ ∈ V.
Hence V is closed under
71
[A2] ๐‘ฟ + ๐’€ = (๐‘ฅ + ๐‘ฅ ′ , ๐‘ฆ + ๐‘ฆ ′ , ๐‘ง + ๐‘ง ′ )
= (๐‘ฅ ′ + ๐‘ฅ, ๐‘ฆ ′ + ๐‘ฆ, ๐‘ง ′ + ๐‘ง)
=๐’€+๐‘ฟ
Hence vector addition is commutative.
[A3] (๐‘ฟ + ๐’€) + ๐’ = (๐‘ฅ + ๐‘ฅ ′ , ๐‘ฆ + ๐‘ฆ ′ , ๐‘ง + ๐‘ง ′ ) + (๐‘ฅ ′′ , ๐‘ฆ ′′ , ๐‘ง ′′ )
= ((๐‘ฅ + ๐‘ฅ ′ ) + ๐‘ฅ ′′ , (๐‘ฆ + ๐‘ฆ ′ ) + ๐‘ฆ ′′ , (๐‘ง + ๐‘ง ′ ) + ๐‘ง′′)
= (๐‘ฅ + (๐‘ฅ ′ + ๐‘ฅ ′′ ), ๐‘ฆ + (๐‘ฆ ′ + ๐‘ฆ ′′ ), ๐‘ง + (๐‘ง ′ + ๐‘ง ′′ ))
= (๐‘ฅ, ๐‘ฆ, ๐‘ง) + (๐‘ฅ ′ + ๐‘ฅ ′′ , ๐‘ฆ′ + ๐‘ฆ ′′ , ๐‘ง ′ + ๐‘ง ′′ )
= ๐‘ฟ + (๐’€ + ๐’)
Thus vector addition is associative.
[A4] Let ๐‘ถ = (0, 0, 0) ∈ V be the zero vector then
๐‘ฟ + ๐‘ถ = (๐‘ฅ, ๐‘ฆ, ๐‘ง) + (0, 0, 0)
= (๐‘ฅ, ๐‘ฆ, ๐‘ง)
=๐‘ฟ
[A5] Let – ๐‘ฟ = (−๐‘ฅ, −๐‘ฆ, −๐‘ง) then
๐‘ฟ + (−๐‘ฟ) = (๐‘ฅ, ๐‘ฆ, ๐‘ง) + (−๐‘ฅ, −๐‘ฆ, −๐‘ง)
= (0, 0, 0)
=๐‘ถ
[M1] ๐‘๐‘ฟ = ๐‘(๐‘ฅ, ๐‘ฆ, ๐‘ง) = (๐‘ฅ, 1, ๐‘ง)
Since ๐‘ฅ and ๐‘ง are real numbers then ๐‘๐‘ฟ ∈ V. Hence V is closed under scalar multiplication.
72
[M2] ๐‘(๐‘ฟ + ๐’€) = ๐‘(๐‘ฅ + ๐‘ฅ ′ , ๐‘ฆ + ๐‘ฆ ′ , ๐‘ง + ๐‘ง ′ )
= (๐‘ฅ + ๐‘ฅ ′ , 1, ๐‘ง + ๐‘ง ′ )
Also,
๐‘๐‘ฟ + ๐‘๐’€ = ๐‘(๐‘ฅ, ๐‘ฆ, ๐‘ง) + ๐‘(๐‘ฅ ′ , ๐‘ฆ ′ , ๐‘ง ′ )
= (๐‘ฅ, 1, ๐‘ง) + (๐‘ฅ ′ , 1, ๐‘ง ′ )
= (๐‘ฅ + ๐‘ฅ ′ , 2, ๐‘ง + ๐‘ง ′ )
Since ๐‘(๐‘ฟ + ๐’€) ≠ ๐‘๐‘ฟ + ๐‘๐’€ then M2 is not satisfied. Hence V under the prescribed
operations is not a vector space.
Theorem 3.3.1: If V is a vector space, then:
a) 0๐‘ฟ = 0, for every ๐‘ฟ in V.
b) ๐‘0 = 0, for every scalar ๐‘.
c) If ๐‘๐‘ฟ = 0, then ๐‘ = 0 or ๐‘ฟ = 0
d) (-1)๐‘ฟ = −๐‘ฟ, for every ๐‘ฟ in V.
SAQ 3-1
Consider the set ๐‘€3,3 (โ„) of all 3 x 3 matrices with entries
in under the usual operations of matrix addition and
scalar multiplication. Show that ๐‘€3,3 (โ„) is a vector space.
ASAQ 3-1
Let ๐ด, ๐ต, and ๐ถ ∈ ๐‘€3,3 (โ„) and ๐‘, ๐‘‘ ∈ โ„. Then
[A1] For every ๐ด, ๐ต ๐‘€3,3 (โ„), we have ๐ด + ๐ต ∈ ๐‘€3,3 (โ„). Hence ๐‘€3,3 (โ„) is closed
under addition.
[A2] For every ๐ด, ๐ต
commutative.
๐‘€3,3 (โ„), we have ๐ด + ๐ต = ๐ต + ๐ด, that is matrix addition is
73
[A3] For every ๐ด, ๐ต, ๐ถ ๐‘€3,3 (โ„), we have (๐ด + ๐ต) + ๐ถ = ๐ด + (๐ต + ๐ถ), that is matrix
addition is associative.
[A4] Let ๐‘‚ be the 3 x 3 zero matrix, then for every ๐ด
๐ด+๐‘‚ =๐‘‚+๐ด=๐ด
[A5] For every ๐ด
๐‘€3,3 (โ„), we have
๐‘€3,3 (โ„), we have ๐ด + (−๐ด) = ๐‘‚.
[M1] For every ๐‘ ∈ โ„ and ๐ด ๐‘€3,3 (โ„), we have ๐‘๐ด
closed under scalar multiplication.
๐‘€3,3 (โ„). Hence ๐‘€3,3 (โ„) is
[M2] For every ๐‘ ∈ โ„ and ๐ด, ๐ต
๐‘€3,3 (โ„), we have ๐‘(๐ด + ๐ต) = ๐‘๐ด + ๐‘๐ต.
[M3] For every ๐‘, ๐‘‘ ∈ โ„ and ๐ด
๐‘€3,3 (โ„), we have (๐‘ + ๐‘‘)๐ด = ๐‘๐ด + ๐‘‘๐ด.
[M4] For every ๐‘, ๐‘‘ ∈ โ„ and ๐ด
๐‘€3,3 (โ„), we have (๐‘๐‘‘)๐ด = ๐‘(๐‘‘๐ด).
[M5] For every ๐ด
๐‘€3,3 (โ„),we have 1๐ด = ๐ด.
Hence ๐‘€3,3 (โ„) under the usual operations of matrix addition and scalar multiplication is a
vector space.
Definition 3.3.2: Let V be a vector space and W a nonempty subset of V. If W is a vector
space under the operations of addition and scalar multiplication defined on V, then W is
called a subspace of V.
Theorem 3.3.2: Let V be a vector space under the operations addition and scalar
multiplication and let W be a nonempty subset of V. Then W is a subspace of V if and only if
the following conditions hold:
a) The zero vector ๐‘ถ belongs to W;
b) For every vector ๐‘ฟ, ๐’€ ๏ƒŽ W, ๐‘ ๏ƒŽ โ„:
i. The sum ๐‘ฟ + ๐’€ ๏ƒŽ W
ii. The multiple ๐‘๐‘ฟ ๏ƒŽ W.
Property (i) in (b) states that W is closed under vector addition, and property (ii) in
(b) states that W is closed under scalar multiplication. Both properties may be combined
into the following equivalent single statement:
(b’) For every ๐‘ฟ, ๐’€ ๏ƒŽ W, ๐‘Ž, ๐‘ ๏ƒŽ
, the linear combination ๐‘Ž๐‘ฟ + ๐‘๐’€ ๏ƒŽ W.
74
REMARK: If V is any vector space, then V automatically contains two subspaces, the set {0}
consisting of the zero vector alone and the whole space V itself. These are called the trivial
subspaces of V.
Example 3: Consider the vector space โ„3 . Let U consists of all vectors in โ„3 whose entries
are equal; that is U = { (a, b, c) : a = b = c }. Show that U is a subspace of โ„3 .
Solution:
(a) The zero vector ( 0, 0, 0 ) belongs to U.
(b) Let ๐‘ฟ = (๐‘Ž, ๐‘Ž, ๐‘Ž) and ๐’€ = (๐‘, ๐‘, ๐‘) be vectors in U and let ๐‘˜, ๐‘‘ ๏ƒŽ
.
๐‘˜๐‘ฟ + ๐‘‘๐’€ = (๐‘˜๐‘Ž, ๐‘˜๐‘Ž, ๐‘˜๐‘Ž) + (๐‘‘๐‘, ๐‘‘๐‘, ๐‘‘๐‘)
= (๐‘˜๐‘Ž + ๐‘‘๐‘, ๐‘˜๐‘Ž + ๐‘‘๐‘, ๐‘˜๐‘Ž + ๐‘‘๐‘)
Clearly, ๐‘˜๐‘ฟ + ๐‘‘๐’€
U hence U is a subspace of โ„3 .
Example 4: Let V = โ„3 and let W = { (a, b, c ): a ๏‚ณ 0}. Determine if W is a subspace of V.
Solution:
(a) The zero vector ( 0, 0, 0 ) ๏ƒŽ W.
(b) Let ๐‘ฟ = (๐‘Ž, ๐‘, ๐‘) and ๐’€ = (๐‘Ž′ , ๐‘ ′ , ๐‘ ′ ) be vectors in W.
๐‘‹ + ๐‘Œ = (๐‘Ž + ๐‘Ž′ , ๐‘ + ๐‘ ′ , ๐‘ + ๐‘ ′ )
Since ๐‘Ž ๏‚ณ 0 and ๐‘Ž′ ๏‚ณ 0 then ๐‘Ž + ๐‘Ž′ ๏‚ณ 0. Hence ๐‘ฟ + ๐’€ ๏ƒŽ W.
(c) Let ๐‘‹ = (1, 2, 3) and ๐‘˜ = −2 then
๐‘˜๐‘ฟ = (−2, −4, −6) ๏ƒ W since −2 < 0.
Therefore, W is not a subspace of V.
75
ACTIVITY
I. In problems 1-4, determine whether the given set together with the given operations is a
vector space.
1. The set of ordered pairs (๐‘Ž, ๐‘) of real numbers with the operations
(๐‘Ž, ๐‘) + (๐‘, ๐‘‘) = (๐‘Ž + ๐‘, ๐‘ + ๐‘‘) and
๐‘˜(๐‘Ž, ๐‘) = (๐‘˜๐‘Ž, 0)
2. The set of all ordered triples of real numbers of the form (0,0, ๐‘ง) with the operations
( 0, 0, ๐‘ง ) + ( 0, 0, ๐‘ง′ ) = ( 0, 0, ๐‘ง + ๐‘ง′ ) and
๐‘ ( 0, 0, ๐‘ง ) = ( 0, 0, ๐‘๐‘ง ).
3. The set of polynomials (in ๐‘ฅ) of degree ๏‚ฃ ๐‘› together with the zero polynomial with
positive constant term.
4. The set of all ordered pairs (๐‘Ž, ๐‘) of real numbers with addition and scalar multiplication
in V defined by
(๐‘Ž, ๐‘) + (๐‘, ๐‘‘) = (๐‘Ž๐‘, ๐‘๐‘‘) and ๐‘˜(๐‘Ž, ๐‘) = (๐‘˜๐‘Ž, ๐‘˜๐‘)
5. Which of the following subsets of โ„3 are subspaces of โ„3 ? The set of all vectors of the
form
(a) (๐‘Ž, ๐‘, ๐‘), where ๐‘Ž = ๐‘ = 0
(b) (๐‘Ž, ๐‘, ๐‘),, where ๐‘Ž = −๐‘
(c) (๐‘Ž, ๐‘, ๐‘),, where ๐‘ = 2๐‘Ž + 1
6. Let V be the set of all 2 x 3 matrices under the usual operations of matrix addition and
scalar multiplication. Which of the following subsets of V are subspaces? The set of all
matrices of the form
๏ƒฉa
(a) ๏ƒช
๏ƒซd
b c๏ƒน
, where ๐‘ = ๐‘Ž + ๐‘.
0 0 ๏ƒบ๏ƒป
๏ƒฉa
(b) ๏ƒช
๏ƒซd
b c๏ƒน
, where ๐‘ > 0
0 0 ๏ƒบ๏ƒป
76
3.4 Linear Combinations and Spanning Sets
Definition 3.4.1: A vector ๐‘ฟ ๏ƒŽ V is said to be a linear combination of the set of vectors
{ ๐‘ฟ๐Ÿ , ๐‘ฟ๐Ÿ , โ‹ฏ , ๐‘ฟ๐’Œ } ๏ƒŒ V if there exists scalars ๐‘1 , ๐‘2 , โ‹ฏ , ๐‘๐‘˜ such that
๐‘ฟ = ๐‘1 ๐‘ฟ๐Ÿ + ๐‘2 ๐‘ฟ๐Ÿ + โ‹ฏ + ๐‘๐‘˜ ๐‘ฟ๐’Œ
We also say that ๐‘ฟ is linearly dependent on the ๐‘ฟ๐’Š .
Example 1. In โ„3 , (−15, −4, 0) is a linear combination of (−3, −1, 4) and (3, 2, 8) since
2(−3, −1, 4) − 3(3, 2, 8) = (−15, −4, 0).
−1
Example 2. [
−1
6
7
−1 0 3
1 3 −1
] is a linear combination of [
] and [
] since
15 16
1 5 6
−2 0 −1
−1 0
3[
1 5
3
1 3
] + 2[
6
−2 0
−1
−1 6
]=[
−1
−1 15
7
]
16
Example 3. In โ„3 , let X1 = (4, 2, 3), X2 = (2, 1, 2), and X3 = ( 2, 1, 0). Determine if
(4, 2, 6) is a linear combination of X1, X2, and X3.
๐‘ฟ=
Solution: X is a linear combination of X1, X2, and X3 if we can find ๐‘1 , ๐‘2, and ๐‘3 ๏ƒŽ
that
such
๐‘1(4, 2, 3) + ๐‘2 (2, 1, 2) + ๐‘3 ( 2, 1, 0)
(4, 2, 6)
Multiplying and adding yields the following linear system.
4๐‘1 + 2๐‘2
2๐‘3 = 4
2๐‘1 + ๐‘2 + ๐‘3 = 2
3๐‘1
2๐‘2
= 6
Row reducing the augmented matrix we have
4
2 −2
[2
1
1
−3 −2 0
4
2]
−6
1
4
๐‘…1
1 1/2 −1/2
๐‘…1 [ 2
1
1
−3 −2
0
1
2]
−6
77
−2๐‘…1 + ๐‘…2
3๐‘…1 + ๐‘…3
1
๐‘…2
−2๐‘…3
2
๐‘…2
1 1/2
๐‘…2
0
[0
๐‘…3
0 −1/2
1 1/2
๐‘…2
[0
0
๐‘…3 0
1
๐‘…3
−1/2
2
−3/2
1
0]
−3
−1/2
1
3
1
0]
6
1 1/2 −1/2
[0
1
3
0
0
1
1
6]
0
This implies that ๐‘1 = 2, ๐‘2 = 6, and ๐‘3 = 0. It can be verified that
2(4, 2, 3) + 6(2, 1, 2) + 0( 2, 1, 0) = (4, 2, 6).
Thus X is a linear combination of X1 = (4, 2, 3), X2 = (2, 1, 2), and X3 = ( 2, 1, 0).
SAQ 3-2
In ๐‘ƒ3 , let ๐‘(๐‘ฅ) = ๐‘ฅ 2 − 2๐‘ฅ, ๐‘ž(๐‘ฅ) = 5๐‘ฅ − 2, and
๐‘Ÿ(๐‘ฅ) = ๐‘ฅ 2 + 3. Determine if ๐‘“(๐‘ฅ) = 2๐‘ฅ 2 + 4๐‘ฅ − 7 is a
linear combination of ๐‘(๐‘ฅ), ๐‘ž(๐‘ฅ), and ๐‘Ÿ(๐‘ฅ).
ASAQ 3-2
๐‘“(๐‘ฅ) is a linear combination of ๐‘(๐‘ฅ), ๐‘ž(๐‘ฅ), and ๐‘Ÿ(๐‘ฅ) if we can find ๐‘1 , ๐‘2 , and ๐‘3 ∈ โ„ such
that
๐‘1 (๐‘ฅ 2 − 2๐‘ฅ) + ๐‘2 (5๐‘ฅ − 2) + ๐‘3 (๐‘ฅ 2 + 3) = 2๐‘ฅ 2 + 4๐‘ฅ − 7
Equating the coefficients of similar terms we obtain the system
๐‘1
+ ๐‘3 = 2
−2๐‘1 + 5๐‘2
=4
−2๐‘2 + 3๐‘3 = −7
Row reducing the augmented matrix we obtain (you must do the calculation here)
78
1 0
[0 1
0 0
1
2/5
1
2
8/5]
−1
Thus the solution to the given system is ๐‘1 = 3, ๐‘2 = 2, ๐‘3 = −1. It can be verified that
3(๐‘ฅ 2 − 2๐‘ฅ) + 2(5๐‘ฅ − 2) − (๐‘ฅ 2 + 3) = 2๐‘ฅ 2 + 4๐‘ฅ − 7
Thus ๐‘“(๐‘ฅ) is a linear combination of ๐‘(๐‘ฅ), ๐‘ž(๐‘ฅ), and ๐‘Ÿ(๐‘ฅ).
Definition 3.4.2: Let S = { ๐‘ฟ๐Ÿ , ๐‘ฟ๐Ÿ , โ‹ฏ , ๐‘ฟ๐’ } be a set of vectors in a vector space V. The set S
spans V or V is spanned by S, if every vector in V is a linear combination of the vector in S.
That is for every ๐‘ฟ ∈ V there are scalars ๐‘1 , ๐‘2 , โ‹ฏ , ๐‘๐‘› such that
๐‘ฟ = ๐‘1 ๐‘ฟ๐Ÿ + ๐‘2 ๐‘ฟ๐Ÿ + โ‹ฏ + ๐‘๐‘› ๐‘ฟ๐’
Example 4. The vectors ๐‘ฟ1 = (1, 0) and ๐‘ฟ๐Ÿ = (0, 1) span โ„2 since every vector (๐‘Ž, ๐‘) in โ„2
can be expressed as a linear combination of ๐‘ฟ1 and ๐‘ฟ๐Ÿ . That is
(๐‘Ž, ๐‘) = ๐‘Ž(1, 0) + ๐‘(0, 1) for every ๐‘Ž, ๐‘ ∈ โ„.
Example 5. Every 2 x 2 matrix can be written as a linear combination of the four matrices
1 0
0 1
0 0
0 0
๐ธ1 = [
], ๐ธ2 = [
], ๐ธ3 = [
], and ๐ธ4 = [
]. That is
0 0
0 0
1 0
0 1
๐‘Ž
[
๐‘
1 0
0
๐‘
] = ๐‘Ž[
]+๐‘[
0 0
0
๐‘‘
1
0
]+๐‘[
0
1
0
0 0
]+๐‘‘[
]
0
0 1
for every ๐‘Ž, ๐‘, ๐‘, ๐‘‘ ∈ โ„. Thus ๐ธ1 , ๐ธ2 , ๐ธ3 , and ๐ธ4 span ๐‘€2,2 .
Example 6. Let V be the vector space โ„3 and let S = { X1 = (1, 2, 3), X2 = (−1, −2, 1),
X3 = (0, 1, 0) }. Does S span โ„3 ?
Solution.
Choose an arbitrary vector X = (๐‘Ž, ๐‘, ๐‘) in โ„3 . The set of vectors S = {๐‘ฟ๐Ÿ , ๐‘ฟ๐Ÿ , ๐‘ฟ๐Ÿ‘ } spans โ„3 if
we can find scalars ๐‘1 , ๐‘2 , and ๐‘3 such that
๐‘1(1, 2, 3) + ๐‘2 (−1, −2, 1) + ๐‘3 (0, 1, 0) = (๐‘Ž, ๐‘, ๐‘)
79
Multiplying and adding we obtain the linear system
๐‘1 − ๐‘2 −
= ๐‘Ž
2๐‘1 − 2๐‘2 + ๐‘3 = ๐‘
3๐‘1 + ๐‘2
= ๐‘
The augmented matrix
1 −1 0
[2 −2 1
3 1 0
๐‘Ž
๐‘ ] is row equivalent to
๐‘
1 0 0
[0 1 0
0 0 1
(๐‘Ž + ๐‘)/4
(−3๐‘Ž + ๐‘)/4] (verify)
−2๐‘Ž + ๐‘
The solution to the given system is
๐‘1 =
๐‘Ž+๐‘
4
, ๐‘2 =
−3๐‘Ž+๐‘
4
, and ๐‘3 = −2๐‘Ž + ๐‘ .
Since there exists ๐‘1 , ๐‘2, and ๐‘3 for every choice of ๐‘Ž, ๐‘, and ๐‘, we say that every vector in
โ„3 is a linear combination of X1 = (1, 2, 3), X2 = (−1, −2, 1), and X3 = (0, 1, 0).
For example, if ๐‘ฟ = (2, −1, 6) then๐‘1 = 2, ๐‘2 = 0 and ๐‘3 = −5. It can be verified that
(2, −1, 6) = 2(1, 2, 3) + 0(−1, −2, 1) − 5(0, 1, 0)
Thus S = {๐‘ฟ๐Ÿ , ๐‘ฟ๐Ÿ , ๐‘ฟ๐Ÿ‘ } spans โ„3 .
Example 7: Does ๐‘ฟ = {−1, 4, 2, 2} belongs to span {๐‘ฟ๐Ÿ , ๐‘ฟ๐Ÿ , ๐‘ฟ๐Ÿ‘ } where X1 = (1, 0, 0, 1), X2 =
(1, −1, 0, 0), and X3 = (0, 1, 2, 1).
Solution: Every vector in span {๐‘ฟ๐Ÿ , ๐‘ฟ๐Ÿ , ๐‘ฟ๐Ÿ‘ } is of the form
๐‘1 ๐‘ฟ๐Ÿ + ๐‘2 ๐‘ฟ๐Ÿ + ๐‘3 ๐‘ฟ๐Ÿ‘
Thus ๐‘ฟ belongs to span {๐‘ฟ๐Ÿ , ๐‘ฟ๐Ÿ , ๐‘ฟ๐Ÿ‘ } if we can write it as a linear combination of
๐‘ฟ๐Ÿ , ๐‘ฟ๐Ÿ , and ๐‘ฟ๐Ÿ‘ . Let ๐‘1 , ๐‘2 , and ๐‘3 be scalars such that
๐‘1 (1, 0, 0, 1) + ๐‘2 (1, −1, 0, 0) + ๐‘3 (0, 1, 2, 1) = (−1, 4, 2, 2)
80
Multiplying and adding we obtain the linear system
๐‘1 + ๐‘2
−๐‘2 + ๐‘3
2๐‘3
๐‘1
+ ๐‘3
=
=
=
=
−1
4
2
2
The third equation implies that ๐‘3 = 1. Substituting it to the second and fourth equations,
we get ๐‘1 = 1 and ๐‘2 = −3. However,
๐‘1 + ๐‘2 = 1 + (−3) = −2 ≠ −1
Thus ๐‘ฟ does not belong to span {๐‘ฟ๐Ÿ , ๐‘ฟ๐Ÿ , ๐‘ฟ๐Ÿ‘ }.
SAQ 3-3
Let S = {X1 = (1, 1, 0), X2 = (1, 3, 2), X3 = (4, 9, 5) } be
vectors in โ„3 . Does S span โ„3 .
ASAQ 3-3
Choose an arbitrary vector X = (๐‘Ž, ๐‘, ๐‘) in โ„3 . The set of vectors S = {๐‘ฟ๐Ÿ , ๐‘ฟ๐Ÿ , ๐‘ฟ๐Ÿ‘ } spans โ„3 if
we can find scalars ๐‘1 , ๐‘2 , and ๐‘3 such that
๐‘1 (1, 1, 0) + ๐‘2 (1, 3, 2) + ๐‘3 (4, 9, 5) = (๐‘Ž, ๐‘, ๐‘)
Multiplying and adding we obtain the linear system
๐‘1 + ๐‘2 + 4๐‘3 = ๐‘Ž
๐‘1 + 3๐‘2 + 9๐‘3 = ๐‘
2๐‘2 + 5๐‘3 = ๐‘
The augmented matrix
1 1
[1 3
0 2
4
9
5
๐‘Ž
1 1
๐‘ ] is row equivalent to [0 2
๐‘
0 0
4
5
0
๐‘Ž
−1 + ๐‘ ](verify)
๐‘Ž−๐‘+๐‘
81
The system has no solution if ๐‘Ž − ๐‘ + ๐‘ ≠ 0. Thus S = {๐‘ฟ๐Ÿ , ๐‘ฟ๐Ÿ , ๐‘ฟ๐Ÿ‘ } does not span โ„3 . For
example, the vector (3, 1, 0) cannot be written as a linear combination of ๐‘ฟ๐Ÿ , ๐‘ฟ๐Ÿ , ๐‘ฟ๐Ÿ‘ .
ACTIVITY
1. Determine whether ๐‘“(๐‘ฅ) = 3๐‘ฅ 2 − 3๐‘ฅ + 1 belongs to span {๐‘(๐‘ฅ), ๐‘ž(๐‘ฅ), ๐‘Ÿ(๐‘ฅ)} where
๐‘(๐‘ฅ) = ๐‘ฅ 2 − ๐‘ฅ, ๐‘ž(๐‘ฅ) = ๐‘ฅ 2 − 2๐‘ฅ + 1 and ๐‘Ÿ(๐‘ฅ) = −๐‘ฅ 2 + 1.
2. Let S = {X1 = ( 6, 4, -2, 4 ), X2 = ( 2, 0, 0, 1 ), X3 = ( 3, 2, -1, 2 ), X4 = ( 5, 6, -3, 2 ),
X5 = ( 0, 4, -2, -1 ) }. Does S spans โ„4 ?
82
MODULE 4
LINEAR INDEPENDENCE
Introduction
In this chapter we will discuss linear dependence and independence of a given set of
vectors, the basis and dimension of a vector space V, the rank of a matrix and how it can be
used to determine whether a matrix is singular or nonsingular and whether a homogeneous
system of equations has a trivial or nontrivial solution.
Objectives
1.
2.
3.
4.
5.
Define linear dependence and independence.
Determine whether a given set of vectors is linearly independent or dependent.
Define and explain basis and dimension of a given vector.
Find a basis for a vector space spanned by a given set of vectors.
Recognize the rank of a matrix and used this information to determine whether a
homogeneous system has a nontrivial solution.
4.1 Definition and Examples
Definition 4.1.1: A set of vectors { ๐‘ฟ๐Ÿ , ๐‘ฟ๐Ÿ , โ‹ฏ , ๐‘ฟ๐’Œ } ๏ƒŒ V is said to be linearly dependent if
there exists scalars ๐‘1 , ๐‘2 , โ‹ฏ , ๐‘๐‘˜ not all of which are zeros such that
๐‘1 ๐‘ฟ๐Ÿ + ๐‘2 ๐‘ฟ๐Ÿ + โ‹ฏ + ๐‘๐‘˜ ๐‘ฟ๐’Œ = ๐ŸŽ
Otherwise, the set is said to be linearly independent. Meaning the set of vectors
{๐‘ฟ๐Ÿ , ๐‘ฟ๐Ÿ , โ‹ฏ , ๐‘ฟ๐’Œ } is linearly independent if ๐‘1 ๐‘ฟ๐Ÿ + ๐‘2 ๐‘ฟ๐Ÿ + โ‹ฏ + ๐‘๐‘˜ ๐‘ฟ๐’Œ = ๐ŸŽ is true if and only
if ๐‘1 = ๐‘2 = โ‹ฏ = ๐‘3 = 0.
Example 1. Let S ={X1 = (1, -1), X2 = (1, 1) } be vectors in โ„2 . Determine whether
S = {X1, X2 } is linearly dependent or linearly independent.
Solution: Let ๐‘1 and ๐‘2 be scalars. Next we form the equation
๐‘1(1, -1) + ๐‘2 (1, 1) = (0, 0)
Multiplying and adding we get the homogeneous system
๐‘1 + ๐‘2 = 0
−๐‘1 + ๐‘2 = 0
Since the only solution to this system is ๐‘1 = ๐‘2 = 0 then S = {X1, X2 } is linearly independent.
83
13
Example 2. Let S = {๐‘ฟ๐Ÿ = (1, 2 3), ๐‘ฟ๐Ÿ = (2, −1, 4), ๐‘ฟ๐Ÿ‘ = (3, − 2 , 4)} be vectors in โ„3 .
Determine whether S = {๐‘ฟ๐Ÿ , ๐‘ฟ๐Ÿ , ๐‘ฟ๐Ÿ‘ } is linearly dependent or linearly independent.
Solution: Let ๐‘1 , ๐‘2 , and ๐‘3 be scalars. Next we form the equation
13
๐‘1 (1, 2, 3) + ๐‘2 (2, −1, 4) + ๐‘3 (3, − 2 , 4) = (0, 0, 0)
Multiplying and adding we obtain the homogeneous system
๐‘1 + 2๐‘2 + 3๐‘3 = 0
13
2๐‘1 − ๐‘2 − 2 ๐‘3 = 0
3๐‘1 + 4๐‘2 + 4๐‘3 = 0
Row reducing the augmented matrix we have
1 2
3
[2 −1 −13/2
3 4
4
0 −2๐‘… + ๐‘…
0] −3๐‘… 1 + ๐‘… 2
1
3
0
−15๐‘…2
2๐‘…2 + ๐‘…3
3
๐‘…2 1 2
[0 −5 −25/2
๐‘…3
0 −2
−5
1 2
๐‘…2 [0 1
0 −2
1 2
๐‘…3 [0 1
0 0
3
5/2
−5
3
5/2
0
0
0]
0
0
0]
0
0
0]
0
Since column 3 has no pivot then ๐‘3 is a free variable. This implies that the system has
infinitely many solutions. In particular if we let ๐‘3 = 2 then ๐‘1 = 4 and ๐‘2 = −5. It can be
verified that
4(1, 2, 3) − 5(2, −1, 4) + 2 (3, −
13
, 4) = (0, 0, 0)
2
Hence ๐‘ฟ๐Ÿ , ๐‘ฟ๐Ÿ , and ๐‘ฟ๐Ÿ‘ are linearly dependent.
REMARKS:
1. Any set of vectors that includes the zero vector is linearly dependent.
Proof:
Let S = {๐‘ฟ๐Ÿ , ๐‘ฟ๐Ÿ , โ‹ฏ , ๐‘ฟ๐’ , ๐‘ถ}. Then
0 โˆ™ ๐‘ฟ๐Ÿ + 0 โˆ™ ๐‘ฟ๐Ÿ + โ‹ฏ + 0 โˆ™ ๐‘ฟ๐’ + 2 โˆ™ ๐‘ถ = 0
Hence S is linearly dependent.
84
2. If X ≠ 0, { X } is linearly independent
3. The zero vector is linearly dependent
SAQ 4-1
Determine whether the vectors ๐‘ฟ๐Ÿ = (1, −1, 2), ๐‘ฟ๐Ÿ = (4, 0, 0),
๐‘ฟ๐Ÿ‘ = (−2, 3, 5) and ๐‘ฟ๐Ÿ’ = (7, 1, 2) are linearly dependent or
linearly independent.
ASAQ 4-1
Let ๐‘1 , ๐‘2, ๐‘3 and ๐‘4 be scalars. Then we form the equation
๐‘1 (1, −1, 2) + ๐‘2 (4, 0, 0) + ๐‘3 (−2, 3, 5) + ๐‘4 (7, 1, 2) = (0, 0, 0)
Multiplying and adding we obtain the homogeneous system
๐‘1 + 4๐‘2 − 2๐‘3 + 7๐‘4 = 0
−๐‘1
+ 3๐‘3 + ๐‘4 = 0
2๐‘1
+ 5๐‘3 + 2๐‘4 = 0
1 4 −2 7
The augmented matrix [−1 0 3 1
2 0 5 2
0
0] is row equivalent to
0
1 4 −2
7
0
2
0] (you must do the calculation here)
[0 1 1/4
0 0
1
4/11 0
The last equation implies that ๐‘4 is a free variable. Hence
๐‘1 = −4๐‘2 + 2๐‘3 − 7๐‘4
1
๐‘2 = −4๐‘3 − 2๐‘4
4
๐‘3 = −11๐‘4
85
If we let ๐‘4 = 11 then ๐‘1 = −1, ๐‘2 = −21, and ๐‘3 = −4. It can be verified that
−1(1, −1, 2) − 21(4, 0, 0) − 4(−2, 3, 5) + 11(7, 1, 2) = (0, 0, 0)
Hence the given vectors are linearly dependent.
A very important theorem about linear dependence or independence of vectors can
now be stated.
Theorem 4.1.1. A set of ๐‘› vectors in โ„๐‘š is always linearly dependent if ๐‘› > ๐‘š.
Corollary 4.1.1. A set of linearly independent vectors in โ„๐‘› contains at most ๐‘› vectors.
Corollary 4.1.1 can be interpreted in this way: if we have ๐‘› linearly independent
vectors, then adding more vectors will make the set of vectors linearly dependent. For
example, the set of vectors {๐‘ฟ๐Ÿ = (1, -1), ๐‘ฟ๐Ÿ = (1, 1)} of Example 1 is linearly independent.
Then by Corollary 4.1.1, the set of vectors {๐‘ฟ๐Ÿ = (1, -1), ๐‘ฟ๐Ÿ = (1, 1), ๐‘ฟ๐Ÿ‘ = (2, 3)} is linearly
dependent.
Theorem 4.1.2: Let S = { ๐‘ฟ๐Ÿ , ๐‘ฟ๐Ÿ , โ‹ฏ , ๐‘ฟ๐’Œ } be a set of nonzero vectors in a vector space V.
Then S is linearly dependent if and only if one of the vectors ๐‘ฟ๐’‹ is a linear combination of
the preceding vectors in S.
Example 3. Let S = {X1 = (1, 1, 0), X2 = (0, 2, 3), X3 = (1, 2, 3), X4 = (3, 6, 6) } be a set of vectors
in โ„3 . Determine if V is linearly dependent. If linearly dependent, express one of the vectors
as a linear combination of the rest.
Solution: By Corollary 4.1.1, we know that S is linearly dependent. Now we form the
equation
๐‘1(1, 1, 0) + ๐‘2 (0, 2, 3) + ๐‘3 (1, 2, 3) + ๐‘4 (3, 6, 6) = (0, 0, 0)
Multiplying and adding we get the homogeneous system
๐‘1
+ ๐‘3 + 3๐‘4 = 0
๐‘1 + 2๐‘2 + 2๐‘3 + 6๐‘4 = 0
3๐‘2 + 3๐‘3 + 6๐‘4 = 0
Row reducing the augmented matrix we obtain
86
1 0 1
[0 1 1
0 0 1
3 0
2 0]
1 0
This implies that ๐‘4 is a free variable. If we let ๐‘4 = ๐‘Ÿ ∈ โ„ and ๐‘Ÿ ≠ 0 then
๐‘3 = −๐‘4
= −๐‘Ÿ
๐‘2 = −๐‘3 − 2๐‘4
= ๐‘Ÿ − 2๐‘Ÿ
= −๐‘Ÿ
๐‘1 = −๐‘3 − 3๐‘4
= ๐‘Ÿ − 3๐‘Ÿ
= −2๐‘Ÿ
Thus the given vectors are linearly dependent and it can be verified that
−2๐‘Ÿ(1, 1, 0) − ๐‘Ÿ(0, 2, 3) − ๐‘Ÿ(1, 2, 3) + ๐‘Ÿ(3, 6, 6) = (0, 0, 0)
By Theorem 4.1.2, we can write one of the vectors as a linear combination of the rest.
Transposing the first three terms to the right side we have
๐‘Ÿ(3, 6, 6) = 2๐‘Ÿ(1, 1, 0) + ๐‘Ÿ(0, 2, 3) + ๐‘Ÿ(1, 2, 3)
(3, 6, 6) = 2(1, 1, 0) + (0, 2, 3) + (1, 2, 3)
(dividing both sides by ๐‘Ÿ)
Clearly, (3, 6, 6) is a linear combination of the preceding vectors.
NOTE: Any vector in the given set can be expressed as a linear combination of the other
vectors.
4.2 Basis and Dimension
Definition 4.2.1: A set of vectors S = {๐‘ฟ๐Ÿ , ๐‘ฟ๐Ÿ , โ‹ฏ , ๐‘ฟ๐’ } in a vector space V is called a basis for
V if S spans V and S is linearly independent.
Example 1. Show that the set S = {X1 = (3, 2, 2), X2 = (-1, 2, 1), X3 = (0, 1, 0) } is a basis for โ„3 .
Solution:
First we have to show that S is linearly independent. Forming the equation
๐‘1(3, 2, 2) + ๐‘2 (−1, 2, 1)+ ๐‘3 (0, 1, 0) = (0, 0, 0)
we obtain the linear system
3๐‘1 − ๐‘2
= 0
2๐‘1 + 2๐‘2 + ๐‘3 = 0
2๐‘1 + ๐‘2
= 0
87
Solving the given system we obtain only the zero solution, ๐‘1 = ๐‘2 = ๐‘3 = 0. This implies
that S is linearly independent. Next we have to show that S spans โ„3 . To show that S spans
โ„3 , let ๐‘ฟ = (๐‘Ž, ๐‘, ๐‘) be any vector in โ„3 . Then we form the equation
๐‘1(3, 2, 2) + ๐‘2 (−1, 2, 1)+ ๐‘3 (0, 1, 0) = (๐‘Ž, ๐‘, ๐‘)
Multiplying and adding we obtain the linear system
3๐‘1 − ๐‘2
= ๐‘Ž
2๐‘1 + 2๐‘2 + ๐‘3 = ๐‘
2๐‘1 + ๐‘2
= ๐‘
a๏€ซc
3c ๏€ญ 2a
2a ๏€ซ 5b ๏€ญ 8c
, c2 ๏€ฝ
, and c3 ๏€ฝ
. Since the above
5
5
5
system is consistent for any choice of ๐‘Ž, ๐‘, and ๐‘, then S spans โ„3 .
Solving the system we get c1 ๏€ฝ
Thus S = { (3, 2, 2), (−1, 2, 1), (0, 1, 0) } } is a basis for โ„3 .
A basis for a vector space V is not unique. In fact it can be easily verified that the set
of vectors {(1, 0, 0), (0, 1, 0), (0, 0, 1)} is also a basis for โ„3 . This is called the natural basis
for โ„3 . Likewise the set of vectors {(1, 0, 0, 0), (0, 1, 0, 0), (0, 0, 1, 0), (0, 0, 0, 1)} forms a
natural basis for โ„4 . In general, the vectors ๐‘‹1 = (1, 0, โ‹ฏ , 0), ๐‘‹2 = (0, 1, โ‹ฏ , 0), โ‹ฏ , ๐‘‹๐‘› =
(0, 0, โ‹ฏ , 1) constitute a basis for โ„๐‘› .
Example 2. The monomials {๐‘ฅ 3 , ๐‘ฅ 2 , ๐‘ฅ, 1} is a natural basis for ๐‘ƒ3 . In general the monomials
{๐‘ฅ ๐‘› , ๐‘ฅ ๐‘›−1 , โ‹ฏ , ๐‘ฅ, 1} constitute a basis for ๐‘ƒ๐‘› .
Example 3. The 2 x 2 matrices {[
1 0 0
],[
0 0 0
1 0 0 0 0
],[
],[
]} is a natural basis for ๐‘€22 .
0 1 0 0 1
SAQ 4-2
Determine whether A = {๐‘ฅ 2 − 1, ๐‘ฅ 2 − 2, ๐‘ฅ 2 − 3} is a basis
for ๐‘ƒ2 .
88
ASAQ 4-2
First we have to determine if A is linearly independent. Let ๐‘1 , ๐‘2, and ๐‘3 be scalars. Then
we form the equation
๐‘1 (๐‘ฅ 2 − 1) + ๐‘2 (๐‘ฅ 2 − 2) + ๐‘3 (๐‘ฅ 2 − 3) = 0
Equating the coefficients of similar terms we obtain the system
๐‘1 + ๐‘2 + ๐‘3 = 0
−๐‘1 − 2 ๐‘2 − 3๐‘3 = 0
The augmented matrix in reduced row echelon form is (verify)
[
1 0
0 1
−1
2
0
]
0
This implies that A is linearly dependent. Thus A is not a basis for ๐‘ƒ2 .
Theorem 4.2.1: If S = { ๐‘ฟ๐Ÿ , ๐‘ฟ๐Ÿ , โ‹ฏ , ๐‘ฟ๐’ } is a basis for a vector space V, then every vector in V
can be written in one and only one way as a linear combination of the vectors in S.
Definition 4.2.2: If the vector space V has a finite basis, then the dimension of V is the
number of vectors in every basis and V is called a finite dimensional vector space.
Otherwise V is called an infinite dimensional vector space. If V = {0}, then V is said to be
zero dimensional. We often write dimV for the dimension of V.
Example 4. Since a basis for โ„3 consists of 3 linearly independent vectors then dimโ„3 = 3.
Likewise dimโ„4 = 4, dimโ„5 = 5 and so on. In general dimโ„๐‘› = ๐‘› since ๐‘› linearly
independent vectors constitute a basis for โ„๐‘› .
Example 5. By Example 2, we see that dim๐‘ƒ3 = 4 and in general the dim๐‘ƒ๐‘› = ๐‘› + 1.
Theorem 4.2.2: Let S = { ๐‘ฟ๐Ÿ , ๐‘ฟ๐Ÿ , โ‹ฏ , ๐‘ฟ๐’ } be a set of nonzero vectors in a vector space V and
let W = spans S. Then some subset of S is a basis for W.
Theorem 4.2.2 means that if W consists of all vectors that can be expressed as a
linear combination of the vectors in S (W=spanS) then some subset of S forms a basis for W.
Thus if S is linearly independent then { ๐‘ฟ๐Ÿ , ๐‘ฟ๐Ÿ , โ‹ฏ , ๐‘ฟ๐’ } is a basis for W. If S is linearly
independent there exists a proper subset of S which forms a basis for W. Let us consider the
next example:
89
Example 6: Let S = {X1 = (1, 2, 2), X2 = (3, 2, 1), X3 = (11, 10, 7), X4 = (7, 6, 4) }. Find a basis for
the subspace W = span S of โ„3 .
Solution: W is a set of all vectors in โ„3 which can be expressed as a linear combination of
the vectors in S. By Theorem 4.1.1 we know that S is linearly dependent. Hence S is not a
basis for W but by Theorem 4.2.2 S contains a proper subset of linearly independent vectors
which forms a basis for W. Next we form the equation
๐‘1(1, 2, 2) + ๐‘2 (3, 2, 1) + ๐‘3 (11, 10, 7) + ๐‘4 (7, 6, 4) = (0, 0, 0)
Equating the corresponding components, we obtain the homogeneous system
๐‘1 + 3๐‘2 + 11๐‘3 + 7๐‘4 = 0
2๐‘1 + 2๐‘2 + 10๐‘3 + 6๐‘4 = 0
2๐‘1 + ๐‘2 + 7๐‘3 + 4๐‘4 = 0
The augmented matrix in reduced row echelon form is (verify)
1 0
[0 1
0 0
2 1 0
3 2 0]
0 0 0
The leading 1s appear in columns 1 and 2, thus {X1, X2} is a basis for W = span S. It means
that {X1, X2} is the smallest set possible that could span W. By Theorem 4.1.2, we can write
๐‘ฟ๐Ÿ‘ and ๐‘ฟ๐Ÿ’ as a linear combination of the two vectors X1 and X2. Thus ๐‘ฟ๐Ÿ‘ and ๐‘ฟ๐Ÿ’ can be
discarded and the remaining vectors still span W. That is
W = span{ ๐‘ฟ๐Ÿ , ๐‘ฟ๐Ÿ , ๐‘ฟ๐Ÿ‘ , ๐‘ฟ๐Ÿ’ } = span{X1, X2}.
Theorem 4.2.3: Suppose that dimV = ๐‘›. If T = { Y1, Y2, …, Yr } is a linearly independent set of
vectors in V, then ๐‘Ÿ ≤ ๐‘›.
Proof: Let ๐‘ฟ๐Ÿ , ๐‘ฟ๐Ÿ , โ‹ฏ , ๐‘ฟ๐’ be a basis for V. If ๐‘Ÿ > ๐‘› then we can find scalars ๐‘1 , ๐‘2 , โ‹ฏ , ๐‘๐‘Ÿ not
all zero such that
๐‘1 ๐’€๐Ÿ + ๐‘2 ๐’€๐Ÿ + โ‹ฏ + ๐‘๐‘Ÿ ๐’€๐’“ = ๐‘ถ
is satisfied. This will contradict the linear independence of the ๐’€๐’Š ′s. Thus ๐‘Ÿ ≤ ๐‘›.
Corollary 4.2.1: If S = { ๐‘ฟ๐Ÿ , ๐‘ฟ๐Ÿ , โ‹ฏ , ๐‘ฟ๐’ } and T = { Y1, Y2, …, Ym } are bases for a vector space,
then ๐‘› = ๐‘š. (If a vector space has one basis with a finite number of elements, then all other
bases are finite and have the same number of elements)
90
Theorem 4.2.4: If S is linearly independent set of vectors in a finite-dimensional vector
space V, then there is a basis T for V, which contains S.
Example 7: In V = โ„4 , the set A = {(1, 1, 0, 0), (0, 0, 1, 1), (1, 0, 1, 0), (0, 1, 0, −1)} is a basis,
and B = { (1, 2, −1, 1), (0, 1, 2, −1) } is linearly independent. Extend B into a basis for V using
A.
Solution:
A ๏ƒˆ B = {(1, 2, −1, 1), (0, 1, 2, −1), (1, 1, 0, 0), (0, 0, 1, 1), (1, 0, 1, 0), (0, 1, 0, −1) }
Delete vectors in A which are linear combination of the preceding vectors.
a. Try (1, 1, 0, 0)
๐‘1(1, 2, −1, 1) + ๐‘2 (0, 1, 2, −1) = (1, 1, 0, 0)
Then
๐‘1
=1
2๐‘1 + ๐‘2 = 1
−๐‘1 + 2๐‘2 = 0
๐‘1 – ๐‘2 = 0
If ๐‘1 = 1 then ๐‘2 = 1 but 2๐‘1 + ๐‘2 = 3 ๏‚น 1. The system of equations has no solution so
(1, 1, 0, 0) is not a linear combination of the preceding vectors. Thus we retain (1, 1, 0, 0).
b. Try (0, 0, 1, 1)
๐‘1 (1, 2, -1, 1) + ๐‘2 (0, 1, 2, -1) = (0, 0, 1, 1)
Then
๐‘1
=0
2๐‘1 + ๐‘2 = 0
−๐‘1 + 2๐‘2 = 1
๐‘1 – ๐‘2 = 1
If ๐‘1 = 0 then ๐‘2 = 0 but –๐‘1 + 2๐‘2 = 0 ๏‚น 1. The system has no solution so (0, 0, 1, 1) is not a
linear combination of the preceding vectors. Thus we retain (0, 0, 1, 1).
Hence, {(1, 2, −1, 1), (0, 1, 2, −1), (1, 1, 0, 0), (0, 0, 1, 1)} is a basis for V.
Theorem 4.2.5: Let V be an ๐‘›-dimensional vector space, and let S = { ๐‘ฟ๐Ÿ , ๐‘ฟ๐Ÿ , โ‹ฏ , ๐‘ฟ๐’ } be a
set of ๐‘› vectors in V.
a. If S is linearly independent, then it is a basis for V.
b. If S spans V, then it is a basis for V.
91
From our previous lesson, we see that the set of all solutions to the homogeneous
system AX = 0, where A is ๐‘š ๐‘ฅ ๐‘› , is a subspace of โ„๐‘› . To find a basis for this solution space
we consider the following example.
Example 8: Find a basis for the solution space W of
๏ƒฉ x1 ๏ƒน
๏ƒฉ1 2 1 2 1 ๏ƒน ๏ƒช ๏ƒบ ๏ƒฉ0 ๏ƒน
๏ƒช1 2 2 1 2 ๏ƒบ ๏ƒช x2 ๏ƒบ ๏ƒช0 ๏ƒบ
๏ƒช
๏ƒบ ๏ƒชx ๏ƒบ ๏€ฝ ๏ƒช ๏ƒบ
๏ƒช 2 4 3 3 3 ๏ƒบ ๏ƒช 3 ๏ƒบ ๏ƒช0 ๏ƒบ
๏ƒช
๏ƒบ ๏ƒช x4 ๏ƒบ ๏ƒช ๏ƒบ
๏ƒซ 0 0 1 ๏€ญ1 ๏€ญ1๏ƒป ๏ƒช ๏ƒบ ๏ƒซ0 ๏ƒป
๏ƒซ x5 ๏ƒป
Solution:
Transform the augmented matrix to reduced row-echelon form using the Gauss-Jordan
reduction method. The augmented matrix in reduced row-echelon form is (verify)
1
0
[
0
0
2
0
0
0
0 3 0
1 −1 0
0 0 1
0 0 0
0
0
]
0
0
The solution is ๐‘1 = −2๐‘  – 3๐‘ก, ๐‘2 = ๐‘ , ๐‘3 = ๐‘ก, ๐‘4 = ๐‘ก, and ๐‘5 = 0 where ๐‘  and ๐‘ก are any real
numbers. Thus every solution is of the form
๏ƒฉ ๏€ญ2 s ๏€ญ 3t ๏ƒน
๏ƒช s ๏ƒบ
๏ƒช
๏ƒบ
๏ƒบ , where ๐‘  and ๐‘ก are real numbers.
X ๏€ฝ๏ƒช t
๏ƒช
๏ƒบ
๏ƒช t
๏ƒบ
๏ƒช๏ƒซ 0 ๏ƒบ๏ƒป
Since W is a solution space then every vector in W can be written in the form of
๏ƒฉ ๏€ญ2 ๏ƒน
๏ƒฉ ๏€ญ3 ๏ƒน
๏ƒช1๏ƒบ
๏ƒช0๏ƒบ
๏ƒช ๏ƒบ
๏ƒช ๏ƒบ
X ๏€ฝ s๏ƒช 0 ๏ƒบ + t๏ƒช 1 ๏ƒบ
๏ƒช ๏ƒบ
๏ƒช ๏ƒบ
๏ƒช0๏ƒบ
๏ƒช1 ๏ƒบ
๏ƒช๏ƒซ 0 ๏ƒบ๏ƒป
๏ƒช๏ƒซ 0 ๏ƒบ๏ƒป
Since ๐‘  and ๐‘ก can take on any values, we first let ๐‘  = 1, ๐‘ก = 0 and let ๐‘  = 0, ๐‘ก = 1, obtaining as
solutions
92
๏ƒฉ ๏€ญ2 ๏ƒน
๏ƒช1๏ƒบ
๏ƒช ๏ƒบ
X1 ๏€ฝ ๏ƒช 0 ๏ƒบ
๏ƒช ๏ƒบ
๏ƒช0๏ƒบ
๏ƒช๏ƒซ 0 ๏ƒบ๏ƒป
๏ƒฉ ๏€ญ3 ๏ƒน
๏ƒช0๏ƒบ
๏ƒช ๏ƒบ
X2 ๏€ฝ ๏ƒช 1 ๏ƒบ
๏ƒช ๏ƒบ
๏ƒช1 ๏ƒบ
๏ƒช๏ƒซ 0 ๏ƒบ๏ƒป
and
Thus S = {X1, X2} belongs to W. Since any vector in W can be written as a linear combination
of the vectors in S then S spans W.
We have to show that S = {X1, X2} is linearly independent. We form the equation
๐‘1(−2, 1, 0, 0, 0 ) + ๐‘2 (−3, 0, 1, 1, 0 ) = (0, 0, 0, 0, 0)
Then
−2๐‘1 − 3๐‘2 = 0
๐‘1
= 0
๐‘2 = 0
The only solution to the above system is ๐‘1 = ๐‘2 = 0, hence S is linearly independent and is
a basis for W. Thus the dimension of W is 2.
SAQ 4-3
Find a basis for the solution space W of the homogeneous
system
๐‘ฅ1 + 2๐‘ฅ2 + 2๐‘ฅ3 − ๐‘ฅ4 + ๐‘ฅ5 = 0
2๐‘ฅ2 + 2๐‘ฅ3 − 2๐‘ฅ4 − ๐‘ฅ5 = 0
2๐‘ฅ1 + 6๐‘ฅ2 + 2๐‘ฅ3 − 4๐‘ฅ4 + ๐‘ฅ5 = 0
๐‘ฅ1 + 4๐‘ฅ2
− 3๐‘ฅ4
=0
What is the dimension of W?
ASAQ 4-3
1
0
The augmented matrix [
2
1
2
2
6
4
2
2
2
0
−1 1 0
−2 −1 0
] is row equivalent (verify) to
−4 1 0
−3 0 0
93
1
0
[
0
0
0
1
0
0
0 1
2
0 −1 −1/2
1 0
0
0 0
0
0
0
] which is in reduced row echelon form.
0
0
Since columns 4 and 5 have no pivots then ๐‘ฅ4 and ๐‘ฅ5 are free variables. If we let ๐‘ฅ4 = ๐‘  and
๐‘ฅ5 = ๐‘ก where ๐‘  and ๐‘ก are real numbers then
๐‘ฅ1 = −๐‘  − 2๐‘ก
1
๐‘ฅ2 = ๐‘  + 2 ๐‘ก
๐‘ฅ3 = 0
๐‘ฅ4 = ๐‘ 
๐‘ฅ5 = ๐‘ก
where ๐‘  and ๐‘ก are real numbers.
Thus all solutions are of the form
−๐‘  − 2๐‘ก
1
๐‘  +2๐‘ก
๐‘‹=
0
๐‘ 
[ ๐‘ก
]
−2
−1
1/2
1
=๐‘  0 +๐‘ก 0
1
0
[0]
[ 1 ]
Since ๐‘  and ๐‘ก can be any real number, we first let ๐‘  = 1, ๐‘ก = 0 and then let ๐‘  = 0, ๐‘ก = 2 to
obtain the solution
−1
−4
1
1
๐‘‹1 = 0 and ๐‘‹2 = 0
1
0
[2]
[0]
which span W. It can be easily verified that ๐‘‹1 and ๐‘‹2 are linearly independent because one
vector is not a multiple of the other. Thus ๐‘‹1 and ๐‘‹2 form a basis for W and dimW = 2. Note
that you may obtain a different basis since ๐‘  and ๐‘ก can take on any values.
94
ACTIVITY
1. Let
a) X1 = ( 4, 2, 1 ), X2 = ( 2, 6, -5 ), X3 = ( 1, -2, 3 )
b) X1 = ( 1, 2, 3 ), X2 = ( 1, 1, 1 ), X3 = ( 1, 0, 1 )
c) X1 = ( 1, 1, 0 ), X2 = ( 0, 2, 3 ), X3 = ( 1, 2, 3 ), X4 = ( 3, 6, 6 )
Which of the given set of vectors in โ„3 is linearly dependent? For those that are, express
one vector as a linear combination of the rest.
3. Determine whether the set
๏ƒฌ ๏ƒฉ1
S ๏€ฝ ๏ƒญ๏ƒช
๏ƒฎ ๏ƒซ0
1๏ƒน ๏ƒฉ0 0๏ƒน ๏ƒฉ1 0๏ƒน ๏ƒฉ0 1๏ƒน ๏ƒผ
,
,
,
๏ƒฝ
0๏ƒบ๏ƒป ๏ƒช๏ƒซ1 1 ๏ƒบ๏ƒป ๏ƒช๏ƒซ0 1 ๏ƒบ๏ƒป ๏ƒช๏ƒซ1 1๏ƒบ๏ƒป ๏ƒพ
is a basis for the vector space V of all 2 x 2 matrices.
4. Find a basis of โ„4 containing the vector (1, 2, 3, 4).
4.3 The Rank of a Matrix
Definition 4.3.1: Let
๏ƒฉ a11 a12 ... a1n ๏ƒน
๏ƒชa
a22 ... a2 n ๏ƒบ๏ƒบ
21
๏ƒช
A๏€ฝ
๏ƒช
๏ƒบ
๏ƒช
๏ƒบ
๏ƒซ๏ƒช am1 am 2 ... amn ๏ƒบ๏ƒป
be an ๐‘š x ๐‘› matrix. The rows of A
X 1 ๏€ฝ (a11 , a12 ,..., a1n )
X 2 ๏€ฝ (a 21 , a 22 ,..., a 2 n )
โ‹ฎ
X m ๏€ฝ (a m1 , a m 2 ,..., a mn )
considered as vectors in โ„๐‘› , span a subspace of โ„๐‘› , called the row space of A. Similarly,
the columns of A,
๏ƒฉ a11 ๏ƒน
๏ƒฉ a12 ๏ƒน
๏ƒฉ a1n ๏ƒน
๏ƒชa ๏ƒบ
๏ƒชa ๏ƒบ
๏ƒชa ๏ƒบ
21 ๏ƒบ
22 ๏ƒบ
2n
๏ƒช
๏ƒช
Y1 ๏€ฝ
, Y2 ๏€ฝ
, … , Yn ๏€ฝ ๏ƒช ๏ƒบ
๏ƒช ๏ƒบ
๏ƒช ๏ƒบ
๏ƒช ๏ƒบ
๏ƒช ๏ƒบ
๏ƒช ๏ƒบ
๏ƒช ๏ƒบ
๏ƒช๏ƒซ am1 ๏ƒบ๏ƒป
๏ƒช๏ƒซ am 2 ๏ƒบ๏ƒป
๏ƒช๏ƒซ amn ๏ƒบ๏ƒป
95
considered as vectors in โ„๐‘š , span a subspace of โ„๐‘š called the column space of A.
๏ƒฉ1 2 0 1 ๏ƒน
Example 1. Let A ๏€ฝ ๏ƒช๏ƒช 2 6 ๏€ญ3 ๏€ญ2 ๏ƒบ๏ƒบ . The rows of A,
๏ƒช๏ƒซ3 10 ๏€ญ6 ๏€ญ5 ๏ƒบ๏ƒป
X1 = (1, 2, 0, 1), X2 = (2, 6, −3, −2), and X3 = (3, 10, −6, −5) are vectors in โ„4 , and these
vectors span a subspace of โ„4 called the row space of A. That is
Row space of A = span { ๐‘‹1 , ๐‘‹2 , ๐‘‹3 }
Similarly, the columns of A, Y1 = (1, 2, 3), Y2 = (2, 6, 10), Y3 = (0, −3, −6) and Y4 = (1, −2, −5)
are vectors in โ„3 , and these vectors span a subspace of โ„3 called the column space of A.
That is
Column space of A = span { ๐‘Œ1 , ๐‘Œ2 , ๐‘Œ3 , ๐‘Œ4 }
Theorem 4.3.1: If A and B are two ๐‘š x ๐‘› row equivalent matrices, then the row spaces of A
and B are equal.
Example 2: Let S = {๐‘ฟ๐Ÿ , ๐‘ฟ๐Ÿ , ๐‘ฟ๐Ÿ‘ , ๐‘ฟ๐Ÿ’ } where ๐‘ฟ๐Ÿ = (1, 2, −1), ๐‘ฟ๐Ÿ = (6, 3, 0), ๐‘ฟ๐Ÿ‘ = (4, −1, 2),
and ๐‘ฟ๐Ÿ’ = (2, −5, 4). Find a basis for the subspace V = spanS of โ„3 .
Solution: V is the row space of the matrix A whose rows are the given vectors
1 2 −1
6 3
0
A=[
]
4 −1 2
2 −5 4
Applying the Gauss-Jordan elimination we obtain the matrix B
1
0
๐ต=[
0
0
0
1
0
0
1/3
−2/3
]
0
0
which is row equivalent to A. By Theorem 4.3.1, the row spaces of A and B are equal. Since
the nonzero rows of B are linearly independent then
(1, 0, 1/3) and (0, 1, −2/3)
96
form a basis for V. It means that all vectors in V can be expressed as a linear combination of
these two vectors. Note that the basis for V is not a subset of S. However, expressing any
vector in V as a linear combination of the basis obtained by this procedure is very simple.
For example, the vectors (2, −5, 4) and (−1, 4, −3) are in V. Since the leading 1s appear in
columns 1 and 2 then
(2, −5, 4) = 2(1, 0, 1/3) – 5(0, 1, −2/3) and
(−1, 4, −3) = −1(1, 0, 1/3) + 4(0, 1, −2/3)
Also note that the dimV = 2 ≠ 3 hence V ≠ โ„3 . Since V ≠ โ„3 then not all vectors in โ„3 can
be expressed as a linear combination of (1, 0, 1/3) and (0, 1, −2/3). It is very easy to see
from our example above that all vectors of the form (2, −5, ๐‘), ๐‘ ≠ 4, are not in V.
Definition 4.3.2: The dimension of the row space of A is called the row rank of A, and the
dimension of the column space of A is called the column rank of A.
1 2
Example 3. Let ๐ด = [2 6
3 10
0
1
−3 −3]. Find the row and column ranks of A.
−6 −7
Solution:
1 0
A is row equivalent (verify) to ๐ต = [0 1
0 0
3
6
−3/2 −5/2] which is in reduced row
0
0
echelon form. The vectors (1, 0, 3, 6) and (0, 1, −3/2, −5/2) form a basis for the row space
of A. Thus the row rank of A is 2.
To find the column rank of A, we form the matrix
1
2
๐ด๐‘‡ = [
0
1
2
3
6 10
]
−3 −6
−3 −7
which is row equivalent (verify) to the matrix
1
0
๐ถ=[
0
0
0 −1
1 2
]
0 0
0 0
97
The vectors (1, 0 −1) and (0, 1, 2) form a basis for the row space of ๐ด๐‘‡ . Thus
1
0
[ 0 ] and [1]
−1
2
form a basis for the column space of A. Hence the column rank of A is 2. Note that the row
rank and column rank of a matrix are equal. This is stated in the next theorem.
Theorem 4.3.2. The row and column ranks of the (๐‘š x ๐‘›) matrix A = ๏ƒฉ๏ƒซ aij ๏ƒน๏ƒป are equal.
The next example will show us how we can apply the method used in Example 3 in
finding a basis for a subspace when the vectors are given in column form.
1 0 2 3
5
2 2 1 2
0
Example 4: Let S = {[ ] , [ ] , [ ] , [ ] , [ ]} . Find a basis for the subspace V = spanS of
1 1 3 1
0
1 2 1 4 −1
โ„4 .
1
2
Solution: Let ๐ด = [
1
1
0
2
1
2
1
0
๐ด๐‘‡ = 2
3
[5
2
1
3
1
2
2
1
2
0
3 5
2 0
]. Next we form the matrix
1 0
4 −1
1 1
1 2
3 1
1 4
0 −1]
Applying elementary row operations we obtain (verify) the matrix
1
0
๐ต๐‘‡ = 0
0
[0
0
1
0
0
0
0
0
1
0
0
−1
3/5
4/5
0
0 ]
The basis for V are the nonzero rows of ๐ต ๐‘‡ written as columns. Thus
98
0
0
1
1
0
0
[ ], [
], and [
]
0
1
0
3/5
4/5
−1
form a basis for V.
Theorem 4.3.3. An ๐‘› x ๐‘› matrix is nonsingular if and only if rank A = ๐‘›.
1
Example 5. Find the rank of ๐ด = [1
1
2 0
1 −3].
3 3
Solution:
Transforming A to reduced row echelon form we obtain (verify)
1 0
[0 1
0 0
−6
3]
0
Since rank of A = 2 < 3 then A is singular. We know from our previous lesson that a matrix is
singular if and only if |๐ด| = 0. It can be verified that
1 2
|1 1
1 3
0
−3| = 0
3
This result is stated in the following corollary and can be used in determining the
rank of an ๐‘› x ๐‘› matrix.
Corollary 4.3.1. If A is an ๐‘› x ๐‘› matrix, then rank A = ๐‘› if and only if A ๏‚น 0.
The next corollary gives another method of testing whether the homogeneous
system ๐ด๐‘‹ = 0 has a trivial or nontrivial solution.
Corollary 4.3.2. The homogeneous system AX = 0 of ๐‘› linear equations in ๐‘› unknowns has a
nontrivial solution if and only if rank A < ๐‘›.
99
Example 6: The 3 x 3 matrix A of Example 5 is singular. Hence the homogeneous system
๐‘ฅ1 + 2๐‘ฅ2
= 0
๐‘ฅ1 + ๐‘ฅ2 − 3๐‘ฅ3 = 0
๐‘ฅ1 + 3๐‘ฅ2 + 3๐‘ฅ3 = 0
has a nontrivial solution.
Corollary 4.3.3. Let S = {๐‘ฟ๐Ÿ , ๐‘ฟ๐Ÿ , โ‹ฏ , ๐‘ฟ๐’ } be a set of vectors in โ„๐‘› and let A be the matrix
whose rows (columns) are the vectors in S. Then S is linearly independent if and only if
|๐ด| ≠ 0.
Example 7: The vectors (1, −2, 3), (2, 4, 7) and (0, −1, 5) are linearly independent because
1
[2
0
−2 3
4 7] = 41 ≠ 0
−1 5
By Theorem 4.2.5, (1, −2, 3), (2, 4, 7) and (0, −1, 5) form a basis for โ„3 .
ACTIVITY
1. Find a basis for the row space and column space of the following matrices:
1 2 3
a. [4 5 6]
7 8 9
2 1 3
b. [2 −1 5
1 1 1
−2
2]
1
2. Let V = P1 and let A = { x + 1, x – 1, 2x + 3 }
a. Show that S spans P1.
b. Find a subset of A which is a basis for P1.
3. Consider the following subset of the vector space of all real-valued functions
S = { cos2t, sin2t, cos2t }
Find a basis for the subspace W = spanS. What is the dimension of W?
100
MODULE 5
LINEAR TRANSFORMATIONS AND MATRICES
Introduction
In this chapter we will discuss a function mapping one vector space to another
vector space, one-to-one and onto linear transformations, the kernel of a linear
transformation and how it can be used to determine if the linear transformation is one-toone, the range of a linear transformation and how it can be used to determine whether the
linear transformation is onto and the matrix of a linear transformation.
Objectives
After going through this chapter, you are expected to be able to do the following:
1. Define and explain linear transformation or linear mapping.
2. Determine whether a function mapping one vector space to another is a linear
transformation.
3. Differentiate between one-to-one linear transformation and onto linear transformation.
4. Define the kernel and range of a linear transformation.
5. Using kernel, distinguish between one-to-one linear transformation and onto linear
transformation.
6. Find the matrix of a linear transformation with respect to some bases.
5.1 Linear Transformations (Linear Mappings)
Definition 5.1.1: Let V and W be vector spaces. A linear transformation L of V into W is a
function assigning a unique vector L(X) in W to each X in V such that:
a. L ( X + Y ) = L (X) + L(Y), for every vector X and Y in V,
b. L ( ๐‘X ) = ๐‘L(X), for every vector X in V and every scalar ๐‘.
Example 1: Let L: โ„3 → โ„3 be defined by
L(x, y, z) = (x, y, 0)
Determine whether L is a linear transformation or not.
Solution: Let X = (๐‘ฅ1 , ๐‘ฆ1 , ๐‘ง1 ) and Y = (๐‘ฅ2 , ๐‘ฆ2 , ๐‘ง2 ) ๏ƒŽ โ„3 then
๐‘‹ + ๐‘Œ = (๐‘ฅ1 + ๐‘ฅ2 , ๐‘ฆ1 + ๐‘ฆ2 , ๐‘ง1 + ๐‘ง2 ).
101
a) We have to show that L(X + Y) = L(X) + L(Y)
L(X + Y) = L(๐‘ฅ1 + ๐‘ฅ2 , ๐‘ฆ1 + ๐‘ฆ2 , ๐‘ง1 + ๐‘ง2 )
= (๐‘ฅ1 + ๐‘ฅ2 , ๐‘ฆ1 + ๐‘ฆ2 , 0)
= (๐‘ฅ1 , ๐‘ฆ1 , 0) + ( ๐‘ฅ2 , ๐‘ฆ2 , 0)
= L(X) + L(Y)
Hence L(X + Y) = L(X) + L(Y)
b) Let ๐‘๐‘ฟ = (๐‘๐‘ฅ1 , ๐‘๐‘ฆ1 , ๐‘๐‘ง1 ). We have to show that L(๐‘X) = ๐‘L(X).
L(๐‘๐‘ฟ) = (๐‘๐‘ฅ1 , ๐‘๐‘ฆ1 , 0)
= ๐‘(๐‘ฅ1 , ๐‘ฆ1 , 0)
= ๐‘L(๐‘ฟ)
Hence L(cX) = cL(X). Since the two conditions are satisfied then L is a linear transformation.
SAQ 5-1
Let L: โ„3
โ„3 be defined by
L(๐‘ฅ, ๐‘ฆ, ๐‘ง) = (๐‘ฅ − ๐‘ฆ, ๐‘ฅ 2 , 2๐‘ง)
Determine whether L is a linear transformation or not.
102
ASAQ 5-1
Let ๐‘ฟ = (๐‘ฅ1 , ๐‘ฆ1 , ๐‘ง1 ) and Y = (๐‘ฅ2 , ๐‘ฆ2 , ๐‘ง2 ) ๏ƒŽ โ„3 then ๐‘‹ + ๐‘Œ = (๐‘ฅ1 + ๐‘ฅ2 , ๐‘ฆ1 + ๐‘ฆ2 , ๐‘ง1 + ๐‘ง2 ).
L(๐‘ฟ + ๐’€) = ((๐‘ฅ1 + ๐‘ฅ2 ) − (๐‘ฆ1 + ๐‘ฆ2 ), (๐‘ฅ1 + ๐‘ฅ2 )2 , 2(๐‘ง1 + ๐‘ง2 ))
= (๐‘ฅ1 + ๐‘ฅ2 − ๐‘ฆ1 − ๐‘ฆ2 , ๐‘ฅ1 2 + 2๐‘ฅ1 ๐‘ฅ2 + ๐‘ฅ2 2 , 2๐‘ง1 + 2๐‘ง2 )
and L(X) + L(Y) = (๐‘ฅ1 − ๐‘ฆ1 , ๐‘ฅ1 2 , 2๐‘ง1 ) + (๐‘ฅ2 − ๐‘ฆ2 , ๐‘ฅ2 2 , 2๐‘ง2 )
= ((๐‘ฅ1 + ๐‘ฅ2 − ๐‘ฆ1 − ๐‘ฆ2 , ๐‘ฅ1 2 + ๐‘ฅ2 2 , 2๐‘ง1 + 2๐‘ง2 )
Since L(๐‘ฟ + ๐’€) ≠ L(๐‘ฟ) + L(๐’€) then L is not a linear transformation.
Theorem 5.1.1: If L: V→ W is a linear transformation, then
L(c1X1 + c2X2 + … + ckXk ) = c1L(X1) + c2L(X2) + … + ckL(Xk)
for any vectors X1, X2, … , Xk in V and any scalars c1, c2, … , ck.
Theorem 5.1.2: Let L: โ„๐‘› → โ„๐‘š be a linear transformation. Then
a. L(0) = 0
b. L(−๐‘ฟ) = − L(๐‘ฟ) for every ๐‘ฟ ∈ โ„๐‘›
c. L ( X – Y ) = L(X) – L(Y)
Proof of Theorem 5.1.2.a.
a.
L(0) = 0
L(0) = L (0 + 0)
L(0) = L(0) + L(0)
L(0) – L(0) = L(0) + L(0) – L(0)
0 = L(0)
103
Proof of Theorem 5.1.2.b.
b. L(−๐‘ฟ) = −L(๐‘ฟ)
L(−๐‘ฟ) = L(−1 โˆ™ ๐‘ฟ)
= −1L(๐‘ฟ)
= −L(๐‘ฟ)
Proof of Theorem 5.1.2.c.
c. L(๐‘ฟ − ๐’€) = L(๐‘ฟ + (−1)๐’€)
= L(๐‘ฟ) + L(−1 โˆ™ ๐’€)
= L(๐‘ฟ) − L(๐’€)
Example 2: Let L: โ„3 → โ„2 be defined by L(๐‘ฅ, ๐‘ฆ, ๐‘ง) = (๐‘ฅ, ๐‘ฆ) be a linear transformation
(verify). It can be verified that
L(0, 0, 0) = (0, 0)
That is, the zero vector in โ„3 is mapped to the zero vector in โ„2 .
Also, let X = (๐‘ฅ1 , ๐‘ฆ1 , ๐‘ง1 ) and Y = (๐‘ฅ2 , ๐‘ฆ2 , ๐‘ง2 ) be any vectors in โ„3 then
๐‘ฟ − ๐’€ = (๐‘ฅ1 − ๐‘ฅ2 , ๐‘ฆ1 − ๐‘ฆ2 , ๐‘ง1 − ๐‘ง2 ). Thus
L(X – Y) = (๐‘ฅ1 − ๐‘ฅ2 , ๐‘ฆ1 − ๐‘ฆ2 )
= (๐‘ฅ1 , ๐‘ฆ1 ) − (๐‘ฅ2 , ๐‘ฆ2 )
= L(X) – L(Y)
Example 3: Let T: โ„2 → โ„2 be the “translation mapping” defined by
T(๐‘ฅ, ๐‘ฆ) = (๐‘ฅ + 1, ๐‘ฆ + 2)
By Theorem 5.1.2 letter a, T is not a linear transformation since
T(0, 0) = (0 + 1, 0 + 2) = (1, 2) ≠ (0, 0)
(The zero vector is not mapped into the zero vector).
104
Theorem 5.1.3: Let L: V→ W be a linear transformation of an n-dimensional vector space V
into a vector space W. Also let S = { X1 , X2, … , Xn } be a basis for V. If X is any vector in V,
then L(X) is completely determined by {L(X1), L(X2), … , L(Xn) }.
Example 4: Let L: โ„3 → โ„3 be a linear transformation for which we know that
(2, −4) , L(0, 1, 0) = (3, −5) and L(0, 0, 1) = (2, 3).
a. What is L(1, −2, 3)?
b) What is L(๐‘Ž, ๐‘, ๐‘)?
Solution: The set {(1, 0, 0), (0, 1, 0), (0, 0, 1)} is a basis for โ„3 , and
(1, −2, 3) = 1(1, 0, 0) – 2(0, 1, 0) + 3(0, 0, 1)
By Theorem 5.1.3:
a)
L(1, −2, 3) = L1(1, 0, 0) – L2(0, 1, 0) + L3(0, 0, 1)
= L(1, 0, 0) – 2L(0, 1, 0) + 3L(0, 0, 1)
= (2, −4) – 2(3, −5) + 3(2, 3)
= (2, 23)
b) We see from part (a) that for any (๐‘Ž, ๐‘, ๐‘) ๏ƒŽ โ„3 , we have
L(๐‘Ž, ๐‘, ๐‘) = ๐‘ŽL(1, 0, 0) + ๐‘L(0, 1, 0) + ๐‘L(0, 0, 1)
= ๐‘Ž(2, −4) + ๐‘(3, −5) + ๐‘(2, 3)
= (2๐‘Ž + 3๐‘ + 2๐‘, −4๐‘Ž −5๐‘ + 3๐‘)
SAQ 5-2
Let L: P2 → P3 be a linear transformation for which we know
that L(1) = 1, L(t) = t2 and L(t2) = t3 + t. Find
(a) L(2t2 – 5t + 3)
(b) L(at2 + bt + c)
L(1, 0,0) =
105
ASAQ2
The set {๐‘ก 2 , ๐‘ก, 1} is a basis for ๐‘ƒ2 . By Theorem 5.1.3
a. L(2๐‘ก 2 − 5๐‘ก + 3) = L2(๐‘ก 2 ) + L(−5)(๐‘ก) + L3(1)
= 2L(๐‘ก 2 ) − 5L(๐‘ก) + 3L(1)
= 2(๐‘ก 3 + ๐‘ก) − 5(๐‘ก 2 ) + 3(1)
= 2๐‘ก 3 + 2๐‘ก − 5๐‘ก 2 + 3
= 2๐‘ก 3 − 5๐‘ก 2 + 2๐‘ก + 3
b. L(๐‘Ž๐‘ก 2 + ๐‘๐‘ก + ๐‘) = ๐‘ŽL(๐‘ก 2 ) + ๐‘L(๐‘ก) + cL(1)
= ๐‘Ž(๐‘ก 3 + ๐‘ก) + ๐‘(๐‘ก 2 ) + ๐‘
= ๐‘Ž๐‘ก 3 + ๐‘Ž๐‘ก + ๐‘๐‘ก 2 + ๐‘
= ๐‘Ž๐‘ก 3 + ๐‘๐‘ก 2 + ๐‘Ž๐‘ก + ๐‘
5.2 The Kernel and Range of a Linear Transformation
Definition 5.2.1: A linear transformation L: V→ W is said to be one-to-one if for all X1, X2 in
V, X1 ≠ X2 implies that L(X1) ≠ L(X2). An equivalent statement is that L is one-to-one if for all
X1, X2 in V, L(X1) = L(X2) implies that X1 = X2.
Example 1. Let L: โ„2 → โ„2 be a linear transformation defined by
L( ๐‘ฅ, ๐‘ฆ ) = ( ๐‘ฅ + ๐‘ฆ, ๐‘ฅ )
Determine if L is one–to–one or not.
Solution:
Let X1 = (๐‘ฅ1 , ๐‘ฆ1 ) and X2 = (๐‘ฅ2 , ๐‘ฆ2 ) be vectors in โ„2 . We have to show that if L(X1) = L(X2) then
X1 = X2.
L(X1) = L(X2)
(๐‘ฅ1 + ๐‘ฆ1 , ๐‘ฅ1 ) = (๐‘ฅ2 + ๐‘ฆ2 , ๐‘ฅ2 )
Equating the corresponding parts we have
๐‘ฅ1 + ๐‘ฆ1 = ๐‘ฅ2 + ๐‘ฆ2
๐‘ฅ1 = ๐‘ฅ2
106
If we subtract the second equation from the first, we get ๐‘ฆ1 = ๐‘ฆ2 which implies that ๐‘ฟ๐Ÿ =
๐‘ฟ๐Ÿ . Thus L is one-to-one.
Example 2. Let L: โ„3 → โ„3 be a linear transformation defined by
L(x, y, z) = (x, y, 0)
Determine if L is one-to-one or not.
Solution:
Let X1 = (4, 5, −3) and X2 = (4, 5, 2) be vectors in โ„3 . We see that X1 ๏‚น X2 but
L(4, 5, −3) = L(4, 5, 2) = (4, 5, 0). Hence L is not one-to-one.
Definition 5.2.2: Let L: V→ W be a linear transformation. The kernel of L, kerL, is the subset
of V consisting of all vectors X such that L(X) = Ow.
Note that kerL is not empty since by Theorem 5.1.2 we know that if L: โ„๐‘› → โ„๐‘š is a linear
transformation then the zero vector in โ„๐‘› is mapped to the zero vector in โ„๐‘š . Thus ๐‘‚ ∈
kerL.
Example 3. Let L: โ„4 → โ„3 be a linear transformation defined by
L(๐‘ฅ, ๐‘ฆ, ๐‘ง, ๐‘ค) = (๐‘ฅ + ๐‘ฆ, ๐‘ง + ๐‘ค, ๐‘ฅ + ๐‘ง)
The vector (1, −1, −1, 1) is in kerL since L(1, −1, −1, 1) = (0, 0, 0) while the vector
(1, 2, 3, −4) is not in kerL because L(1, 2, 3, −4) = ( 3, 1, 5 ) ≠ (0, 0, 0). Thus all vectors X in
โ„4 such that L(X) = (0, 0, 0) are in kerL.
Example 4. Let L be the linear transformation of Example 3. The kerL consists of all vectors
X in โ„4 such that L(X) = 0. If we let X = (๐‘ฅ, ๐‘ฆ, ๐‘ง, ๐‘ค) then
L(๐‘ฅ, ๐‘ฆ, ๐‘ง, ๐‘ค) = (0, 0, 0)
(๐‘ฅ + ๐‘ฆ, ๐‘ง + ๐‘ค, ๐‘ฅ + ๐‘ง) = (0, 0, 0)
Equating the corresponding parts we obtain the homogeneous system
๐‘ฅ+๐‘ฆ
=0
๐‘ง + ๐‘ค=0
๐‘ฅ +๐‘ง
=0
107
The augmented matrix in reduced echelon form (verify) is
1
[0
0
0 0 −1 0
1 0 1 0]
0 1 1 0
This implies that ๐‘ค is a free variable and can take on any value. Hence
๐‘ฅ=๐‘Ÿ
๐‘ฆ = −๐‘Ÿ
๐‘ง = −๐‘Ÿ
๐‘ค = ๐‘Ÿ where ๐‘Ÿ ∈ โ„
Thus kerL consists of all vectors of the form (๐‘Ÿ, −๐‘Ÿ, −๐‘Ÿ, ๐‘Ÿ) where ๐‘Ÿ is any real number. That
is
ker L = { (๐‘Ÿ, −๐‘Ÿ, −๐‘Ÿ, ๐‘Ÿ): ๐‘Ÿ ๏ƒŽ โ„}
= { ๐‘Ÿ(1, −1, −1, 1): ๐‘Ÿ ๏ƒŽ โ„}
Hence ker L = span { (1, −1, −1, 1)}
Since (1, −1, −1, 1) is linearly independent then it forms a basis for kerL. Thus
dim(kerL) = 1.
SAQ 5-3
Let L: โ„4 → โ„2 be defined by
๏ƒฆ๏ƒฉx๏ƒน๏ƒถ
๏ƒง๏ƒช ๏ƒบ๏ƒท
y
๏ƒฉx ๏€ญ y๏ƒน
L๏ƒง ๏ƒช ๏ƒบ๏ƒท ๏€ฝ ๏ƒช
๏ƒง ๏ƒช z ๏ƒบ ๏ƒท ๏ƒซ z ๏€ญ w๏ƒบ๏ƒป
๏ƒง๏ƒง ๏ƒช ๏ƒบ ๏ƒท๏ƒท
๏ƒจ ๏ƒซ w๏ƒป ๏ƒธ
Find kerL.
108
ASAQ 5-3
ker L = { (๐‘ฅ, ๐‘ฆ, ๐‘ง, ๐‘ค) : (๐‘ฅ − ๐‘ฆ, ๐‘ง − ๐‘ค) = (0, 0) }
Equating the corresponding parts we get:
๐‘ฅ − ๐‘ฆ = 0 which implies that ๐‘ฅ = ๐‘ฆ
and
๐‘ง − ๐‘ค = 0 which implies that ๐‘ง = ๐‘ค
If we let ๐‘ฆ = ๐‘Ÿ ๏ƒŽ โ„ and ๐‘ค = ๐‘  ๏ƒŽ โ„ then
kerL = { (๐‘Ÿ, ๐‘Ÿ, ๐‘ , ๐‘ ) : ๐‘Ÿ and ๐‘  ๏ƒŽ โ„ }
= { ๐‘Ÿ (1, 1, 0, 0) + ๐‘ (0, 0, 1, 1) }
Hence kerL = span {(1, 1, 0, 0), (0, 0, 1, 1)}
Since {(1, 1, 0, 0), (0, 0, 1, 1)} is linearly independent (one vector is not a scalar multiple of
the other) then it forms a basis for kerL. Hence dim(kerL) = 2.
Theorem 5.2.1: If L: V→ W is a linear transformation, then kerL is a subspace of V.
Definition 5.2.3: If L: V→ W is a linear transformation, then the range of L, denoted by range
L, is the set of all vectors in W that are images, under L, of vectors in V. Thus a vector Y is in
range L if we can find some vector X in V such that L(X) = Y. If range L = W, we say that L is
onto.
Example 5: Let L: โ„3 → โ„2 be a linear transformation defined by
L(๐‘ฅ, ๐‘ฆ, ๐‘ง) = (๐‘ฅ + ๐‘ฆ, ๐‘ฆ − ๐‘ง)
The range or image of L (ImL) is:
ImL = { (๐‘ฅ + ๐‘ฆ, ๐‘ฆ − ๐‘ง) }
= { ๐‘ฅ (1, 0) + ๐‘ฆ (1, 1) + ๐‘ง (0, -1) : ๐‘ฅ, ๐‘ฆ, ๐‘ง ๏ƒŽ โ„ }
Thus ImL = span {(1, 0), (1, 1), (0, −1)}. By Theorem 4.1.1 {(1, 0), (1, 1), (0, −1)} is linearly
dependent. Since (0, −1) is a linear combination of the other vectors such as
109
(0, −1) = 1 (1, 0) + (−1)(1, 1)
then we can delete (0, −1) from the set and the remaining vectors still span ImL. The set
{(1, 0), (1, 1)} is linearly independent therefore forms a basis for ImL. Hence dim(ImL) = 2.
Theorem 5.2.2: Let L: V→ W be a linear transformation
(i) L is a monomorphism (one-to-one) if and only if dim(kerL) = 0.
(ii) L is an epimorphism (onto) if and only if dim(ImL) = dim W.
Example 6. In ASAQ 5-3, since dim(kerL) = 2 then L is not a monomorphism (one-to-one)
and in Example 5, since dim(ImL) = 2 = dimโ„2 then L is an epimorphism (onto).
Theorem 5.2.3: If L: V→ W is a linear transformation, then range L is a subspace of W.
Theorem 5.2.4: If L: V→ W is a linear transformation, then
dim(kerL) + dim(range L) = dim V
Example 7: Verify Theorem 5.2.4 using the linear transformation of Example 5.
Solution:
kerL = { (๐‘ฅ, ๐‘ฆ, ๐‘ง) : (๐‘ฅ + ๐‘ฆ, ๐‘ฆ − ๐‘ง) = (0, 0) }
Equating the corresponding parts we get:
๐‘ฅ + ๐‘ฆ = 0 which implies that ๐‘ฅ = −๐‘ฆ
and
๐‘ฆ − ๐‘ง = 0 which implies that ๐‘ง = ๐‘ฆ ;
If we let ๐‘ฆ = ๐‘Ÿ, ๐‘Ÿ ๏ƒŽ โ„, then
kerL = { (−๐‘Ÿ, ๐‘Ÿ, ๐‘Ÿ) : ๐‘Ÿ ๏ƒŽ โ„ }
= { ๐‘Ÿ(−1, 1, 1) : ๐‘Ÿ ๏ƒŽ โ„ }
Thus kerL = span { (−1, 1, 1) } which is linearly independent hence forms a basis for kerL.
Thus dim(kerL) = 1.
110
From the previous example, we know that dim(ImL) = 2. Therefore
dim(kerL) + dim(ImL) = dim โ„3
1
+
2
= 3
Remark: Let L: V → W be a linear transformation. Then the rank of L is defined to be the
dimension of its image, and the nullity of L is defined to be the dimension of its kernel.
rank (L) = dim(ImL) and nullity (L) = dim(kerL)
Thus the preceding theorem yields the following formula for L when V has a finite
dimension:
rank(L) + nullity(L) = dim V.
Corollary 5.2.1: Let L: V→ W be a linear transformation and dim V = dim W,
a. If L is one-to-one, then it is onto.
b. If L is onto, then it is one-to-one.
SAQ 5-4
Let L: โ„3 →โ„4 be defined by
L(๐‘ฅ, ๐‘ฆ, ๐‘ง) = (๐‘ฅ + ๐‘ง, ๐‘ฆ − ๐‘ฅ, ๐‘ฆ + ๐‘ง, ๐‘ฅ + ๐‘ฆ + 2๐‘ง)
Verify Theorem 5.2.7
111
ASAQ 5-4
L(๐‘ฅ, ๐‘ฆ, ๐‘ง) = (0, 0, 0, 0)
(๐‘ฅ + ๐‘ง, ๐‘ฆ − ๐‘ฅ, ๐‘ฆ + ๐‘ง, ๐‘ฅ + ๐‘ฆ + 2๐‘ง) = (0, 0, 0, 0)
Equating the corresponding parts we obtain the homogeneous system
๐‘ฅ
+๐‘ง =0
−๐‘ฅ + ๐‘ฆ
=0
๐‘ฆ +๐‘ง =0
๐‘ฅ + ๐‘ฆ + 2๐‘ง = 0
The augmented matrix in reduced row echelon form (verify) is
1
0
[
0
0
0
1
0
0
1
1
0
0
0
0
]
0
0
Thus ๐‘ฅ = −๐‘ง, ๐‘ฆ = −๐‘ง, where ๐‘ง can be any real number. If we let ๐‘ง = ๐‘Ÿ ∈ โ„ then
kerL = {(−๐‘Ÿ, −๐‘Ÿ, ๐‘Ÿ): ๐‘Ÿ ∈ โ„}
= {๐‘Ÿ(−1, −1, 1): ๐‘Ÿ ∈ โ„}
Hence kerL = span {(−1, −1, 1)}. Since (−1, −1, 1) is linearly independent then it is a basis
for kerL. Thus
nullity of L = dim(kerL) = 1.
Solving for the image of L, we have
Im(L) = {(๐‘ฅ + ๐‘ง, ๐‘ฆ − ๐‘ฅ, ๐‘ฆ + ๐‘ง, ๐‘ฅ + ๐‘ฆ + 2๐‘ง)}
= {๐‘ฅ(1, −1, 0, 1) + ๐‘ฆ(0, 1, 1, 1) + ๐‘ง(1, 0, 1, 2): ๐‘ฅ, ๐‘ฆ, ๐‘ง ∈ โ„}
Since (1, 0, 1, 2) = (1, −1, 0, 1) + (0, 1, 1, 1) then (1, 0, 1, 2) is a linear combination of the
preceding vectors and may be deleted from the spanning set. Thus
Im(L) = span {(1, −1, 0, 1), (0, 1, 1, 1)}
Since {(1, −1, 0, 1), (0, 1, 1, 1)} is linearly independent (one vector is not a scalar multiple of
the other) then it is a basis for Im(L). Therefore
Rank of L = dim(ImL) = 2
112
Now it can be verified that
Nullity of L + rank of L = dim(โ„3 )
1
+ 2
= 3
ACTIVITY
1. Which of the following are linear transformations?
(a) L(x, y) = (x2 + x, y – y2)
(b) L(x, y) = (x – y, 0, 2x + 3)
๏ƒฆ ๏ƒฉ x ๏ƒน ๏ƒถ ๏ƒฉ2x ๏€ญ 3 y ๏ƒน
๏ƒง
๏ƒท
(c) L ๏ƒง ๏ƒช๏ƒช y ๏ƒบ๏ƒบ ๏ƒท ๏€ฝ ๏ƒช๏ƒช3 y ๏€ญ 2 z ๏ƒบ๏ƒบ
๏ƒง ๏ƒช z ๏ƒบ ๏ƒท ๏ƒช 2z ๏ƒบ
๏ƒป
๏ƒจ๏ƒซ ๏ƒป๏ƒธ ๏ƒซ
2. Let L: โ„3 → โ„4 be defined by
L(x, y, z) = (x + y + z, x + 2y – 3z, 2x + 3y – 2z, 3x + 4y – z)
(a) Find a basis for and the dim(kerL),
(b) Find a basis for and the dim(ImL),
(c) Verify Theorem 5.2.4.
3. Let L: โ„3 → โ„3 be defined by
L(x, y z) = (x + z, x + y + 2z, 2x + y + 3z )
(a) Is L one-to-one?
(b) Is L onto?
4. Let L: P2 → P2 be the linear transformation defined by L(at2 + bt + c) = (a + c)t2 +
(b + c)t.
(a) Is t2 – t – 1 in kerL?
(d) Find a basis for kerL.
(b) Is t2 + t -1 in kerL?
(e) Find a basis for ImL.
2
(c) Is 2t – t in range L?
113
5.3 The Matrix of a Linear Transformation
Coordinate Vectors
Definition 5.3.1: Let V be an n-dimensional vector space with basis S = { X1 , X2, … , Xn }. If
X = a1X1 + a2X2 + … + anXn
is any vector in V, then the vector
๏ƒฉ a1 ๏ƒน
๏ƒชa ๏ƒบ
X
๏€ฝ
๏› ๏S ๏ƒช๏ƒช 2 ๏ƒบ๏ƒบ
๏ƒช ๏ƒบ
๏ƒช๏ƒซ an ๏ƒบ๏ƒป
in โ„๐‘› is called the coordinate vector of X with respect to the basis S. The components of
[ X ] S are called the coordinates of X with respect to S.
๏ƒฌ๏ƒฉ 1 ๏ƒน ๏ƒฉ 2๏ƒน ๏ƒผ
Example 1. Let S = ๏ƒญ๏ƒช ๏ƒบ , ๏ƒช ๏ƒบ ๏ƒฝ be a basis for โ„2 . Find the coordinate vectors of the
๏ƒฎ๏ƒซ๏€ญ1๏ƒป ๏ƒซ 3๏ƒป ๏ƒพ
following vectors with respect to S.
๏ƒฉ ๏€ญ3 ๏ƒน
(a) ๏ƒช ๏ƒบ
๏ƒซ ๏€ญ7 ๏ƒป
๏ƒฉ12 ๏ƒน
(b) ๏ƒช ๏ƒบ
๏ƒซ13๏ƒป
Solution:
๏ƒฉ ๏€ญ3 ๏ƒน
(a) Let X = ๏ƒช ๏ƒบ . To find [X]S we must find c1 and c2 such that
๏ƒซ ๏€ญ7 ๏ƒป
๏ƒฉ ๏€ญ3 ๏ƒน
๏ƒฉ2๏ƒน
๏ƒฉ1๏ƒน
๏ƒช ๏€ญ7 ๏ƒบ = c1 ๏ƒช ๏ƒบ + c2 ๏ƒช 3 ๏ƒบ
๏ƒซ ๏ƒป
๏ƒซ ๏ƒป
๏ƒซ ๏€ญ1๏ƒป
Multiplying, adding, and equating the corresponding parts, we get
c1 + 2c2 = − 3
− c1 + 3c2 = − 7
The solution is c1 = 1 and c2 = −2. Thus the coordinate vector of X with respect to the
basis S is
๏ƒฉ1๏ƒน
๏ƒช ๏€ญ2 ๏ƒบ
๏ƒซ ๏ƒป
114
Letter (b) is left as an exercise.
๏ƒฌ
๏ƒฉ1 ๏ƒน
๏ƒฉ0๏ƒน ๏ƒผ
Example 2. Let S = ๏ƒญ E1 ๏€ฝ ๏ƒช ๏ƒบ , E2 ๏€ฝ ๏ƒช ๏ƒบ ๏ƒฝ be the natural basis for โ„2 . Then
๏ƒซ0๏ƒป
๏ƒซ1 ๏ƒป ๏ƒพ
๏ƒฎ
๏ƒฉ ๏€ญ3 ๏ƒน
๏ƒฉ1 ๏ƒน
๏ƒฉ0 ๏ƒน
๏ƒช ๏€ญ7 ๏ƒบ = −3 ๏ƒช 0 ๏ƒบ + (−7) ๏ƒช 1 ๏ƒบ
๏ƒซ ๏ƒป
๏ƒซ ๏ƒป
๏ƒซ ๏ƒป
๏ƒฉ ๏€ญ3 ๏ƒน
Hence [X]S = ๏ƒช ๏ƒบ . Note that [X]S is the original vector itself when the natural basis is
๏ƒซ ๏€ญ7 ๏ƒป
used. This result is true in general.
Theorem 5.3.1: Let L: V→ W be a linear transformation of an ๐‘›-dimensional vector space V
into an ๐‘š-dimensional vector space W ( ๐‘› ≠ 0 and ๐‘š ≠ 0 ) and let S = { X1 , X2, … , Xn} and T =
{Y1, Y2, …, Ym } be bases for V and W, respectively. Then the ๐‘š x ๐‘› matrix A, whose jth
column is the coordinate vector [L(Xj)]T of L(Xj) with respect to T, is associated with L and
has the following property:
If Y = L(X) for some X in V, then [Y]T = A[X]S ,
where [X]S and [Y]T are the coordinate vectors of X and Y with respect to the respective
bases S and T. Moreover, A is the only matrix with this property.
Procedure for computing the matrix of a linear transformation L: V→ W with respect to the
bases S = { X1 , X2, … , Xn} and T = { Y1, Y2, …, Ym } for V and W, respectively:
STEP 1: Compute L(Xj) for j = 1, 2, … , ๐‘›.
STEP 2: Find the coordinate vector [L(Xj)]T of L(Xj) with respect to the basis T. This means
that we have to express L(Xj) as a linear combination of the vectors in T.
STEP 3: The matrix A of L with respect to S and T is formed by choosing [L(Xj)]T as the jth
column of A.
Definition 5.3.2: The matrix of Theorem 5.3.1 is called the matrix of L with respect to the
bases S and T. The equation [L(X)]T = A[X]S is called the representation of L with respect to S
and T. We also say that the said equation represents L with respect to S and T.
115
Example 3: Let L: โ„2 → โ„2 be defined by
๏ƒฆ ๏ƒฉ x๏ƒน ๏ƒถ
๏ƒฉx ๏€ซ 2 y๏ƒน
L๏ƒง ๏ƒช ๏ƒบ๏ƒท = ๏ƒช
๏ƒบ
๏ƒซ2x ๏€ญ y ๏ƒป
๏ƒจ ๏ƒซ y๏ƒป ๏ƒธ
๏ƒฌ
๏ƒฉ1 ๏ƒน
๏ƒฉ0๏ƒน ๏ƒผ
Let S = ๏ƒญ X1 ๏€ฝ ๏ƒช ๏ƒบ , X 2 ๏€ฝ ๏ƒช ๏ƒบ ๏ƒฝ be the natural basis for โ„2 and let
๏ƒซ0๏ƒป
๏ƒซ1 ๏ƒป ๏ƒพ
๏ƒฎ
๏ƒฌ
๏ƒฉ ๏€ญ1๏ƒน
๏ƒฉ 2๏ƒน ๏ƒผ
T = ๏ƒญY1 ๏€ฝ ๏ƒช ๏ƒบ , Y2 ๏€ฝ ๏ƒช ๏ƒบ ๏ƒฝ be another basis for โ„2 . Find the matrix of L with respect to S and
๏ƒซ2๏ƒป
๏ƒซ0๏ƒป ๏ƒพ
๏ƒฎ
๏ƒฆ ๏ƒฉ1 ๏ƒน ๏ƒถ
T. Compute L ๏ƒง ๏ƒช ๏ƒบ ๏ƒท using the matrix of L.
๏ƒจ ๏ƒซ2๏ƒป ๏ƒธ
Solution:
๏ƒฆ ๏ƒฉ1 ๏ƒน ๏ƒถ
๏ƒฉ1 ๏ƒน
๏ƒฉ ๏€ญ1๏ƒน
๏ƒฉ 2๏ƒน
L ๏ƒง ๏ƒช ๏ƒบ ๏ƒท = ๏ƒช ๏ƒบ ๏€ฝ a1 ๏ƒช ๏ƒบ ๏€ซ a2 ๏ƒช ๏ƒบ
๏ƒซ2๏ƒป
๏ƒซ2๏ƒป
๏ƒซ0๏ƒป
๏ƒจ ๏ƒซ0 ๏ƒป ๏ƒธ
1 = ๏€ญa1 ๏€ซ 2a2
2 = 2a1
๏ƒฉ1๏ƒน
The solution is a1 ๏€ฝ 1 and a2 ๏€ฝ 1, so ๏› L( X1 )๏T = ๏ƒช ๏ƒบ
๏ƒซ1๏ƒป
๏ƒฆ ๏ƒฉ0 ๏ƒน ๏ƒถ
๏ƒฉ2๏ƒน
๏ƒฉ ๏€ญ1๏ƒน
๏ƒฉ 2๏ƒน
L ๏ƒง ๏ƒช ๏ƒบ ๏ƒท = ๏ƒช ๏ƒบ ๏€ฝ a1 ๏ƒช ๏ƒบ ๏€ซ a2 ๏ƒช ๏ƒบ
๏ƒซ ๏€ญ1๏ƒป
๏ƒซ2๏ƒป
๏ƒซ0๏ƒป
๏ƒจ ๏ƒซ1 ๏ƒป ๏ƒธ
2 = ๏€ญa1 ๏€ซ 2a2
−1 = 2a1
๏ƒฉ ๏€ญ1/ 2 ๏ƒน
The solution is a1 ๏€ฝ ๏€ญ1/ 2 and a2 ๏€ฝ 3/ 4 , so ๏› L( X 2 )๏T = ๏ƒช
๏ƒบ.
๏ƒซ 3/ 4 ๏ƒป
Thus the matrix A of L with respect to S and T is
๏ƒฉ1 ๏€ญ1/ 2 ๏ƒน
A๏€ฝ๏ƒช
๏ƒบ
๏ƒซ1 3 / 4 ๏ƒป
๏ƒฉ ๏ƒฆ ๏ƒฉ1 ๏ƒน ๏ƒถ ๏ƒน
๏ƒฉ ๏ƒฉ1 ๏ƒน ๏ƒน
To compute for ๏ƒช L ๏ƒง ๏ƒช ๏ƒบ ๏ƒท ๏ƒบ using A, we first compute for ๏ƒช ๏ƒช ๏ƒบ ๏ƒบ . Since S is the natural
๏ƒช๏ƒซ ๏ƒจ ๏ƒซ 2๏ƒป ๏ƒธ ๏ƒบ๏ƒปT
๏ƒซ ๏ƒซ2๏ƒป ๏ƒป S
basis for โ„2 then
116
๏ƒฉ ๏ƒฉ1 ๏ƒน ๏ƒน
๏ƒฉ1 ๏ƒน
๏ƒช๏ƒช ๏ƒบ๏ƒบ = ๏ƒช ๏ƒบ
๏ƒซ2๏ƒป
๏ƒซ ๏ƒซ2๏ƒป ๏ƒป S
So,
๏ƒฉ ๏ƒฆ ๏ƒฉ1 ๏ƒน ๏ƒถ ๏ƒน
๏ƒฉ1 ๏€ญ1/ 2 ๏ƒน
๏ƒช L ๏ƒง ๏ƒช ๏ƒบ ๏ƒท๏ƒบ = ๏ƒช
๏ƒบ
๏ƒซ1 3 / 4 ๏ƒป
๏ƒซ๏ƒช ๏ƒจ ๏ƒซ 2๏ƒป ๏ƒธ ๏ƒป๏ƒบT
๏ƒฉ1 ๏ƒน
๏ƒฉ 0 ๏ƒน
๏ƒช 2 ๏ƒบ = ๏ƒช5 / 2 ๏ƒบ
๏ƒซ
๏ƒป
๏ƒซ ๏ƒป
ACTIVITY
๏ƒฌ ๏ƒฉ 1 ๏ƒน ๏ƒฉ 0 ๏ƒน ๏ƒฉ1 ๏ƒน ๏ƒผ
๏ƒฏ
๏ƒฏ
1. Let S = ๏ƒญ ๏ƒช๏ƒช ๏€ญ1๏ƒบ๏ƒบ , ๏ƒช๏ƒช 2 ๏ƒบ๏ƒบ , ๏ƒช๏ƒช 0 ๏ƒบ๏ƒบ ๏ƒฝ be a basis for โ„3 . Find the coordinate vector of each of the
๏ƒฏ ๏ƒช 2 ๏ƒบ ๏ƒช1 ๏ƒบ ๏ƒช0๏ƒบ ๏ƒฏ
๏ƒฎ๏ƒซ ๏ƒป ๏ƒซ ๏ƒป ๏ƒซ ๏ƒป ๏ƒพ
following vectors with respect to S.
๏ƒฉ1 ๏ƒน
(a) ๏ƒช๏ƒช 4 ๏ƒบ๏ƒบ
๏ƒช๏ƒซ 2 ๏ƒบ๏ƒป
๏ƒฉ3๏ƒน
(b) ๏ƒช๏ƒช ๏€ญ4 ๏ƒบ๏ƒบ
๏ƒช๏ƒซ 3 ๏ƒบ๏ƒป
๏ƒฉ1๏ƒน
(c) ๏ƒช๏ƒช ๏€ญ2 ๏ƒบ๏ƒบ
๏ƒช๏ƒซ 4 ๏ƒบ๏ƒป
๏ƒฉx ๏€ญ 2 y๏ƒน
๏ƒฆ
๏ƒถ
x
๏ƒฉ
๏ƒน
2. Let L: โ„2 → โ„3 be defined by L ๏ƒง ๏ƒช ๏ƒบ ๏ƒท ๏€ฝ ๏ƒช๏ƒช 2 x ๏€ซ y ๏ƒบ๏ƒบ . Let S and T be the natural bases for
๏ƒจ ๏ƒซ y๏ƒป ๏ƒธ ๏ƒช x ๏€ซ y ๏ƒบ
๏ƒซ
๏ƒป
๏ƒฌ ๏ƒฉ 1 ๏ƒน ๏ƒฉ0 ๏ƒน ๏ƒผ
โ„ and โ„ , respectively. Also, let S’ = ๏ƒญ๏ƒช ๏ƒบ , ๏ƒช ๏ƒบ ๏ƒฝ and T’ =
๏ƒฎ๏ƒซ๏€ญ1๏ƒป ๏ƒซ1๏ƒป ๏ƒพ
2
3
๏ƒฌ ๏ƒฉ1 ๏ƒน ๏ƒฉ 0 ๏ƒน ๏ƒฉ 1 ๏ƒน ๏ƒผ
๏ƒฏ๏ƒช ๏ƒบ ๏ƒช ๏ƒบ ๏ƒช ๏ƒบ ๏ƒฏ
๏ƒญ ๏ƒช1 ๏ƒบ , ๏ƒช1 ๏ƒบ , ๏ƒช ๏€ญ1๏ƒบ ๏ƒฝ be bases for
๏ƒฏ ๏ƒช0 ๏ƒบ ๏ƒช1 ๏ƒบ ๏ƒช 1 ๏ƒบ ๏ƒฏ
๏ƒฎ๏ƒซ ๏ƒป ๏ƒซ ๏ƒป ๏ƒซ ๏ƒป ๏ƒพ
โ„2 and โ„3 , respectively. Find the matrix of L with respect to
(a) S and T
(b) S’ and T’
3. Let L: P1 → P3 be defined by L(p(t)) = t2p(t). Let S = { t, t + 1 } and
T = { t3 , t2 – 1, t, t + 1 } be bases for P1 and P3, respectively. Find the matrix of L
with respect to S and T. Compute [ L(–3t + 3) ]T using the matrix of L.
117
MODULE 6
EIGENVALUES AND EIGENVECTORS
Introduction
In this section we will discuss the concepts of eigenvalues, eigenvectors, and
eigenspaces, algebraic and geometric multiplicity of an eigenvalue, Hamilton-Cayley
theorem, similar matrices and diagonalizable matrices.
Objectives
At the end of this chapter, you are expected to be able to do the following:
1. Define eigenvalues and eigenvectors.
2. Compute for the eigenvalues and the associated eigenvectors of a matrix representing a
linear transformation.
3. Determine the algebraic and geometric multiplicities of the eigenvalues.
3. Identify the properties of a diagonalizable matrix.
4. Determine the matrix of transition P so that ๐‘ƒ −1 ๐ด๐‘ƒ is a diagonal matrix, given a
matrix A representing a linear transformation of a vector space into itself.
6.1 Characteristic Polynomial
Definition 6.1.1: If A ๏€ฝ [ a ij ] is an ๐‘› x ๐‘› matrix, the polynomial matrix
๐‘ฅ๐ผ − ๐ด = ๐ถ
is called the characteristic matrix of ๐ด. The determinant of ๐ถ is called the characteristic
polynomial of ๐ด. The equation det๐ถ = 0 is called the characteristic equation of ๐ด.
1
Example 1: Let ๐ด = [
4
−3
]. Then
−2
๐‘ฅ๐ผ2 − ๐ด = ๐ถ = [
= [
is called the characteristic matrix of ๐ด.
๐‘ฅ
0
0
1
]−[
๐‘ฅ
4
๐‘ฅ−1
−4
3
]
๐‘ฅ+2
−3
]
−2
118
The determinant of ๐ถ
๐‘ฅ−1
|
−4
3
| = (๐‘ฅ − 1)(๐‘ฅ + 2) − (−12)
๐‘ฅ+2
= ๐‘ฅ 2 + ๐‘ฅ + 10
is called the characteristic polynomial of ๐ด and the equation
๐‘ฅ 2 + ๐‘ฅ + 10 = 0
is called the characteristic equation of ๐ด.
6.2 Hamilton-Cayley Theorem
Let ๐‘“(๐‘ฅ) = ๐‘Ž0 ๐‘ฅ ๐‘› + ๐‘Ž1 ๐‘ฅ ๐‘›−1 + … + ๐‘Ž๐‘›−1 ๐‘ฅ + ๐‘Ž๐‘› be a polynomial in ๐‘ฅ with real
coefficients. If ๐ด is an ๐‘› x ๐‘› matrix then ๐‘“(๐ด) is equal to the matrix
๐‘Ž0 ๐ด๐‘› + ๐‘Ž1 ๐ด๐‘›−1 + … + ๐‘Ž๐‘›−1 ๐ด + ๐‘Ž๐‘› ๐ผ๐‘›
Note that we replace the constant term by ๐‘Ž๐‘› ๐ผ๐‘› so that each term of ๐‘“(๐ด) is a matrix.
Theorem 6.2.1. (Hamilton-Cayley Theorem) If ๐ด is an ๐‘› x ๐‘› matrix and ๐‘“(๐‘ฅ) is its
characteristic polynomial, then ๐‘“(๐ด) = 0.
2 3
Example 1: Let ๐ด = [
]. The characteristic matrix of ๐ด is
−1 4
๐‘ฅ − 2 −3
๐ถ=[
]
1
๐‘ฅ−4
Next we compute for the characteristic polynomial of ๐ด.
๐‘“(๐‘ฅ) = det๐ถ
๐‘ฅ−2
๐‘“(๐‘ฅ) = |
1
−3
|
๐‘ฅ−4
= (๐‘ฅ − 2)(๐‘ฅ − 4) − (−3)
= ๐‘ฅ 2 − 6๐‘ฅ + 11
It can be verified that the
constant term is equal to
the determinant of ๐ด.
119
Computing for ๐‘“(๐ด) we have
๐‘“(๐ด) = ๐ด2 − 6๐ด + 11๐ผ2
2 3 2
=[
][
−1 4 −1
1
3
2 3
] − 6[
] + 11 [
0
4
−1 4
1 18
12 18
11
= [
]−[
]+[
−6 13
−6 24
0
0
= [
0
0
]
1
0
]
11
0
]
0
Hence ๐ด satisfies its characteristic equation, that is ๐‘“(๐ด) = 0. Since det๐ด = 11 ≠ 0 then ๐ด
is nonsingular so ๐ด−1 exists. By the Hamilton-Cayley Theorem, we form the equation
๐ด2 − 6๐ด + 11๐ผ2 = 0
11๐ผ2 = 6๐ด − ๐ด2
๐ผ๐Ÿ =
๐ผ๐Ÿ =
1
11
(6๐ด − ๐ด2 )
1
11
(๐ด)(6๐ผ − ๐ด)
Therefore,
1
๐ด−1 = 11 (6๐ผ − ๐ด)
4 −3
]
1 2
4/11 −3/11
= [
]
1/11 2/11
1
= 11 [
120
SAQ 6-1
Find the characteristic polynomial for the matrix
2 −2 3
[1 1
1]
1 3 −1
and
a) Show by direct substitution that this matrix
satisfies its characteristic equation.
b) Find ๐ด−1 .
ASAQ 6-1
The characteristic matrix of ๐ด is
๐‘ฅ−2
[ −1
−1
2
−3
๐‘ฅ−1
−1 ]
−3 ๐‘ฅ + 1
and its characteristic polynomial is
๐‘“(๐‘ฅ) = (๐‘ฅ − 2)(๐‘ฅ − 1)(๐‘ฅ + 1) − 7 − [3(๐‘ฅ − 1) − 2(๐‘ฅ + 1) + 3(๐‘ฅ − 2)]
= ๐‘ฅ 3 − 2๐‘ฅ 2 − 5๐‘ฅ + 6
Next we form the equation
๐‘“(๐ด) = ๐ด3 − 2๐ด2 − 5๐ด + 6๐ผ3 = 0
(1)
where
2 −2 3 2 −2 3
5
๐ด2 = ๐ด โˆ™ ๐ด = [1 1
1 ] [1 1
1 ] = [4
1 3 −1 1 3 −1
4
5 3 1 2 −2
๐ด3 = ๐ด2 โˆ™ ๐ด = [4 2 3] [1 1
4 −2 7 1 3
Substituting these to (1), we have
3 1
2 3] and
−2 7
3
14 −4
1 ] = [13 3
−1
13 11
17
11]
3
121
14
๐‘“(๐ด) = [13
13
0 0
= [0 0
0 0
−4 17
2 −2 3
6 0
5 3 1
]
−
2
[
]
−
5
[
]
+
[
3 11
1 1
1
0 6
4 2 3
11 3
1 3 −1
0 0
4 −2 7
0
0]
6
0
0]
0
Thus ๐ด satisfies its characteristic equation.
b. To solve for ๐ด−1 , we form the equation
๐ด3 − 2๐ด2 − 5๐ด + 6๐ผ3 = 0
6๐ผ3 = −๐ด3 + 2๐ด2 + 5๐ด
1
๐ผ3 = 6 (−๐ด3 + 2๐ด2 + 5๐ด)
1
๐ผ3 = 6 (๐ด)(−๐ด2 + 2๐ด + 5๐ผ3 )
Therefore
1
๐ด−1 = 6 (−๐ด2 + 2๐ด + 5๐ผ3 )
2
1 −5 −3 −1
= [[−4 −2 −3] + 2 [1
6
1
−4 2 −7
−2 3
5 0 0
]
+
[
1
1
0 5 0]]
3 −1
0 0 5
4 −7 5
= 6 [−2 5 −1]
−2 8 −4
2/3 −7/6 5/6
−1/3
5/6 −1/6]
=[
−1/3 4/3 −2/3
1
6.3 Eigenvalues, Eigenvectors, and Eigenspaces
Definition 6.3.1: Let ๐ด be an ๐‘› x ๐‘› matrix. The real number ๐œ† is called an eigenvalue of ๐ด if
there exists a nonzero vector ๐‘ฟ in โ„๐‘› such that
๐ด๐‘ฟ = ๐œ†๐‘ฟ
(1)
122
Every nonzero vector ๐‘ฟ satisfying (1) is called an eigenvector of ๐ด associated with the
eigenvalue ๐œ†.
Example 1: Let ๐ด = [
[
2 −1
]. Since
−2 3
2 −1 1
1
][ ] = 4[ ]
−2 3 −2
−2
then 4 is an eigenvalue of ๐ด and [
4.
1
] is the eigenvector associated with the eigenvalue ๐œ† =
−2
Theorem 6.3.1: The eigenvalues of ๐ด are the real roots of the characteristic polynomial of
A.
1 0
Example 2. Let ๐ด = [−1 3
3 2
๐‘ฅ−1
| 1
−3
0
0 ]. The characteristic polynomial of ๐ด is
−2
0
๐‘ฅ−3
−2
0
0 | = (๐‘ฅ − 1)(๐‘ฅ − 3)(๐‘ฅ + 2)
๐‘ฅ+2
By Theorem 6.3.1, the eigenvalues of ๐ด are the roots of the characteristic equation
(๐‘ฅ − 1)(๐‘ฅ − 3)(๐‘ฅ + 2) = 0
Thus ๐œ†1 = 1, ๐œ†2 = 3, and ๐œ†3 = −2 are the eigenvalues of ๐ด.
To find the eigenvector associated with ๐œ†1 = 1, we form the system (1๐ผ3 − ๐ด)๐‘‹ = 0:
0
0 0 ๐‘ฅ1
0
๐‘ฅ
[ 1 −2 0] [ 2 ] = [0]
−3 −2 3 ๐‘ฅ3
0
6
A solution to this system is {(4 ๐‘Ÿ, 8 ๐‘Ÿ, ๐‘Ÿ) : ๐‘Ÿ ∈ โ„}. Thus if ๐‘Ÿ = 8 then ๐‘‹1 = [3] is the
8
eigenvector associated with ๐œ†1 = 1.
3
3
123
To find the eigenvector associated with ๐œ†2 = 3, we form the system (3๐ผ3 − ๐ด)๐‘‹ = 0:
2
0 0 ๐‘ฅ1
0
๐‘ฅ
[1
0 0] [ 2 ] = [0]
−3 −2 5 ๐‘ฅ3
0
0
A solution to this system is {(0, 2 ๐‘Ÿ, ๐‘Ÿ) : ๐‘Ÿ ∈ โ„}. Thus if ๐‘Ÿ = 2 then ๐‘‹2 = [5] is the
2
eigenvector associated with ๐œ†2 = 3.
5
To find the eigenvector associated with ๐œ†3 = −2, we form the system (−2๐ผ3 − ๐ด)๐‘‹ = 0:
−3 0 0 ๐‘ฅ1
0
๐‘ฅ
[ 1 −5 0] [ 2 ] = [0]
−3 −2 0 ๐‘ฅ3
0
0
A solution to this system is {(0, 0, ๐‘Ÿ): ๐‘Ÿ ∈ โ„}. Thus ๐‘‹2 = [0] is the eigenvector associated
1
with ๐œ†3 = −2.
๏ƒฉ 2 1๏ƒน
Example 3: Let A ๏€ฝ ๏ƒช
๏ƒบ.
๏ƒซ ๏€ญ1 3๏ƒป
The characteristic polynomial of ๐ด is
๐‘ฅ−2
|
1
−1
| = (๐‘ฅ − 2)(๐‘ฅ − 3) − (−1)
๐‘ฅ−3
= ๐‘ฅ 2 − 5๐‘ฅ + 7
Since ๐‘ฅ 2 − 5๐‘ฅ + 7 = 0 has no real roots then A has no eigenvalues.
Definition 6.3.2: Let L: V→ V be a linear transformation and let ๏ฌ be an eigenvalue of L.
Denote by S( ๏ฌ ) the set of all eigenvectors associated with ๏ฌ , together with the zero vector.
Then S( ๏ฌ ) is called the eigenspace of L associated with ๏ฌ .
Definition 6.3.3: Let ๏ฌ be an eigenvalue of L.
(a) The geometric multiplicity of ๏ฌ is the dimension of S( ๏ฌ ).
(b) The algebraic multiplicity of ๏ฌ is its multiplicity as a root of the characteristic polynomial
f( ๏ฌ ).
124
Example 4. Let L be a linear transformation represented by the matrix
2 0
0
๐ด = [3 −1 0 ]
0 4 −1
Determine the geometric and algebraic multiplicities of its eigenvalues.
Solution: The characteristic polynomial of A is
๐‘ฅ−2
0
0
๐‘“(๐‘ฅ) = | −3 ๐‘ฅ + 1
0 |
0
−4
๐‘ฅ+1
= (๐‘ฅ − 2)(๐‘ฅ + 1)(๐‘ฅ + 1)
= (๐‘ฅ − 2)(๐‘ฅ + 1)2
Thus the eigenvalues of A are ๐œ†1 = 2 with algebraic multiplicity 1 and ๐œ†2 = ๐œ†3 = −1 with
algebraic multiplicity 2.
To find the geometric multiplicity of ๐œ†1 = 2, we have to determine the dimension of the
eigenspace S(2) associated with ๐œ†1 = 2. Substituting 2 for x in the characteristic matrix of A
we obtain
0
0 0
[−3 3 0]
0 −4 3
This matrix in reduced row echelon form is
1
[0
0
3
The solution is {(4 ๐‘Ÿ,
3
4
0 −3/4
1 −3/4]
0
0
๐‘Ÿ, ๐‘Ÿ) : ๐‘Ÿ ∈ โ„}. Thus
3
S(2) = {(4 ๐‘Ÿ,
3
4
๐‘Ÿ, ๐‘Ÿ) : ๐‘Ÿ ∈ โ„}
= span {(3, 3, 4)}
Hence ๐œ†1 = 2 has geometric multiplicity 1.
125
Similarly, substituting −1 for x in the characteristic matrix we obtain
−3 0 0
[−3 0 0]
0 −4 0
This matrix in reduced row echelon form is
1 0
[0 1
0 0
0
0]
0
Thus
S(−1) = {(0, 0, ๐‘Ÿ): ๐‘Ÿ ∈ โ„}
= span{(0, 0, 1)}
Hence ๐œ†2 = ๐œ†3 = −1 has geometric multiplicity 1.
Theorem 6.3.2. Let ๐œ† be an eigenvalue. Then the geometric multiplicity of ๐œ† does not
exceed its algebraic multiplicity.
Theorem 6.3.3. If the eigenvalues ๐œ†1 , … , ๐œ†๐‘› are all different and {๐œ‰1 , … , ๐œ‰2 } is a set of
eigenvectors, ๐œ‰๐‘– corresponding to ๐œ†๐‘– , then the set {๐œ‰1 , … , ๐œ‰2 } is linearly independent.
Example 5. Let A be the matrix of Example 2. The three distinct eigenvalues of A are
6
0
0
๐œ†1 = 1, ๐œ†2 = 3, and ๐œ†3 = −2 and the corresponding eigenvectors are [3], [5], and [0]. It
8
2
1
6
0
0
can be verified that [3], [5], and [0] are linearly independent.
8
2
1
126
SAQ 6-2
3
2
Let ๐ด = [ 1
4
−2 −4
2
1 ].
−1
Determine the algebraic and geometric multiplicities of its
eigenvalues.
ASAQ 6-2
The characteristic polynomial of A is
๐‘ฅ−3
−2
−2
๐‘“(๐‘ฅ) = | −1 ๐‘ฅ − 4 −1 |
2
4
๐‘ฅ+1
= ๐‘ฅ 3 − 6๐‘ฅ 2 + 11๐‘ฅ − 6
= (๐‘ฅ − 1)(๐‘ฅ − 2)(๐‘ฅ − 3)
Thus the eigenvalues of A are 1, 2, and 3 all of which have algebraic multiplicity 1.
To solve for the geometric multiplicity of ๐œ† = 1, we substitute 1 for x in the characteristic
matrix obtaining the matrix
−2 −2 −2
[−1 −3 −1]
2
4
2
This matrix in reduced row echelon form is
1 0
[0 1
0 0
Thus
S(1) = {(−๐‘Ÿ, 0, ๐‘Ÿ): ๐‘Ÿ ∈ โ„
= span{(−1, 0, 1)}
Hence ๐œ† = 1 has geometric multiplicity 1.
1
0]
0
127
To solve for the geometric multiplicity of ๐œ† = 2, we substitute 2 for x in the characteristic
matrix obtaining the matrix
−1 −2 −2
[−1 −2 −1]
2
4
3
This matrix in reduced row echelon form is equal to
1 2
[0 0
0 0
0
1]
0
Thus
S(2) = {(−2๐‘Ÿ, ๐‘Ÿ, 0): ๐‘Ÿ ∈ โ„
= span{(−2, 1, 0)}
Hence ๐œ† = 2 has geometric multiplicity 1.
To solve for the geometric multiplicity of ๐œ† = 3, we substitute 3 for x in the characteristic
matrix obtaining the matrix
0 −2 −2
[−1 −1 −1]
2
4
4
This matrix in reduced row echelon form is
1 0
[0 1
0 0
Thus
S(3) = {(0, −๐‘Ÿ, ๐‘Ÿ): ๐‘Ÿ ∈ โ„
= span{(0, −1, 1)}
Hence ๐œ† = 3 has geometric multiplicity 1.
0
1]
0
128
6.4 Diagonalization
Definition 6.4.1: A matrix B is similar to a matrix A if there is a nonsingular matrix P such
that
B = P-1AP
Example 1: Let ๐ด = [
๐‘ƒ−1 ๐ด๐‘ƒ we have
1 −1
−1 −1
1
] and ๐‘ƒ = [
]. Then ๐‘ƒ−1 = [
2 4
2
1
−2
๐ต=[
= [
1
]. If we let ๐ต =
−1
1
1 1 −1 −1 −1
][
][
]
−2 −1 2 4
2
1
3 0
]
0 2
By Definition 6.4.1, B is similar to A.
REMARKS:
1. A is similar to A.
2. If B is similar to A, then A is similar to B.
3. If A is similar to B and B is similar to C, then A is similar to C.
Theorem 6.4.1. Similar matrices have the same characteristic polynomial.
Theorem 6.4.2. Similar matrices have the same eigenvalues and eigenvectors.
Example 2. Let A and B be the matrices of Example 1. The characteristic polynomial of A is
๐‘ฅ−1
1
๐‘“(๐‘ฅ) = |
| = ๐‘ฅ 2 − 5๐‘ฅ + 6
−2 ๐‘ฅ − 4
and the characteristic polynomial of B is
๐‘ฅ−3
0
๐‘“(๐‘ฅ) = |
| = ๐‘ฅ 2 − 5๐‘ฅ + 6
0
๐‘ฅ−2
Thus A and B have the same characteristic polynomial. Consequently, A and B also have the
same eigenvalues and eigenvectors.
129
Definition 6.4.2: We shall say that the matrix A is diagonalizable if it is similar to a diagonal
matrix. In this case we also say that A can be diagonalized.
Example 3. Let A be the matrix of Example 1. Since A is similar to a diagonal matrix then A is
diagonalizable.
Theorem 6.4.3: An ๐‘› x ๐‘› matrix A is diagonalizable if and only if it has ๐‘› linearly independent
eigenvectors. In this case A is similar to a diagonal matrix D, with P-1AP = D, whose diagonal
elements are the eigenvalues of A, while P is a matrix whose columns are ๐‘› linearly
independent eigenvectors of A.
Theorem 6.4.4: A matrix A is diagonalizable if all the roots of its characteristics polynomial
are real and distinct.
Example 4. Let ๐ด = [
1 4
]. The characteristic polynomial of A is
1 −2
๐‘ฅ−1
๐‘“(๐‘ฅ) = |
−1
−4
| = ๐‘ฅ 2 + ๐‘ฅ − 6 = (๐‘ฅ + 3)(๐‘ฅ − 2)
๐‘ฅ+2
Since the eigenvalues are real and distinct then A is diagonalizable.
Example 5. Let ๐ด be the matrix of SAQ6-2. The distinct eigenvalues of A are 1, 2, and 3
−1
−2
hence A is diagonalizable. The linearly independent eigenvectors of A are [ 0 ] , [ 1 ] ,
1
0
0
and [−1]. Thus by Theorem 6.4.3
1
−1 −2 0
1
2
2
๐‘ƒ=[ 0
1 −1] and ๐‘ƒ−1 = [−1 −1 −1] (verify)
1
0
1
−1 −2 −1
Then
๐ท = ๐‘ƒ−1 ๐ด๐‘ƒ
1
= [−1
−1
2
2
3
2
2 −1 −2 0
−1 −1] [ 1
4
1 ][ 0
1 −1]
−2 −1 −2 −4 −1 1
0
1
1
2
2 −1 −2 0
= [−2 −2 −2] [ 0
1 −1]
−3 −6 −3 1
0
1
130
1 0
= [0 2
0 0
0
0]
3
Note that D is a diagonal matrix whose diagonal elements are the eigenvalues of A.
ACTIVITY
๏ƒฉ2 ๏€ญ 2 3 ๏ƒน
1. Let A ๏€ฝ ๏ƒช๏ƒช0 3 ๏€ญ 2๏ƒบ๏ƒบ . Find the characteristic polynomial, eigenvalues, and
๏ƒช๏ƒซ0 ๏€ญ 1 2 ๏ƒบ๏ƒป
eigenvectors of A.
๏ƒฉ3 ๏€ญ 2 1 ๏ƒน
2. Let A ๏€ฝ ๏ƒช๏ƒช0 2 0๏ƒบ๏ƒบ . Find a nonsingular matrix P such that P-1AP is diagonal.
๏ƒช๏ƒซ0 0 0๏ƒบ๏ƒป
3. Let L be a linear transformation represented by the matrix
๏ƒฉ0 0 0 ๏ƒน
A ๏€ฝ ๏ƒช๏ƒช0 1 0 ๏ƒบ๏ƒบ
๏ƒช๏ƒซ1 0 1 ๏ƒบ๏ƒป
Determine the geometric and algebraic multiplicities of its eigenvalues.
131
MODULE 7
INNER PRODUCT SPACES
Introduction
In this section we will discuss inner product, inner product spaces, orthonormal
bases in โ„๐‘› , the Gram-Schmidth orthogonalization process, diagonalization of symmetric
matrix, and quadratic forms.
Objectives
After studying this module, you should be able to:
1. Explain the steps on the diagonalization of the matrix.
2. Enumerate the properties of a diagonizable symmetric matrix.
3. Perform diagonalization of a symmetric matrix by orthogonalization.
7.1 Inner Product in โ„๐’
Definition 7.1.1. The length or magnitude or norm of the vector ๐‘‹ = (๐‘ฅ1 , ๐‘ฅ2 , … , ๐‘ฅ๐‘› ) in โ„๐‘›
is
โ€–๐‘‹โ€– = √๐‘ฅ1 2 + ๐‘ฅ2 2 + … + ๐‘ฅ๐‘› 2
The above formula is also use to define the distance from the point (๐‘ฅ1 , ๐‘ฅ2 , … , ๐‘ฅ๐‘› ) to the
origin.
Example 1. Let ๐‘‹ = (1, 2, −3) be a vector in โ„3 . The length of ๐‘‹ is
โ€–๐‘‹โ€– = √12 + 22 + (−3)2 = √14
Definition 7.1.2. If ๐‘‹ = (๐‘ฅ1 , ๐‘ฅ2 , … , ๐‘ฅ๐‘› ) and ๐‘Œ = (๐‘ฆ1 , ๐‘ฆ2 , … , ๐‘ฆ๐‘› ) are vectors in โ„๐‘› , then their
inner product (also called dot product) is given by
๐‘‹ โˆ™ ๐‘Œ = ๐‘ฅ1 ๐‘ฆ1 + ๐‘ฅ2 ๐‘ฆ2 + โ‹ฏ ๐‘ฅ๐‘› ๐‘ฆ๐‘› .
Example 2. Let ๐‘‹ = (1, 0, 1) and ๐‘Œ = (2, 3, −1). Then their inner product is
๐‘‹ โˆ™ ๐‘Œ = 1 โˆ™ 2 + 0 โˆ™ 3 + 1(−1) = 1
132
Theorem 7.1.1. (Properties of Inner Product) If ๐‘‹, ๐‘Œ and ๐‘ are vectors in โ„๐‘› and ๐‘ is a
scalar, then:
a. ๐‘‹ โˆ™ ๐‘‹ = โ€–๐‘‹โ€–2 ≥ 0, with equality if and only if ๐‘‹ = 0.
b. ๐‘‹ โˆ™ ๐‘Œ = ๐‘Œ โˆ™ ๐‘‹
c. (๐‘‹ + ๐‘Œ) โˆ™ ๐‘ = ๐‘‹ โˆ™ ๐‘ + ๐‘Œ โˆ™ ๐‘
d. (๐‘๐‘‹) โˆ™ ๐‘Œ = ๐‘‹ โˆ™ (๐‘๐‘Œ) = ๐‘(๐‘‹ โˆ™ ๐‘Œ)
NOTE: A vector space V along with an inner product or scalar product operation satisfying
all these properties is called an inner product space.
Example 3.
product.
โ„๐‘› is an inner product space since it satisfies all the properties of inner
Definition 7.1.3. The angle between two nonzero vectors ๐‘‹ and ๐‘Œ is defined as the unique
number ๐œƒ, 0 ≤ ๐œƒ ≤ ๐œ‹, such that
๐‘๐‘œ๐‘ ๐œƒ =
๐‘‹โˆ™๐‘Œ
โ€–๐‘‹โ€–โ€–๐‘Œโ€–
Example 4. Let ๐‘‹ = (0, 0, 1, 1) and ๐‘Œ = (1, 0, 1, 0). Then
โ€–๐‘‹โ€– = √2, โ€–๐‘Œโ€– = √2 and ๐‘‹ โˆ™ ๐‘Œ = 1
Thus
๐‘๐‘œ๐‘ ๐œƒ =
1
(√2)(√2)
1
๐‘๐‘œ๐‘ ๐œƒ = 2
1
๐œƒ = cos−1 (2)
๐œƒ = 60๐‘œ
Definition 7.1.4. Two nonzero vectors ๐‘‹ and ๐‘Œ in โ„๐‘› are said to be orthogonal if ๐‘‹ โˆ™ ๐‘Œ = 0.
If one of the vectors is the zero vector, we agree to say that the vectors are orthogonal.
They are said to be parallel if |๐‘‹ โˆ™ ๐‘Œ| = โ€–๐‘‹โ€–โ€–๐‘Œโ€–. They are in the same direction if ๐‘‹ โˆ™ ๐‘Œ =
โ€–๐‘‹โ€–โ€–๐‘Œโ€–. That is, they are orthogonal if ๐‘๐‘œ๐‘ ๐œƒ = 0, parallel if ๐‘๐‘œ๐‘ ๐œƒ = ±1 and in the same
direction if ๐‘๐‘œ๐‘ ๐œƒ = 1.
133
Example 5. Let ๐‘‹1 = (4, 2, 6, −8), ๐‘‹2 = (−2, 3, −1, −1) and ๐‘‹3 = (−2, −1, −3, 4). Then
๐‘‹1 โˆ™ ๐‘‹2 = 4(−2) + 2 โˆ™ 3 + 6(−1) + (−8)(−1) = 0,
๐‘‹2 โˆ™ ๐‘‹3 = (−2)(−2) + 3(−1) + (−1)(−3) + (−1)(4) = 0 and
๐‘‹1 โˆ™ ๐‘‹3 = 4(−2) + 2(−1) + 6(−3) + (−8)(4) = −60
This shows that ๐‘‹1 and ๐‘‹2 are orthogonal and ๐‘‹2 and ๐‘‹3 are also orthogonal.
Moreover, โ€–๐‘‹1 โ€– = 2√30 and โ€–๐‘‹3 โ€– = √30 then
|๐‘‹1 โˆ™ ๐‘‹3 | = โ€–๐‘‹1 โ€–โ€–๐‘‹3 โ€–
|−60| = (2√30)(√30)
60 = 60
This implies that ๐‘‹1 and ๐‘‹3 are parallel but not in the same direction.
Definition 7.1.5. A unit vector ๐‘ˆ in โ„๐‘› is a vector of unit length. If ๐‘‹ is a nonzero vector,
then the vector
๐‘ˆ= [
1
โ€–๐‘‹โ€–
]๐‘‹
is a unit vector in the direction of ๐‘‹.
Example 6. Consider the vector ๐‘‹ = (0, 4, 2, 3). Since โ€–๐‘‹โ€– = √29 then the vector
๐‘ˆ=
1
√29
(0, 4, 2, 3) = (0,
Is a unit vector in the direction of ๐‘‹.
4
√29
,
2
√29
,
3
)
√29
134
72. Orthonormal Bases in โ„๐’
Definition 7.2.1. A set ๐‘† = {๐‘ฅ1 , ๐‘ฅ2 , … , ๐‘ฅ๐‘˜ } in โ„๐‘› is called orthogonal if any two distinct
vectors in S are orthogonal. An orthonormal set of vectors is an orthogonal set of unit
vectors.
Example 1. Let ๐‘‹1 = (1, 2, −1, 1), ๐‘‹2 = (0, −1, −2, 0) and ๐‘‹3 = (1, 0, 0, −1). Since
๐‘‹1 โˆ™ ๐‘‹2 = 0, ๐‘‹1 โˆ™ ๐‘‹3 = 0 and ๐‘‹2 โˆ™ ๐‘‹3 = 0 then the set {๐‘‹1, ๐‘‹2 , ๐‘‹3 } is orthogonal. The
vectors
๐‘Œ1 = (
1
,
2
√7 √7
,−
1
,
1
) , ๐‘Œ2 = (0, −
√7 √7
1
√5
,−
2
√5
, 0) and ๐‘Œ3 = (
1
√2
, 0, 0, −
1
)
√2
are unit vectors in the direction of ๐‘‹1 , ๐‘‹2 and ๐‘‹3 , respectively. Thus {๐‘Œ1 , ๐‘Œ2 , ๐‘Œ3 } is an
orthonormal set and span the same subspace as {๐‘‹1 , ๐‘‹2 , ๐‘‹3 }, that is, span{๐‘‹1 , ๐‘‹2 , ๐‘‹3 } =
span{๐‘Œ1 , ๐‘Œ2 , ๐‘Œ3 }.
Notice that an orthonormal set is just a set of orthogonal vectors in which each
vector has been normalized to unit length.
Theorem 7.2.1. Let ๐‘† = {๐‘ฅ1 , ๐‘ฅ2 , … , ๐‘ฅ๐‘˜ } be an orthogonal set of nonzero vectors in โ„๐‘› . Then
S is linearly independent.
Example 2. The vectors ๐‘‹1 , ๐‘‹2 and ๐‘‹3 of Example 1 are orthogonal hence they are linearly
independent.
Corollary 7.2.1. An orthonormal set of vectors in โ„๐‘› is linearly independent.
Example 3. The set of vectors {๐‘Œ1 , ๐‘Œ2 , ๐‘Œ3 } of Example 1 is orthonormal hence it is linearly
independent.
Gram-Schmidt Orthogonalization Process
Gram-Schmidth process is a procedure by which an orthonormal set of vectors is
obtained from a linearly independent set of vectors in an inner product space.
135
Definition 7.2.2. (Gram-Schmidt Process) Let W be a nonzero subspace of โ„๐‘› with basis ๐‘† =
{๐‘‹1 , ๐‘‹2 , … , ๐‘‹๐‘š }. Then there exists an orthonormal basis ๐‘‡ = {๐‘1 , ๐‘2 , … , ๐‘๐‘š } for W.
Steps in the Gram-Schmidth process:
STEP 1. Let ๐‘Œ1 = ๐‘‹1.
STEP 2. Compute the vectors ๐‘Œ2 , ๐‘Œ3 , … , ๐‘Œ๐‘š , by the formula
๐‘‹๐‘˜ โˆ™๐‘Œ๐‘—
๐‘Œ๐‘˜ = ๐‘‹๐‘˜ − ∑๐‘˜−1
๐‘—=1 ( ๐‘Œ
๐‘— โˆ™๐‘Œ๐‘—
) ๐‘Œ๐‘— for 2 ≤ ๐‘˜ ≤ ๐‘š
The set of vectors ๐‘† ′ = {๐‘Œ1 , ๐‘Œ2 , … , ๐‘Œ๐‘š } is an orthogonal set.
STEP 3. Normalize each vector in ๐‘†′ to obtain an orthonormal basis for W.
Example 4. Use the Gram-Schmidth process to transform the basis {(1, 1, 1), (0, 1, 1),
(1, 2, 3)} for โ„3 into an orthonormal basis for โ„3 .
Solution: We apply the Gram-Schmidth process to obtain 3 vectors ๐‘1 , ๐‘2 , ๐‘3 which also
span โ„3 and orthogonal to each other.
STEP 1. Let ๐‘Œ1 = ๐‘‹1 = (1, 1, 1).
STEP 2. We now compute for ๐‘Œ2 and ๐‘Œ3 :
๐‘‹2 โˆ™ ๐‘Œ1
๐‘Œ2 = ๐‘‹2 − (
)๐‘Œ
๐‘Œ1 โˆ™ ๐‘Œ1 1
2
= (0, 1, 1) − ( ) (1, 1, 1)
3
2 1 1
= (− , , )
3 3 3
and
๐‘‹3 โˆ™ ๐‘Œ1
๐‘‹3 โˆ™ ๐‘Œ2
๐‘Œ3 = ๐‘‹3 − (
) ๐‘Œ1 − (
)๐‘Œ
๐‘Œ1 โˆ™ ๐‘Œ1
๐‘Œ2 โˆ™ ๐‘Œ2 2
6
1
2 1 1
= (1, 2, 3) − ( ) (1, 1, 1) − (
) (− , , )
3
6/9
3 3 3
136
3
2 1 1
= (1, 2, 3) − (2, 2, 2) − (− , , )
2
3 3 3
1 1
= (0, − , )
2 2
2 1 1
1 1
The set of vectors ๐‘† ′ = {(1, 1, 1), (− 3 , 3 , 3) , (0, − 2 , 2)} is orthogonal.
STEP 3. We normalize each vector found in STEP 2.
Let
๐‘1 =
๐‘Œ1
โ€–๐‘Œ1 โ€–
1
๐‘2 =
๐‘Œ2
1
=
โ€–๐‘Œ2 โ€–
√6/9
=
=
√3
3
๐‘Œ3
โ€–๐‘Œ3 โ€–
=
=
1
,
1
,
1
√3 √3 √3
) , (−
1
,
1
)
2 1 1
2 1 1
√6
2
√2
,
,
1
)
1 1
√2/4
2
1
,
√6 √6 √6
1
2
,
√3 √3 √3
(− 3 , 3 , 3)
(0, − 2 , 2)
1 1
(0, − 2 , 2)
= (0, −
Then ๐‘‡ = {(
1
(− 3 , 3 , 3)
= (−
๐‘3 =
(1, 1, 1) = (
1
1
,
1
)
√2 √2
1
1
1
, ) , (0, − √2 , √2)} is an orthonormal basis for โ„3 .
√6 √6 √6
137
SAQ 7-1
Use the Gram-Schmidt process to find an orthonormal basis for the
subspace W of โ„4 with basis {(1, 1, -1, 0), (0, 2, 0, 1), (-1, 0, 0, 1)}.
ASAQ 7-1
STEP 1. Let ๐‘Œ1 = ๐‘‹1 = (1, 1, −1, 0)
STEP 2. We now solve for ๐‘Œ2 and ๐‘Œ3 :
๐‘‹2 โˆ™ ๐‘Œ1
๐‘Œ2 = ๐‘‹2 − (
)๐‘Œ
๐‘Œ1 โˆ™ ๐‘Œ1 1
2
= (0, 2, 0, 1) − (1, 1, −1, 0)
3
2 4 2
= (− , , , 1)
3 3 3
๐‘‹3 โˆ™ ๐‘Œ1
๐‘‹3 โˆ™ ๐‘Œ2
๐‘Œ3 = ๐‘‹3 − (
) ๐‘Œ1 − (
)๐‘Œ
๐‘Œ1 โˆ™ ๐‘Œ1
๐‘Œ2 โˆ™ ๐‘Œ2 2
1
5/3
2 4 2
= (−1, 0, 0,1) − (− ) (1, 1, −1, 0) −
(− , , , 1)
3
11/3
3 3 3
= (−
4
3
7 6
,− ,− , )
11 11 11 11
2 4 2
4
3
7
6
Thus ๐‘† ′ = {(1, 1, −1, 0), (− 3 , 3 , 3 , 1) , (− 11 , − 11 , − 11 , 11)} is an orthogonal basis for a
subspace W of โ„4 .
138
STEP 3. We normalize each vector in ๐‘†′.
Let ๐‘1 =
๐‘2 =
๐‘Œ1
โ€–๐‘Œ1 โ€–
1
=
(1, 1, -1, 0) = (
√3
,
1
√3 √3
,−
1
√3
,0 )
๐‘Œ2
1
2 4 2
=
(− 3 , 3 , 3 , 1)
โ€–๐‘Œ2 โ€–
√33/9
=
3
2 4 2
๐‘Œ3
โ€–๐‘Œ3 โ€–
1
=
11
1 1
3 3
4
,
,
2
,
3
)
√33 √33 √33 √33
3
7
6
(− 11 , − 11 , − 11 , 11)
4
√110
= (−
2
4
√110/121
=
(− 3 , 3 , 3 , 1)
√33
= (−
๐‘3 =
1
3
7
6
(− 11 , − 11 , − 11 , 11)
4
√110
,−
3
√110
1
3
,−
7
,
6
√110 √110
)
2
4
2
3
4
3
7
6
, , , ) , (− √110 , − √110 , − √110 , √110)} is
33 √33 √33 √33
Thus ๐‘‡ = {(√ , √ , − √ , 0 ) , (− √
an orthonormal basis for the subspace W of โ„4
ACTIVITY
1. Which of the following are orthonormal sets of vectors?
1
1
1 1 1
a. ( , 0, − ) , ( , , ) , (0, 1, 0)
√2
√2
√3 √3 √3
b. (0, 2, 2, 1), (−1, 1, 2, 2), (0, 1, −2, 1)
1
c. (
√6
,−
2
√6
, 0,
1
) , (−
√6
1
√3
,−
1
√3
, 0, −
1
√3
) , (0, 0, 1, 0)
139
2. Use the Gram-Schmidt process to find an orthonormal basis for the subspace of โ„4
with basis {(1, -1, 0, 1), (2, 0, 0, -1), (0, 0, 1, 0)}.
7.3 Diagonalization of Symmetric Matrix
Theorem 7.3.1. All roots of the characteristic polynomial of a symmetric matrix are real
numbers.
−1 0
Example 1. Let ๐ด = [ 0 2
0 0
0
0]. The characteristic polynomial of A is
3
๐‘ฅ+1
๐‘“(๐‘ฅ) = | 0
0
0
0
๐‘ฅ−2
0 |
0
๐‘ฅ−3
= (๐‘ฅ + 1)(๐‘ฅ − 2)(๐‘ฅ − 3)
Clearly the roots of ๐‘“(๐‘ฅ) are all real numbers.
Corollary 7.3.1. If A is a symmetric matrix all of whose eigenvalues are distinct, then A is
diagonalizable.
Example 2. Let A be the matrix of Example 1. The eigenvalues ๐œ†1 = −1, ๐œ†2 = 2 and ๐œ†3 = 3
are real and distinct. Hence A is diagonalizable.
Theorem 7.3.2. If A is a symmetric matrix, then the eigenvectors that belong to distinct
eigenvalues of A are orthogonal.
Example 3. Let A be the matrix of Example 1. To find the associated eigenvectors, we form
the system (๐œ†๐ผ3 − ๐ด)๐‘‹ = 0 and solve for x. Thus
1
0
0
๐‘‹1 = [0], ๐‘‹2 = [1] and X 3 = [0]
0
0
1
140
are the eigenvectors associated with ๐œ†1 = −1, ๐œ†2 = 2 and ๐œ†3 = 3, respectively. It can be
verified that {๐‘‹1 , ๐‘‹2 , ๐‘‹3 } is an orthogonal set of vectors in โ„3 2.
Definition 7.3.1. A nonsingular matrix A is called orthogonal if ๐ด−1 = ๐ด๐‘‡ .
1
0
0
Example 4. Let ๐ด = [0 1/√2 −1/√2].
0 −1/√2 −1/√2
−1
Since ๐ด
1
= [0
0
0
0
1/√2 −1/√2] = ๐ด๐‘‡ then A is an orthogonal matrix.
−1/√2 −1/√2
Theorem 7.3.3. The ๐‘› x ๐‘› matrix A is orthogonal if and only if the columns (and rows) of A
form an orthonormal set of vectors in โ„๐‘› .
Example 5. Let A be the matrix of Example 4. The columns (and rows) of A are of unit length
and are mutually orthogonal. Thus A is orthogonal.
Theorem 7.3.4. If A is symmetric ๐‘› x ๐‘› matrix , then there exists an orthogonal matrix P
such that ๐‘ƒ−1 ๐ด๐‘ƒ = ๐‘ƒ๐‘‡ ๐ด๐‘ƒ = ๐ท, a diagonal matrix, where the columns of P consist of the
linearly independent eigenvectors of A and the diagonal elements of D are the eigenvalues
of A associated with these eigenvectors.
0 −1 −1
Example 6. Let ๐ด = [−1 0 −1]. Find an orthogonal matrix P and a diagonal matrix D
−1 −1 0
such that ๐ท = ๐‘ƒ−1 ๐ด๐‘ƒ = ๐‘ƒ๐‘‡ ๐ด๐‘ƒ.
Solution: The characteristic polynomial of A is
๐‘ฅ
๐‘“(๐‘ฅ) = |1
1
1 1
๐‘ฅ 1| = (๐‘ฅ − 1)2 (๐‘ฅ + 2)
1 ๐‘ฅ
141
The eigenvalues of A are ๐œ†1 = 1, ๐œ†2 = 1 and ๐œ†3 = −2. To find the eigenvector associated
with ๐œ†1 = 1, we solve for the homogeneous system (๐œ†1 ๐ผ3 − ๐ด)๐‘‹ = 0:
1 1 1 ๐‘ฅ1
0
[1 1 1] [๐‘ฅ2 ] = [0]
1 1 1 ๐‘ฅ3
0
(1)
The solution to this system is (verify)
−๐‘Ÿ − ๐‘ 
[ ๐‘Ÿ ] where ๐‘Ÿ, ๐‘  ∈ โ„
๐‘ 
Thus a basis for the solution space of (1) consists of the eigenvectors
−1
−1
๐‘‹1 = [ 1 ] and ๐‘‹2 = [ 0 ]
0
1
However, the two vectors are not orthogonal so we apply the Gram-Schmidt process to
obtain an orthonormal basis for the solution space of (1).
Let ๐‘Œ1 = ๐‘‹1 = (−1, 1, 0). Then
๐‘‹2 โˆ™ ๐‘Œ1
๐‘Œ2 = ๐‘‹2 − (
)๐‘Œ
๐‘Œ1 โˆ™ ๐‘Œ1 1
1
= (−1, 0, 1) − (−1, 1, 0)
2
1 1
= (− , − , 1)
2 2
Normalizing these two vectors we get
๐‘1 =
๐‘‹1
โ€–๐‘‹1 โ€–
=
1
√2
(−1, 1, 0)
2
1
1
๐‘‹
๐‘2 = โ€–๐‘‹2 โ€– =
(− 2 , − 2 , 1)
√6
2
= (−
1
√6
,−
1
,
2
)
√6 √6
142
Thus {๐‘1 , ๐‘2 } is an orthonormal basis of eigenvectors for the solution space of (1).
Next we find the eigenvector associated with ๐œ†3 = −2 by solving the homogeneous system
(−2๐ผ3 − ๐ด)๐‘‹ = 0:
−2 1
1 ๐‘ฅ1
0
[ 1 −2 1 ] [๐‘ฅ2 ] = [0]
1
1 −2 ๐‘ฅ3
0
(2)
The solution to this system is (verify)
๐‘Ÿ
[๐‘Ÿ] where ๐‘Ÿ ∈ โ„
๐‘Ÿ
Thus a basis for the solution space of (2) consists of the eigenvector
1
๐‘‹3 = [1]
1
Normalizing this we get
๐‘‹3
โ€–๐‘‹3 โ€–
๐‘3 =
=
1
√3
=(
1
(1, 1, 1)
,
1
,
1
)
√3 √3 √3
Note that ๐‘3 is orthogonal to both ๐‘1 and ๐‘2 thus {๐‘1 , ๐‘2 , ๐‘3 } is an orthonormal basis of
โ„3 consisting of eigenvectors of A. By Theorem 7.3.4,
−
๐‘ƒ=
1
√2
1
√2
[ 0
1
1
√6
1
√3
1
√6
2
1
−
−
√6
1
− √2
1
and ๐‘ƒ −1 = ๐‘ƒ๐‘‡ = −
√6
√3
√3]
1
[ √3
1
√2
1
− √6
1
√3
0
2
√6
1
√ 3]
143
Hence
๐ท = ๐‘ƒ๐‘‡ ๐ด๐‘ƒ
1
− √2
1
6
= −
√
1
[ √3
−
1
0
√2
1
− √6
1
√3
1
1
√2
1
√2
1
= − √6 − √6
2
0
2
[−1
√6
1 −1
√3]
2
0
2
√6
2
− √1
2
−1 −1
0 −1]
−1 0
1
√2
[
−
1
√2
1
√2
[− √3 − √3 − √3] [ 0
0
1
1
√6
1
√3
1
√6
2
√3
1
−
−
√6
− √1
6
1
−√
6
2
√6
1
√3
1
√3
1
√3]
√3]
1 0 0
= [0 1 0 ]
0 0 −2
Note that D is a diagonal matrix whose main diagonal consists of the eigenvalues of A.
SAQ 7-2
1
Let ๐ด = [0
0
0
0
3 −2]. Find an orthogonal matrix P and a
−2 3
diagonal matrix D such that ๐ท = ๐‘ƒ−1 ๐ด๐‘ƒ = ๐‘ƒ๐‘‡ ๐ด๐‘ƒ.
144
ASAQ 7-2
The characteristic polynomial of A is
๐‘ฅ−1
0
0
๐‘“(๐‘ฅ) = | 0
๐‘ฅ−3
2 |
0
2
๐‘ฅ−3
= (๐‘ฅ − 1)(๐‘ฅ − 3)(๐‘ฅ − 3) − 4(๐‘ฅ − 1)
= (๐‘ฅ − 1)2 (๐‘ฅ − 5)
The eigenvalues of A are ๐œ†1 = 1, ๐œ†2 = 1 and ๐œ†3 = 5. To find the eigenvector associated
with ๐œ†1 = 1, we solve for the homogeneous system (๐œ†1 ๐ผ3 − ๐ด)๐‘‹ = 0:
0 0
0 ๐‘ฅ1
0
[0 −2 2 ] [๐‘ฅ2 ] = [0]
0 2 −2 ๐‘ฅ3
0
(1)
The solution to this system is (verify)
๐‘Ÿ
[๐‘ ] where ๐‘Ÿ, ๐‘  ∈ โ„
๐‘ 
Thus a basis for the solution space of (1) consists of the eigenvectors
1
0
๐‘‹1 = [0] and ๐‘‹2 = [1]
0
1
Note that ๐‘‹1 and ๐‘‹2 are already orthogonal. Next we normalize ๐‘‹2:
0
๐‘‹2
๐‘2 =
= [1/√2]
โ€–๐‘‹2 โ€–
1/√2
Thus {๐‘‹1 , ๐‘2 } is an orthonormal basis of eigenvectors for the solution space of (1).
Next we find the eigenvector associated with ๐œ†3 = 5 by solving the homogeneous system
(5๐ผ3 − ๐ด)๐‘‹ = 0:
145
4 0 0 ๐‘ฅ1
0
[0 2 2] [๐‘ฅ2 ] = [0]
0 2 2 ๐‘ฅ3
0
(2)
The solution to this system is (verify)
0
[−๐‘Ÿ] ๐‘Ÿ ∈ โ„
๐‘Ÿ
Thus a basis for the solution space of (2) consists of the eigenvector
0
๐‘‹3 = [−1]
1
Normalizing ๐‘‹3 we obtain
0
๐‘‹3
๐‘3 =
= [−1/√2]
โ€–๐‘‹3 โ€–
1/√2
Since ๐‘3 is orthogonal to both ๐‘‹1 and ๐‘2 then {๐‘‹1 , ๐‘2 , ๐‘3 } is an orthonormal basis of โ„3
consisting of the eigenvectors of A. Thus
1
0
0
1
0
0
−1
๐‘‡
๐‘ƒ = [0 1/√2 −1/√2] and ๐‘ƒ = ๐‘ƒ = [0 1/√2 1/√2]
0 1/√2 1/√2
0 −1/√2 1/√2
Hence
1
0
0
1
๐ท = [0 1/√2 1/√2] [0
0 −1/√2 1/√2 0
1
= [0
0
0
0
1
1/√2 1/√2] [0
−5/√2 5/√2 0
1 0 0
= [0 1 0 ]
0 0 5
0
0 1
3 −2] [0
−2 3 0
0
0
1/√2 −1/√2]
1/√2 1/√2
0
0
1/√2 −1/√2]
1/√2 1/√2
146
ACTIVITY
−3 0
1. Let ๐ด = [ 0 −2
−1 0
−1
0 ]. Find an orthogonal matrix P and a diagonal matrix D such
−3
that ๐ท = ๐‘ƒ −1 ๐ด๐‘ƒ = ๐‘ƒ๐‘‡ ๐ด๐‘ƒ.
REFERENCES:
1. Kolman. Bernard. Introductory Linear Algebra; 4th edition, New York: Mcmillan
Publishing Company, 1988.
2. Lipschutz, Seymour. Linear Algebra; SI (Metric Edition), Singapore: McGraw-Hill
International Book Company, 1981.
3. Johnson and Riess. Introduction to Linear Algebra; Phil: Addison-Wesley Publishing
Company, Inc. 1981.
4. Perry, William L. Elementary Linear Algebra; USA: McGraw-Hill, Inc. 1988.
Download