1 MODULE 1 MATRICES OVER A FIELD ๐ญ Introduction In this lesson, we will discuss the different operations on matrices and its properties, the transpose of a matrix, the different types of matrices and the special types of square matrices. Objectives 1. 2. 3. 4. 5. After going through this chapter, you are expected to be able to do the following: Define matrix. Give the different types of matrices. Perform fundamental operations on matrices. Find the transpose of a given matrix. Identify special types of square matrices. 1.1 Definition of a Field Definition 1.1.1: By a field ๐น we mean a nonempty set of elements with two laws of combination, which we call addition and multiplication, satisfying the following conditions: ๐น1 Closure Properties To every pair of elements ๐, ๐ ∈ ๐น, ๐ + ๐ ∈ ๐น To every pair of elements ๐, ๐ ∈ ๐น, ๐๐ ∈ ๐น ๐น2 Commutative Laws ๐+๐ =๐+๐ ๐๐ = ๐๐ ∀๐, ๐ ∈ ๐น ๐น3 Associative Laws (๐ + ๐) + ๐ = ๐ + (๐ + ๐) (๐๐)๐ = ๐(๐๐) ∀๐, ๐, ๐ ∈ ๐น ๐น4 Distributive Laws (๐ + ๐)๐ = ๐๐ + ๐๐ ๐(๐ + ๐) = ๐๐ + ๐๐ ∀๐, ๐, ๐ ∈ ๐น ๐น5 Identity Elements ∃0 ∈ ๐น such that ๐ + 0 = ๐ ∀๐ ∈ ๐น ∃1 ∈ ๐น, 1 ≠ 0 such that ๐ โ 1 = ๐ ∀๐ ∈ ๐น 2 ๐น6 Inverse Elements ∀๐ ∈ ๐น, ∃ − ๐ ∈ ๐น such that ๐ + (−๐) = 0 ∀๐ ∈ ๐น, ๐ ≠ 0, ∃๐−1 ∈ ๐น such that ๐๐ −1 = 1 The elements of ๐น are called scalars. The set of real numbers โ and complex numbers โ are examples of fields under the usual addition and multiplication in these sets 1.2 Definition of a Matrix Definition 1.2.1: A matrix over a field ๐น is a rectangular array of elements from ๐น arranged in ๐ horizontal rows and ๐ vertical columns: ๐11 ๐12 … ๐1๐ ๐21 ๐22 … ๐2๐ ๐ด=[ โฎ โฎ โฎ ] ๐๐1 ๐๐2 … ๐๐๐ (1) The ith row of A is [๐๐1 ๐๐2 … ๐1๐ ] (1 ≤ ๐ ≤ ๐); the jth column of A is ๐1๐ ๐2๐ [ โฎ ] (1 ≤ ๐ ≤ ๐). ๐๐๐ The symbol ๐๐,๐ (๐น) denotes the collection of all ๐ x ๐ matrices over ๐น. We usually denote matrices by capital letters. If ๐ is equal to ๐, then the matrix is called a square matrix of order ๐. Example 1: Below are examples of matrices of different sizes. 2 −5 i) ๐ด = [ ] is a 2 x 2 matrix 3 4 3 ii) ๐ต = [−1 0 (A is a square matrix of order 2) 4 −8 −5 5 3 1 ] is a 3 x 4 matrix 7 2 −4 −8 The second row is [−1 5 3 1] and the third column is [ 3 ] 2 3 iii) ๐ถ = [ 1 2 1/2 0 ] is a 2 x 3 matrix 3 4/3 We will use the notation ๐ด = [๐๐๐ ] to denote a matrix A with entries ๐๐๐ . The subscripts ๐ and ๐ will be used to denote the position of an entry ๐๐๐ in a matrix. Thus, the ๐๐th element or entry of a matrix A is the number appearing in the ๐th row and ๐th column of A. 1 0 1 0 4 Example 2: Let ๐ด = [ ] and ๐ต = [5 1 −2 2 3 3 2 −1 4 ]. −2 Then A is a 2 x 3 matrix with ๐12 = 0 (the element appearing in the first row and the second column), ๐13 = 4, and ๐23 = 3; B is a 3 x 3 matrix with ๐12 = 0, ๐13 = −1 , ๐21 = 5 , ๐23 = 4, ๐31 = 3 , and ๐32 = 2 . If A is a square matrix of order ๐, then the numbers ๐11 , ๐22 , … , ๐๐๐ form the main diagonal. Thus in matrix B above, the elements ๐11 = 1, ๐22 = 1, and ๐33 = −2 form the main diagonal. Definition 1.2.2: (Equality of Matrices) Two ๐ x ๐ matrix ๐ด = [๐๐๐ ] and ๐ต = [๐๐๐ ] are said to be equal if ๐๐๐ = ๐๐๐ , 1 ≤ ๐ ≤ ๐ , 1 ≤ ๐ ≤ ๐ , that is, if corresponding elements agree. Example 3: 1 2 ๐ด=[ 2 −5 −3 1 2 ] and ๐ต = [ 4 2 ๐ฅ ๐ค ] are equal if ๐ฅ = −5 and ๐ค = −3. 4 1.3 Matrix Operations and Its Properties Definition 1.3.1: (Addition of Matrices) If ๐ด = [๐๐๐ ] and ๐ต = [๐๐๐ ] are ๐ x ๐ matrices, then the sum of A and B is the ๐ x ๐ matrix ๐ถ = [๐๐๐ ] defined by [๐๐๐ ] = [๐๐๐ ] + [๐๐๐ ] (1 ๏ฃ ๐ ≤ ๐, 1 ๏ฃ ๐ ≤ ๐) That is, C is obtained by adding the corresponding elements of A and B. 4 ๏ฉ1 ๏ญ2 4 ๏น ๏ฉ0 ๏ญ2 4 ๏น Example 1: Let A ๏ฝ ๏ช and B ๏ฝ ๏ช ๏บ ๏บ ๏ซ3 ๏ญ1 5 ๏ป ๏ซ1 ๏ญ3 1 ๏ป Then ๐ด + ๐ต = [ 1+0 3+1 (−2) + (−2) 4 + 4 1 ]= [ (−1) + (−3) 5 + 1 4 −4 8 ] −4 6 Note that only matrices with the same size can be added. Definition 1.3.2: (Scalar Multiplication) If ๐ด = [๐๐๐ ] is an ๐ x ๐ matrix and ๐ is a real number, then the scalar multiple of A by ๐, ๐A, is the ๐ x ๐ matrix ๐ต = [๐๐๐ ] , where bij ๏ฝ ra ij (1 ๏ฃ ๐ ≤ ๐, 1 ๏ฃ ๐ ≤ ๐) That is, B is obtained by multiplying each element of A by ๐. ๏ญ 2๏น ๏ฉ4 3 Example 2: Let ๐ = −3 and A ๏ฝ ๏ช ๏บ . Then ๏ซ2 ๏ญ 5 0 ๏ป 4 ๐๐ด = −3 [ 2 −3(4) −3(3) 3 −2 ] =[ −3(2) −3(−5) −5 0 −3(−2) −12 −9 6 ]= [ ] −3(0) −6 15 0 Definition 1.3.3: (Matrix Multiplication) If ๐ด = [๐๐๐ ] is an ๐ x ๐ matrix and ๐ต = [๐๐๐ ] is an ๐ x ๐ matrix, then the product of A and B is the ๐ x ๐ matrix ๐ถ = [๐๐๐ ], defined by ๐๐๐ = ∑๐๐=1 ๐๐๐ ๐๐๐ To illustrate this, let us consider an ๐ x ๐ matrix ๐ด and an ๐ x ๐ matrix ๐ต. ๐th column of ๐ต ๐th row of ๐ด ๐11 ๐21 โฎ ๐๐1 โฎ [ ๐๐1 ๐12 ๐22 โฎ ๐๐2 โฎ ๐๐2 … ๐1๐ ๐11 ๐12 … ๐1๐ … ๐1๐ … ๐2๐ ๐21 ๐22 … ๐2๐ … ๐2๐ โฎ โฎ โฎ โฎ โฎ … ๐๐๐ ๐๐1 ๐๐2 … ๐๐๐ … ๐๐๐ โฎ [ ] … ๐๐๐ ] Note that the ๐th row of ๐ด and the ๐th column of ๐ต must have the same number of components for product AB to be defined. 5 If we let AB = C then the ๐๐th element of AB is ๐๐๐ = [๐๐1 ๐๐2 Example 3: If ๐ด = [ ๐1๐ ๐ … ๐๐๐ ] 2๐ = ๐๐1 ๐๐๐ + ๐๐2 ๐2๐ + … + ๐๐๐ ๐๐๐ โฎ [๐๐๐ ] 1 3 −2 6 ] and ๐ต = [ ] then −2 5 4 7 −2 ๐11 = [1 3] [ ] = 1(−2) + 3(4) = 10 4 −2 ๐21 = [−2 5] [ ] = (−2)(−2) + 5(4) = 24 4 6 ๐12 = [1 3] [ ] = 1(6) + 3(7) = 27 7 6 ๐22 = [−2 5] [ ] = (−2)(6) + 5(7) = 23 7 10 Thus ๐ด๐ต = [ 24 Similarly, ๐ต๐ด = [ 27 ] 23 (−2)(1) + 6(−2) (−2)(3) + 6(5) −2 6 1 3 ][ ]= [ ] 4(1) + 7(−2) 4(3) + 7(5) 4 7 −2 5 −14 = [ −10 24 ] 47 Example 3 shows that matrix multiplication, in general, is not commutative. 7 −1 2 0 5 Example 4: Let ๐ด = [ ] and ๐ต = [−2 5 4 −3 1 3 2 Find a) AB and b) BA. 1 6 4 −4]. −3 2 6 Solution: a. Note that the number of columns of A is equal to the number of rows of B hence the product AB is defined. By Definition 1.3.3, AB is a 2 x 4 matrix. 2 ๐ด๐ต = [ 4 = [ = [ 7 −1 0 5 ] [−2 5 −3 1 3 2 14 + 0 + 15 28 + 6 + 3 29 8 37 −17 1 6 4 −4] −3 2 (−2) + 0 + 10 (−4) + (−15) + 2 2 + 0 + (−15) 4 + (−12) + (−3) 12 + 0 + 10 ] 24 + 12 + 2 −13 22 ] −11 38 b. The product BA is not defined because the number of columns of B is not equal to number of rows of A. the SAQ 1-1 −2 5 3 −1 ] and ๐ต = [−4 3]. Find 1 −4 3 1 a. ๐ด๐ต and ๐ต๐ด 1 Let ๐ด = [ 2 Solve SAQ1-1 in your notebook and compare your answer with ASAQ1-1. ASAQ 1-1 The number of columns of ๐ด is equal to the number of rows of ๐ต hence ๐ด๐ต is defined. Similarly, the number of columns of ๐ต is equal to the number of rows of ๐ด thus ๐ต๐ด is also defined. Multiplying we have ๏ฉ1(๏ญ2) ๏ซ 3(๏ญ4) ๏ซ (๏ญ1)(3) 1(5) ๏ซ 3(3) ๏ซ (๏ญ1)(1) ๏น AB ๏ฝ ๏ช ๏บ ๏ซ 2(๏ญ2) ๏ซ 1(๏ญ4) ๏ซ (๏ญ4)(3) 2(5) ๏ซ 1(3) ๏ซ (๏ญ4)(1) ๏ป ๏ฉ ๏ญ17 13๏น ๏ฝ๏ช ๏บ ๏ซ ๏ญ20 9 ๏ป 7 and ๏ฉ ๏ญ2 5๏น ๏ฉ8 ๏ญ1 ๏ญ18๏น ๏ฉ1 3 ๏ญ1๏น ๏ช ๏ช ๏บ BA ๏ฝ ๏ช ๏ญ4 3๏บ ๏ช ๏ฝ ๏ช 2 ๏ญ9 ๏ญ8 ๏บ๏บ ๏บ 2 1 ๏ญ4๏ป ๏ช๏ซ3 1 ๏บ๏ป ๏ซ ๏ช๏ซ5 10 ๏ญ7 ๏บ๏ป Remarks: If ๐ด = [๐๐๐ ] is an ๐ x ๐ matrix and ๐ต = [๐๐๐ ] is an ๐ x ๐ matrix then, 1. BA may not be defined as in Example 4.b; this will take place if ๐ ≠ ๐. 2. If BA is defined, which means ๐ = ๐, then BA is ๐ x ๐ while AB is ๐ x ๐; thus if ๐ ≠ ๐, AB and BA are of different size (see ASAQ1-1). 3. If AB and BA are both of the same size, they may be equal. 4. If AB and BA are both of the same size, they may be unequal (see Example 3). Definition 1.3.4: (Additive Inverse of a Matrix) Let ๐ด = [๐๐๐ ]. Then – ๐ด is the matrix obtained by replacing the elements of ๐ด by their additive inverses; that is −๐ด = −[๐๐๐ ] = [−๐๐๐ ] Example 5. Let ๐ด = [ −2 5 1/2 ] then 3 2/3 −17 –๐ด = [ 2 −5 −1/2 ] −3 −2/3 17 Definition 1.3.5: (The Zero Matrix) An ๐ x ๐ matrix ๐ด = [๐๐๐ ] is called a zero matrix and is denoted by 0 if ๐๐๐ = 0 , 1 ๏ฃ ๐ ≤ ๐, 1 ๏ฃ ๐ ≤ ๐; that is all entries are equal to zero. Properties of Matrix Operations Theorem 1.3.1: (Properties of Matrix Addition) Let A, B, and C be ๐ x ๐ matrices then: (a) A + B = B + A (commutative law for matrix addition) (b) A + ( B + C ) = ( A + B ) + C (associative law for matrix addition) (c) A + O = A (The matrix O is the ๐ x ๐ zero matrix) (d) there is a unique ๐ x ๐ matrix (−๐ด) such that ๐ด + (−๐ด) = ๐ We will prove part (a) and leave the proof of the remaining parts as an exercise. 8 Proof of part (a): Let ๐ด = [๐๐๐ ] and ๐ต = [๐๐๐ ] be ๐ x ๐ matrices. By Definition 1.3.1 (1 ๏ฃ ๐ ≤ ๐, 1 ๏ฃ ๐ ≤ ๐) ๐ด + ๐ต = ๐๐๐ + ๐๐๐ = ๐๐๐ + ๐๐๐ since a, b ∈ ๐ then a + b = b + a for any real numbers a and b. = ๐ต + ๐ด Therefore, matrix addition is commutative. Example 6: To illustrate part (a) of Theorem 1.3.1, let ๏ฉ 4 ๏ญ1๏น ๏ฉ 4 6๏น A๏ฝ๏ช and B ๏ฝ ๏ช ๏บ ๏บ ๏ซ3 2 ๏ป ๏ซ10 2 ๏ป Then, ๏ฉ 4 ๏ญ1๏น ๏ฉ 4 6 ๏น ๏ฉ 8 5 ๏น A๏ซ B ๏ฝ ๏ช ๏บ๏ซ๏ช ๏บ๏ฝ๏ช ๏บ ๏ซ 3 2 ๏ป ๏ซ10 2๏ป ๏ซ13 4๏ป and ๏ฉ 4 6 ๏น ๏ฉ 4 ๏ญ1๏น ๏ฉ 8 5 ๏น B๏ซ A๏ฝ ๏ช ๏บ๏ซ๏ช ๏บ๏ฝ๏ช ๏บ ๏ซ10 2 ๏ป ๏ซ 3 2 ๏ป ๏ซ13 4๏ป Example 7: To illustrate part (c) of Theorem 1.3.1, let ๏ฉ1 2 ๏ญ1๏น ๏ฉ0 0 0 ๏น ๏ช ๏บ A ๏ฝ ๏ช 3 2 ๏ญ2๏บ and O ๏ฝ ๏ช๏ช0 0 0๏บ๏บ . Note that O is the 3 x 3 zero matrix. ๏ช๏ซ 4 5 3 ๏บ๏ป ๏ช๏ซ0 0 0๏บ๏ป Then, 1 ๐ด + ๐ = [3 4 2 −1 0 0 2 −2] + [0 0 5 3 0 0 0 1 0] = [3 0 4 Example 8: To illustrate part (d) of Theorem 1.3.1, let ๏ฉ2 2 3 ๏น ๏ฉ ๏ญ2 ๏ญ2 ๏ญ3๏น A๏ฝ๏ช and ๏ญ A ๏ฝ ๏ช ๏บ ๏ญ2 ๏บ๏ป ๏ซ3 ๏ญ1 2 ๏ป ๏ซ ๏ญ3 1 2 −1 2 −2] = ๐ด 5 3 9 Then, ๏ฉ 2 2 3 ๏น ๏ฉ ๏ญ2 ๏ญ2 ๏ญ3๏น ๏ฉ0 0 0 ๏น A ๏ซ (๏ญ A) ๏ฝ ๏ช ๏ฝ ๏บ๏ซ๏ช ๏ญ2๏บ๏ป ๏ช๏ซ0 0 0 ๏บ๏ป ๏ซ3 ๏ญ1 2๏ป ๏ซ ๏ญ3 1 Theorem 1.3.2: (Properties of Matrix Multiplication) Let ๐ด = [๐๐๐ ] be an ๐ x ๐ matrix, ๐ต = ๐๐๐ be an ๐ x ๐ matrix, and ๐ถ = ๐๐๐ be a ๐ x ๐ matrix. Then (a) (๐ด๐ต)๐ถ = ๐ด(๐ต๐ถ) (associative law for matrix multiplication) (b) ๐ด(๐ต + ๐ถ) = ๐ด๐ต + ๐ด๐ถ (left distributive law for matrix multiplication) (c) (๐ด + ๐ต)๐ถ = ๐ด๐ถ + ๐ต๐ถ (right distributive law for matrix multiplication) Proof of part (a): Note that AB is an ๐ x ๐ matrix hence (AB)C is an ๐ x ๐ matrix. Similarly, BC is an ๐ x ๐ matrix hence A(BC) is an ๐ x ๐ matrix. We see that (AB)C and A(BC) have the same size. Now we must show that the corresponding parts of (AB)C and A(BC) are equal. Let AB = D = [๐๐๐ ]. Then ๐๐๐ = ∑๐๐=1 ๐๐๐ ๐๐๐ The ๐๐th component of (AB)C = DC is ∑๐๐ =1 ๐๐๐ ๐๐ ๐ = ∑๐๐ =1(∑๐๐=1 ๐๐๐ ๐๐๐ )๐๐ ๐ ๐ = ∑๐ =1 ∑๐๐=1 ๐๐๐ ๐๐๐ ๐๐ ๐ Now we let BC = E = [๐๐๐ ] then ๐ ๐๐๐ = ∑๐ =1 ๐๐๐ ๐๐ ๐ Thus the ๐๐th component of A(BC) = AE is ∑๐๐=1 ๐๐๐ ๐๐๐ = ∑๐๐=1 ∑๐๐ =1 ๐๐๐ ๐๐๐ ๐๐ ๐ = ๐๐th component of (AB)C. This shows that (๐ด๐ต)๐ถ = ๐ด(๐ต๐ถ). 10 Example 9: To illustrate part (a) of Theorem 1.3.2 we let ๏ฉ1 ๏ฉ 2 3 ๏ญ1 0 ๏น ๏ช2 ๏ฉ3 ๏ญ2 4 ๏น ๏ช ๏บ ๏ช A๏ฝ๏ช ๏บ , B ๏ฝ ๏ช0 2 ๏ญ2 2๏บ , and C ๏ฝ ๏ช0 ๏ญ 1 ๏ญ 2 3 ๏ซ ๏ป ๏ช๏ซ1 0 ๏ญ1 3 ๏บ๏ป ๏ช ๏ซ ๏ญ2 0 ๏ญ2 ๏น ๏ญ3 0 ๏บ๏บ 1 ๏ญ3 ๏บ ๏บ 0 0 ๏ป Then, ๏ฉ 8 ๏ญ10 ๏ญ1๏น ๏ญ18 ๏ญ11๏น ๏ฉ3 ๏ญ2 4๏น ๏ช ๏บ ๏ฝ ๏ฉ4 A( BC ) ๏ฝ ๏ช 0 ๏ญ 8 6 ๏บ ๏ช ๏บ ๏ช ๏บ ๏ซ ๏ญ1 ๏ญ2 3 ๏ป ๏ช ๏ญ5 ๏ญ1 1 ๏บ ๏ซ ๏ญ23 23 ๏ญ8 ๏ป ๏ซ ๏ป and ๏ฉ1 ๏ญ3 8 ๏น ๏ช๏ช 2 ๏ฉ10 5 ( AB)C ๏ฝ ๏ช ๏บ ๏ซ1 ๏ญ7 2 5๏ป ๏ช0 ๏ช ๏ซ ๏ญ2 0 ๏ญ2 ๏น ๏ญ3 0 ๏บ๏บ ๏ฉ 4 ๏ญ18 ๏ญ11๏น ๏ฝ๏ช 1 ๏ญ3 ๏บ ๏ซ ๏ญ23 23 ๏ญ8 ๏บ๏ป ๏บ 0 0 ๏ป This shows that matrix multiplication is associative. Proof of part (b): Let ๐ด = [๐๐๐ ] be an ๐ x ๐ matrix and let ๐ต = [๐๐๐ ] and ๐ถ = [๐๐๐ ] be ๐ x ๐ matrices. Then the ๐๐th element of B + C is ๐๐๐ + ๐๐๐ and the ๐๐th element of A(B + C) is ∑๐๐=1 ๐๐๐ (๐๐๐ + ๐๐๐ ) = ∑๐๐=1 ๐๐๐ ๐๐๐ + ∑๐๐=1 ๐๐๐ ๐๐๐ = ๐๐th component of AB + ๐๐th component of AC Thus, A(B + C) = AB + AC. The proof of part (c) is similar to the proof of part (b) and is left as an exercise. 0๏น ๏ฉ1 ๏ฉ0 2 ๏น ๏ฉ2 2 4๏น ๏ช ๏บ Example 10: Let A ๏ฝ ๏ช , B ๏ฝ ๏ช ๏ญ2 3 ๏บ , and C ๏ฝ ๏ช๏ช ๏ญ1 1 ๏บ๏บ ๏บ ๏ซ ๏ญ4 ๏ญ1 3 ๏ป ๏ช๏ซ5 1 ๏บ๏ป ๏ช๏ซ 2 0 ๏บ๏ป 11 Then 2๏น ๏ฉ1 ๏ฉ2 2 4๏น ๏ช ๏ฉ 24 16 ๏น A( B ๏ซ C ) ๏ฝ ๏ช ๏ญ3 4๏บ๏บ ๏ฝ ๏ช ๏บ ๏บ ๏ช ๏ซ ๏ญ4 ๏ญ1 3 ๏ป ๏ช7 1 ๏บ ๏ซ 20 ๏ญ9๏ป ๏ซ ๏ป and ๏ฉ18 10๏น ๏ฉ6 6 ๏น ๏ฉ 24 16 ๏น AB ๏ซ AC ๏ฝ ๏ช ๏บ๏ซ๏ช ๏บ๏ฝ๏ช ๏บ ๏ซ13 0 ๏ป ๏ซ7 ๏ญ9๏ป ๏ซ 20 ๏ญ9๏ป Theorem 1.3.3: (Properties of Scalar Multiplication) If r and s are real numbers and A and B are matrices then (a) r(sA) = (rs)A (b) (r + s)A = rA + sA (c) r(A + B) = rA + rB (d) A(rB) = r(AB) = (rA)B We prove part (c) of the theorem and leave the proof of the other parts as an exercise. Proof: Let ๐ด = [๐๐๐ ] and ๐ต = [๐๐๐ ] be ๐ x ๐ matrices. Then the ๐๐th entry of ๐(๐ด + ๐ต) = ๐(๐๐๐ + ๐๐๐ ) definition of scalar multiplication = ๐๐๐๐ + ๐๐๐๐ distributive property of real numbers = ๐๐th entry of ๐๐ด + ๐๐th entry of ๐๐ต definition of scalar multiplication Hence ๐(๐ด + ๐ต) = ๐๐ด + ๐๐ต. Example 11: To illustrate part (d) of Theorem 1.3.3 we let Then, 3 1 3 2 ๐ = −3, ๐ด = [ ], and ๐ต = [2 2 −1 3 0 ๐ด(๐๐ต) = [ 3 −4 1 3 2 ] ((−3) [2 1 ]) 2 −1 3 0 5 −9 12 1 3 2 = [ ] [−6 −3 ] 2 −1 3 0 −15 −4 1] 5 12 −27 −27 = [ ] −12 −18 and 3 −4 1 3 2 ๐(๐ด๐ต) = −3 ([ ] [2 1 ]) 2 −1 3 0 5 = −3 [ 9 9 ] 4 6 −27 −27 = [ ] −12 −18 Hence ๐ด(๐๐ต) = ๐(๐ด๐ต). 1.4 Transpose of a Matrix Definition 1.4.1: (The Transpose of a Matrix) If ๐ด = [๐๐๐ ] is an ๐ x ๐ matrix, then the ๐ x ๐ matrix AT ๏ฝ [aij ] , where T aij ๏ฝ a ji T (1 ๏ฃ ๐ ≤ ๐, 1 ๏ฃ ๐ ≤ ๐) is called the transpose of A. Thus the transpose of A is obtained by interchanging the rows and columns of A. 4 −2 −3 Example 1: Let ๐ด = [ ], ๐ต = [3 0 −5 −2 4 0 ๐ด๐ = [−2 −5] −3 −2 2 ], and ๐ถ = [ −1]. Then 5 −1 4 Note that the first row of A became the first column of ๐ด๐ and the second row of A became the second column of ๐ด๐ . Similarly 3 ๐ต ๐ = [ 5 ] and ๐ถ ๐ = [2 −1 −1 4] 13 Theorem 1.4.1: (Properties of Transpose) If r is a scalar and A and B are matrices, then (a) (AT)T = A (b) (A + B)T = AT + BT (c) (AB)T = BTAT (d) (rA)T = rAT Proof of part (a): Let ๐ด = [๐๐๐ ]. Then the ๐๐th entry of A is ๐๐๐ = ๐๐๐๐ definition of transpose of a matrix = (๐๐ )๐๐๐ definition of a transpose of a matrix ๐ ๐ Hence (๐ด ) = ๐ด. Proof of part (b): Let [(๐ด + ๐ต)๐ ]๐๐ be the ๐๐th component of (๐ด + ๐ต)๐ and (๐ด + ๐ต)๐๐ denote the ๐๐th component of ๐ด + ๐ต. Then [(๐ด + ๐ต)๐ ]๐๐ = (๐ด + ๐ต)๐๐ definition of a transpose = ๐๐๐ + ๐๐๐ definition of addition of matrices = (๐ด๐ )๐๐ + (๐ต ๐ )๐๐ definition of a transpose = ๐๐th entry of ๐ด๐ + ๐๐th entry of ๐ต ๐ Thus (๐ด + ๐ต)๐ = ๐ด๐ + ๐ต ๐ . The proof of parts (c) and (d) is left as an exercise. Example 2: To illustrate part (c) of Theorem 1.4.1 we let ๏ฉ0 ๏ญ1๏น B ๏ฝ ๏ช๏ช 2 ๏ญ2๏บ๏บ ๏ช๏ซ3 ๏ญ1๏บ๏ป 2 ๏น ๏ฉ1 ๏ฉ0 2 3 ๏น ๏ช ๏ฉ0 7 ๏น ๏บ ๏ฝ ๏ฉ0 7 ๏น T T T ๏ญ 3 ๏ญ 1 Then ( AB) ๏ฝ ๏ช and B A ๏ฝ ๏ช ๏บ ๏บ ๏บ ๏ช ๏บ ๏ช ๏ซ 3 ๏ญ3๏ป ๏ซ ๏ญ1 ๏ญ2 ๏ญ1๏ป ๏ช 2 3 ๏บ ๏ซ3 ๏ญ3๏ป ๏ซ ๏ป ๏ฉ1 ๏ญ3 2๏น A๏ฝ๏ช ๏บ and ๏ซ 2 ๏ญ1 3 ๏ป 14 1.5 Special Types of Square Matrices Definition 1.5.1: A square matrix ๐ด = [๐๐๐ ] for which every term off the main diagonal is zero, that is, ๐๐๐ = 0 for ๐ ≠ ๐, is called a diagonal matrix. Example 1: The following are diagonal matrices. ๏ฉ ๏ญ5 0 0๏น ๏ฉ3 0 ๏น and H ๏ฝ ๏ช๏ช 0 ๏ญ8 0 ๏บ๏บ G๏ฝ๏ช ๏บ ๏ซ0 ๏ญ1๏ป ๏ช๏ซ 0 0 3๏บ๏ป Definition 1.5.2: A diagonal matrix ๐ด = [๐๐๐ ], for which all terms on the main diagonal are equal, that is, ๐๐๐ = ๐ for ๐ = ๐ and ๐๐๐ = 0 for ๐ ≠ ๐, is called a scalar matrix. Example 2: The following are examples of scalar matrices. 3 ๐ต = [0 0 0 0 −7 0 ] 3 0] and ๐ถ = [ 0 −7 0 3 Definition 1.5.3: A matrix ๐ด = [๐๐๐ ] is called upper triangular if a ij ๏ฝ 0 for i > j. It is called lower triangular if a ij ๏ฝ 0 for i < j. Definition 1.5.4: A matrix ๐ด = [๐๐๐ ] is called strictly upper triangular if a ij ๏ฝ 0 for i is called strictly lower triangular if a ij ๏ฝ 0 for i Example 3: 0 ๐ท = [0 0 8 0 0 j. It j. ๏ฉ๏ญ 2 0 0๏น ๏ฉ1 3 5 ๏น 0 0 0 ๏ช ๏บ ๏ช ๏บ Let A = ๏ช0 2 ๏ญ 1๏บ , B = ๏ช 5 ๏ญ 1 0๏บ , ๐ถ = [2 0 0] and 3 7 0 ๏ช๏ซ 6 ๏ญ 3 4๏บ๏ป ๏ช๏ซ0 0 ๏ญ 4๏บ๏ป 4 −9] 0 Matrix A is upper triangular since all entries below the main diagonal are zero. Matrix B is lower triangular since all entries above the main diagonal are zero while C and D are strictly lower triangular and strictly upper triangular, respectively. Strictly upper/lower triangular matrices are upper/ lower triangular matrices with all entries on the main diagonal equal to zero. 15 There are two important types of square matrices that can be defined in terms of the transpose operation. These are symmetric and skew symmetric matrices which we define as follows: Definition 1.5.5: A matrix ๐ด = [๐๐๐ ] is called symmetric if AT = A. ๏ฉ 1 ๏ญ2 ๏ญ3๏น 1 −2 −3 ๏ช ๏บ ๐ Example 4: Let A ๏ฝ ๏ช ๏ญ2 5 0 ๏บ then ๐ด = [−2 5 0 ]. Since ๐ด๐ = ๐ด then A is −3 0 4 ๏ช๏ซ ๏ญ3 0 4 ๏บ๏ป symmetric. SAQ 1-2 Let ๐ด and ๐ต be symmetric matrices. Show that ๐ด + ๐ต is symmetric. ASAQ 1-2 (๐ด + ๐ต)๐ = ๐ด๐ + ๐ต ๐ =๐ด+๐ต Part (b) of Theorem 1.4.1 Definition of symmetric matrix Hence ๐ด + ๐ต is symmetric. Definition 1.5.6: A matrix ๐ด = [๐๐๐ ] is called skew symmetric if ๐ด๐ = −๐ด. ๏ฉ 0 ๏ญ3 2 ๏น Example 5: The matrix B ๏ฝ ๏ช๏ช 3 0 4 ๏บ๏บ is skew symmetric because ๏ช๏ซ ๏ญ2 ๏ญ4 0 ๏บ๏ป 0 3 ๐ต ๐ = [−3 0 2 4 −2 −4] = −๐ต. 0 16 SAQ 1-3 Let ๐ด and ๐ต be skew-symmetric matrices. Show that ๐ด + ๐ต is skew-symmetric. ASAQ 1-3 (๐ด + ๐ต)๐ = ๐ด๐ + ๐ต ๐ = (−๐ด) + (−๐ต) = −(๐ด + ๐ต) Part (b) of Theorem 1.4.1 Definition of skew-symmetric matrix Thus ๐ด + ๐ต is skew-symmetric. Definition 1.5.7: (The Identity Matrix) The ๐ x ๐ matrix ๐ผ๐ = [๐๐๐ ] defined by ๐๐๐ = 1 if ๐ = ๐, ๐๐๐ = 0 if ๐ ≠ ๐, is called the ๐ x ๐ identity matrix of order ๐. Example 6: Below are examples of identity matrices. ๏ฉ1 0 0 ๏น I 3 ๏ฝ ๏ช๏ช0 1 0๏บ๏บ ๏ซ๏ช0 0 1 ๏บ๏ป ๏ฉ1 ๏ช0 I4 ๏ฝ ๏ช ๏ช0 ๏ช ๏ซ0 is a 3 x 3 identity matrix of order 3 0 0 0๏น 1 0 0 ๏บ๏บ is a 4 x 4 identity matrix of order 4 0 1 0๏บ ๏บ 0 0 1๏ป Note that an identity matrix is a scalar matrix where all the elements on the main diagonal are 1. 17 Remarks: The identity matrix ๐ผ๐ functions for ๐ x ๐ matrices the way the number 1 functions for real numbers. In other words, the identity matrix ๐ผ๐ is actually a multiplicative identity for ๐ x ๐ matrices. That is ๐ด๐ผ๐ = ๐ผ๐ ๐ด = ๐ด for every ๐ x ๐ matrix A. ๏ฉ4 0 1๏น Example 7: Let A ๏ฝ ๏ช๏ช 5 ๏ญ3 2 ๏บ๏บ . Then ๏ช๏ซ 2 ๏ญ1 4 ๏บ๏ป ๏ฉ 4 0 1 ๏น ๏ฉ1 0 0 ๏น ๏ฉ 4 0 1 ๏น AI3 ๏ฝ ๏ช๏ช5 ๏ญ3 2๏บ๏บ ๏ช๏ช0 1 0๏บ๏บ ๏ฝ ๏ช๏ช5 ๏ญ3 2 ๏บ๏บ ๏ฝ A ๏ช๏ซ 2 ๏ญ1 4๏บ๏ป ๏ช๏ซ0 0 1 ๏บ๏ป ๏ช๏ซ 2 ๏ญ1 4๏บ๏ป and ๏ฉ1 0 0 ๏น ๏ฉ 4 0 1 ๏น ๏ฉ 4 0 1 ๏น I 3 A ๏ฝ ๏ช๏ช0 1 0๏บ๏บ ๏ช๏ช5 ๏ญ3 2๏บ๏บ ๏ฝ ๏ช๏ช5 ๏ญ3 2๏บ๏บ ๏ฝ A ๏ช๏ซ0 0 1 ๏บ๏ป ๏ช๏ซ 2 ๏ญ1 4๏บ๏ป ๏ช๏ซ 2 ๏ญ1 4๏บ๏ป Hence, AI3 = I3A = A Suppose that A is a square matrix. If p is a positive integer, then we define 1. Ap = A.A…A p factors If A is n x n, we also define 2. A0 = In 3. ApAq = Ap+q 4. (Ap)q = Apq NOTE: 1. The rule (AB)P = APBP holds only if AB = BA. 2. The rule “if AB = 0 then either A = 0 or B = 0” is not true for matrices. 18 ๏ฉ2 Example 8: Let A ๏ฝ ๏ช ๏ซ ๏ญ8 Then, ๏ฉ2 AB ๏ฝ ๏ช ๏ซ ๏ญ8 ๏ญ3๏น ๏ฉ3 6๏น and B ๏ฝ ๏ช ๏บ ๏บ 12 ๏ป ๏ซ 2 4๏ป ๏ญ3๏น ๏ฉ 3 6 ๏น ๏ฉ0 0๏น ๏ฝ 12 ๏บ๏ป ๏ช๏ซ 2 4๏บ๏ป ๏ช๏ซ0 0๏บ๏ป ACTIVITY 1 1. Let ๐ด = [ 3 2 0 3 5 1 2 4 3 5 ], ๐ต = [3 4 ], ๐ถ = [−2 7 2 ], ๐ท = [ ], −1 5 2 −4 1 −5 6 4 −3 2 4 5 −5 4 ๐ธ = [0 1 4], and ๐น = [ ]. 2 3 3 −2 1 If possible, compute a. ( BT + A )C b. AB c. 2D – 3F d) ( C + E )T e) AB + DF f) ( 3C – 2E )TB 2. Let ๐ด and ๐ต be skew-symmetric matrices, show ๐ด๐ต is symmetric if and only if ๐ด๐ต = ๐ต๐ด. 3. Let ๐ด be an ๐ x ๐ matrix. Show that a. ๐ด + ๐ด๐ is symmetric. b. ๐ด − ๐ด๐ is skew-symmetric. c. ๐ด๐ด๐ and ๐ด๐ ๐ด are symmetric. ๏ฉ 4 2๏น 4. Let A ๏ฝ ๏ช ๏บ . Find ๏ซ1 3 ๏ป a) A2 + 3A b) 2A3 + 3A2 + 4A + 5I2 19 MODULE 2 LINEAR EQUATIONS AND MATRICES Introduction In this lesson we will discuss the Gauss-Jordan reduction method and the Gaussian Elimination method and their application to the solution of linear systems, inverse of a matrix and the practical method for finding the inverse of a matrix, determinants and its properties, cofactor expansion and Cramer’s rule. Objectives After going through the lessons in this chapter, you are expected to be able to do the following: 1. Explain solution of system of linear equations by substitution and by elimination through simultaneous equations. 2. Transform a given matrix into a row-echelon and reduced row-echelon form. 3. Solve systems of linear equations using the Gauss-Jordan reduction method and the Gaussian Elimination method. 4. Define homogenous systems. 5. Differentiate between singular and non-singular matrices. 6. Enumerate the properties of the Inverse. 7. Find the inverse of a given matrix. 8. Solve systems of linear equations using the inverse of a matrix. 9. Define permutation. 10. Evaluate the determinant of a matrix using permutation. 11. Reduce the problem of evaluating determinants by co-factor expansion. 12. Find the inverse of a matrix using determinants and co-factor expansion. 13. Solve systems of linear equations by using the Cramer’s Rule. 2.1 Solutions of Systems of Linear Equations In this section we introduce a systematic technique of eliminating unknowns in solving systems of linear equations. To illustrate this technique, let us consider the following system: Example 1: 3๐ฅ1 + 2๐ฅ2 − ๐ฅ3 = −2 2๐ฅ1 − 3๐ฅ2 + 2๐ฅ3 = 14 ๐ฅ1 + 2๐ฅ2 + 3๐ฅ3 = 6 (1) 20 Interchanging the first and third equations gives us ๐ฅ1 + 2๐ฅ2 + 3๐ฅ3 = 6 2๐ฅ1 − 3๐ฅ2 + 2๐ฅ3 = 14 3๐ฅ1 + 2๐ฅ2 − ๐ฅ3 = −2 (2) If we multiply the first equation in (2) by – 2 and add the result to the second equation we get −2๐ฅ1 − 4๐ฅ2 − 6๐ฅ3 = −12 2๐ฅ1 − 3๐ฅ2 + 2๐ฅ3 = 14 −7๐ฅ2 − 4๐ฅ3 = 2 The new equation obtained may replace either the first or second equation in system (2), (the two equations used to obtain it). Next let us multiply the first equation in (2) by – 3 and add the result to the third equation. This gives us −3๐ฅ1 − 6๐ฅ2 − 9๐ฅ3 = −18 3๐ฅ1 + 2๐ฅ2 − ๐ฅ3 = −2 or −5๐ฅ2 − 10๐ฅ3 = −20 ๐ฅ2 + 2๐ฅ3 = 4 (dividing both sides of the equation by –5 ) The new equation obtained may replace either the first or the third equation in system (2). Thus if we replace the second equation in (2) by −7๐ฅ2 − 4๐ฅ3 = 2 and the third equation by ๐ฅ2 + 2๐ฅ3 = 4, we obtain the new system: ๐ฅ1 + 2๐ฅ2 + 3๐ฅ3 = 6 −7๐ฅ2 − 4๐ฅ3 = 2 ๐ฅ2 + 2๐ฅ3 = 4 (3) Note that the variable ๐ฅ1 has been eliminated from the second and third equations of system (3). This new system is equivalent to the original system (1). Next we multiply the third equation of (3) by 7 and add the result to the second equation. This gives us 7 ๐ฅ2 + 14๐ฅ3 = 28 −7๐ฅ2 − 4๐ฅ3 = 2 or 10๐ฅ3 = 30 ๐ฅ3 = 3 (dividing both sides of the equation by 10) 21 The new equation obtained may replace either the second or third equation in (3). Replacing the third equation we get an equivalent system ๐ฅ1 + 2๐ฅ2 + 3๐ฅ3 = 6 −7๐ฅ2 − 4๐ฅ3 = 2 ๐ฅ3 = 3 (4) To solve for the variable ๐ฅ2 we substitute ๐ฅ3 = 3 to the second equation in (4): −7๐ฅ2 – 4(3) = 2 ๐ฅ2 = −2 To solve for the variable ๐ฅ1 , we substitute ๐ฅ2 = −2 and ๐ฅ3 = 3 to the first equation in (4): ๐ฅ1 + 2(−2) + 3(3) = 6 ๐ฅ1 = 1 Thus the solution to the system (1) is the ordered triple (1, 2, 3). The method used here is called the Gaussian elimination. The objective of Gaussian elimination is to reduce a given system to triangular or echelon (staircase pattern) form and then use back substitution to find the solution of the system. The Gaussian elimination was modified by Carl Friedrich Gauss (1777 – 1855) and Wilhelm Jordan (1842 – 1899). They called it the Gauss-Jordan elimination. To illustrate it, let us start from system (3) of Example 1. Example 2: ๐ฅ1 + 2๐ฅ2 + 3๐ฅ3 = 6 −7๐ฅ2 − 4๐ฅ3 = 2 ๐ฅ2 + 2๐ฅ3 = 4 (3) Interchange the second and third equations in (3): ๐ฅ1 + 2๐ฅ2 + 3๐ฅ3 = 6 ๐ฅ2 + 2๐ฅ3 = 4 −7๐ฅ2 − 4๐ฅ3 = 2 (4) Now let us multiply the second equation in (4) by – 2 and add the result to the first equation. This gives us ๐ฅ1 + 2๐ฅ2 + 3๐ฅ3 = 6 −2 ๐ฅ2 − 4๐ฅ3 = −8 ๐ฅ1 − ๐ฅ3 = −2 22 Next we multiply the second equation in (4) by 7 and add the result to the third equation. This gives us 7 ๐ฅ2 + 14๐ฅ3 −7๐ฅ2 − 4๐ฅ3 10๐ฅ3 or ๐ฅ3 = = = = 28 2 30 3 Replacing the first equation in (4) by ๐ฅ1 − ๐ฅ3 = −2 and the third equation by ๐ฅ3 = 3, we obtain the new system ๐ฅ1 − ๐ฅ3 = −2 ๐ฅ2 + 2๐ฅ3 = 4 ๐ฅ3 = 3 (5) Note that the variable ๐ฅ2 has been eliminated from the first and third equations of (5). This time let us add the first equation in (5) to the third equation. This gives us ๐ฅ1 ๐ฅ1 − ๐ฅ3 = −2 ๐ฅ3 = 3 = 1 Finally we multiply the third equation in (5) by – 2 and add the result to the second equation. ๐ฅ2 + 2๐ฅ3 = 4 −2๐ฅ3 = 3 ๐ฅ2 = −2 System (5) now becomes ๐ฅ1 ๐ฅ2 = 1 = −2 ๐ฅ3 = 3 Thus the solution to the system is the ordered triple (1, -2, 3). The Gauss-Jordan elimination is an algorithm that reduces a given system to reduced row echelon form or row canonical form. This method is less efficient than the Gaussian elimination in solving systems of equations however, it is well suited for calculating the inverse of a matrix which we will discuss in a later section. 23 2.2 Elementary Row Operations A system of linear equations can be placed in matrix form. This notation is very efficient especially when we are dealing with large systems. To begin, let us consider an ๐ x ๐ system of linear equations ๐11 ๐ฅ1 + ๐12 ๐ฅ2 + … + ๐1๐ ๐ฅ๐ = ๐1 ๐21 ๐ฅ1 + ๐22 ๐ฅ2 + … + ๐2๐ ๐ฅ๐ = ๐2 โฎ ๐๐1 ๐ฅ1 + ๐๐2 ๐ฅ2 + … + ๐๐๐ ๐ฅ๐ = ๐๐ (1) The coefficients of this linear system can be written in a rectangular array having ๐ rows and ๐ columns, and we designate this array as ๐ด: ๐11 ๐12 … ๐1๐ ๐21 ๐22 … ๐2๐ โฎ โฎ โฎ ๐ด= ๐๐1 ๐๐2 … ๐๐๐ [ ] This ๐ x ๐ matrix ๐ด is called the coefficient matrix for the given system (1) above. If we add another column in ๐ด to include the constants b1, b2, …, bm , we will have a matrix that expresses compactly all the relevant information contained in (1). Such a matrix is called the augmented matrix for (1), and is usually denoted as [AโB]. If we let C = [AโB] be the augmented matrix for the system (1), then C is the ๐ x ( ๐+1 ) matrix given by ๐11 ๐12 … ๐1๐ ๐1 ๐21 ๐22 … ๐2๐ ๐2 C= โฎ โฎ โฎ โฎ ๐๐1 ๐๐2 … ๐๐๐ ๐๐ [ ] We also write ๐ด๐ = ๐ต, where ๐11 ๐12 … ๐1๐ ๐21 ๐22 … ๐2๐ โฎ โฎ โฎ ๐ด= and ๐๐1 ๐๐2 … ๐๐๐ [ ] represent the coefficients and ๐ฅ1 ๐ฅ2 ๐=[ โฎ ] ๐ฅ๐ ๐1 ๐2 ๐ต=[ ] โฎ ๐๐ 24 represents the variables. For example, the array 1 −3 0 2 [0 4 5 −1 2 −1 3 0 2 −3] 4 represents the system of 3 linear equations ๐ฅ1 − 3๐ฅ2 + 2๐ฅ4 = 2 4๐ฅ2 + 5๐ฅ3 − ๐ฅ4 = −3 2๐ฅ1 − ๐ฅ2 + 3๐ฅ3 = 4 The task of this section is to manipulate the augmented matrix representing a given linear system into reduced row echelon form. But before we continue, let us introduce some terminology. We have seen from Examples 1 and 2 that three important operations were applied to solve the given system. These are: multiplying or dividing both sides of an equation by a nonzero number, adding a multiple of one equation to another equation, and interchanging two equations. These three operations when applied to the rows of an augmented matrix are called elementary row operations and will result in the matrix of an equivalent system. Definition 2.2.1: An elementary row operation on an ๐ x ๐ matrix ๐ด = [๐๐๐ ] is any one of the following operations: 1. Interchange rows ๐ and ๐ of A. 2. Multiply row ๐ of ๐ด by any nonzero real number. 3. Add ๐ times row ๐ of ๐ด to row ๐ of ๐ด, ๐ ≠ ๐. The process of applying elementary row operations to simplify an augmented matrix is called row reduction. The following notations will be used to denote the elementary row operation used: 1. ๐ ๐ ๐ ๐ means interchanging the ๐th row and the ๐th row 2. ๐๐ ๐ ๐ ๐ means that the ๐th row is replace with ๐ times the ๐th row. 3. ๐๐ ๐ + ๐ ๐ ๐ ๐ means that the ๐th row is replaced with the sum of the ๐th row and the ๐th row multiplied by ๐. Before we illustrate the three elementary operations, let us define first a matrix in reduced row echelon form. 25 Definition 2.2.2: An ๐ x ๐ matrix is said to be in reduced row echelon form when it satisfies the following properties: a) All rows consisting entirely of zeros, if any, are at the bottom of the matrix. b) The first nonzero entry in each row that does not consist entirely of zeros is a 1, called the leading entry(pivot) of its row. c) If rows ๐ and ๐ + 1 are two successive rows that do not consist entirely of zeros, then the leading entry of row ๐ + 1 is to the right of the leading entry of row ๐. d) If a column contains a leading entry of some row, then all other entries in that column are zero. Note that a matrix in reduced row echelon form might not have any rows that consist entirely of zeros. Example 1: Below are matrices in reduced row echelon form: ๏ฉ1 0 0 ๏ญ4 ๏น A ๏ฝ ๏ช๏ช0 1 0 3 ๏บ๏บ ๏ช๏ซ0 0 1 2 ๏บ๏ป ๏ท The first nonzero entry in each row is 1 called the pivot. ๏ท The leading entry (pivot) for row 2 is to the right of the leading entry for row 1 and the leading entry for row 3 is to the right of the leading entry for row 2. ๏ท All other entries for the columns containing a leading entry are zero. ๏ท There are entries other than zero in column 4 since there is no pivot in that column. ๏ฉ1 2 0 0๏น B ๏ฝ ๏ช๏ช0 0 1 0๏บ๏บ ๏ช๏ซ0 0 0 1 ๏บ๏ป ๏ท The leading entry for row 2 appears in column 3 and the leading entry for row 3 appears in column 4. ๏ท Column 2 does not contain a pivot. ๏ฉ1 ๏ช0 C๏ฝ๏ช ๏ช0 ๏ช ๏ซ0 ๏ท All rows consisting of zeros are at the bottom of the matrix. ๏ท Columns 1 and 3 contain a pivot hence all other entries in those columns are zero. 0 0 1/ 3๏น 0 1 0 ๏บ๏บ 0 0 0 ๏บ ๏บ 0 0 0 ๏ป 26 Example 2: Below are matrices not in reduced row echelon form: ๏ฉ1 2 0 4๏น D ๏ฝ ๏ช๏ช0 0 0 0๏บ๏บ ๏ช๏ซ0 0 1 3 ๏บ๏ป ๏ท Fails property (a), the row consisting entirely of zeros must be at the bottom of the matrix. ๏ท Fails property (b), the first nonzero entry in row 2 is not equal to 1. ๏ท Fails property (d), column 3 contains a leading entry but the other entries in that column are not equal to zero. 4๏น ๏ฉ1 0 3 E ๏ฝ ๏ช๏ช0 2 ๏ญ 2 5 ๏บ๏บ ๏ช๏ซ0 0 1 2๏บ๏ป ๏ฉ1 ๏ช0 F๏ฝ๏ช ๏ช0 ๏ช ๏ซ0 0 3 4๏น 1 ๏ญ 2 5๏บ๏บ 1 2 2๏บ ๏บ 0 0 0๏ป ๏ท Fails property (c), the leading entry of the third row must be to the right of the leading entry of row 2. Example 3: The three elementary operations are illustrated below: ๏ฉ0 1 0 2 ๏น Let A ๏ฝ ๏ช๏ช2 3 0 ๏ญ 4๏บ๏บ ๏ช๏ซ2 1 6 9 ๏บ๏ป (a) Interchanging rows 1 and 3 of A, (๐ 1 ๐ 3 ) , we obtain: ๏ฉ2 1 6 9 ๏น B ๏ฝ ๏ช๏ช2 3 0 ๏ญ 4๏บ๏บ ๏ช๏ซ0 1 0 2 ๏บ๏ป (b) Replacing row 1 of B by ½ times row 1 of B, ( 12๐ 1 1 9 1 2 3 2 ๐ถ = [ 2 3 0 − 4] 0 1 0 2 ๐ 1 ) , we obtain: 27 (c) Replacing row 2 by the sum of -2 times row 1 of ๐ถ plus row 2 of ๐ถ, (−2๐ 1 + ๐ 2 ๐ 2 ) we obtain: 1 9 1 2 3 2 ๐ท = [ 0 2 − 6 − 13] 0 1 0 2 Definition 2.2.3: An ๐ x ๐ matrix ๐ด is said to be row equivalent to an ๐ x ๐ matrix ๐ต if ๐ต can be obtained by applying a finite sequence of elementary row operations to ๐ด. Hence in Example 3, ๐ด is row equivalent to ๐ท because ๐ท is obtained from ๐ด after applying a series of elementary row operations on A. NOTE: 1. Every matrix is row equivalent to itself. 2. If A is row equivalent to B, then B is row equivalent to A. 3. If A is row equivalent to B and B is row equivalent to C, then A is row equivalent to C. Theorem 2.2.1: Every nonzero ๐ x ๐ matrix is row equivalent to a unique matrix in reduced row echelon form. 2.3 The Gauss-Jordan Method The Gauss-Jordan method is a systematic technique for applying matrix row transformations in an attempt to reduce a matrix to diagonal form also called reduced row echelon form. The following are steps in using the Gauss-Jordan method to put a matrix into reduced row echelon form: STEP 1: Form the augmented matrix [ AโB ]. STEP 2: Obtain 1 as the first element of the first column. STEP 3: Use the first row to transform the remaining entries in the first column to 0. STEP 4: Obtain 1 as the second entry in the second column. STEP 5: Use the second row to transform the remaining entries in the second column to 0. STEP 6: Continue in this manner as far as possible. 28 ๏ฉ0 0 1 2 ๏น Example 1: Transform A ๏ฝ ๏ช๏ช2 3 0 ๏ญ 2๏บ๏บ to reduced row echelon form. ๏ช๏ซ3 3 6 ๏ญ 9 ๏บ๏ป To transform A to reduced row echelon form, we perform Gauss-Jordan elimination method. The sequence of elementary row operations, beginning with A, follows. Interchange rows 1 and 3 of ๐ด to obtain ๐ด1 0 0 1 2 ๐ด = [ 2 3 0 −2 ] 3 3 6 −9 ๐ 1 ๐ 3 3 3 6 −9 [2 3 0 −2] = ๐ด1 0 0 1 2 (Multiply row 1 of ๐ด1 by 1/3 to obtain ๐ด2 ) 1 3 ๐ 1 ๐ 1 1 [2 0 1 2 −3 3 0 −2] = ๐ด2 0 1 2 (Multiply row 1 of ๐ด2 by -2 and add to row 2 to obtain ๐ด3 ) −2๐ 1 + ๐ 2 ๐ 2 1 [0 0 1 2 −3 1 −4 4 ] = ๐ด3 0 1 2 (Multiply row 2 of ๐ด3 by -1 and add to row 1 to obtain ๐ด4 ) 1๐ 2 + ๐ 1 ๐ 1 1 0 [0 1 0 0 6 −7 −4 4 ] = ๐ด4 1 2 Multiply row 3 of ๐ด4 by −6 and add to row 1; multiply row 3 of ๐ด4 by 4 and add to row 2 to obtain ๐ด5 ) −6๐ 3 + ๐ 1 4๐ 3 + ๐ 2 ๐ 1 ๐ 3 1 0 0 −19 [0 1 0 12 ] = ๐ด5 0 0 1 2 ๐ด5 is already in reduced row echelon form. We now apply these results to the solution of linear system. 29 Theorem 2.3.1. Let AX = B and CX = D be two linear systems each of ๐ equations in ๐ unknowns. If the augmented matrices [ AโB ] and [ CโD ] of these systems are row equivalent, then both linear systems have exactly the same solutions. Example 2. Solve the linear system x + y + 2z = -1 x – 2y + z = -5 3x + y + z = 3 by Gauss-Jordan reduction. Solution: Let the augmented matrix be 1 ๐ด = [1 3 1 2 −2 1 1 1 −1 −5] 3 Applying the Gauss-Jordan elimination we have 1 1 2 [1 −2 1 3 1 1 −1 ๐ + ๐ 1 2 −5] 3๐ + ๐ 1 3 3 1 − 3 ๐ 2 −๐ 2 + ๐ 1 2๐ 2 + ๐ 3 3 − 13 ๐ 3 5 − 3 ๐ 3 + ๐ 1 1 − 3 ๐ 3 + ๐ 2 ๐ 2 ๐ 3 1 [0 0 1 2 −3 −1 −2 −5 ๐ 2 1 [0 0 1 2 1 1/3 −2 −5 1 [0 0 0 5/3 −7/3 1 1/3 4/3 ] 0 −13/3 26/3 ๐ 1 ๐ 3 ๐ 3 ๐ 1 ๐ 2 1 [0 0 1 [0 0 0 1 0 0 1 0 −1 −4] 6 5/3 1/3 1 0 0 1 −1 4/3] 6 −7/3 4/3 ] −2 1 2] −2 The last matrix is in reduced row echelon form and is row equivalent to the augmented matrix representing the given system. As an augmented matrix, it represents the system x y z = 1 = 2 = -2 30 Thus the solution set is {(1, 2, - 2)}. The given system is consistent. Example 3: Solve the linear system ๐ฅ1 + ๐ฅ2 − ๐ฅ3 = 7 4๐ฅ1 − ๐ฅ2 + 5๐ฅ3 = 4 6๐ฅ1 + ๐ฅ2 + 3๐ฅ3 = 0 Solution: Let the augmented matrix be 1 [4 6 1 −1 −1 5 1 3 7 4] 0 Applying the Gauss-Jordan elimination we have 1 1 −1 [4 −1 5 6 1 3 7 −4๐ + ๐ 4] −6๐ 1 + ๐ 2 1 3 0 1 ๐ 2 + ๐ 1 −๐ 2 + ๐ 3 5 1 − 5 ๐ 2 ๐ 2 1 [ ๐ 3 0 0 1 −1 −5 9 −5 9 ๐ 1 ๐ 3 1 0 4/5 [0 −5 9 0 0 0 ๐ 2 1 [0 0 0 1 0 7 −24] −42 11/5 −24 ] −18 4/5 −9/5 0 11/5 24/5 ] −18 The last equation reads 0๐ฅ1 + 0๐ฅ2 + 0๐ฅ3 = −18, which is impossible since 0 ≠ −18. Hence the given system has no solution. In this case the system is said to be inconsistent. Example 4: Solve the system 2๐ฅ1 + 6๐ฅ2 − 4๐ฅ3 + 2๐ฅ4 = 4 ๐ฅ1 − ๐ฅ3 + ๐ฅ4 = 5 −3๐ฅ1 + 2๐ฅ2 − 2๐ฅ3 = −2 Solution: Applying the Gauss-Jordan elimination to the augmented matrix, we have 2 6 −4 [ 1 0 −1 −3 2 −2 2 1 0 4 5] −2 ๐ 1 ๐ 2 1 0 −1 [ 2 6 −4 −3 2 −2 1 2 0 5 4] −2 31 ๐ 2 ๐ 3 1 0 [ 0 6 0 2 ๐ 2 ๐ 2 1 [ 0 0 −2๐ 2 + ๐ 3 ๐ 3 1 0 [ 0 1 0 0 ๐ 3 1 1 0 −1 0 [ 0 1 −1/3 0 0 1 −9/13 −2๐ 1 + ๐ 2 3๐ 1 + ๐ 3 1 6 3 − 13 ๐ 3 ๐ 3 + ๐ 1 1 3 ๐ 3 + ๐ 2 ๐ 1 ๐ 2 −1 −2 −5 1 0 3 0 −1 1 −1/3 2 −5 5 −6] 13 1 0 3 5 −1] 13 −1 1 −1/3 0 −13/3 3 5 −1] 15 1 0 0 4/13 [ 0 1 0 3/13 0 0 1 −9/13 5 −1 ] −45/13 20/13 −28/13] −45/13 Column 4 has no pivot hence ๐ฅ4 is a free variable. If we let ๐ฅ4 = ๐, ๐ ∈ ๐ , then we obtain: ๐ฅ1 = 20 13 − 4 13 ๐ , ๐ฅ2 = − 28 13 20 Thus the solution set is in the form of ( 13 − − 3 13 4 13 ๐ , and ๐ฅ3 = − ๐, − 28 13 − 3 13 45 13 ๐, − + 45 13 9 13 + ๐. 9 13 ๐, ๐), ๐ ∈ ๐ . The given system has infinitely many solution. Corollary 2.3.1. If A and C are row equivalent ๐ x ๐ matrices, then the linear systems AX = 0 and CX = 0 have exactly the same solutions. SAQ 2-1 Consider the system 2๐ฅ1 − ๐ฅ2 + 3๐ฅ3 = ๐ 3๐ฅ1 + ๐ฅ2 − 5๐ฅ3 = ๐ −5๐ฅ1 − 5๐ฅ2 + 21๐ฅ3 = ๐ Show that the system is inconsistent if ๐ ≠ 2๐ − 3๐. 32 ASAQ 2-1 Reducing the augmented matrix of the given system to row echelon form, we have 2 −1 3 [3 1 −5 −5 −5 21 ๐ ๐] ๐ 1 −1/2 ๐ 1 [ 3 1 −5 −5 1 ๐ 2 1 −3๐ 1 + ๐ 2 5๐ 1 + ๐ 3 ๐ 2 ๐ 3 1 [0 0 3/2 −5 21 −1/2 3/2 5/2 −19/2 −15/2 57/2 ๐/2 ๐ ] ๐ ๐/2 −3๐+2๐ 2 5๐+2๐ ] 2 3๐ 2 + ๐ 3 ๐ 3 1 [0 0 −1/2 3/2 5/2 −19/2 0 0 ๐/2 −3๐+2๐ 2 ] −2๐ + 3๐ + ๐ The last equation reads 0๐ฅ1 + 0๐ฅ2 + 0๐ฅ3 = −2๐ + 3๐ + ๐. The given system has no solution if −2๐ + 3๐ + ๐ ≠ 0. Thus the system is inconsistent if ๐ ≠ 2๐ − 3๐. 2.4 Homogeneous Systems A linear system of the form a11 x1 ๏ซ a12 x2 ๏ซ ... ๏ซ a1n xn ๏ฝ 0 a21 x1 ๏ซ a22 x2 ๏ซ ... ๏ซ a2 n xn ๏ฝ 0 (2) am1 x1 ๏ซ am 2 x2 ๏ซ ... ๏ซ amn xn ๏ฝ 0 is called a homogeneous system. We can write it in matrix form as ๐ด๐ = 0 The solution to the homogeneous system (2) is called the trivial solution if x1 ๏ฝ x2 ๏ฝ ... ๏ฝ xn ๏ฝ 0 and a solution x1 , x2 ,..., x n in which not all the xi are zero is called a nontrivial solution. 33 Example 1. Consider the homogeneous system ๐ฅ + 2๐ฆ + ๐ง = 0 2๐ฅ + 5๐ฆ + 3๐ง = 0 3๐ฅ + 7๐ฆ + 5๐ง = 0 The augmented matrix of this system is 1 2 [2 5 3 7 1 3 5 0 0] 0 which is row equivalent to (you must do the calculation here) 1 0 [0 1 0 0 0 0 1 0 0] 0 Hence the solution to the given homogeneous system is ๐ฅ = ๐ฆ = ๐ง = 0 (a trivial solution). SAQ 2-2 Find the solution of the homogeneous system ๐ฅ + ๐ฆ + ๐ง + ๐ค = 0 2๐ฅ + ๐ฆ ๐ง = 0 3๐ฅ + ๐ฆ + ๐ง + 2๐ค = 0 ASAQ 2-2 The augmented matrix of this system is 1 1 1 1 [2 1 −1 0 3 1 1 2 0 0] 0 34 Transforming it to reduced row echelon form we have 1 1 1 1 [2 1 −1 0 3 1 1 2 0 −2๐ + ๐ 0] −3๐ 1 + ๐ 2 1 3 0 −๐ 2 1 ๐ 2 1 1 [ ๐ 3 0 −1 −3 0 −2 −2 ๐ 2 1 [0 0 1 1 1 3 −2 −2 1 0 −2 0] −1 0 1 0 2 0] −1 0 ๐ 1 1 [ ๐ 3 0 0 0 −2 1 3 0 4 −1 0 2 0] 3 0 ๐ 4 3 1 ๐ 3 [0 0 0 −2 1 3 0 1 −1 0 2 0] 3/4 0 2๐ 3 + ๐ 1 −3๐ 3 + ๐ 2 1 ๐ 1 [ ๐ 2 0 0 −๐ 2 + ๐ 1 2๐ 2 + ๐ 3 1 0 0 1 0 0 1 1/2 0 −1/4 0] 3/4 0 The fourth column has no pivot hence ๐ค is a free variable. If we let ๐ค = ๐ , ๐ ∈ ๐ , then the solution of the given homogeneous system is 1 ๐ฅ = −2๐ ๐ฆ = 1 4 ๐ 3 ๐ง = −4๐ ๐ค = ๐ 1 1 3 where ๐ is any real number. Thus the solution set is in the form of {(− 2 ๐ , 4 ๐ , − 4 ๐ , ๐ )}, ๐ ∈ ๐ . This shows that the system has a nontrivial solution. This result is generalized in the next theorem. Theorem 2.4.1: A homogeneous system of ๐ equations in ๐ unknowns always has a nontrivial solution if ๐ < ๐, that is , if the number of unknowns exceeds the number of equations. 35 ACTIVITY 1. Find all solutions to the given linear systems. a. ๐ฅ1 − ๐ฅ2 − ๐ฅ3 = 0 b. 2๐ฅ1 + 3๐ฅ2 − ๐ฅ3 = 0 2๐ฅ1 + ๐ฅ2 + 2๐ฅ3 = 0 −4๐ฅ1 + 2๐ฅ2 + ๐ฅ3 = 0 ๐ฅ1 − 4๐ฅ2 − 5๐ฅ3 = 0 7๐ฅ1 + 3๐ฅ2 − 9๐ฅ3 = 0 c. ๐ฅ1 + ๐ฅ2 + ๐ฅ3 + ๐ฅ4 = 6 2๐ฅ1 − ๐ฅ3 − ๐ฅ4 = 4 3 ๐ฅ3 + 6๐ฅ4 = 3 ๐ฅ1 − ๐ฅ4 = 5 2. Find all values of ๐ for which the resulting system has (a) no solution, (b) a unique solution, and (c) infinitely many solutions. ๐ฅ1 + ๐ฅ2 − ๐ฅ3 = 3 ๐ฅ1 − ๐ฅ2 + 3๐ฅ3 = 4 ๐ฅ1 + ๐ฅ2 + (๐2 − 10)๐ฅ3 = ๐ 2.5 The Inverse of a Matrix When working with real numbers, we know that a number ๐ times its inverse is one provided that ๐ ≠ 0. Thus the equation ๐๐ฅ = ๐ could be solved for ๐ฅ by dividing both sides (or multiplying both sides by 1/๐) of the equation by ๐ to get ๐ฅ = ๐/๐ provided that ๐ ≠ 0. However the matrix equation ๐ด๐ = ๐ต cannot be divided by the matrix ๐ด on both sides because there is no matrix division. This section is concerned on finding a matrix whose function is similar to the inverse of a real number ๐. We like to find a matrix inverse such that ๐ด times its inverse is equal to “one”. The matrix equivalent of “one” is called the “identity matrix”. Definition 2.5.1: An ๐ x ๐ matrix ๐ด is called nonsingular (or invertible) if there exists an ๐ x ๐ matrix ๐ต such that ๐ด๐ต = ๐ต๐ด = ๐ผ๐ The matrix ๐ต is called an inverse of ๐ด. If there exists no such matrix ๐ต, then ๐ด is called singular (or non-invertible). 36 −2 Example 1: Let ๐ด = [ 3 −4/11 1 ] and ๐ต = [ 3/11 4 −2 1 −4/11 ๐ด๐ต = [ ][ 3 4 3/11 −4/11 ๐ต๐ด = [ 3/11 1/11 ] then, 2/11 1/11 1 0 ] = [ ] = ๐ผ2 2/11 0 1 1/11 −2 ][ 2/11 3 1 1 ]= [ 4 0 and 0 ] = ๐ผ2 1 Since ๐ด๐ต = ๐ต๐ด = ๐ผ2 , we conclude that ๐ต is an inverse of ๐ด and that ๐ด is nonsingular. 4 Example 2: Let ๐ด = [ −3 Solution: Let ๐ด−1 = [ ๐ ๐ 2 ], find ๐ด−1 . 1 ๐ ] then ๐ ๐ด๐ด−1 = ๐ผ2 [ [ 4 2 ๐ ][ −3 1 ๐ 4๐ + 2๐ −3๐ + ๐ 1 0 ๐ ]= [ ] 0 1 ๐ 4๐ + 2๐ 1 ]= [ 0 −3๐ + ๐ 0 ] 1 Two matrices are equal if and only if their corresponding parts are equal. Hence if we equate the corresponding parts, we get 4๐ + 2๐ = 1 −3๐ + ๐ = 0 and 4๐ + 2๐ = 0 −3๐ + ๐ = 1 Solving the system we have a = 1/10, b = -1/5, c = 3/10, and d = 2/5. Thus ๐ด−1 = [ 1/10 3/10 −1/5 ] 2/5 You can always check your answer by taking the product ๐ด๐ด−1 and making sure that the answer is the identity matrix ๐ผ2 . There is a simple procedure for finding the inverse of a 2 x 2 matrix. It can be done easily as follows: Let ๐ด = [ ๐ ๐ ๐ค ๐ ]. We are looking for a matrix [ ๐ฆ ๐ ๐ฅ ๐ง ] such that 37 [ [ ๐ ๐ ๐๐ค + ๐๐ฆ ๐๐ค + ๐๐ฆ ๐ ๐ค ][ ๐ ๐ฆ ๐ฅ 1 0 ๐ง ] = [0 1] ๐๐ฅ + ๐๐ง 1 0 ] = [ ] ๐๐ฅ + ๐๐ง 0 1 Equating the corresponding parts we have ๐๐ค + ๐๐ฆ = 1 (1) ๐๐ค + ๐๐ฆ = 0 (2) ๐๐ฅ + ๐๐ง = 0 (3) ๐๐ฅ + ๐๐ง = 1 (4) Multiplying equation (1) by ๐ and equation (2) by ๐ we get: ๐๐๐ค + ๐๐๐ฆ = ๐ ๐๐๐ค + ๐๐๐ฆ = 0 (1a) (2a) Subtracting equation (1a) from equation (2a) and solving for ๐ฆ gives us: ๐ฆ= −๐ ๐๐−๐๐ Multiplying equation (3) by ๐ and equation (4) by ๐ we obtain: ๐๐๐ฅ + ๐๐๐ง = 0 ๐๐๐ฅ + ๐๐๐ง = ๐ (3a) (4a) Subtracting equation (3a) from equation (4a) and solving for ๐ง gives us: ๐ง= ๐ ๐๐−๐๐ In a similar manner we can solve for ๐ค by multiplying equation (1) by ๐ and equation (2) by ๐: ๐ ๐ค= ๐๐−๐๐ Finally we multiply equation (3) by ๐ and equation (4) by ๐ to solve for ๐ฅ: ๐ฅ= −๐ ๐๐−๐๐ 38 Thus, the inverse of matrix ๐ด = [ ๐ −๐ [๐๐−๐๐ −๐ ๐๐−๐๐ ๐ ] ๐๐−๐๐ ๐๐−๐๐ = ๐ ๐ ๐ ] is the matrix ๐ 1 ๐ [ ๐๐−๐๐ −๐ −๐ ] provided that ๐๐ − ๐๐ ≠ 0. ๐ 4 2 Example 3: Solve the inverse of matrix ๐ด = [ ] using the formula. −3 1 Solution: Let ๐ด = [ 4 2 ๐ ]=[ −3 1 ๐ ๐ด−1 = 1 1 [ 10 3 ๐ ] then ๐๐ − ๐๐ = 4(1) − 2(−3) = 10. Thus ๐ 1/10 −2 ] = [ 3/10 4 −1/5 ] 2/5 Note that this procedure only works for 2 x 2 matrices. In general, the procedure in finding the inverse of any ๐ x ๐ matrix is as follows: 1. Form the ๐ x 2๐ partitioned matrix [ ๐ดโ๐ผ๐ ] obtained by adjoining the identity matrix ๐ผ๐ to the given matrix ๐ด. 2. Use elementary row operations to transform the matrix obtained in Step 1 to reduced row echelon form. Remember that any elementary row operation we do to a row of ๐ด we also do to the corresponding row of ๐ผ๐ . 3. The series of elementary row operations which reduces ๐ด to ๐ผ๐ will reduce ๐ผ๐ to ๐ด−1 . If ๐ด cannot be reduced to ๐ผ๐ then ๐ด is singular and ๐ด−1 does not exist. 1 2 Example 4. Find the inverse of the matrix ๐ด = [2 5 1 0 3 3]. 8 Solution: We form the 3 x 6 partitioned matrix [A โI3 ] by adjoining the 3 x 3 identity matrix to ๐ด and transform it to reduced row echelon form by applying the Gauss-Jordan elimination. ๐ด 1 [2 1 2 5 0 ๐ผ3 3 3 8 1 0 0 0 1 0 0 −2๐ + ๐ 0] −๐ 1+ ๐ 2 1 3 1 ๐ 2 ๐ 3 39 1 2 [0 1 0 −2 3 −3 5 9 −3 −1 1 0 0 −2 1 0] −1 0 1 −2๐ 2 + ๐ 1 2๐ 2 + ๐ 3 1 [0 0 0 1 0 1 [0 0 0 9 1 −3 0 1 5 −2 0 −2 1 0] 5 −2 −1 1 [0 0 0 1 0 −40 13 5 0 0 1 5 −2 −2 1 −5 2 0 0] −๐ 3 1 ๐ 1 ๐ 3 ๐ 3 −9๐ 3 + ๐ 1 3๐ 3 + ๐ 2 ๐ 1 ๐ 2 16 9 −5 −3]. −2 −1 −40 Since ๐ด has been transformed to ๐ผ3 then ๐ด−1 = [ 13 5 16 −5 −2 9 −3]. −1 You can check your answer by taking the product ๐ด๐ด−1 and making sure that the answer is the identity matrix ๐ผ3 . 2 1 Example 5. Find the inverse of the matrix ๐ด = [ 1 −2 −3 −1 −1 −3]. 2 Solution: We form the 3 x 6 matrix [A โI3 ] and transform it to reduced row echelon form (you must do the calculation here): 2 1 −1 1 [ 1 −2 −3 0 −3 −1 2 0 0 0 1 0 ] is row equivalent to [ 1 0 0 1 0 1 0 0 2/5 1/5 0 −1 1/5 −2/5 0 ] 1 0 −1/5 −1/35 −1/7 Since ๐ด is row equivalent to a matrix which has a row consisting of zeros, then ๐ด cannot be reduced to ๐ผ3 . Thus ๐ด has no inverse and we say that ๐ด is a singular matrix. Example 4 shows that an inverse exists if the partitioned matrix [๐ด ๐ผ๐ ] can be reduced to [๐ผ๐ ๐ต] where ๐ต = ๐ด−1 . That is, A is row equivalent to ๐ผ๐ . This is stated in the following theorem. Theorem 2.5.1: An ๐ x ๐ matrix is nonsingular if and only if it is row equivalent to ๐ผ๐ . Theorem 2.5.2: If a matrix has an inverse, then the inverse is unique. 40 Proof: Suppose ๐ด has two inverses, say ๐ต and C. Then by Definition 2.5.1, ๐ด๐ต = ๐ต๐ด = ๐ผ๐ and ๐ด๐ถ = ๐ถ๐ด = ๐ผ๐ . We now have ๐ต(๐ด๐ถ) = (๐ต๐ด)๐ถ Matrix multiplication is associative Then ๐ต = ๐ต๐ผ๐ = ๐ต(๐ด๐ถ) = (๐ต๐ด)๐ถ = ๐ผ๐ ๐ถ = ๐ถ By transitivity, we conclude that ๐ต = ๐ถ and the theorem is proved. Theorem 2.5.3: ( Properties of the Inverse ) a) If ๐ด is a nonsingular matrix, then ๐ด−1 is nonsingular and (๐ด−1 )−1 = ๐ด b) If ๐ด and ๐ต are nonsingular matrices, then ๐ด๐ต is nonsingular and (๐ด๐ต)−1 = ๐ต −1 ๐ด−1 c) If A is a nonsingular matrix, then (๐ด๐ )−1 = (๐ด−1 )๐ We prove part (b) of the theorem and leave the proof of parts (a) and (c) as an exercise. Proof of part (b): We have to show that (๐ด๐ต)(๐ต−1 ๐ด−1 ) = (๐ต −1 ๐ด−1 )(๐ด๐ต) = ๐ผ๐ , that is the inverse of ๐ด๐ต = ๐ต −1 ๐ด−1 . (๐ด๐ต)(๐ต−1 ๐ด−1 ) = ๐ด(๐ต๐ต −1 )๐ด−1 Matrix multiplication is associative = ๐ด(๐ผ๐ )๐ด−1 Definition of an inverse of a matrix = ๐ด๐ด−1 ๐ผ๐ is a multiplicative identity for an ๐ x ๐ matrix = ๐ผ๐ Definition of an inverse of a matrix Similarly (๐ต −1 ๐ด−1 )(๐ด๐ต) = ๐ต −1 (๐ด−1 ๐ด)๐ต Matrix multiplication is associative = ๐ต −1 (๐ผ๐ )๐ต Definition of an inverse of a matrix = ๐ต −1 ๐ต ๐ผ๐ is a multiplicative identity for an ๐ x ๐ matrix = ๐ผ๐ Definition of an inverse of a matrix 41 Hence (๐ด๐ต)−1 = ๐ต −1 ๐ด−1. Corollary 2.5.1: If ๐ด1 , ๐ด2 , โฏ , ๐ด๐ are ๐ x ๐ nonsingular matrices, then ๐ด1 ๐ด2 โฏ ๐ด๐ is nonsingular and (๐ด1 ๐ด2 โฏ ๐ด๐ )−1 = ๐ด๐ −1 ๐ด๐−1 −1 โฏ ๐ด1 −1 Theorem 2.5.4: Suppose that ๐ด and ๐ต are ๐ x ๐ matrices a) If ๐ด๐ต = ๐ผ๐ then ๐ต๐ด = ๐ผ๐ . b) If ๐ต๐ด = ๐ผ๐ then ๐ด๐ต = ๐ผ๐ . 2.6 Linear Systems and Inverses Suppose that matrix ๐ด is invertible. Consider the system ๐ด๐ = ๐ต where ๐ฅ1 ๐1 ๐ฅ2 ๐ ๐ = [ โฎ ] and ๐ต = [ 2 ] are ๐ x 1 matrices where ๐ฅ1 , ๐ฅ2 , โฏ , ๐ฅ๐ are variables and โฎ ๐ฅ๐ ๐๐ ๐1 , ๐2 , โฏ , ๐๐ are real numbers. Since ๐ด is invertible then ๐ด−1 exists and ๐ด๐ = ๐ต ๐ด−1 ๐ด๐ = ๐ด−1 ๐ต Multiplying both sides by ๐ด−1 ๐ผ๐ ๐ = ๐ด−1 ๐ต Definition of an inverse of a matrix ๐ = ๐ด−1 ๐ต ๐ผ๐ is the multiplicative identity for an ๐ x ๐ matrix Hence ๐ = ๐ด−1 ๐ต is the unique solution of the system. Example 1. Let ๐ด = [ 1 1 5 ], ๐ต = [ ]. Solve ๐ด๐ = ๐ต by using the inverse of A. 1 2 7 Solution: ๐ด−1 = [ 2 −1 ] (verify) −1 1 42 Hence ๐ = ๐ด−1 ๐ต 2 −1 5 3 ๐=[ ][ ]= [ ] −1 1 7 2 Hence the unique solution is ๐ = (3, 2). Theorem 2.6.1: If ๐ด is an ๐ x ๐ matrix, the homogeneous system ๐ด๐ = 0 has a nontrivial solution if and only if A is singular. Example 2: Consider the homogeneous system 2๐ฅ1 + ๐ฅ2 − ๐ฅ3 = 0 ๐ฅ1 − 2๐ฅ2 − 3๐ฅ3 = 0 −3๐ฅ1 − ๐ฅ2 + 2๐ฅ3 = 0 2 1 −1 where the coefficient matrix ๐ด = [ 1 −2 −3] is the singular matrix of Example 5 of −3 −1 2 the previous section. The augmented matrix 2 1 −1 [ 1 −2 −3 −3 −1 2 0 1 0 0] is row equivalent to [0 1 0 0 0 −1 1 0 0 0] 0 which implies that ๐ฅ1 = ๐, ๐ฅ2 = −๐, and ๐ฅ3 = ๐, where ๐ ∈ ๐ . Thus the system has a nontrivial solution. Theorem 2.6.2: If ๐ด is an ๐ x ๐ matrix, then ๐ด is nonsingular if and only if the linear system ๐ด๐ = ๐ต has a unique solution for every ๐ x 1 matrix ๐ต. 43 SAQ 2-3 1 Let ๐ด = [1 2 ๐ต 1 −1 −1 2 ]. Find the inverse of ๐ด then solve ๐ด๐ = ๐ต where 1 −1 1 is the 3 x 1 matrix (a) [−2], 3 4 (b) [−3], and 5 5 (c) [ 7 ] −4 ASAQ 2-3 To solve for ๐ด−1 let us form the 3 x 6 matrix [๐ด ๐ผ3 ] and transform it to reduced row echelon form. 1 1 −1 [1 −1 2 2 1 −1 1 0 0 −๐ + ๐ 1 2 0 1 0] −2๐ + ๐ 1 3 0 0 1 ๐ 2 −๐ 2 ๐ 2 1 [ ๐ 3 0 0 1 −1 −2 3 −1 1 1 0 −1 1 −2 0 0 0] 1 1 1 −1 ๐ 3 [0 −1 1 0 −2 3 1 0 0 −2 0 1] −1 1 0 1 1 −1 ๐ 2 [0 1 −1 0 −2 3 1 2 −1 0 0 0 −1] 1 0 −1 0 2 0 3 1 1 −1] −2 −๐ 2 + ๐ 1 2๐ 2 + ๐ 3 ๐ 1 ๐ 3 1 0 [0 1 0 0 0 −1 1 ๐ 3 + ๐ 2 ๐ 2 1 0 0 [0 1 0 0 0 1 −1 0 1 5 1 −3] 3 1 −2 −1 0 1 Thus ๐ด−1 = [ 5 1 −3]. Using this to solve the linear system ๐ด๐ = ๐ต for ๐ where 3 1 −2 44 1 (a) ๐ต = [−2] we have 3 ๐ = ๐ด−1 ๐ต −1 0 = [5 1 3 1 1 1 −3] [−2] −2 3 2 = [−6] −5 4 (b) ๐ต = [−3] we have 5 ๐ = ๐ด−1 ๐ต 4 −1 0 1 = [ 5 1 −3] [−3] 3 1 −2 5 1 = [2] −1 5 (c) ๐ต = [ 7 ] we have −4 ๐ = ๐ด−1 ๐ต −1 0 1 5 = [ 5 1 −3] [ 7 ] 3 1 −2 −4 −9 = [ 44 ] 30 This only shows that when a matrix is nonsingular, then the system ๐ด๐ = ๐ต has a unique solution for every ๐ x 1 matrix ๐ต. 45 Based on the preceding theorems and examples, it can be noted that the following statements are equivalent: 1. The matrix ๐ด is invertible (nonsingular). 2. The system ๐ด๐ = 0 has only the trivial solution. 3. The matrices ๐ด and ๐ผ๐ are row equivalent. 4. The system ๐ด๐ = ๐ต has a unique solution for every ๐ x 1 matrix ๐ต. ACTIVITY 1. Find the inverse of the following matrices: ๏ฉ 1 2 ๏ญ3 4 ๏น ๏ฉ1 1 1๏น ๏ช๏ญ 4 2 1 3 ๏บ๏บ a. ๏ช๏ช0 2 3๏บ๏บ b. ๏ช ๏ช 3 0 0 ๏ญ 3๏บ ๏ช๏ซ5 5 1๏บ๏ป ๏ช ๏บ ๏ซ 2 0 ๏ญ2 3 ๏ป 1 4 1 0 5 2. Let ๐ต = [2 1 4] and ๐ท = [0 2 3 5 3 4 7 2 1]. Calculate ๐ต๐ท and ๐ท−1 ๐ต −1 . Verify 3 that (๐ต๐ท)(๐ท−1 ๐ต−1 ) = ๐ผ3 . 3. Find all values of a ๏ฉ1 A ๏ฝ ๏ช๏ช1 ๏ช๏ซ1 -1 exists. What is A ? for which the inverse of 1 0๏น 0 0 ๏บ๏บ 2 a ๏บ๏ป 4. Show that if ๐ด, ๐ต, and ๐ถ are invertible matrices, then ๐ด๐ต๐ถ is invertible and (๐ด๐ต๐ถ)−1 = ๐ถ −1 ๐ต −1 ๐ด−1 . 5. Prove parts (a) and (c) of Theorem 2.5.4. 46 2.7 Determinants Definition 2.7.1: Let S = {1, 2, … , n } be the set of integers from 1 to n, arranged in ascending order. A rearrangement j1 j 2 ... j n of the elements of S is called a permutation of S. The set of all permutations of S is denoted by ๐๐ . For example, {1, 2} has two permutations namely ๐2 = { (12), (21) } and {1, 2, 3} has six permutations namely ๐3 = { (123), (132), (213), (231), (312), (321) }. The number of permutations of the set {1, 2, …, ๐} can be determined without writing a list as in the example above. Notice that there are ๐ possible positions to be filled. There are ๐ choices for the first position, ๐ − 1 for the second, ๐ − 2 for the third, and only one element for the ๐th position. Thus the total number of permutations of ๐ elements is ๐(๐ − 1)(๐ − 2) … 2.1 = ๐! A permutation j1 j 2 ... j n of S = { 1, 2, … , n } is said to have an inversion if a larger integer jr precedes a smaller one j s . For example, consider the permutation (23541). The following pairs of numbers form an inversion: 21, 31, 54, 51. A permutation is called even if the total number of inversions in it is even and odd if the total number of inversions in it is odd. Example 1: The even permutations in S3 are: 123 ( There is no inversion ) 231 ( There are two inversions; 21 and 31 ) 312 ( There are two inversions; 31 and 32 ) The odd permutations in S3 are: 132 ( There is one inversion; 32 ) 213 ( There is one inversion; 21 ) 321 ( There are three inversions; 32, 31, and 21 ) Note that the number of odd permutations is equal to the number of even permutations. Definition 2.7.2: Let A ๏ฝ [ a ij ] be an ๐ x ๐ matrix. We define the determinant of ๐ด (written โ๐ดโ) by A ๏ฝ ๏ฅ ๏ฑ a1 j1a2 j2 ...an jn , where the summation ranges over all permutations j1 j 2 ... j n of the set S = {1, 2, …, n}. 47 Note that each term ±๐1๐1 ๐2๐2 … ๐๐๐๐ of |๐ด| is a product of ๐ elements of ๐ด such that one and only one element comes from each row and one and only one element comes from each column. Thus if the factors come from successive rows then the first number in the subscripts are in the natural order 1, 2, …, ๐. Since all rows are different then the subscripts ๐1 , ๐2 , … , ๐๐ are distinct. This means that {๐1 , ๐2 , … , ๐๐ } is a permutation of {1, 2, … , ๐ }. The sign of the term ๐1๐1 ๐2๐2 … ๐๐๐๐ is if the permutation ๐1 , ๐2 , … , ๐๐ is even otherwise, the sign is . ๐11 Example 2: Let ๐ด = [๐ 21 ๐12 ๐22 ]. To illustrate Definition 2.7.2, let us consider exactly one element from each row and each column. There are two possibilities: ๐11 [ ∗ ∗ ๐22 ] (1) ∗ [๐ ๐12 ∗ ] (2) 21 In (1), the second subscripts form the even permutation 12 hence the sign of the product ๐11 ๐22 is positive. In (2), the second subscripts form the odd permutation 21 hence the sign of the product ๐12 ๐21 is negative. Thus โAโ= a11a22 ๏ญ a12 a21 . The determinant in Example 2 can be solved by writing the terms ๐1___ ๐2___ and ๐1___ ๐2___ (there are only two terms because ๐2 has only two permutations) Next we fill in the blanks with the two permutations of ๐2 . ๐11 ๐22 and ๐12 ๐21 We assign a sign for the even permutation 12 and a Taking the sum of the two terms we have โAโ= a11a22 ๏ญ a12 a21 . sign for the odd permutation 21. 48 Example 3: Let ๐ด = [ 2 −4 ] then 3 5 |๐ด| = (2)(5) − (−4)(3) = 22 ๏ฉ a11 Example 4: Let A ๏ฝ ๏ช๏ชa 21 ๏ช๏ซa31 terms a12 a 22 a32 a13 ๏น a 23 ๏บ๏บ . Since ๐3 has 6 permutations then we write the 6 a33 ๏บ๏ป ๐1___ ๐2___ ๐3___ , ๐1___ ๐2___ ๐3___, ๐1___ ๐2___ ๐3___, ๐1___ ๐2___ ๐3___, ๐1___ ๐2___ ๐3___, and ๐1___ ๐2___ ๐3___ Next we fill in the blanks with the six permutations of ๐3 . The even permutations are 123, 231, and 312 and the odd permutations are 213, 132, and 321. Thus |๐ด| = ๐11 ๐22 ๐33 + ๐12 ๐23 ๐31 + ๐13 ๐21 ๐32 − ๐12 ๐21 ๐33 − ๐11 ๐23 ๐32 − ๐13 ๐22 ๐31 We can also obtain the determinant of ๐ด by augmenting the first two columns as shown below: ๐11 ๐12 ๐13 ๐11 ๐12 ๐21 ๐22 ๐23 ๐21 ๐22 ๐31 ๐32 ๐33 ๐31 ๐32 − − − + + + We form the product of each of the three entries joined by the line from left to right and precede each product by a plus sign. Next we form the product of each of the three entries joined by the line from right to left and precede each product by a minus sign. 1 2 Example 5: Let ๐ด = [−2 1 3 −1 1 2 3 |๐ด| = |−2 1 4 3 −1 2 1 −2 3 3 4]. Then 2 2 1| −1 = (1)(1)(2) + (2)(4)(3) + (3)(-2)(-1) – (3)(1)(3) – (1)(4)(-1) – (2)(-2)(2) = 35 49 2.8 Properties of Determinants Theorem 2.8.1: The determinant of a matrix ๐ด and its transpose are equal. Example 1. Let A be the matrix of Example 5. Then 1 ๐ด = [2 3 −2 3 1 −1] 4 2 ๐ The determinant of AT is 1 −2 3 ๐ด = |2 1 −1 3 4 2 ๐ 1 −2 2 1| 3 4 = (1)(1)(2) + (-2)(-1)(3) + (3)(2)(4) – (3)(1)(3) – (1)(-1)(4) – (-2)(2)(2) = 35 Because of this property we can now replace “row” by “column” in the succeeding theorems about the determinants of a matrix ๐ด. Theorem 2.8.2: Let ๐ต be the matrix obtained from a matrix ๐ด by (i) multiplying a row (column) of ๐ด by a scalar ๐; then |๐ต| = ๐|๐ด|. (ii) interchanging two rows(columns) of |๐ด|; then |๐ต| = −|๐ด|. (iii) adding a multiple of a row(column) of ๐ด to another; then |๐ต| = |๐ด|. 2 Example 2. To illustrate part (i) of Theorem 2.8.2, let ๐ต = [ 4 factor of the entries in column 1, then 5 ]. Since 2 is a common −7 2 5 1 5 ๏ฝ2 ๏ฝ 2(๏ญ7 ๏ญ 10) ๏ฝ ๏ญ34 4 ๏ญ7 2 ๏ญ7 ๏ฉ 2 ๏ญ1๏น ๏ฉ5 ๏ญ2๏น Example 3: To illustrate part (ii) of Theorem 2.8.2, let A ๏ฝ ๏ช and B ๏ฝ ๏ช ๏บ ๏บ . Then ๏ซ5 ๏ญ2๏ป ๏ซ 2 ๏ญ1๏ป A ๏ฝ 4 + 5 = 1 and B = 5 + 4 = 1. This shows that |๐ต| = −|๐ด|. 50 6 Example 4: To illustrate part (iii) of theorem 2.8.2, let ๐ด = [−1 3 A ๏ฉ 6 9 ๏ญ12๏น ๏ช ๏ญ1 0 2 ๏บ 3๐ + ๐ 2 3 ๏ช ๏บ ๏ช๏ซ 3 0 ๏ญ8 ๏บ๏ป ๐ 3 9 −12 0 2 ]. 0 −8 B ๏ฉ 6 9 ๏ญ12๏น ๏ช ๏ญ1 0 2 ๏บ ๏ช ๏บ ๏ช๏ซ 0 0 ๏ญ2 ๏บ๏ป |๐ด| = 54 − 72 = −18 and |๐ต| = 0 − 18 = −18. Thus |๐ด| = |๐ต|. Theorem 2.8.3: Let ๐ด be a square matrix. (i) If two rows (columns) of ๐ด are equal, then โ๐ดโ = 0. (ii) If a row (column) of ๐ด consists entirely of zeros, then โ๐ดโ = 0. Proof of part (i): Suppose the rows ๐ and ๐ of ๐ด are equal. Interchange rows ๐ and ๐ of ๐ด to obtain a matrix ๐ต. By theorem 2.8.2 (ii), |๐ต| = −|๐ด|. Since rows ๐ and ๐ are equal then ๐ต = ๐ด, so |๐ต| = |๐ด|. Thus |๐ด| = −|๐ด| (By substitution) Hence |๐ด| = 0. ๏ฉ 3 2 1๏น Example 5: Let A ๏ฝ ๏ช๏ช ๏ญ1 0 4 ๏บ๏บ (row 1 is equal to row 3).The determinant of A is ๏ช๏ซ 3 2 1 ๏บ๏ป 3 2 1 A ๏ฝ ๏ญ1 0 4 = 3(0)(4) + 2(4)(3) +(-1)(2)(1)] – 3(2)(4) - 2(-1)(1) - 1(0)(3) 3 2 1 = 0 + 24 – 2 – 24 + 2 – 0 = 0 51 Proof of part (ii): Let the ๐th row(column) of ๐ด consist entirely of zero. Since each term in the |๐ด| contains a factor from each row(column) then each term in the |๐ด| is equal to zero. Hence |๐ด| = 0. Theorem 2.8.4: If a matrix A ๏ฝ [ a ij ] is upper (lower) triangular, then A ๏ฝ a11a22 ...ann that is, the determinant of a triangular matrix is the product of the elements on the main diagonal. Example 6: Evaluate the determinant of each matrix by applying theorem 2.8.4. 2 (a) ๐ด = [0 1 4 2 (b) ๐ต = [ 3 −2 −2 0 8 −2 3 −4 −4 2 ] −1 5 3 −4 1 5] 1 −3 6 4 Solution: (a) We transform matrix ๐ด into triangular form by applying elementary operations and taking note of the corresponding changes in the determinant. 2 |0 1 3 −4 −4 2 | −1 5 ๐ 1 ๐ 3 −2๐ 1 + ๐ 3 1 ๐ 4 2 5๐ 2 + ๐ 3 1 = − |0 2 (By theorem 2.8.2.ii) ๐ 3 1 −1 5 = − |0 −4 2 | 0 5 −14 (By theorem 2.8.2.iii) ๐ 2 1 = (−4) |0 0 (By theorem 2.8.2.i) ๐ 3 1 −1 5 1/2 | = ( −4) |0 −1 0 0 −23/2 The last matrix is in triangular form thus 23 −1 5 −4 2 | 3 −4 |๐ด| = (−4)(1)(−1) (− ) = −46 2 −1 5 −1 1/2 | 5 −14 (By theorem 2.8.2.iii) 52 4 2 3 −2 (b) | −2 0 8 −2 3 −4 1 5| ๐ 1 1 −3 6 4 ๐ 3 2๐ 1 + ๐ 3 4๐ 1 + ๐ 4 1 − ๐ 1 2 1 0 = −(−2) | 3 −2 0 2 0 −2 ๐ 1 −๐ 3 + ๐ 4 ๐ 2 ๐ 3 ๐ 4 ๐ 4 1 −3 1 5 | (By theorem 2.8.2.ii) 3 −4 6 4 −2 0 1 −3 3 −2 1 5 | (By theorem 2.8.2.iii) = −| 0 2 5 −10 0 −2 10 −8 ๐ 3 ๐ 4 −3๐ 1 + ๐ 2 ๐ 2 + ๐ 3 −๐ 2 + ๐ 4 −2 0 3 −2 = −| 4 2 8 −2 1 0 = 2 | 0 −2 0 2 0 −2 −1/2 3/2 1 5 | (By theorem 2.8.2.i) 5 −10 10 −8 −1/2 5/2 5 10 1 0 −1/2 0 −2 5/2 = 2| 15/2 0 0 15/2 0 0 3/2 1/2 |(By theorem 2.8.2.iii) −10 −8 3/2 1/2 |(By theorem 2.8.2.iii) −19/2 −17/2 3/2 1 0 −1/2 5/2 1/2 = 2 | 0 −2 |(By theorem 2.8.2.iii) 0 0 15/2 −19/2 0 0 0 1 15 Thus |๐ต| = 2(1)(−2) ( 2 ) (1) = −30. Theorem 2.8.5: The determinant of a product of two matrices is the product of their determinants; that is AB ๏ฝ A B Corollary 2.8.1: If A is nonsingular, then A ๏น 0 and A ๏ญ1 ๏ฝ 1 . A 53 ๏ฉ 2 ๏ญ4๏น ๏ฉ 5 / 22 2 /11๏น Example 7. Let A ๏ฝ ๏ช then A๏ญ1 ๏ฝ ๏ช ๏บ. ๏บ ๏ซ ๏ญ3 / 22 1/11 ๏ป ๏ซ3 5 ๏ป The determinant of A is โAโ= 10 – (-12) = 22 and the determinant of A-1 is 1 A-1โ= (5/22)(1/11) – (2/11)(-3/22) = 1/22. Hence |๐ด−1 | = . |๐ด| ACTIVITY 1. Evaluate the determinants of the following matrices. ๏ฉ 4 3 2๏น a. A ๏ฝ ๏ช๏ช3 ๏ญ 2 5๏บ๏บ ๏ช๏ซ2 4 6๏บ๏ป ๏ฉ 3 ๏ญ 1 2๏น b. B ๏ฝ ๏ช๏ช4 5 6๏บ๏บ ๏ช๏ซ7 1 2๏บ๏ป 4 ๏น ๏ฉ ๏ญ1 2 ๏ช c. ๏ช 4 ๏ญ8 ๏ญ16 ๏บ๏บ ๏ช๏ซ 3 0 5 ๏บ๏ป 2. Find all values of ๏ฌ for which a. ๏ฌ ๏ญ2 3 2 ๏ฝ0 ๏ฌ ๏ญ3 1 0 ๏ญ1 b. ๏ฌ I3 ๏ญ A ๏ฝ 0 where A ๏ฝ 2 0 1 0 0 ๏ญ1 3. The matrix ๐ด is called idempotent if ๐ด2 = ๐ด. What are the possible values for the |๐ด| if ๐ด is idempotent? 4. Prove Corollary 2.8.1. 2.9 Minors and Cofactor Expansion For large matrices such as ๐ ≥ 4, evaluating the determinants using the permutation formula could be very tedious. We see from the preceding section that transforming a matrix into a triangular matrix could make the computation easier. Another method that is also efficient in the computation of determinants of large matrices is by means of cofactor expansion. 54 Definition 2.9.1: Let A ๏ฝ [ a ij ] be an ๐ x ๐ matrix. Let M ij be the (๐ -1) x (๐ – 1) submatrix of ๐ด obtained by deleting the ๐th row and the ๐th column of ๐ด. The determinant M ij is called the minor of ๐๐๐ . The cofactor Aij of ๐๐๐ is defined as Aij ๏ฝ (๏ญ1) i ๏ซ j M ij ๏ฉ 5 ๏ญ1 6 ๏น Example 1: Let A ๏ฝ ๏ช๏ช 3 4 2 ๏บ๏บ . If we delete the first row and second column of ๐ด we ๏ช๏ซ7 2 1 ๏บ๏ป 3 2 obtain the 2 x 2 submatrix ๐12 = [ ]. The determinant of ๐12 is the minor of ๐12 = 7 1 −1. Thus M 12 ๏ฝ 3 2 ๏ฝ 3 ๏ญ 14 ๏ฝ ๏ญ11 . 7 1 Likewise, if we delete the second row and the third column of ๐ด we obtain the 2 x 2 5 −1 submatrix ๐23 = [ ]. The determinant |๐23 | is called the minor of ๐23 = 2. Thus 7 2 M 23 ๏ฝ 5 ๏ญ1 ๏ฝ 10 ๏ญ (๏ญ7) ๏ฝ 17 . 7 2 By Definition 2.9.1, the cofactor of ๐12 is ๐ด12 = (−1)1+2 |๐12 | = −(−11) = 11 Similarly the cofactor of ๐23 is ๐ด23 = (−1)2+3 |๐23 | = (−1)(17) = −17 55 Theorem 2.9.1: Let A ๏ฝ [ a ij ] be an ๐ x ๐ matrix. Then for each 1 ≤ ๐ ≤ ๐, A ๏ฝ ai1 Ai1 ๏ซ ai 2 Ai 2 ๏ซ ... ๏ซ ain Ain (expansion of โ๐ดโ about the ๐th row) and for each 1 ≤ ๐ ≤ ๐, A ๏ฝ a1 j A1 j ๏ซ a2 j A2 j ๏ซ ... ๏ซ anj Anj (expansion of โ๐ดโ about the ๐th column) 2 2 −3 1 0 1 2 −1 Example 2: Evaluate the determinant of ๐ด = [ ] by cofactor expansion. 3 −1 4 1 2 3 0 0 Solution: You can expand about any row or column of your choice, however it is best to expand about the fourth row because it has the most number of zeros. Thus if we expand about the fourth row we obtain |๐ด| = ๐41 ๐ด41 + ๐42 ๐ด42 + ๐43 ๐ด43 + ๐44 ๐ด44 = 2๐ด41 + 3๐ด42 + 0๐ด43 + 0๐ด44 Notice that the cofactor of a zero entry need not be calculated, so if we can get another zero on the fourth row then the computation would be a lot easier. Let us apply elementary operation on the determinant of ๐ด and take note of the changes in the determinant. 2 2 −3 1 0 1 2 −1 | | 3 −1 4 1 2 3 0 0 1 ๐ถ 2 1 −3๐ถ1 + ๐ถ2 ๐ถ1 1 2 −3 1 0 1 2 −1 = 2| | 3/2 − 1 4 1 1 3 0 0 By theorem 2.8.2.i 1 −1 −3 1 0 1 2 −1 ๐ถ2 = 2 | | By theorem 2.8.2.iii 3/2 − 11/2 4 1 1 0 0 0 1 −1 −3 1 0 1 2 −1 Now let us evaluate the determinant | | . Expanding about the fourth 3/2 − 11/2 4 1 1 0 0 0 row we have 56 1 −1 −3 1 0 1 2 −1 | | = (1)๐ด41 + 0๐ด42 + 0๐ด43 + 0๐ด44 3/2 − 11/2 4 1 1 0 0 0 = (1)(−1)4+1 −1 −3 1 | 111 2 −1| −2 4 1 −1 −3 1 = − | 111 2 −1| −2 4 1 −1 Next we evaluate | 111 − 2 −3 2 4 1 −1| by expanding about the third row. (Note, you can 1 expand about any row or column of your choice.) −1 | 111 − 2 −3 2 4 1 −1| = (− 11) (−1)3+1 |−3 2 2 1 1 −1 1 | + (4)(−1)3+2 | | −1 1 −1 −1 −3 + (1)(−1)3+3 | | 1 2 11 = (− 2 ) (1) + (−4)(0) + (1)(1) = −9/2 Substituting, we have 1 −1 −3 1 9 0 1 2 −1 | | = − (− 2) = 9/2 3/2 − 11/2 4 1 1 0 0 0 57 Hence 2 2 −3 1 9 0 1 2 −1 | | = 2( ) = 9 3 −1 4 1 2 2 3 0 0 SAQ 2-4 ๏ฉ4 ๏ญ 4 2 1 ๏น ๏ช1 2 0 3 ๏บ๏บ ๏ช Let A ๏ฝ . Evaluate the determinant by ๏ช2 0 3 4๏บ ๏ช ๏บ ๏ซ0 ๏ญ 3 2 1 ๏ป (a) reducing ๐ด into a triangular matrix, and (b) expanding about the second row. ASAQ 2-4 (a) Reducing ๐ด into a triangular matrix using elementary row operation we have 4 −4 1 2 | 2 0 0 −3 2 0 3 2 1 3 | ๐ 1 4 1 −4๐ 1 + ๐ 2 −2๐ 1 + ๐ 3 3๐ถ4 + ๐ถ2 −2๐ถ4 + ๐ถ3 1 2 4 −4 ๐ 2 = − | 2 0 0 −3 0 2 3 2 3 1 | 4 1 By theorem 2.8.2.ii 0 3 2 − 11 | 3 −2 2 1 By theorem 2.8.2.iii 1 11 − 6 3 ๐ถ2 0 − 45 24 − 11 = −| | ๐ถ3 0 − 10 7 − 2 0 0 0 1 By theorem 2.8.2.iii 1 2 ๐ 2 0 − 12 = −| ๐ 3 0 −4 0 −3 58 1 9 ๐ 2 −2๐ 2 + ๐ 3 −6 1 11 24/9 ๐ 2 = −9 |0 −5 0 −10 7 0 0 0 1 11 ๐ 3 = −9 | 0 −5 0 −10 0 0 3 −11/9 | −2 1 By theorem 2.8.2.i −6 3 24/9 −11/9 | By theorem 2.8.2.iii 5/3 4/9 0 1 5 Hence |๐ด| = −9(1)(−5) (3) (1) = 75 (b) Before we expand about the second row, let us introduce more zeros on the second row by applying elementary column operations on the determinant. 4 −4 1 2 | 2 0 0 −3 2 0 3 2 1 3 −2๐ถ1 + ๐ถ2 | 4 −3๐ถ1 + ๐ถ4 1 4 − 12 ๐ถ2 1 0 = | ๐ถ4 2 −4 0 −3 2 − 11 0 0 | 3 −2 2 1 By theorem 2.8.2.iii Expanding about the 2nd row we have 4 − 12 1 0 | 2 −4 0 −3 2 − 11 0 0 | = 1๐ด21 3 −2 2 1 −12 2 −11 = 1(−1)2+1 | −4 3 −2 | −3 2 1 Next we introduce zeros on the third row by using column operations. −12 2 −11 3๐ถ + ๐ถ 3 1 − | −4 3 −2 | −2๐ถ3 + ๐ถ2 −3 2 1 −45 24 ๐ถ1 = − |−10 7 ๐ถ2 0 0 −11 −2 | 1 −45 24 = −(−1)3+3 | | expanding about the −10 7 3rd row = −(−315 + 240) = 75 59 SAQ 2-5 Prove: If the ๐th row of A is multiplied by a scalar ๐, then the determinant of ๐ด is multiplied by ๐. ASAQ 2-5 ๐11 ๐21 | โฎ Let |๐ต| = ๐๐ | ๐1 โฎ ๐๐1 ๐12 ๐22 โฎ ๐๐๐2 โฎ ๐๐2 โฏ ๐1๐ โฏ ๐2๐ โฎ | โฏ ๐๐๐๐ | โฎ โฏ ๐๐๐ If we expand about the ๐th row, we have |๐ต| = ๐๐๐1 ๐ด๐1 + ๐๐๐2 ๐ด๐2 + โฏ + ๐๐๐๐ ๐ด๐๐ = ๐(๐๐1 ๐ด๐1 + ๐๐2 ๐ด๐2 + โฏ + ๐๐๐ ๐ด๐๐ ) = ๐|๐ด| Definition 2.9.2: If A ๏ฝ [ a ij ] is an ๐ x ๐ matrix, the adjoint of ๐ด denoted by adj๐ด, is the transpose of the matrix of cofactors. Thus, ๏ฉ A11 ๏ชA adjA ๏ฝ ๏ช 12 ๏ช ๏ช ๏ซ๏ช A1n A21 ... An1 ๏น A22 ... An 2 ๏บ๏บ ๏บ ๏บ A2 n ... Ann ๏ป๏บ 60 ๏ฉ 2 1 3๏น Example 3: Let A ๏ฝ ๏ช๏ช ๏ญ1 2 0 ๏บ๏บ . Compute adj ๐ด. ๏ช๏ซ 3 ๏ญ2 1 ๏บ๏ป Solution: 2 0 1 3 A11 ๏ฝ (๏ญ1)1๏ซ1 ๏ฝ2 A21 ๏ฝ (๏ญ1) 2๏ซ1 ๏ฝ ๏ญ7 ๏ญ2 1 ๏ญ2 1 A31 ๏ฝ (๏ญ1)3๏ซ1 1 3 ๏ฝ ๏ญ6 2 0 A12 ๏ฝ (๏ญ1)1๏ซ 2 ๏ญ1 0 ๏ฝ1 3 1 A22 ๏ฝ (๏ญ1) 2๏ซ 2 2 3 ๏ฝ ๏ญ7 3 1 A32 ๏ฝ (๏ญ1)3๏ซ 2 2 3 ๏ฝ ๏ญ3 ๏ญ1 0 A13 ๏ฝ (๏ญ1)1๏ซ3 ๏ญ1 2 ๏ฝ ๏ญ4 3 ๏ญ2 A23 ๏ฝ (๏ญ1) 2๏ซ3 2 1 ๏ฝ7 3 ๏ญ2 A33 ๏ฝ (๏ญ1)3๏ซ3 2 1 ๏ฝ5 ๏ญ1 2 Hence, ๏ฉ 2 ๏ญ7 ๏ญ6 ๏น adj ๐ด = ๏ช๏ช 1 ๏ญ7 ๏ญ3๏บ๏บ ๏ช๏ซ ๏ญ4 7 5 ๏บ๏ป Theorem 2.9.2: If A ๏ฝ [ a ij ] is an ๐ x ๐ matrix, then ๐ด(๐๐๐๐ด) = (๐๐๐๐ด)๐ด = |๐ด|๐ผ๐ ๏ฉ 2 1 3๏น Example 4: Let A ๏ฝ ๏ช๏ช ๏ญ1 2 0 ๏บ๏บ be the matrix of Example 3. Compute |๐ด| using Theorem ๏ช๏ซ 3 ๏ญ2 1 ๏บ๏ป 2.9.2. Solution: ๏ฉ 2 1 3๏น ๏ฉ 2 ๏ญ7 ๏ญ6๏น ๏ฉ ๏ญ7 0 0 ๏น ๏ฉ1 0 0 ๏น ๏ช ๏บ ๏ช ๏บ ๏ช ๏บ ๐ด(adj ๐ด) = ๏ช ๏ญ1 2 0๏บ ๏ช 1 ๏ญ7 ๏ญ3๏บ = ๏ช 0 ๏ญ7 0 ๏บ ๏ฝ ๏ญ7 ๏ช๏ช0 1 0 ๏บ๏บ ๏ซ๏ช 3 ๏ญ2 1 ๏บ๏ป ๏ช๏ซ ๏ญ4 7 5 ๏ป๏บ ๏ซ๏ช 0 0 ๏ญ7 ๏ป๏บ ๏ซ๏ช0 0 1 ๏ป๏บ Thus, A ๏ฝ ๏ญ7 61 Corollary 2.9.1: If A is ๐ x ๐ matrix and A ๏น 0 , then A ๏ญ1 ๏ฝ 1 (adjA) A ๏ฉ 2 1 3๏น Example 5: Let A ๏ฝ ๏ช๏ช ๏ญ1 2 0 ๏บ๏บ be the matrix of Example 3, find ๐ด−1 . ๏ช๏ซ 3 ๏ญ2 1 ๏บ๏ป Solution: From previous examples we have ๏ฉ 2 ๏ญ7 ๏ญ6 ๏น |๐ด| = −7 and adj ๐ด = ๏ช 1 ๏ญ7 ๏ญ3๏บ ๏ช ๏บ ๏ช๏ซ ๏ญ4 7 5 ๏บ๏ป Hence, −1 ๐ด ๏ฉ 2 ๏ญ7 ๏ญ6 ๏น = −7 ๏ช๏ช 1 ๏ญ7 ๏ญ3๏บ๏บ ๏ช๏ซ ๏ญ4 7 5 ๏บ๏ป 1 −2/7 = [−1/7 4/7 1 6/7 1 3/7 ] −1 −5/7 You can check your answer by taking the product ๐ด๐ด−1 and making sure that the answer is the identity matrix ๐ผ3 . Theorem 2.9.3: A matrix A is nonsingular if and only if A ๏น 0 . Corollary 2.9.2: If A is an ๐ x ๐ matrix, then the homogeneous system AX = 0 has a nontrivial solution if and only if A ๏ฝ 0 . 62 Theorem 2.9.4: (Cramer’s Rule) Let a11 x1 ๏ซ a12 x2 ๏ซ ... ๏ซ a1n xn ๏ฝ b1 a21 x1 ๏ซ a22 x2 ๏ซ ... ๏ซ a2n xn ๏ฝ b2 an1 x1 ๏ซ an 2 x2 ๏ซ ... ๏ซ ann xn ๏ฝ bn be a linear system of ๐ equations in ๐ unknowns and let A ๏ฝ [ a ij ] be the coefficient matrix so that we can write the given system as AX = B, where ๏ฉb1 ๏น ๏ชb ๏บ B ๏ฝ ๏ช 2๏บ ๏ช ๏บ ๏ช ๏บ ๏ซ๏ชbn ๏ป๏บ If A ๏น 0 , then the system has the unique solution x1 ๏ฝ A1 A , x2 ๏ฝ A2 A , … , xn ๏ฝ An A , where ๐ด๐ is the matrix obtained by replacing the ๐th column of ๐ด by ๐ต. Example 6: Using Cramer’s Rule, determine the solution of the linear system 3x + y – z = 4 - x + y + 3z = 0 x + 2y + z = 1 Solution: The determinant of the coefficient matrix is (you must do the calculation here) 3 1 ๏ญ1 A ๏ฝ ๏ญ 1 1 3 = −8 and 1 2 1 4 1 ๏ญ1 4 A1 ๏ฝ 0 1 3 = −16 (๐ด1 is obtained by replacing the first column of ๐ด by ๐ต = [0]) 1 1 2 1 63 3 4 ๏ญ1 A2 ๏ฝ ๏ญ 1 0 1 1 3 3 =8, 1 1 4 A3 ๏ฝ ๏ญ 1 1 0 = −8 1 2 1 4 (๐ด2 is obtained by replacing the second column of ๐ด by ๐ต = [0]) 1 4 (๐ด3 is obtained by replacing the third column of ๐ด by ๐ต = [0]) 1 By Cramer’s Rule, x๏ฝ A1 A ๏ฝ A2 A3 ๏ญ 8 8 ๏ญ 16 ๏ฝ ๏ฝ ๏ญ1, and z ๏ฝ ๏ฝ 2, y ๏ฝ ๏ฝ ๏ฝ1 A ๏ญ8 ๏ญ8 A ๏ญ8 Thus the solution set is {(2, -1, 1)}. NOTE: Cramer’s rule is only applicable to the case where we have ๐ equations in ๐ unknowns and where the coefficient matrix ๐ด is nonsingular. Cramer’s rule becomes computationally inefficient for ๐ > 4, and it is better to use the Gauss – Jordan method. ACTIVITY I. Solve the following system using Cramer’s Rule. 1. 2๐ฅ1 + 5๐ฅ2 − ๐ฅ3 = −1 4๐ฅ1 + ๐ฅ2 + 3๐ฅ3 = 3 −2๐ฅ1 + 2๐ฅ2 = 0 2. ๐ฅ1 + ๐ฅ2 + ๐ฅ3 + ๐ฅ4 = 6 2๐ฅ1 − ๐ฅ3 − ๐ฅ4 = 4 3 ๐ฅ3 + 6๐ฅ4 = 3 ๐ฅ1 − ๐ฅ4 = 5 1 a a2 3. Show that 1 b b 2 = (b-a)(c-a)(c-b). Also, find the inverse of ๐ด by using the formula 1 c c2 1 ๐ด−1 = |๐ด| adj ๐ด. 2 2 4. Find the determinant of ๐ด = 4 1 [2 1 3 5 0 1 0 1 2 1 0 −1 3 2 5 3 −3 . 2 4 1 0] 64 MODULE 3 VECTOR SPACES OVER A FIELD Introduction In this chapter we will discuss vectors in โ๐ and its basic properties, properties and structure of a vector space and subspace, linear combination and spanning sets. Objectives At the end of this chapter, you are expected to be able to do the following: 1. Define vector, vector spaces and subspaces. 2. Enumerate and explain the properties of vector spaces and subspaces. 3. Determine whether a set with given operations is a vector space. 4. Determine whether the given subsets of โ๐ are subspaces. 5. Define linear combination and spanning. 6. Check whether a given set of vectors spans a vector space V. 3.1 Vectors in the Plane Vectors are measurable quantities which has magnitude and direction. Example of vectors are velocity, force, and acceleration. Vectors in โ๐ A vector on the plane โ๐ can be described as an ordered pair ๐ = (๐ฅ1 , ๐ฅ2 ) where ๐ฅ1 , ๐ฅ2 ∈ โ. It can also be denoted by a 2 x 1 matrix ๐ฅ1 ๐ = [๐ฅ ] 2 With X we associate the directed line segment with initial point at the origin O and the โโโโโ . terminal point at ๐(๐ฅ1 , ๐ฅ2 ), denoted by ๐๐ The direction of a directed line segment is the angle made with the positive X-axis and the magnitude of a directed line segment is its length. 65 Vector Operations Definition 3.1.1: Let ๐ = (๐ฅ1 , ๐ฅ2 ) and ๐ = (๐ฆ1 , ๐ฆ2 ) be two vectors in the plane. The sum of the vectors X and Y, denoted by ๐ฟ + ๐, is the vector (๐ฅ1 + ๐ฆ1 , ๐ฅ2 + ๐ฆ2 ). Example 1: Let X = (2, 3) and Y = (-4, 1). Then X + Y = (-2, 4) Definition 3.1.2: If X = (๐ฅ, ๐ฆ) and ๐ is a scalar (a real number), then the scalar multiple ๐๐ฟ of ๐ฟ by ๐ is the vector (๐๐ฅ, ๐๐ฆ). If ๐ > 0, then ๐๐ฅ is in the same direction as ๐ฟ, whereas if ๐ < 0, then ๐๐ฅ is in the opposite direction. Example 2: Let c = 3, d = -2, and X = (3, -2). Then cX = (9, -6) and dX = (-6, 4) NOTE: 1. ๐ = (0, 0) is called the zero vector. 2. X + ๐ = X 3. X + (-1)X = ๐ 4. (-1)X = X; the negative of X 5. X + (-1)Y = X – Y; the difference between X and Y. 3.2. ๐-Vectors Definition 3.2.1: An ๐-vector is an ๐ x 1 matrix ๐ฅ1 ๐ฅ2 ๐ฟ=[โฎ ] ๐ฅ๐ where ๐ฅ1 , ๐ฅ2 , … , ๐ฅ๐ are real numbers, which are called the components of X. The set of all ๐-vectors is denoted by โ๐ and is called ๐-space. 66 Definition 3.2.2: Let ๐ฟ = (๐ฅ1 , ๐ฅ2 , … , ๐ฅ๐ ) and ๐ = (๐ฆ1 , ๐ฆ2 , … , ๐ฆ๐ ) be two vectors in โ๐ . The sum of the vectors X and Y is the vector (๐ฅ1 + ๐ฆ1 , ๐ฅ2 + ๐ฆ2 , … , ๐ฅ๐ + ๐ฆ๐ ) and is denoted by X + Y. Example 1: If X = (1, 2, -3) and Y = (0, 1, -2) are vectors in R3, then X + Y = (1+0, 2+1, -3 + (-2) ) = ( 1, 3, -5 ). Definition 3.2.3: If ๐ฟ = (๐ฅ1 , ๐ฅ2 , … , ๐ฅ๐ ) is a vector in โ๐ and ๐ is a scalar, then the scalar multiple ๐๐ฟ of ๐ฟ by ๐ is the vector (๐๐ฅ1 , ๐๐ฅ2 , … , ๐๐ฅ๐ ) Example 2: If X = ( 2, 1, 5, -2 ) is a vector in โ4 and c = -3, then cX = (-3) ( 2, 1, 5, -2 ) = (-6, -3, -15, 6 ) The operations of vector addition and scalar multiplication satisfy the following properties: Theorem 3.2.1: Let X, Y, and Z be any vectors in โ๐ ; let ๐ and d be any scalars. Then I. X + Y is a vector in โ๐ (that is, โ๐ is closed under the operation of vector addition). a. X + Y = Y + X b. X + ( Y + Z ) = ( X + Y ) +Z c. There is a unique vector in โ๐ , 0 = (0, 0, …, 0) such that X + 0 = 0 + X = X. d. There is a unique vector – X, -X = (- x1 , - x2 , … , - xn ) such that X + (- X) = 0 67 II. cX is a vector in โ๐ a. ( X + Y ) = cX + cY b. ( c + d )X = cX + dX c. c(dX) = (cd)X d. 1X = X 3.3 Vector Spaces and Subspaces Definition 3.3.1: A vector space V over a field ๐น is a nonempty set of elements, called vectors, together with two operations called vector addition and scalar multiplication satisfying the following properties: [A1] (Closure under addition) To every pair of vectors X, Y ∈ V, then X + Y ∈ V. [A2] (Commutative law of vector addition) X+ Y= Y+X ∀ X , Y ๏ V. [A3] (Associative law of vector addition) X + ( Y + Z ) = ( X + Y ) + Z ∀ X, Y, and Z ๏ V. [A4] (Zero Element) ∃๐ถ ∈ V such that X + ๐ถ = ๐ถ + X = X ∀ X ๏ V. [A5] (Negative Elements) ∀ X ๏ V, ∃ X ∈ V such that X + ( X) = ๐ถ [M1] (Closure under scalar multiplication) To every X ๏ V and ๐ ∈ ๐น, then ๐ X ๏ V. [M2] (Distributive Law) c ( X + Y ) = c X + c Y, ∀๐ ∈ ๐น , ∀ X, Y ๏ V. 68 [M3] (Distributive Law) (๐ + ๐ ) X = ๐ X + ๐ X, ∀๐, ๐ ∈ ๐น, ∀ X ๏ V. [M4] (Associative Law of Scalar Multiplication) (๐๐ )X = ๐(๐ X ), ∀๐, ๐ ∈ ๐น, ∀ X ๏ V. [M5] (Identity Element) 1 X = X, ∀ X ๏ V. Example 1. Consider the set ๐๐ of all polynomials of degree ≤ ๐ together with the zero polynomial. Show that ๐๐ is a vector space. Solution: We have to show that all properties are satisfied. A polynomial in ๐๐ is expressible as ๐(๐ฅ) = ๐0 ๐ฅ ๐ + ๐1 ๐ฅ ๐−1 + … + ๐๐−1 ๐ฅ + ๐๐ where ๐0 , ๐1 , … , ๐๐ are real numbers. Let ๐(๐ฅ), ๐(๐ฅ), and ๐(๐ฅ) ∈ ๐๐ and ๐, ๐ ∈ โ. Then [A1] ๐(๐ฅ) + ๐(๐ฅ) = (๐0 + ๐0 )๐ฅ ๐ + (๐1 + ๐1 )๐ฅ ๐−1 + โฏ + (๐๐−1 + ๐๐−1 )๐ฅ + (๐๐ + ๐๐ ) Clearly, the sum of two polynomials of degree ≤ ๐ is another polynomial with degree ≤ ๐. Hence ๐๐ is closed under addition. [A2] ๐(๐ฅ) + ๐(๐ฅ) = (๐0 + ๐0 )๐ฅ ๐ + (๐1 + ๐1 )๐ฅ ๐−1 + โฏ + (๐๐−1 + ๐๐−1 )๐ฅ + (๐๐ + ๐๐ ) = (๐0 + ๐0 )๐ฅ ๐ + (๐1 + ๐1 )๐ฅ ๐−1 + โฏ + (๐๐−1 + ๐๐−1 )๐ฅ + (๐๐ + ๐๐ ) = ๐(๐ฅ) + ๐(๐ฅ) Thus addition is commutative. [A3] [๐(๐ฅ) + ๐(๐ฅ)] + ๐(๐ฅ) = [(๐0 + ๐0 )๐ฅ ๐ + (๐1 + ๐1 )๐ฅ ๐−1 + โฏ + (๐๐−1 + ๐๐−1 )๐ฅ +(๐๐ + ๐๐ )] + ๐0 ๐ฅ ๐ + ๐1 ๐ฅ ๐−1 + … + ๐๐−1 ๐ฅ + ๐๐ = [(๐0 + ๐0 ) + ๐0 ]๐ฅ ๐ + [(๐1 + ๐1 )+๐1 ]๐ฅ ๐−1 + โฏ + [(๐๐−1 + ๐๐−1 ) + ๐๐−1 ]๐ฅ +[(๐๐ + ๐๐ ) + ๐๐ ] = [๐0 + (๐0 + ๐0 )]๐ฅ ๐ + [๐1 + (๐1 + ๐1 )]๐ฅ ๐−1 + โฏ + [๐๐−1 + (๐๐−1 + ๐๐−1 )]๐ฅ + [๐๐ + (๐๐ + ๐๐ )] 69 = (๐0 ๐ฅ ๐ + ๐1 ๐ฅ ๐−1 + … + ๐๐−1 ๐ฅ + ๐๐ ) + [(๐0 + ๐0 )๐ฅ ๐ + (๐1 + ๐1 )๐ฅ ๐−1 + โฏ + (๐๐−1 + ๐๐−1 )๐ฅ + (๐๐ + ๐๐ )] = ๐(๐ฅ) + [๐(๐ฅ) + ๐(๐ฅ)] Thus vector addition is associative. [A4] Let ๐ = 0๐ฅ ๐ + 0๐ฅ ๐−1 + … + 0๐ฅ + 0 be the zero polynomial then ๐(๐ฅ) + ๐ = ๐ + ๐(๐ฅ) = ๐(๐ฅ) [A5] Let – ๐(๐ฅ) = −๐0 ๐ฅ ๐ − ๐1 ๐ฅ ๐−1 − โฏ − ๐๐−1 ๐ฅ − ๐๐ , then ๐(๐ฅ) + (−๐(๐ฅ)) = ๐ [M1] ๐๐(๐ฅ) = ๐๐0 ๐ฅ ๐ + ๐๐1 ๐ฅ ๐−1 + … + ๐๐๐−1 ๐ฅ + ๐๐๐ where ๐๐0 , ๐๐1 , … , ๐๐๐ ∈ โ . Clearly , the product of a real number and a polynomial is also a polynomial. Hence ๐๐ is closed under scalar multiplication. [M2] ๐[๐(๐ฅ) + ๐(๐ฅ)] = ๐[(๐0 + ๐0 )๐ฅ ๐ + (๐1 + ๐1 )๐ฅ ๐−1 + โฏ + (๐๐−1 + ๐๐−1 )๐ฅ +(๐๐ + ๐๐ )] = ๐(๐0 + ๐0 )๐ฅ ๐ + ๐(๐1 + ๐1 )๐ฅ ๐−1 + โฏ + ๐(๐๐−1 + ๐๐−1 )๐ฅ + ๐(๐๐ + ๐๐ ) = (๐๐0 + ๐๐0 )๐ฅ ๐ + (๐๐1 + ๐๐1 )๐ฅ ๐−1 + โฏ + (๐๐๐−1 + ๐๐๐−1 )๐ฅ + (๐๐๐ + ๐๐๐ ) = (๐๐0 ๐ฅ ๐ + ๐๐1 ๐ฅ ๐−1 + … + ๐๐๐−1 ๐ฅ + ๐๐๐ ) + (๐๐0 ๐ฅ ๐ + ๐๐1 ๐ฅ ๐−1 + … + ๐๐๐−1 ๐ฅ + ๐๐๐ ) = ๐๐(๐ฅ) + ๐๐(๐ฅ) [M3] (๐ + ๐)๐(๐ฅ) = (๐ + ๐)(๐0 ๐ฅ ๐ + ๐1 ๐ฅ ๐−1 + … + ๐๐−1 ๐ฅ + ๐๐ ) = (๐ + ๐)๐0 ๐ฅ ๐ + (๐ + ๐)๐1 ๐ฅ ๐−1 + … + (๐ + ๐)๐๐−1 ๐ฅ + (๐ + ๐)๐๐ 70 = (๐๐0 + ๐๐0 )๐ฅ ๐ + (๐๐1 + ๐๐1 )๐ฅ ๐−1 + โฏ + (๐๐๐−1 + ๐๐๐−1 )๐ฅ + (๐๐๐ + ๐๐๐ ) = (๐๐0 ๐ฅ ๐ + ๐๐1 ๐ฅ ๐−1 + … + ๐๐๐−1 ๐ฅ + ๐๐๐ ) + ๐๐0 ๐ฅ ๐ + ๐๐1 ๐ฅ ๐−1 + … + ๐๐๐−1 ๐ฅ + ๐๐๐ = ๐๐(๐ฅ) + ๐๐(๐ฅ) [M4] (๐๐)๐(๐ฅ) = (๐๐)๐0 ๐ฅ ๐ + (๐๐)๐1 ๐ฅ ๐−1 + … + (๐๐)๐๐−1 ๐ฅ + (๐๐)๐๐ = ๐(๐๐0 )๐ฅ ๐ + ๐(๐๐1 )๐ฅ ๐−1 + … + ๐(๐๐๐−1 )๐ฅ + ๐(๐๐๐ ) = ๐(๐๐0 ๐ฅ ๐ + ๐๐1 ๐ฅ ๐−1 + … + ๐๐๐−1 ๐ฅ + ๐๐๐ ) = ๐(๐๐(๐ฅ)) [M5] 1๐(๐ฅ) = 1๐0 ๐ฅ ๐ + 1๐1 ๐ฅ ๐−1 + … + 1๐๐−1 ๐ฅ + 1๐๐ = ๐0 ๐ฅ ๐ + ๐1 ๐ฅ ๐−1 + … + ๐๐−1 ๐ฅ + ๐๐ = ๐(๐ฅ) Since all ten properties are satisfied, then ๐๐ is a vector space. Example 2. Let V be the set of all ordered triples of real numbers (๐ฅ, ๐ฆ, ๐ง) with the operations (๐ฅ, ๐ฆ, ๐ง) + (๐ฅ ′ , ๐ฆ ′ , ๐ง ′ ) = (๐ฅ + ๐ฅ ′ , ๐ฆ + ๐ฆ ′ , ๐ง + ๐ง ′ ) ๐(๐ฅ, ๐ฆ, ๐ง) = (๐ฅ, 1, ๐ง) Determine if V is a vector space. Solution: To prove that V is a vector space, we have to show that all ten properties are satisfied. Let ๐ฟ = (๐ฅ, ๐ฆ, ๐ง), ๐ = (๐ฅ ′ , ๐ฆ ′ , ๐ง ′ ) and ๐ = (๐ฅ ′′ , ๐ฆ ′′ , ๐ง ′′ ) [A1] V and ๐, ๐ ∈ โ. Then ๐ฟ + ๐ = (๐ฅ + ๐ฅ ′ , ๐ฆ + ๐ฆ ′ , ๐ง + ๐ง ′ ) Since ๐ฅ + ๐ฅ ′ , ๐ฆ + ๐ฆ ′ , and ๐ง + ๐ง ′ addition. ∈ โ, then ๐ฟ + ๐ ∈ V. Hence V is closed under 71 [A2] ๐ฟ + ๐ = (๐ฅ + ๐ฅ ′ , ๐ฆ + ๐ฆ ′ , ๐ง + ๐ง ′ ) = (๐ฅ ′ + ๐ฅ, ๐ฆ ′ + ๐ฆ, ๐ง ′ + ๐ง) =๐+๐ฟ Hence vector addition is commutative. [A3] (๐ฟ + ๐) + ๐ = (๐ฅ + ๐ฅ ′ , ๐ฆ + ๐ฆ ′ , ๐ง + ๐ง ′ ) + (๐ฅ ′′ , ๐ฆ ′′ , ๐ง ′′ ) = ((๐ฅ + ๐ฅ ′ ) + ๐ฅ ′′ , (๐ฆ + ๐ฆ ′ ) + ๐ฆ ′′ , (๐ง + ๐ง ′ ) + ๐ง′′) = (๐ฅ + (๐ฅ ′ + ๐ฅ ′′ ), ๐ฆ + (๐ฆ ′ + ๐ฆ ′′ ), ๐ง + (๐ง ′ + ๐ง ′′ )) = (๐ฅ, ๐ฆ, ๐ง) + (๐ฅ ′ + ๐ฅ ′′ , ๐ฆ′ + ๐ฆ ′′ , ๐ง ′ + ๐ง ′′ ) = ๐ฟ + (๐ + ๐) Thus vector addition is associative. [A4] Let ๐ถ = (0, 0, 0) ∈ V be the zero vector then ๐ฟ + ๐ถ = (๐ฅ, ๐ฆ, ๐ง) + (0, 0, 0) = (๐ฅ, ๐ฆ, ๐ง) =๐ฟ [A5] Let – ๐ฟ = (−๐ฅ, −๐ฆ, −๐ง) then ๐ฟ + (−๐ฟ) = (๐ฅ, ๐ฆ, ๐ง) + (−๐ฅ, −๐ฆ, −๐ง) = (0, 0, 0) =๐ถ [M1] ๐๐ฟ = ๐(๐ฅ, ๐ฆ, ๐ง) = (๐ฅ, 1, ๐ง) Since ๐ฅ and ๐ง are real numbers then ๐๐ฟ ∈ V. Hence V is closed under scalar multiplication. 72 [M2] ๐(๐ฟ + ๐) = ๐(๐ฅ + ๐ฅ ′ , ๐ฆ + ๐ฆ ′ , ๐ง + ๐ง ′ ) = (๐ฅ + ๐ฅ ′ , 1, ๐ง + ๐ง ′ ) Also, ๐๐ฟ + ๐๐ = ๐(๐ฅ, ๐ฆ, ๐ง) + ๐(๐ฅ ′ , ๐ฆ ′ , ๐ง ′ ) = (๐ฅ, 1, ๐ง) + (๐ฅ ′ , 1, ๐ง ′ ) = (๐ฅ + ๐ฅ ′ , 2, ๐ง + ๐ง ′ ) Since ๐(๐ฟ + ๐) ≠ ๐๐ฟ + ๐๐ then M2 is not satisfied. Hence V under the prescribed operations is not a vector space. Theorem 3.3.1: If V is a vector space, then: a) 0๐ฟ = 0, for every ๐ฟ in V. b) ๐0 = 0, for every scalar ๐. c) If ๐๐ฟ = 0, then ๐ = 0 or ๐ฟ = 0 d) (-1)๐ฟ = −๐ฟ, for every ๐ฟ in V. SAQ 3-1 Consider the set ๐3,3 (โ) of all 3 x 3 matrices with entries in under the usual operations of matrix addition and scalar multiplication. Show that ๐3,3 (โ) is a vector space. ASAQ 3-1 Let ๐ด, ๐ต, and ๐ถ ∈ ๐3,3 (โ) and ๐, ๐ ∈ โ. Then [A1] For every ๐ด, ๐ต ๐3,3 (โ), we have ๐ด + ๐ต ∈ ๐3,3 (โ). Hence ๐3,3 (โ) is closed under addition. [A2] For every ๐ด, ๐ต commutative. ๐3,3 (โ), we have ๐ด + ๐ต = ๐ต + ๐ด, that is matrix addition is 73 [A3] For every ๐ด, ๐ต, ๐ถ ๐3,3 (โ), we have (๐ด + ๐ต) + ๐ถ = ๐ด + (๐ต + ๐ถ), that is matrix addition is associative. [A4] Let ๐ be the 3 x 3 zero matrix, then for every ๐ด ๐ด+๐ =๐+๐ด=๐ด [A5] For every ๐ด ๐3,3 (โ), we have ๐3,3 (โ), we have ๐ด + (−๐ด) = ๐. [M1] For every ๐ ∈ โ and ๐ด ๐3,3 (โ), we have ๐๐ด closed under scalar multiplication. ๐3,3 (โ). Hence ๐3,3 (โ) is [M2] For every ๐ ∈ โ and ๐ด, ๐ต ๐3,3 (โ), we have ๐(๐ด + ๐ต) = ๐๐ด + ๐๐ต. [M3] For every ๐, ๐ ∈ โ and ๐ด ๐3,3 (โ), we have (๐ + ๐)๐ด = ๐๐ด + ๐๐ด. [M4] For every ๐, ๐ ∈ โ and ๐ด ๐3,3 (โ), we have (๐๐)๐ด = ๐(๐๐ด). [M5] For every ๐ด ๐3,3 (โ),we have 1๐ด = ๐ด. Hence ๐3,3 (โ) under the usual operations of matrix addition and scalar multiplication is a vector space. Definition 3.3.2: Let V be a vector space and W a nonempty subset of V. If W is a vector space under the operations of addition and scalar multiplication defined on V, then W is called a subspace of V. Theorem 3.3.2: Let V be a vector space under the operations addition and scalar multiplication and let W be a nonempty subset of V. Then W is a subspace of V if and only if the following conditions hold: a) The zero vector ๐ถ belongs to W; b) For every vector ๐ฟ, ๐ ๏ W, ๐ ๏ โ: i. The sum ๐ฟ + ๐ ๏ W ii. The multiple ๐๐ฟ ๏ W. Property (i) in (b) states that W is closed under vector addition, and property (ii) in (b) states that W is closed under scalar multiplication. Both properties may be combined into the following equivalent single statement: (b’) For every ๐ฟ, ๐ ๏ W, ๐, ๐ ๏ , the linear combination ๐๐ฟ + ๐๐ ๏ W. 74 REMARK: If V is any vector space, then V automatically contains two subspaces, the set {0} consisting of the zero vector alone and the whole space V itself. These are called the trivial subspaces of V. Example 3: Consider the vector space โ3 . Let U consists of all vectors in โ3 whose entries are equal; that is U = { (a, b, c) : a = b = c }. Show that U is a subspace of โ3 . Solution: (a) The zero vector ( 0, 0, 0 ) belongs to U. (b) Let ๐ฟ = (๐, ๐, ๐) and ๐ = (๐, ๐, ๐) be vectors in U and let ๐, ๐ ๏ . ๐๐ฟ + ๐๐ = (๐๐, ๐๐, ๐๐) + (๐๐, ๐๐, ๐๐) = (๐๐ + ๐๐, ๐๐ + ๐๐, ๐๐ + ๐๐) Clearly, ๐๐ฟ + ๐๐ U hence U is a subspace of โ3 . Example 4: Let V = โ3 and let W = { (a, b, c ): a ๏ณ 0}. Determine if W is a subspace of V. Solution: (a) The zero vector ( 0, 0, 0 ) ๏ W. (b) Let ๐ฟ = (๐, ๐, ๐) and ๐ = (๐′ , ๐ ′ , ๐ ′ ) be vectors in W. ๐ + ๐ = (๐ + ๐′ , ๐ + ๐ ′ , ๐ + ๐ ′ ) Since ๐ ๏ณ 0 and ๐′ ๏ณ 0 then ๐ + ๐′ ๏ณ 0. Hence ๐ฟ + ๐ ๏ W. (c) Let ๐ = (1, 2, 3) and ๐ = −2 then ๐๐ฟ = (−2, −4, −6) ๏ W since −2 < 0. Therefore, W is not a subspace of V. 75 ACTIVITY I. In problems 1-4, determine whether the given set together with the given operations is a vector space. 1. The set of ordered pairs (๐, ๐) of real numbers with the operations (๐, ๐) + (๐, ๐) = (๐ + ๐, ๐ + ๐) and ๐(๐, ๐) = (๐๐, 0) 2. The set of all ordered triples of real numbers of the form (0,0, ๐ง) with the operations ( 0, 0, ๐ง ) + ( 0, 0, ๐ง′ ) = ( 0, 0, ๐ง + ๐ง′ ) and ๐ ( 0, 0, ๐ง ) = ( 0, 0, ๐๐ง ). 3. The set of polynomials (in ๐ฅ) of degree ๏ฃ ๐ together with the zero polynomial with positive constant term. 4. The set of all ordered pairs (๐, ๐) of real numbers with addition and scalar multiplication in V defined by (๐, ๐) + (๐, ๐) = (๐๐, ๐๐) and ๐(๐, ๐) = (๐๐, ๐๐) 5. Which of the following subsets of โ3 are subspaces of โ3 ? The set of all vectors of the form (a) (๐, ๐, ๐), where ๐ = ๐ = 0 (b) (๐, ๐, ๐),, where ๐ = −๐ (c) (๐, ๐, ๐),, where ๐ = 2๐ + 1 6. Let V be the set of all 2 x 3 matrices under the usual operations of matrix addition and scalar multiplication. Which of the following subsets of V are subspaces? The set of all matrices of the form ๏ฉa (a) ๏ช ๏ซd b c๏น , where ๐ = ๐ + ๐. 0 0 ๏บ๏ป ๏ฉa (b) ๏ช ๏ซd b c๏น , where ๐ > 0 0 0 ๏บ๏ป 76 3.4 Linear Combinations and Spanning Sets Definition 3.4.1: A vector ๐ฟ ๏ V is said to be a linear combination of the set of vectors { ๐ฟ๐ , ๐ฟ๐ , โฏ , ๐ฟ๐ } ๏ V if there exists scalars ๐1 , ๐2 , โฏ , ๐๐ such that ๐ฟ = ๐1 ๐ฟ๐ + ๐2 ๐ฟ๐ + โฏ + ๐๐ ๐ฟ๐ We also say that ๐ฟ is linearly dependent on the ๐ฟ๐ . Example 1. In โ3 , (−15, −4, 0) is a linear combination of (−3, −1, 4) and (3, 2, 8) since 2(−3, −1, 4) − 3(3, 2, 8) = (−15, −4, 0). −1 Example 2. [ −1 6 7 −1 0 3 1 3 −1 ] is a linear combination of [ ] and [ ] since 15 16 1 5 6 −2 0 −1 −1 0 3[ 1 5 3 1 3 ] + 2[ 6 −2 0 −1 −1 6 ]=[ −1 −1 15 7 ] 16 Example 3. In โ3 , let X1 = (4, 2, 3), X2 = (2, 1, 2), and X3 = ( 2, 1, 0). Determine if (4, 2, 6) is a linear combination of X1, X2, and X3. ๐ฟ= Solution: X is a linear combination of X1, X2, and X3 if we can find ๐1 , ๐2, and ๐3 ๏ that such ๐1(4, 2, 3) + ๐2 (2, 1, 2) + ๐3 ( 2, 1, 0) (4, 2, 6) Multiplying and adding yields the following linear system. 4๐1 + 2๐2 2๐3 = 4 2๐1 + ๐2 + ๐3 = 2 3๐1 2๐2 = 6 Row reducing the augmented matrix we have 4 2 −2 [2 1 1 −3 −2 0 4 2] −6 1 4 ๐ 1 1 1/2 −1/2 ๐ 1 [ 2 1 1 −3 −2 0 1 2] −6 77 −2๐ 1 + ๐ 2 3๐ 1 + ๐ 3 1 ๐ 2 −2๐ 3 2 ๐ 2 1 1/2 ๐ 2 0 [0 ๐ 3 0 −1/2 1 1/2 ๐ 2 [0 0 ๐ 3 0 1 ๐ 3 −1/2 2 −3/2 1 0] −3 −1/2 1 3 1 0] 6 1 1/2 −1/2 [0 1 3 0 0 1 1 6] 0 This implies that ๐1 = 2, ๐2 = 6, and ๐3 = 0. It can be verified that 2(4, 2, 3) + 6(2, 1, 2) + 0( 2, 1, 0) = (4, 2, 6). Thus X is a linear combination of X1 = (4, 2, 3), X2 = (2, 1, 2), and X3 = ( 2, 1, 0). SAQ 3-2 In ๐3 , let ๐(๐ฅ) = ๐ฅ 2 − 2๐ฅ, ๐(๐ฅ) = 5๐ฅ − 2, and ๐(๐ฅ) = ๐ฅ 2 + 3. Determine if ๐(๐ฅ) = 2๐ฅ 2 + 4๐ฅ − 7 is a linear combination of ๐(๐ฅ), ๐(๐ฅ), and ๐(๐ฅ). ASAQ 3-2 ๐(๐ฅ) is a linear combination of ๐(๐ฅ), ๐(๐ฅ), and ๐(๐ฅ) if we can find ๐1 , ๐2 , and ๐3 ∈ โ such that ๐1 (๐ฅ 2 − 2๐ฅ) + ๐2 (5๐ฅ − 2) + ๐3 (๐ฅ 2 + 3) = 2๐ฅ 2 + 4๐ฅ − 7 Equating the coefficients of similar terms we obtain the system ๐1 + ๐3 = 2 −2๐1 + 5๐2 =4 −2๐2 + 3๐3 = −7 Row reducing the augmented matrix we obtain (you must do the calculation here) 78 1 0 [0 1 0 0 1 2/5 1 2 8/5] −1 Thus the solution to the given system is ๐1 = 3, ๐2 = 2, ๐3 = −1. It can be verified that 3(๐ฅ 2 − 2๐ฅ) + 2(5๐ฅ − 2) − (๐ฅ 2 + 3) = 2๐ฅ 2 + 4๐ฅ − 7 Thus ๐(๐ฅ) is a linear combination of ๐(๐ฅ), ๐(๐ฅ), and ๐(๐ฅ). Definition 3.4.2: Let S = { ๐ฟ๐ , ๐ฟ๐ , โฏ , ๐ฟ๐ } be a set of vectors in a vector space V. The set S spans V or V is spanned by S, if every vector in V is a linear combination of the vector in S. That is for every ๐ฟ ∈ V there are scalars ๐1 , ๐2 , โฏ , ๐๐ such that ๐ฟ = ๐1 ๐ฟ๐ + ๐2 ๐ฟ๐ + โฏ + ๐๐ ๐ฟ๐ Example 4. The vectors ๐ฟ1 = (1, 0) and ๐ฟ๐ = (0, 1) span โ2 since every vector (๐, ๐) in โ2 can be expressed as a linear combination of ๐ฟ1 and ๐ฟ๐ . That is (๐, ๐) = ๐(1, 0) + ๐(0, 1) for every ๐, ๐ ∈ โ. Example 5. Every 2 x 2 matrix can be written as a linear combination of the four matrices 1 0 0 1 0 0 0 0 ๐ธ1 = [ ], ๐ธ2 = [ ], ๐ธ3 = [ ], and ๐ธ4 = [ ]. That is 0 0 0 0 1 0 0 1 ๐ [ ๐ 1 0 0 ๐ ] = ๐[ ]+๐[ 0 0 0 ๐ 1 0 ]+๐[ 0 1 0 0 0 ]+๐[ ] 0 0 1 for every ๐, ๐, ๐, ๐ ∈ โ. Thus ๐ธ1 , ๐ธ2 , ๐ธ3 , and ๐ธ4 span ๐2,2 . Example 6. Let V be the vector space โ3 and let S = { X1 = (1, 2, 3), X2 = (−1, −2, 1), X3 = (0, 1, 0) }. Does S span โ3 ? Solution. Choose an arbitrary vector X = (๐, ๐, ๐) in โ3 . The set of vectors S = {๐ฟ๐ , ๐ฟ๐ , ๐ฟ๐ } spans โ3 if we can find scalars ๐1 , ๐2 , and ๐3 such that ๐1(1, 2, 3) + ๐2 (−1, −2, 1) + ๐3 (0, 1, 0) = (๐, ๐, ๐) 79 Multiplying and adding we obtain the linear system ๐1 − ๐2 − = ๐ 2๐1 − 2๐2 + ๐3 = ๐ 3๐1 + ๐2 = ๐ The augmented matrix 1 −1 0 [2 −2 1 3 1 0 ๐ ๐ ] is row equivalent to ๐ 1 0 0 [0 1 0 0 0 1 (๐ + ๐)/4 (−3๐ + ๐)/4] (verify) −2๐ + ๐ The solution to the given system is ๐1 = ๐+๐ 4 , ๐2 = −3๐+๐ 4 , and ๐3 = −2๐ + ๐ . Since there exists ๐1 , ๐2, and ๐3 for every choice of ๐, ๐, and ๐, we say that every vector in โ3 is a linear combination of X1 = (1, 2, 3), X2 = (−1, −2, 1), and X3 = (0, 1, 0). For example, if ๐ฟ = (2, −1, 6) then๐1 = 2, ๐2 = 0 and ๐3 = −5. It can be verified that (2, −1, 6) = 2(1, 2, 3) + 0(−1, −2, 1) − 5(0, 1, 0) Thus S = {๐ฟ๐ , ๐ฟ๐ , ๐ฟ๐ } spans โ3 . Example 7: Does ๐ฟ = {−1, 4, 2, 2} belongs to span {๐ฟ๐ , ๐ฟ๐ , ๐ฟ๐ } where X1 = (1, 0, 0, 1), X2 = (1, −1, 0, 0), and X3 = (0, 1, 2, 1). Solution: Every vector in span {๐ฟ๐ , ๐ฟ๐ , ๐ฟ๐ } is of the form ๐1 ๐ฟ๐ + ๐2 ๐ฟ๐ + ๐3 ๐ฟ๐ Thus ๐ฟ belongs to span {๐ฟ๐ , ๐ฟ๐ , ๐ฟ๐ } if we can write it as a linear combination of ๐ฟ๐ , ๐ฟ๐ , and ๐ฟ๐ . Let ๐1 , ๐2 , and ๐3 be scalars such that ๐1 (1, 0, 0, 1) + ๐2 (1, −1, 0, 0) + ๐3 (0, 1, 2, 1) = (−1, 4, 2, 2) 80 Multiplying and adding we obtain the linear system ๐1 + ๐2 −๐2 + ๐3 2๐3 ๐1 + ๐3 = = = = −1 4 2 2 The third equation implies that ๐3 = 1. Substituting it to the second and fourth equations, we get ๐1 = 1 and ๐2 = −3. However, ๐1 + ๐2 = 1 + (−3) = −2 ≠ −1 Thus ๐ฟ does not belong to span {๐ฟ๐ , ๐ฟ๐ , ๐ฟ๐ }. SAQ 3-3 Let S = {X1 = (1, 1, 0), X2 = (1, 3, 2), X3 = (4, 9, 5) } be vectors in โ3 . Does S span โ3 . ASAQ 3-3 Choose an arbitrary vector X = (๐, ๐, ๐) in โ3 . The set of vectors S = {๐ฟ๐ , ๐ฟ๐ , ๐ฟ๐ } spans โ3 if we can find scalars ๐1 , ๐2 , and ๐3 such that ๐1 (1, 1, 0) + ๐2 (1, 3, 2) + ๐3 (4, 9, 5) = (๐, ๐, ๐) Multiplying and adding we obtain the linear system ๐1 + ๐2 + 4๐3 = ๐ ๐1 + 3๐2 + 9๐3 = ๐ 2๐2 + 5๐3 = ๐ The augmented matrix 1 1 [1 3 0 2 4 9 5 ๐ 1 1 ๐ ] is row equivalent to [0 2 ๐ 0 0 4 5 0 ๐ −1 + ๐ ](verify) ๐−๐+๐ 81 The system has no solution if ๐ − ๐ + ๐ ≠ 0. Thus S = {๐ฟ๐ , ๐ฟ๐ , ๐ฟ๐ } does not span โ3 . For example, the vector (3, 1, 0) cannot be written as a linear combination of ๐ฟ๐ , ๐ฟ๐ , ๐ฟ๐ . ACTIVITY 1. Determine whether ๐(๐ฅ) = 3๐ฅ 2 − 3๐ฅ + 1 belongs to span {๐(๐ฅ), ๐(๐ฅ), ๐(๐ฅ)} where ๐(๐ฅ) = ๐ฅ 2 − ๐ฅ, ๐(๐ฅ) = ๐ฅ 2 − 2๐ฅ + 1 and ๐(๐ฅ) = −๐ฅ 2 + 1. 2. Let S = {X1 = ( 6, 4, -2, 4 ), X2 = ( 2, 0, 0, 1 ), X3 = ( 3, 2, -1, 2 ), X4 = ( 5, 6, -3, 2 ), X5 = ( 0, 4, -2, -1 ) }. Does S spans โ4 ? 82 MODULE 4 LINEAR INDEPENDENCE Introduction In this chapter we will discuss linear dependence and independence of a given set of vectors, the basis and dimension of a vector space V, the rank of a matrix and how it can be used to determine whether a matrix is singular or nonsingular and whether a homogeneous system of equations has a trivial or nontrivial solution. Objectives 1. 2. 3. 4. 5. Define linear dependence and independence. Determine whether a given set of vectors is linearly independent or dependent. Define and explain basis and dimension of a given vector. Find a basis for a vector space spanned by a given set of vectors. Recognize the rank of a matrix and used this information to determine whether a homogeneous system has a nontrivial solution. 4.1 Definition and Examples Definition 4.1.1: A set of vectors { ๐ฟ๐ , ๐ฟ๐ , โฏ , ๐ฟ๐ } ๏ V is said to be linearly dependent if there exists scalars ๐1 , ๐2 , โฏ , ๐๐ not all of which are zeros such that ๐1 ๐ฟ๐ + ๐2 ๐ฟ๐ + โฏ + ๐๐ ๐ฟ๐ = ๐ Otherwise, the set is said to be linearly independent. Meaning the set of vectors {๐ฟ๐ , ๐ฟ๐ , โฏ , ๐ฟ๐ } is linearly independent if ๐1 ๐ฟ๐ + ๐2 ๐ฟ๐ + โฏ + ๐๐ ๐ฟ๐ = ๐ is true if and only if ๐1 = ๐2 = โฏ = ๐3 = 0. Example 1. Let S ={X1 = (1, -1), X2 = (1, 1) } be vectors in โ2 . Determine whether S = {X1, X2 } is linearly dependent or linearly independent. Solution: Let ๐1 and ๐2 be scalars. Next we form the equation ๐1(1, -1) + ๐2 (1, 1) = (0, 0) Multiplying and adding we get the homogeneous system ๐1 + ๐2 = 0 −๐1 + ๐2 = 0 Since the only solution to this system is ๐1 = ๐2 = 0 then S = {X1, X2 } is linearly independent. 83 13 Example 2. Let S = {๐ฟ๐ = (1, 2 3), ๐ฟ๐ = (2, −1, 4), ๐ฟ๐ = (3, − 2 , 4)} be vectors in โ3 . Determine whether S = {๐ฟ๐ , ๐ฟ๐ , ๐ฟ๐ } is linearly dependent or linearly independent. Solution: Let ๐1 , ๐2 , and ๐3 be scalars. Next we form the equation 13 ๐1 (1, 2, 3) + ๐2 (2, −1, 4) + ๐3 (3, − 2 , 4) = (0, 0, 0) Multiplying and adding we obtain the homogeneous system ๐1 + 2๐2 + 3๐3 = 0 13 2๐1 − ๐2 − 2 ๐3 = 0 3๐1 + 4๐2 + 4๐3 = 0 Row reducing the augmented matrix we have 1 2 3 [2 −1 −13/2 3 4 4 0 −2๐ + ๐ 0] −3๐ 1 + ๐ 2 1 3 0 −15๐ 2 2๐ 2 + ๐ 3 3 ๐ 2 1 2 [0 −5 −25/2 ๐ 3 0 −2 −5 1 2 ๐ 2 [0 1 0 −2 1 2 ๐ 3 [0 1 0 0 3 5/2 −5 3 5/2 0 0 0] 0 0 0] 0 0 0] 0 Since column 3 has no pivot then ๐3 is a free variable. This implies that the system has infinitely many solutions. In particular if we let ๐3 = 2 then ๐1 = 4 and ๐2 = −5. It can be verified that 4(1, 2, 3) − 5(2, −1, 4) + 2 (3, − 13 , 4) = (0, 0, 0) 2 Hence ๐ฟ๐ , ๐ฟ๐ , and ๐ฟ๐ are linearly dependent. REMARKS: 1. Any set of vectors that includes the zero vector is linearly dependent. Proof: Let S = {๐ฟ๐ , ๐ฟ๐ , โฏ , ๐ฟ๐ , ๐ถ}. Then 0 โ ๐ฟ๐ + 0 โ ๐ฟ๐ + โฏ + 0 โ ๐ฟ๐ + 2 โ ๐ถ = 0 Hence S is linearly dependent. 84 2. If X ≠ 0, { X } is linearly independent 3. The zero vector is linearly dependent SAQ 4-1 Determine whether the vectors ๐ฟ๐ = (1, −1, 2), ๐ฟ๐ = (4, 0, 0), ๐ฟ๐ = (−2, 3, 5) and ๐ฟ๐ = (7, 1, 2) are linearly dependent or linearly independent. ASAQ 4-1 Let ๐1 , ๐2, ๐3 and ๐4 be scalars. Then we form the equation ๐1 (1, −1, 2) + ๐2 (4, 0, 0) + ๐3 (−2, 3, 5) + ๐4 (7, 1, 2) = (0, 0, 0) Multiplying and adding we obtain the homogeneous system ๐1 + 4๐2 − 2๐3 + 7๐4 = 0 −๐1 + 3๐3 + ๐4 = 0 2๐1 + 5๐3 + 2๐4 = 0 1 4 −2 7 The augmented matrix [−1 0 3 1 2 0 5 2 0 0] is row equivalent to 0 1 4 −2 7 0 2 0] (you must do the calculation here) [0 1 1/4 0 0 1 4/11 0 The last equation implies that ๐4 is a free variable. Hence ๐1 = −4๐2 + 2๐3 − 7๐4 1 ๐2 = −4๐3 − 2๐4 4 ๐3 = −11๐4 85 If we let ๐4 = 11 then ๐1 = −1, ๐2 = −21, and ๐3 = −4. It can be verified that −1(1, −1, 2) − 21(4, 0, 0) − 4(−2, 3, 5) + 11(7, 1, 2) = (0, 0, 0) Hence the given vectors are linearly dependent. A very important theorem about linear dependence or independence of vectors can now be stated. Theorem 4.1.1. A set of ๐ vectors in โ๐ is always linearly dependent if ๐ > ๐. Corollary 4.1.1. A set of linearly independent vectors in โ๐ contains at most ๐ vectors. Corollary 4.1.1 can be interpreted in this way: if we have ๐ linearly independent vectors, then adding more vectors will make the set of vectors linearly dependent. For example, the set of vectors {๐ฟ๐ = (1, -1), ๐ฟ๐ = (1, 1)} of Example 1 is linearly independent. Then by Corollary 4.1.1, the set of vectors {๐ฟ๐ = (1, -1), ๐ฟ๐ = (1, 1), ๐ฟ๐ = (2, 3)} is linearly dependent. Theorem 4.1.2: Let S = { ๐ฟ๐ , ๐ฟ๐ , โฏ , ๐ฟ๐ } be a set of nonzero vectors in a vector space V. Then S is linearly dependent if and only if one of the vectors ๐ฟ๐ is a linear combination of the preceding vectors in S. Example 3. Let S = {X1 = (1, 1, 0), X2 = (0, 2, 3), X3 = (1, 2, 3), X4 = (3, 6, 6) } be a set of vectors in โ3 . Determine if V is linearly dependent. If linearly dependent, express one of the vectors as a linear combination of the rest. Solution: By Corollary 4.1.1, we know that S is linearly dependent. Now we form the equation ๐1(1, 1, 0) + ๐2 (0, 2, 3) + ๐3 (1, 2, 3) + ๐4 (3, 6, 6) = (0, 0, 0) Multiplying and adding we get the homogeneous system ๐1 + ๐3 + 3๐4 = 0 ๐1 + 2๐2 + 2๐3 + 6๐4 = 0 3๐2 + 3๐3 + 6๐4 = 0 Row reducing the augmented matrix we obtain 86 1 0 1 [0 1 1 0 0 1 3 0 2 0] 1 0 This implies that ๐4 is a free variable. If we let ๐4 = ๐ ∈ โ and ๐ ≠ 0 then ๐3 = −๐4 = −๐ ๐2 = −๐3 − 2๐4 = ๐ − 2๐ = −๐ ๐1 = −๐3 − 3๐4 = ๐ − 3๐ = −2๐ Thus the given vectors are linearly dependent and it can be verified that −2๐(1, 1, 0) − ๐(0, 2, 3) − ๐(1, 2, 3) + ๐(3, 6, 6) = (0, 0, 0) By Theorem 4.1.2, we can write one of the vectors as a linear combination of the rest. Transposing the first three terms to the right side we have ๐(3, 6, 6) = 2๐(1, 1, 0) + ๐(0, 2, 3) + ๐(1, 2, 3) (3, 6, 6) = 2(1, 1, 0) + (0, 2, 3) + (1, 2, 3) (dividing both sides by ๐) Clearly, (3, 6, 6) is a linear combination of the preceding vectors. NOTE: Any vector in the given set can be expressed as a linear combination of the other vectors. 4.2 Basis and Dimension Definition 4.2.1: A set of vectors S = {๐ฟ๐ , ๐ฟ๐ , โฏ , ๐ฟ๐ } in a vector space V is called a basis for V if S spans V and S is linearly independent. Example 1. Show that the set S = {X1 = (3, 2, 2), X2 = (-1, 2, 1), X3 = (0, 1, 0) } is a basis for โ3 . Solution: First we have to show that S is linearly independent. Forming the equation ๐1(3, 2, 2) + ๐2 (−1, 2, 1)+ ๐3 (0, 1, 0) = (0, 0, 0) we obtain the linear system 3๐1 − ๐2 = 0 2๐1 + 2๐2 + ๐3 = 0 2๐1 + ๐2 = 0 87 Solving the given system we obtain only the zero solution, ๐1 = ๐2 = ๐3 = 0. This implies that S is linearly independent. Next we have to show that S spans โ3 . To show that S spans โ3 , let ๐ฟ = (๐, ๐, ๐) be any vector in โ3 . Then we form the equation ๐1(3, 2, 2) + ๐2 (−1, 2, 1)+ ๐3 (0, 1, 0) = (๐, ๐, ๐) Multiplying and adding we obtain the linear system 3๐1 − ๐2 = ๐ 2๐1 + 2๐2 + ๐3 = ๐ 2๐1 + ๐2 = ๐ a๏ซc 3c ๏ญ 2a 2a ๏ซ 5b ๏ญ 8c , c2 ๏ฝ , and c3 ๏ฝ . Since the above 5 5 5 system is consistent for any choice of ๐, ๐, and ๐, then S spans โ3 . Solving the system we get c1 ๏ฝ Thus S = { (3, 2, 2), (−1, 2, 1), (0, 1, 0) } } is a basis for โ3 . A basis for a vector space V is not unique. In fact it can be easily verified that the set of vectors {(1, 0, 0), (0, 1, 0), (0, 0, 1)} is also a basis for โ3 . This is called the natural basis for โ3 . Likewise the set of vectors {(1, 0, 0, 0), (0, 1, 0, 0), (0, 0, 1, 0), (0, 0, 0, 1)} forms a natural basis for โ4 . In general, the vectors ๐1 = (1, 0, โฏ , 0), ๐2 = (0, 1, โฏ , 0), โฏ , ๐๐ = (0, 0, โฏ , 1) constitute a basis for โ๐ . Example 2. The monomials {๐ฅ 3 , ๐ฅ 2 , ๐ฅ, 1} is a natural basis for ๐3 . In general the monomials {๐ฅ ๐ , ๐ฅ ๐−1 , โฏ , ๐ฅ, 1} constitute a basis for ๐๐ . Example 3. The 2 x 2 matrices {[ 1 0 0 ],[ 0 0 0 1 0 0 0 0 ],[ ],[ ]} is a natural basis for ๐22 . 0 1 0 0 1 SAQ 4-2 Determine whether A = {๐ฅ 2 − 1, ๐ฅ 2 − 2, ๐ฅ 2 − 3} is a basis for ๐2 . 88 ASAQ 4-2 First we have to determine if A is linearly independent. Let ๐1 , ๐2, and ๐3 be scalars. Then we form the equation ๐1 (๐ฅ 2 − 1) + ๐2 (๐ฅ 2 − 2) + ๐3 (๐ฅ 2 − 3) = 0 Equating the coefficients of similar terms we obtain the system ๐1 + ๐2 + ๐3 = 0 −๐1 − 2 ๐2 − 3๐3 = 0 The augmented matrix in reduced row echelon form is (verify) [ 1 0 0 1 −1 2 0 ] 0 This implies that A is linearly dependent. Thus A is not a basis for ๐2 . Theorem 4.2.1: If S = { ๐ฟ๐ , ๐ฟ๐ , โฏ , ๐ฟ๐ } is a basis for a vector space V, then every vector in V can be written in one and only one way as a linear combination of the vectors in S. Definition 4.2.2: If the vector space V has a finite basis, then the dimension of V is the number of vectors in every basis and V is called a finite dimensional vector space. Otherwise V is called an infinite dimensional vector space. If V = {0}, then V is said to be zero dimensional. We often write dimV for the dimension of V. Example 4. Since a basis for โ3 consists of 3 linearly independent vectors then dimโ3 = 3. Likewise dimโ4 = 4, dimโ5 = 5 and so on. In general dimโ๐ = ๐ since ๐ linearly independent vectors constitute a basis for โ๐ . Example 5. By Example 2, we see that dim๐3 = 4 and in general the dim๐๐ = ๐ + 1. Theorem 4.2.2: Let S = { ๐ฟ๐ , ๐ฟ๐ , โฏ , ๐ฟ๐ } be a set of nonzero vectors in a vector space V and let W = spans S. Then some subset of S is a basis for W. Theorem 4.2.2 means that if W consists of all vectors that can be expressed as a linear combination of the vectors in S (W=spanS) then some subset of S forms a basis for W. Thus if S is linearly independent then { ๐ฟ๐ , ๐ฟ๐ , โฏ , ๐ฟ๐ } is a basis for W. If S is linearly independent there exists a proper subset of S which forms a basis for W. Let us consider the next example: 89 Example 6: Let S = {X1 = (1, 2, 2), X2 = (3, 2, 1), X3 = (11, 10, 7), X4 = (7, 6, 4) }. Find a basis for the subspace W = span S of โ3 . Solution: W is a set of all vectors in โ3 which can be expressed as a linear combination of the vectors in S. By Theorem 4.1.1 we know that S is linearly dependent. Hence S is not a basis for W but by Theorem 4.2.2 S contains a proper subset of linearly independent vectors which forms a basis for W. Next we form the equation ๐1(1, 2, 2) + ๐2 (3, 2, 1) + ๐3 (11, 10, 7) + ๐4 (7, 6, 4) = (0, 0, 0) Equating the corresponding components, we obtain the homogeneous system ๐1 + 3๐2 + 11๐3 + 7๐4 = 0 2๐1 + 2๐2 + 10๐3 + 6๐4 = 0 2๐1 + ๐2 + 7๐3 + 4๐4 = 0 The augmented matrix in reduced row echelon form is (verify) 1 0 [0 1 0 0 2 1 0 3 2 0] 0 0 0 The leading 1s appear in columns 1 and 2, thus {X1, X2} is a basis for W = span S. It means that {X1, X2} is the smallest set possible that could span W. By Theorem 4.1.2, we can write ๐ฟ๐ and ๐ฟ๐ as a linear combination of the two vectors X1 and X2. Thus ๐ฟ๐ and ๐ฟ๐ can be discarded and the remaining vectors still span W. That is W = span{ ๐ฟ๐ , ๐ฟ๐ , ๐ฟ๐ , ๐ฟ๐ } = span{X1, X2}. Theorem 4.2.3: Suppose that dimV = ๐. If T = { Y1, Y2, …, Yr } is a linearly independent set of vectors in V, then ๐ ≤ ๐. Proof: Let ๐ฟ๐ , ๐ฟ๐ , โฏ , ๐ฟ๐ be a basis for V. If ๐ > ๐ then we can find scalars ๐1 , ๐2 , โฏ , ๐๐ not all zero such that ๐1 ๐๐ + ๐2 ๐๐ + โฏ + ๐๐ ๐๐ = ๐ถ is satisfied. This will contradict the linear independence of the ๐๐ ′s. Thus ๐ ≤ ๐. Corollary 4.2.1: If S = { ๐ฟ๐ , ๐ฟ๐ , โฏ , ๐ฟ๐ } and T = { Y1, Y2, …, Ym } are bases for a vector space, then ๐ = ๐. (If a vector space has one basis with a finite number of elements, then all other bases are finite and have the same number of elements) 90 Theorem 4.2.4: If S is linearly independent set of vectors in a finite-dimensional vector space V, then there is a basis T for V, which contains S. Example 7: In V = โ4 , the set A = {(1, 1, 0, 0), (0, 0, 1, 1), (1, 0, 1, 0), (0, 1, 0, −1)} is a basis, and B = { (1, 2, −1, 1), (0, 1, 2, −1) } is linearly independent. Extend B into a basis for V using A. Solution: A ๏ B = {(1, 2, −1, 1), (0, 1, 2, −1), (1, 1, 0, 0), (0, 0, 1, 1), (1, 0, 1, 0), (0, 1, 0, −1) } Delete vectors in A which are linear combination of the preceding vectors. a. Try (1, 1, 0, 0) ๐1(1, 2, −1, 1) + ๐2 (0, 1, 2, −1) = (1, 1, 0, 0) Then ๐1 =1 2๐1 + ๐2 = 1 −๐1 + 2๐2 = 0 ๐1 – ๐2 = 0 If ๐1 = 1 then ๐2 = 1 but 2๐1 + ๐2 = 3 ๏น 1. The system of equations has no solution so (1, 1, 0, 0) is not a linear combination of the preceding vectors. Thus we retain (1, 1, 0, 0). b. Try (0, 0, 1, 1) ๐1 (1, 2, -1, 1) + ๐2 (0, 1, 2, -1) = (0, 0, 1, 1) Then ๐1 =0 2๐1 + ๐2 = 0 −๐1 + 2๐2 = 1 ๐1 – ๐2 = 1 If ๐1 = 0 then ๐2 = 0 but –๐1 + 2๐2 = 0 ๏น 1. The system has no solution so (0, 0, 1, 1) is not a linear combination of the preceding vectors. Thus we retain (0, 0, 1, 1). Hence, {(1, 2, −1, 1), (0, 1, 2, −1), (1, 1, 0, 0), (0, 0, 1, 1)} is a basis for V. Theorem 4.2.5: Let V be an ๐-dimensional vector space, and let S = { ๐ฟ๐ , ๐ฟ๐ , โฏ , ๐ฟ๐ } be a set of ๐ vectors in V. a. If S is linearly independent, then it is a basis for V. b. If S spans V, then it is a basis for V. 91 From our previous lesson, we see that the set of all solutions to the homogeneous system AX = 0, where A is ๐ ๐ฅ ๐ , is a subspace of โ๐ . To find a basis for this solution space we consider the following example. Example 8: Find a basis for the solution space W of ๏ฉ x1 ๏น ๏ฉ1 2 1 2 1 ๏น ๏ช ๏บ ๏ฉ0 ๏น ๏ช1 2 2 1 2 ๏บ ๏ช x2 ๏บ ๏ช0 ๏บ ๏ช ๏บ ๏ชx ๏บ ๏ฝ ๏ช ๏บ ๏ช 2 4 3 3 3 ๏บ ๏ช 3 ๏บ ๏ช0 ๏บ ๏ช ๏บ ๏ช x4 ๏บ ๏ช ๏บ ๏ซ 0 0 1 ๏ญ1 ๏ญ1๏ป ๏ช ๏บ ๏ซ0 ๏ป ๏ซ x5 ๏ป Solution: Transform the augmented matrix to reduced row-echelon form using the Gauss-Jordan reduction method. The augmented matrix in reduced row-echelon form is (verify) 1 0 [ 0 0 2 0 0 0 0 3 0 1 −1 0 0 0 1 0 0 0 0 0 ] 0 0 The solution is ๐1 = −2๐ – 3๐ก, ๐2 = ๐ , ๐3 = ๐ก, ๐4 = ๐ก, and ๐5 = 0 where ๐ and ๐ก are any real numbers. Thus every solution is of the form ๏ฉ ๏ญ2 s ๏ญ 3t ๏น ๏ช s ๏บ ๏ช ๏บ ๏บ , where ๐ and ๐ก are real numbers. X ๏ฝ๏ช t ๏ช ๏บ ๏ช t ๏บ ๏ช๏ซ 0 ๏บ๏ป Since W is a solution space then every vector in W can be written in the form of ๏ฉ ๏ญ2 ๏น ๏ฉ ๏ญ3 ๏น ๏ช1๏บ ๏ช0๏บ ๏ช ๏บ ๏ช ๏บ X ๏ฝ s๏ช 0 ๏บ + t๏ช 1 ๏บ ๏ช ๏บ ๏ช ๏บ ๏ช0๏บ ๏ช1 ๏บ ๏ช๏ซ 0 ๏บ๏ป ๏ช๏ซ 0 ๏บ๏ป Since ๐ and ๐ก can take on any values, we first let ๐ = 1, ๐ก = 0 and let ๐ = 0, ๐ก = 1, obtaining as solutions 92 ๏ฉ ๏ญ2 ๏น ๏ช1๏บ ๏ช ๏บ X1 ๏ฝ ๏ช 0 ๏บ ๏ช ๏บ ๏ช0๏บ ๏ช๏ซ 0 ๏บ๏ป ๏ฉ ๏ญ3 ๏น ๏ช0๏บ ๏ช ๏บ X2 ๏ฝ ๏ช 1 ๏บ ๏ช ๏บ ๏ช1 ๏บ ๏ช๏ซ 0 ๏บ๏ป and Thus S = {X1, X2} belongs to W. Since any vector in W can be written as a linear combination of the vectors in S then S spans W. We have to show that S = {X1, X2} is linearly independent. We form the equation ๐1(−2, 1, 0, 0, 0 ) + ๐2 (−3, 0, 1, 1, 0 ) = (0, 0, 0, 0, 0) Then −2๐1 − 3๐2 = 0 ๐1 = 0 ๐2 = 0 The only solution to the above system is ๐1 = ๐2 = 0, hence S is linearly independent and is a basis for W. Thus the dimension of W is 2. SAQ 4-3 Find a basis for the solution space W of the homogeneous system ๐ฅ1 + 2๐ฅ2 + 2๐ฅ3 − ๐ฅ4 + ๐ฅ5 = 0 2๐ฅ2 + 2๐ฅ3 − 2๐ฅ4 − ๐ฅ5 = 0 2๐ฅ1 + 6๐ฅ2 + 2๐ฅ3 − 4๐ฅ4 + ๐ฅ5 = 0 ๐ฅ1 + 4๐ฅ2 − 3๐ฅ4 =0 What is the dimension of W? ASAQ 4-3 1 0 The augmented matrix [ 2 1 2 2 6 4 2 2 2 0 −1 1 0 −2 −1 0 ] is row equivalent (verify) to −4 1 0 −3 0 0 93 1 0 [ 0 0 0 1 0 0 0 1 2 0 −1 −1/2 1 0 0 0 0 0 0 0 ] which is in reduced row echelon form. 0 0 Since columns 4 and 5 have no pivots then ๐ฅ4 and ๐ฅ5 are free variables. If we let ๐ฅ4 = ๐ and ๐ฅ5 = ๐ก where ๐ and ๐ก are real numbers then ๐ฅ1 = −๐ − 2๐ก 1 ๐ฅ2 = ๐ + 2 ๐ก ๐ฅ3 = 0 ๐ฅ4 = ๐ ๐ฅ5 = ๐ก where ๐ and ๐ก are real numbers. Thus all solutions are of the form −๐ − 2๐ก 1 ๐ +2๐ก ๐= 0 ๐ [ ๐ก ] −2 −1 1/2 1 =๐ 0 +๐ก 0 1 0 [0] [ 1 ] Since ๐ and ๐ก can be any real number, we first let ๐ = 1, ๐ก = 0 and then let ๐ = 0, ๐ก = 2 to obtain the solution −1 −4 1 1 ๐1 = 0 and ๐2 = 0 1 0 [2] [0] which span W. It can be easily verified that ๐1 and ๐2 are linearly independent because one vector is not a multiple of the other. Thus ๐1 and ๐2 form a basis for W and dimW = 2. Note that you may obtain a different basis since ๐ and ๐ก can take on any values. 94 ACTIVITY 1. Let a) X1 = ( 4, 2, 1 ), X2 = ( 2, 6, -5 ), X3 = ( 1, -2, 3 ) b) X1 = ( 1, 2, 3 ), X2 = ( 1, 1, 1 ), X3 = ( 1, 0, 1 ) c) X1 = ( 1, 1, 0 ), X2 = ( 0, 2, 3 ), X3 = ( 1, 2, 3 ), X4 = ( 3, 6, 6 ) Which of the given set of vectors in โ3 is linearly dependent? For those that are, express one vector as a linear combination of the rest. 3. Determine whether the set ๏ฌ ๏ฉ1 S ๏ฝ ๏ญ๏ช ๏ฎ ๏ซ0 1๏น ๏ฉ0 0๏น ๏ฉ1 0๏น ๏ฉ0 1๏น ๏ผ , , , ๏ฝ 0๏บ๏ป ๏ช๏ซ1 1 ๏บ๏ป ๏ช๏ซ0 1 ๏บ๏ป ๏ช๏ซ1 1๏บ๏ป ๏พ is a basis for the vector space V of all 2 x 2 matrices. 4. Find a basis of โ4 containing the vector (1, 2, 3, 4). 4.3 The Rank of a Matrix Definition 4.3.1: Let ๏ฉ a11 a12 ... a1n ๏น ๏ชa a22 ... a2 n ๏บ๏บ 21 ๏ช A๏ฝ ๏ช ๏บ ๏ช ๏บ ๏ซ๏ช am1 am 2 ... amn ๏บ๏ป be an ๐ x ๐ matrix. The rows of A X 1 ๏ฝ (a11 , a12 ,..., a1n ) X 2 ๏ฝ (a 21 , a 22 ,..., a 2 n ) โฎ X m ๏ฝ (a m1 , a m 2 ,..., a mn ) considered as vectors in โ๐ , span a subspace of โ๐ , called the row space of A. Similarly, the columns of A, ๏ฉ a11 ๏น ๏ฉ a12 ๏น ๏ฉ a1n ๏น ๏ชa ๏บ ๏ชa ๏บ ๏ชa ๏บ 21 ๏บ 22 ๏บ 2n ๏ช ๏ช Y1 ๏ฝ , Y2 ๏ฝ , … , Yn ๏ฝ ๏ช ๏บ ๏ช ๏บ ๏ช ๏บ ๏ช ๏บ ๏ช ๏บ ๏ช ๏บ ๏ช ๏บ ๏ช๏ซ am1 ๏บ๏ป ๏ช๏ซ am 2 ๏บ๏ป ๏ช๏ซ amn ๏บ๏ป 95 considered as vectors in โ๐ , span a subspace of โ๐ called the column space of A. ๏ฉ1 2 0 1 ๏น Example 1. Let A ๏ฝ ๏ช๏ช 2 6 ๏ญ3 ๏ญ2 ๏บ๏บ . The rows of A, ๏ช๏ซ3 10 ๏ญ6 ๏ญ5 ๏บ๏ป X1 = (1, 2, 0, 1), X2 = (2, 6, −3, −2), and X3 = (3, 10, −6, −5) are vectors in โ4 , and these vectors span a subspace of โ4 called the row space of A. That is Row space of A = span { ๐1 , ๐2 , ๐3 } Similarly, the columns of A, Y1 = (1, 2, 3), Y2 = (2, 6, 10), Y3 = (0, −3, −6) and Y4 = (1, −2, −5) are vectors in โ3 , and these vectors span a subspace of โ3 called the column space of A. That is Column space of A = span { ๐1 , ๐2 , ๐3 , ๐4 } Theorem 4.3.1: If A and B are two ๐ x ๐ row equivalent matrices, then the row spaces of A and B are equal. Example 2: Let S = {๐ฟ๐ , ๐ฟ๐ , ๐ฟ๐ , ๐ฟ๐ } where ๐ฟ๐ = (1, 2, −1), ๐ฟ๐ = (6, 3, 0), ๐ฟ๐ = (4, −1, 2), and ๐ฟ๐ = (2, −5, 4). Find a basis for the subspace V = spanS of โ3 . Solution: V is the row space of the matrix A whose rows are the given vectors 1 2 −1 6 3 0 A=[ ] 4 −1 2 2 −5 4 Applying the Gauss-Jordan elimination we obtain the matrix B 1 0 ๐ต=[ 0 0 0 1 0 0 1/3 −2/3 ] 0 0 which is row equivalent to A. By Theorem 4.3.1, the row spaces of A and B are equal. Since the nonzero rows of B are linearly independent then (1, 0, 1/3) and (0, 1, −2/3) 96 form a basis for V. It means that all vectors in V can be expressed as a linear combination of these two vectors. Note that the basis for V is not a subset of S. However, expressing any vector in V as a linear combination of the basis obtained by this procedure is very simple. For example, the vectors (2, −5, 4) and (−1, 4, −3) are in V. Since the leading 1s appear in columns 1 and 2 then (2, −5, 4) = 2(1, 0, 1/3) – 5(0, 1, −2/3) and (−1, 4, −3) = −1(1, 0, 1/3) + 4(0, 1, −2/3) Also note that the dimV = 2 ≠ 3 hence V ≠ โ3 . Since V ≠ โ3 then not all vectors in โ3 can be expressed as a linear combination of (1, 0, 1/3) and (0, 1, −2/3). It is very easy to see from our example above that all vectors of the form (2, −5, ๐), ๐ ≠ 4, are not in V. Definition 4.3.2: The dimension of the row space of A is called the row rank of A, and the dimension of the column space of A is called the column rank of A. 1 2 Example 3. Let ๐ด = [2 6 3 10 0 1 −3 −3]. Find the row and column ranks of A. −6 −7 Solution: 1 0 A is row equivalent (verify) to ๐ต = [0 1 0 0 3 6 −3/2 −5/2] which is in reduced row 0 0 echelon form. The vectors (1, 0, 3, 6) and (0, 1, −3/2, −5/2) form a basis for the row space of A. Thus the row rank of A is 2. To find the column rank of A, we form the matrix 1 2 ๐ด๐ = [ 0 1 2 3 6 10 ] −3 −6 −3 −7 which is row equivalent (verify) to the matrix 1 0 ๐ถ=[ 0 0 0 −1 1 2 ] 0 0 0 0 97 The vectors (1, 0 −1) and (0, 1, 2) form a basis for the row space of ๐ด๐ . Thus 1 0 [ 0 ] and [1] −1 2 form a basis for the column space of A. Hence the column rank of A is 2. Note that the row rank and column rank of a matrix are equal. This is stated in the next theorem. Theorem 4.3.2. The row and column ranks of the (๐ x ๐) matrix A = ๏ฉ๏ซ aij ๏น๏ป are equal. The next example will show us how we can apply the method used in Example 3 in finding a basis for a subspace when the vectors are given in column form. 1 0 2 3 5 2 2 1 2 0 Example 4: Let S = {[ ] , [ ] , [ ] , [ ] , [ ]} . Find a basis for the subspace V = spanS of 1 1 3 1 0 1 2 1 4 −1 โ4 . 1 2 Solution: Let ๐ด = [ 1 1 0 2 1 2 1 0 ๐ด๐ = 2 3 [5 2 1 3 1 2 2 1 2 0 3 5 2 0 ]. Next we form the matrix 1 0 4 −1 1 1 1 2 3 1 1 4 0 −1] Applying elementary row operations we obtain (verify) the matrix 1 0 ๐ต๐ = 0 0 [0 0 1 0 0 0 0 0 1 0 0 −1 3/5 4/5 0 0 ] The basis for V are the nonzero rows of ๐ต ๐ written as columns. Thus 98 0 0 1 1 0 0 [ ], [ ], and [ ] 0 1 0 3/5 4/5 −1 form a basis for V. Theorem 4.3.3. An ๐ x ๐ matrix is nonsingular if and only if rank A = ๐. 1 Example 5. Find the rank of ๐ด = [1 1 2 0 1 −3]. 3 3 Solution: Transforming A to reduced row echelon form we obtain (verify) 1 0 [0 1 0 0 −6 3] 0 Since rank of A = 2 < 3 then A is singular. We know from our previous lesson that a matrix is singular if and only if |๐ด| = 0. It can be verified that 1 2 |1 1 1 3 0 −3| = 0 3 This result is stated in the following corollary and can be used in determining the rank of an ๐ x ๐ matrix. Corollary 4.3.1. If A is an ๐ x ๐ matrix, then rank A = ๐ if and only if A ๏น 0. The next corollary gives another method of testing whether the homogeneous system ๐ด๐ = 0 has a trivial or nontrivial solution. Corollary 4.3.2. The homogeneous system AX = 0 of ๐ linear equations in ๐ unknowns has a nontrivial solution if and only if rank A < ๐. 99 Example 6: The 3 x 3 matrix A of Example 5 is singular. Hence the homogeneous system ๐ฅ1 + 2๐ฅ2 = 0 ๐ฅ1 + ๐ฅ2 − 3๐ฅ3 = 0 ๐ฅ1 + 3๐ฅ2 + 3๐ฅ3 = 0 has a nontrivial solution. Corollary 4.3.3. Let S = {๐ฟ๐ , ๐ฟ๐ , โฏ , ๐ฟ๐ } be a set of vectors in โ๐ and let A be the matrix whose rows (columns) are the vectors in S. Then S is linearly independent if and only if |๐ด| ≠ 0. Example 7: The vectors (1, −2, 3), (2, 4, 7) and (0, −1, 5) are linearly independent because 1 [2 0 −2 3 4 7] = 41 ≠ 0 −1 5 By Theorem 4.2.5, (1, −2, 3), (2, 4, 7) and (0, −1, 5) form a basis for โ3 . ACTIVITY 1. Find a basis for the row space and column space of the following matrices: 1 2 3 a. [4 5 6] 7 8 9 2 1 3 b. [2 −1 5 1 1 1 −2 2] 1 2. Let V = P1 and let A = { x + 1, x – 1, 2x + 3 } a. Show that S spans P1. b. Find a subset of A which is a basis for P1. 3. Consider the following subset of the vector space of all real-valued functions S = { cos2t, sin2t, cos2t } Find a basis for the subspace W = spanS. What is the dimension of W? 100 MODULE 5 LINEAR TRANSFORMATIONS AND MATRICES Introduction In this chapter we will discuss a function mapping one vector space to another vector space, one-to-one and onto linear transformations, the kernel of a linear transformation and how it can be used to determine if the linear transformation is one-toone, the range of a linear transformation and how it can be used to determine whether the linear transformation is onto and the matrix of a linear transformation. Objectives After going through this chapter, you are expected to be able to do the following: 1. Define and explain linear transformation or linear mapping. 2. Determine whether a function mapping one vector space to another is a linear transformation. 3. Differentiate between one-to-one linear transformation and onto linear transformation. 4. Define the kernel and range of a linear transformation. 5. Using kernel, distinguish between one-to-one linear transformation and onto linear transformation. 6. Find the matrix of a linear transformation with respect to some bases. 5.1 Linear Transformations (Linear Mappings) Definition 5.1.1: Let V and W be vector spaces. A linear transformation L of V into W is a function assigning a unique vector L(X) in W to each X in V such that: a. L ( X + Y ) = L (X) + L(Y), for every vector X and Y in V, b. L ( ๐X ) = ๐L(X), for every vector X in V and every scalar ๐. Example 1: Let L: โ3 → โ3 be defined by L(x, y, z) = (x, y, 0) Determine whether L is a linear transformation or not. Solution: Let X = (๐ฅ1 , ๐ฆ1 , ๐ง1 ) and Y = (๐ฅ2 , ๐ฆ2 , ๐ง2 ) ๏ โ3 then ๐ + ๐ = (๐ฅ1 + ๐ฅ2 , ๐ฆ1 + ๐ฆ2 , ๐ง1 + ๐ง2 ). 101 a) We have to show that L(X + Y) = L(X) + L(Y) L(X + Y) = L(๐ฅ1 + ๐ฅ2 , ๐ฆ1 + ๐ฆ2 , ๐ง1 + ๐ง2 ) = (๐ฅ1 + ๐ฅ2 , ๐ฆ1 + ๐ฆ2 , 0) = (๐ฅ1 , ๐ฆ1 , 0) + ( ๐ฅ2 , ๐ฆ2 , 0) = L(X) + L(Y) Hence L(X + Y) = L(X) + L(Y) b) Let ๐๐ฟ = (๐๐ฅ1 , ๐๐ฆ1 , ๐๐ง1 ). We have to show that L(๐X) = ๐L(X). L(๐๐ฟ) = (๐๐ฅ1 , ๐๐ฆ1 , 0) = ๐(๐ฅ1 , ๐ฆ1 , 0) = ๐L(๐ฟ) Hence L(cX) = cL(X). Since the two conditions are satisfied then L is a linear transformation. SAQ 5-1 Let L: โ3 โ3 be defined by L(๐ฅ, ๐ฆ, ๐ง) = (๐ฅ − ๐ฆ, ๐ฅ 2 , 2๐ง) Determine whether L is a linear transformation or not. 102 ASAQ 5-1 Let ๐ฟ = (๐ฅ1 , ๐ฆ1 , ๐ง1 ) and Y = (๐ฅ2 , ๐ฆ2 , ๐ง2 ) ๏ โ3 then ๐ + ๐ = (๐ฅ1 + ๐ฅ2 , ๐ฆ1 + ๐ฆ2 , ๐ง1 + ๐ง2 ). L(๐ฟ + ๐) = ((๐ฅ1 + ๐ฅ2 ) − (๐ฆ1 + ๐ฆ2 ), (๐ฅ1 + ๐ฅ2 )2 , 2(๐ง1 + ๐ง2 )) = (๐ฅ1 + ๐ฅ2 − ๐ฆ1 − ๐ฆ2 , ๐ฅ1 2 + 2๐ฅ1 ๐ฅ2 + ๐ฅ2 2 , 2๐ง1 + 2๐ง2 ) and L(X) + L(Y) = (๐ฅ1 − ๐ฆ1 , ๐ฅ1 2 , 2๐ง1 ) + (๐ฅ2 − ๐ฆ2 , ๐ฅ2 2 , 2๐ง2 ) = ((๐ฅ1 + ๐ฅ2 − ๐ฆ1 − ๐ฆ2 , ๐ฅ1 2 + ๐ฅ2 2 , 2๐ง1 + 2๐ง2 ) Since L(๐ฟ + ๐) ≠ L(๐ฟ) + L(๐) then L is not a linear transformation. Theorem 5.1.1: If L: V→ W is a linear transformation, then L(c1X1 + c2X2 + … + ckXk ) = c1L(X1) + c2L(X2) + … + ckL(Xk) for any vectors X1, X2, … , Xk in V and any scalars c1, c2, … , ck. Theorem 5.1.2: Let L: โ๐ → โ๐ be a linear transformation. Then a. L(0) = 0 b. L(−๐ฟ) = − L(๐ฟ) for every ๐ฟ ∈ โ๐ c. L ( X – Y ) = L(X) – L(Y) Proof of Theorem 5.1.2.a. a. L(0) = 0 L(0) = L (0 + 0) L(0) = L(0) + L(0) L(0) – L(0) = L(0) + L(0) – L(0) 0 = L(0) 103 Proof of Theorem 5.1.2.b. b. L(−๐ฟ) = −L(๐ฟ) L(−๐ฟ) = L(−1 โ ๐ฟ) = −1L(๐ฟ) = −L(๐ฟ) Proof of Theorem 5.1.2.c. c. L(๐ฟ − ๐) = L(๐ฟ + (−1)๐) = L(๐ฟ) + L(−1 โ ๐) = L(๐ฟ) − L(๐) Example 2: Let L: โ3 → โ2 be defined by L(๐ฅ, ๐ฆ, ๐ง) = (๐ฅ, ๐ฆ) be a linear transformation (verify). It can be verified that L(0, 0, 0) = (0, 0) That is, the zero vector in โ3 is mapped to the zero vector in โ2 . Also, let X = (๐ฅ1 , ๐ฆ1 , ๐ง1 ) and Y = (๐ฅ2 , ๐ฆ2 , ๐ง2 ) be any vectors in โ3 then ๐ฟ − ๐ = (๐ฅ1 − ๐ฅ2 , ๐ฆ1 − ๐ฆ2 , ๐ง1 − ๐ง2 ). Thus L(X – Y) = (๐ฅ1 − ๐ฅ2 , ๐ฆ1 − ๐ฆ2 ) = (๐ฅ1 , ๐ฆ1 ) − (๐ฅ2 , ๐ฆ2 ) = L(X) – L(Y) Example 3: Let T: โ2 → โ2 be the “translation mapping” defined by T(๐ฅ, ๐ฆ) = (๐ฅ + 1, ๐ฆ + 2) By Theorem 5.1.2 letter a, T is not a linear transformation since T(0, 0) = (0 + 1, 0 + 2) = (1, 2) ≠ (0, 0) (The zero vector is not mapped into the zero vector). 104 Theorem 5.1.3: Let L: V→ W be a linear transformation of an n-dimensional vector space V into a vector space W. Also let S = { X1 , X2, … , Xn } be a basis for V. If X is any vector in V, then L(X) is completely determined by {L(X1), L(X2), … , L(Xn) }. Example 4: Let L: โ3 → โ3 be a linear transformation for which we know that (2, −4) , L(0, 1, 0) = (3, −5) and L(0, 0, 1) = (2, 3). a. What is L(1, −2, 3)? b) What is L(๐, ๐, ๐)? Solution: The set {(1, 0, 0), (0, 1, 0), (0, 0, 1)} is a basis for โ3 , and (1, −2, 3) = 1(1, 0, 0) – 2(0, 1, 0) + 3(0, 0, 1) By Theorem 5.1.3: a) L(1, −2, 3) = L1(1, 0, 0) – L2(0, 1, 0) + L3(0, 0, 1) = L(1, 0, 0) – 2L(0, 1, 0) + 3L(0, 0, 1) = (2, −4) – 2(3, −5) + 3(2, 3) = (2, 23) b) We see from part (a) that for any (๐, ๐, ๐) ๏ โ3 , we have L(๐, ๐, ๐) = ๐L(1, 0, 0) + ๐L(0, 1, 0) + ๐L(0, 0, 1) = ๐(2, −4) + ๐(3, −5) + ๐(2, 3) = (2๐ + 3๐ + 2๐, −4๐ −5๐ + 3๐) SAQ 5-2 Let L: P2 → P3 be a linear transformation for which we know that L(1) = 1, L(t) = t2 and L(t2) = t3 + t. Find (a) L(2t2 – 5t + 3) (b) L(at2 + bt + c) L(1, 0,0) = 105 ASAQ2 The set {๐ก 2 , ๐ก, 1} is a basis for ๐2 . By Theorem 5.1.3 a. L(2๐ก 2 − 5๐ก + 3) = L2(๐ก 2 ) + L(−5)(๐ก) + L3(1) = 2L(๐ก 2 ) − 5L(๐ก) + 3L(1) = 2(๐ก 3 + ๐ก) − 5(๐ก 2 ) + 3(1) = 2๐ก 3 + 2๐ก − 5๐ก 2 + 3 = 2๐ก 3 − 5๐ก 2 + 2๐ก + 3 b. L(๐๐ก 2 + ๐๐ก + ๐) = ๐L(๐ก 2 ) + ๐L(๐ก) + cL(1) = ๐(๐ก 3 + ๐ก) + ๐(๐ก 2 ) + ๐ = ๐๐ก 3 + ๐๐ก + ๐๐ก 2 + ๐ = ๐๐ก 3 + ๐๐ก 2 + ๐๐ก + ๐ 5.2 The Kernel and Range of a Linear Transformation Definition 5.2.1: A linear transformation L: V→ W is said to be one-to-one if for all X1, X2 in V, X1 ≠ X2 implies that L(X1) ≠ L(X2). An equivalent statement is that L is one-to-one if for all X1, X2 in V, L(X1) = L(X2) implies that X1 = X2. Example 1. Let L: โ2 → โ2 be a linear transformation defined by L( ๐ฅ, ๐ฆ ) = ( ๐ฅ + ๐ฆ, ๐ฅ ) Determine if L is one–to–one or not. Solution: Let X1 = (๐ฅ1 , ๐ฆ1 ) and X2 = (๐ฅ2 , ๐ฆ2 ) be vectors in โ2 . We have to show that if L(X1) = L(X2) then X1 = X2. L(X1) = L(X2) (๐ฅ1 + ๐ฆ1 , ๐ฅ1 ) = (๐ฅ2 + ๐ฆ2 , ๐ฅ2 ) Equating the corresponding parts we have ๐ฅ1 + ๐ฆ1 = ๐ฅ2 + ๐ฆ2 ๐ฅ1 = ๐ฅ2 106 If we subtract the second equation from the first, we get ๐ฆ1 = ๐ฆ2 which implies that ๐ฟ๐ = ๐ฟ๐ . Thus L is one-to-one. Example 2. Let L: โ3 → โ3 be a linear transformation defined by L(x, y, z) = (x, y, 0) Determine if L is one-to-one or not. Solution: Let X1 = (4, 5, −3) and X2 = (4, 5, 2) be vectors in โ3 . We see that X1 ๏น X2 but L(4, 5, −3) = L(4, 5, 2) = (4, 5, 0). Hence L is not one-to-one. Definition 5.2.2: Let L: V→ W be a linear transformation. The kernel of L, kerL, is the subset of V consisting of all vectors X such that L(X) = Ow. Note that kerL is not empty since by Theorem 5.1.2 we know that if L: โ๐ → โ๐ is a linear transformation then the zero vector in โ๐ is mapped to the zero vector in โ๐ . Thus ๐ ∈ kerL. Example 3. Let L: โ4 → โ3 be a linear transformation defined by L(๐ฅ, ๐ฆ, ๐ง, ๐ค) = (๐ฅ + ๐ฆ, ๐ง + ๐ค, ๐ฅ + ๐ง) The vector (1, −1, −1, 1) is in kerL since L(1, −1, −1, 1) = (0, 0, 0) while the vector (1, 2, 3, −4) is not in kerL because L(1, 2, 3, −4) = ( 3, 1, 5 ) ≠ (0, 0, 0). Thus all vectors X in โ4 such that L(X) = (0, 0, 0) are in kerL. Example 4. Let L be the linear transformation of Example 3. The kerL consists of all vectors X in โ4 such that L(X) = 0. If we let X = (๐ฅ, ๐ฆ, ๐ง, ๐ค) then L(๐ฅ, ๐ฆ, ๐ง, ๐ค) = (0, 0, 0) (๐ฅ + ๐ฆ, ๐ง + ๐ค, ๐ฅ + ๐ง) = (0, 0, 0) Equating the corresponding parts we obtain the homogeneous system ๐ฅ+๐ฆ =0 ๐ง + ๐ค=0 ๐ฅ +๐ง =0 107 The augmented matrix in reduced echelon form (verify) is 1 [0 0 0 0 −1 0 1 0 1 0] 0 1 1 0 This implies that ๐ค is a free variable and can take on any value. Hence ๐ฅ=๐ ๐ฆ = −๐ ๐ง = −๐ ๐ค = ๐ where ๐ ∈ โ Thus kerL consists of all vectors of the form (๐, −๐, −๐, ๐) where ๐ is any real number. That is ker L = { (๐, −๐, −๐, ๐): ๐ ๏ โ} = { ๐(1, −1, −1, 1): ๐ ๏ โ} Hence ker L = span { (1, −1, −1, 1)} Since (1, −1, −1, 1) is linearly independent then it forms a basis for kerL. Thus dim(kerL) = 1. SAQ 5-3 Let L: โ4 → โ2 be defined by ๏ฆ๏ฉx๏น๏ถ ๏ง๏ช ๏บ๏ท y ๏ฉx ๏ญ y๏น L๏ง ๏ช ๏บ๏ท ๏ฝ ๏ช ๏ง ๏ช z ๏บ ๏ท ๏ซ z ๏ญ w๏บ๏ป ๏ง๏ง ๏ช ๏บ ๏ท๏ท ๏จ ๏ซ w๏ป ๏ธ Find kerL. 108 ASAQ 5-3 ker L = { (๐ฅ, ๐ฆ, ๐ง, ๐ค) : (๐ฅ − ๐ฆ, ๐ง − ๐ค) = (0, 0) } Equating the corresponding parts we get: ๐ฅ − ๐ฆ = 0 which implies that ๐ฅ = ๐ฆ and ๐ง − ๐ค = 0 which implies that ๐ง = ๐ค If we let ๐ฆ = ๐ ๏ โ and ๐ค = ๐ ๏ โ then kerL = { (๐, ๐, ๐ , ๐ ) : ๐ and ๐ ๏ โ } = { ๐ (1, 1, 0, 0) + ๐ (0, 0, 1, 1) } Hence kerL = span {(1, 1, 0, 0), (0, 0, 1, 1)} Since {(1, 1, 0, 0), (0, 0, 1, 1)} is linearly independent (one vector is not a scalar multiple of the other) then it forms a basis for kerL. Hence dim(kerL) = 2. Theorem 5.2.1: If L: V→ W is a linear transformation, then kerL is a subspace of V. Definition 5.2.3: If L: V→ W is a linear transformation, then the range of L, denoted by range L, is the set of all vectors in W that are images, under L, of vectors in V. Thus a vector Y is in range L if we can find some vector X in V such that L(X) = Y. If range L = W, we say that L is onto. Example 5: Let L: โ3 → โ2 be a linear transformation defined by L(๐ฅ, ๐ฆ, ๐ง) = (๐ฅ + ๐ฆ, ๐ฆ − ๐ง) The range or image of L (ImL) is: ImL = { (๐ฅ + ๐ฆ, ๐ฆ − ๐ง) } = { ๐ฅ (1, 0) + ๐ฆ (1, 1) + ๐ง (0, -1) : ๐ฅ, ๐ฆ, ๐ง ๏ โ } Thus ImL = span {(1, 0), (1, 1), (0, −1)}. By Theorem 4.1.1 {(1, 0), (1, 1), (0, −1)} is linearly dependent. Since (0, −1) is a linear combination of the other vectors such as 109 (0, −1) = 1 (1, 0) + (−1)(1, 1) then we can delete (0, −1) from the set and the remaining vectors still span ImL. The set {(1, 0), (1, 1)} is linearly independent therefore forms a basis for ImL. Hence dim(ImL) = 2. Theorem 5.2.2: Let L: V→ W be a linear transformation (i) L is a monomorphism (one-to-one) if and only if dim(kerL) = 0. (ii) L is an epimorphism (onto) if and only if dim(ImL) = dim W. Example 6. In ASAQ 5-3, since dim(kerL) = 2 then L is not a monomorphism (one-to-one) and in Example 5, since dim(ImL) = 2 = dimโ2 then L is an epimorphism (onto). Theorem 5.2.3: If L: V→ W is a linear transformation, then range L is a subspace of W. Theorem 5.2.4: If L: V→ W is a linear transformation, then dim(kerL) + dim(range L) = dim V Example 7: Verify Theorem 5.2.4 using the linear transformation of Example 5. Solution: kerL = { (๐ฅ, ๐ฆ, ๐ง) : (๐ฅ + ๐ฆ, ๐ฆ − ๐ง) = (0, 0) } Equating the corresponding parts we get: ๐ฅ + ๐ฆ = 0 which implies that ๐ฅ = −๐ฆ and ๐ฆ − ๐ง = 0 which implies that ๐ง = ๐ฆ ; If we let ๐ฆ = ๐, ๐ ๏ โ, then kerL = { (−๐, ๐, ๐) : ๐ ๏ โ } = { ๐(−1, 1, 1) : ๐ ๏ โ } Thus kerL = span { (−1, 1, 1) } which is linearly independent hence forms a basis for kerL. Thus dim(kerL) = 1. 110 From the previous example, we know that dim(ImL) = 2. Therefore dim(kerL) + dim(ImL) = dim โ3 1 + 2 = 3 Remark: Let L: V → W be a linear transformation. Then the rank of L is defined to be the dimension of its image, and the nullity of L is defined to be the dimension of its kernel. rank (L) = dim(ImL) and nullity (L) = dim(kerL) Thus the preceding theorem yields the following formula for L when V has a finite dimension: rank(L) + nullity(L) = dim V. Corollary 5.2.1: Let L: V→ W be a linear transformation and dim V = dim W, a. If L is one-to-one, then it is onto. b. If L is onto, then it is one-to-one. SAQ 5-4 Let L: โ3 →โ4 be defined by L(๐ฅ, ๐ฆ, ๐ง) = (๐ฅ + ๐ง, ๐ฆ − ๐ฅ, ๐ฆ + ๐ง, ๐ฅ + ๐ฆ + 2๐ง) Verify Theorem 5.2.7 111 ASAQ 5-4 L(๐ฅ, ๐ฆ, ๐ง) = (0, 0, 0, 0) (๐ฅ + ๐ง, ๐ฆ − ๐ฅ, ๐ฆ + ๐ง, ๐ฅ + ๐ฆ + 2๐ง) = (0, 0, 0, 0) Equating the corresponding parts we obtain the homogeneous system ๐ฅ +๐ง =0 −๐ฅ + ๐ฆ =0 ๐ฆ +๐ง =0 ๐ฅ + ๐ฆ + 2๐ง = 0 The augmented matrix in reduced row echelon form (verify) is 1 0 [ 0 0 0 1 0 0 1 1 0 0 0 0 ] 0 0 Thus ๐ฅ = −๐ง, ๐ฆ = −๐ง, where ๐ง can be any real number. If we let ๐ง = ๐ ∈ โ then kerL = {(−๐, −๐, ๐): ๐ ∈ โ} = {๐(−1, −1, 1): ๐ ∈ โ} Hence kerL = span {(−1, −1, 1)}. Since (−1, −1, 1) is linearly independent then it is a basis for kerL. Thus nullity of L = dim(kerL) = 1. Solving for the image of L, we have Im(L) = {(๐ฅ + ๐ง, ๐ฆ − ๐ฅ, ๐ฆ + ๐ง, ๐ฅ + ๐ฆ + 2๐ง)} = {๐ฅ(1, −1, 0, 1) + ๐ฆ(0, 1, 1, 1) + ๐ง(1, 0, 1, 2): ๐ฅ, ๐ฆ, ๐ง ∈ โ} Since (1, 0, 1, 2) = (1, −1, 0, 1) + (0, 1, 1, 1) then (1, 0, 1, 2) is a linear combination of the preceding vectors and may be deleted from the spanning set. Thus Im(L) = span {(1, −1, 0, 1), (0, 1, 1, 1)} Since {(1, −1, 0, 1), (0, 1, 1, 1)} is linearly independent (one vector is not a scalar multiple of the other) then it is a basis for Im(L). Therefore Rank of L = dim(ImL) = 2 112 Now it can be verified that Nullity of L + rank of L = dim(โ3 ) 1 + 2 = 3 ACTIVITY 1. Which of the following are linear transformations? (a) L(x, y) = (x2 + x, y – y2) (b) L(x, y) = (x – y, 0, 2x + 3) ๏ฆ ๏ฉ x ๏น ๏ถ ๏ฉ2x ๏ญ 3 y ๏น ๏ง ๏ท (c) L ๏ง ๏ช๏ช y ๏บ๏บ ๏ท ๏ฝ ๏ช๏ช3 y ๏ญ 2 z ๏บ๏บ ๏ง ๏ช z ๏บ ๏ท ๏ช 2z ๏บ ๏ป ๏จ๏ซ ๏ป๏ธ ๏ซ 2. Let L: โ3 → โ4 be defined by L(x, y, z) = (x + y + z, x + 2y – 3z, 2x + 3y – 2z, 3x + 4y – z) (a) Find a basis for and the dim(kerL), (b) Find a basis for and the dim(ImL), (c) Verify Theorem 5.2.4. 3. Let L: โ3 → โ3 be defined by L(x, y z) = (x + z, x + y + 2z, 2x + y + 3z ) (a) Is L one-to-one? (b) Is L onto? 4. Let L: P2 → P2 be the linear transformation defined by L(at2 + bt + c) = (a + c)t2 + (b + c)t. (a) Is t2 – t – 1 in kerL? (d) Find a basis for kerL. (b) Is t2 + t -1 in kerL? (e) Find a basis for ImL. 2 (c) Is 2t – t in range L? 113 5.3 The Matrix of a Linear Transformation Coordinate Vectors Definition 5.3.1: Let V be an n-dimensional vector space with basis S = { X1 , X2, … , Xn }. If X = a1X1 + a2X2 + … + anXn is any vector in V, then the vector ๏ฉ a1 ๏น ๏ชa ๏บ X ๏ฝ ๏ ๏S ๏ช๏ช 2 ๏บ๏บ ๏ช ๏บ ๏ช๏ซ an ๏บ๏ป in โ๐ is called the coordinate vector of X with respect to the basis S. The components of [ X ] S are called the coordinates of X with respect to S. ๏ฌ๏ฉ 1 ๏น ๏ฉ 2๏น ๏ผ Example 1. Let S = ๏ญ๏ช ๏บ , ๏ช ๏บ ๏ฝ be a basis for โ2 . Find the coordinate vectors of the ๏ฎ๏ซ๏ญ1๏ป ๏ซ 3๏ป ๏พ following vectors with respect to S. ๏ฉ ๏ญ3 ๏น (a) ๏ช ๏บ ๏ซ ๏ญ7 ๏ป ๏ฉ12 ๏น (b) ๏ช ๏บ ๏ซ13๏ป Solution: ๏ฉ ๏ญ3 ๏น (a) Let X = ๏ช ๏บ . To find [X]S we must find c1 and c2 such that ๏ซ ๏ญ7 ๏ป ๏ฉ ๏ญ3 ๏น ๏ฉ2๏น ๏ฉ1๏น ๏ช ๏ญ7 ๏บ = c1 ๏ช ๏บ + c2 ๏ช 3 ๏บ ๏ซ ๏ป ๏ซ ๏ป ๏ซ ๏ญ1๏ป Multiplying, adding, and equating the corresponding parts, we get c1 + 2c2 = − 3 − c1 + 3c2 = − 7 The solution is c1 = 1 and c2 = −2. Thus the coordinate vector of X with respect to the basis S is ๏ฉ1๏น ๏ช ๏ญ2 ๏บ ๏ซ ๏ป 114 Letter (b) is left as an exercise. ๏ฌ ๏ฉ1 ๏น ๏ฉ0๏น ๏ผ Example 2. Let S = ๏ญ E1 ๏ฝ ๏ช ๏บ , E2 ๏ฝ ๏ช ๏บ ๏ฝ be the natural basis for โ2 . Then ๏ซ0๏ป ๏ซ1 ๏ป ๏พ ๏ฎ ๏ฉ ๏ญ3 ๏น ๏ฉ1 ๏น ๏ฉ0 ๏น ๏ช ๏ญ7 ๏บ = −3 ๏ช 0 ๏บ + (−7) ๏ช 1 ๏บ ๏ซ ๏ป ๏ซ ๏ป ๏ซ ๏ป ๏ฉ ๏ญ3 ๏น Hence [X]S = ๏ช ๏บ . Note that [X]S is the original vector itself when the natural basis is ๏ซ ๏ญ7 ๏ป used. This result is true in general. Theorem 5.3.1: Let L: V→ W be a linear transformation of an ๐-dimensional vector space V into an ๐-dimensional vector space W ( ๐ ≠ 0 and ๐ ≠ 0 ) and let S = { X1 , X2, … , Xn} and T = {Y1, Y2, …, Ym } be bases for V and W, respectively. Then the ๐ x ๐ matrix A, whose jth column is the coordinate vector [L(Xj)]T of L(Xj) with respect to T, is associated with L and has the following property: If Y = L(X) for some X in V, then [Y]T = A[X]S , where [X]S and [Y]T are the coordinate vectors of X and Y with respect to the respective bases S and T. Moreover, A is the only matrix with this property. Procedure for computing the matrix of a linear transformation L: V→ W with respect to the bases S = { X1 , X2, … , Xn} and T = { Y1, Y2, …, Ym } for V and W, respectively: STEP 1: Compute L(Xj) for j = 1, 2, … , ๐. STEP 2: Find the coordinate vector [L(Xj)]T of L(Xj) with respect to the basis T. This means that we have to express L(Xj) as a linear combination of the vectors in T. STEP 3: The matrix A of L with respect to S and T is formed by choosing [L(Xj)]T as the jth column of A. Definition 5.3.2: The matrix of Theorem 5.3.1 is called the matrix of L with respect to the bases S and T. The equation [L(X)]T = A[X]S is called the representation of L with respect to S and T. We also say that the said equation represents L with respect to S and T. 115 Example 3: Let L: โ2 → โ2 be defined by ๏ฆ ๏ฉ x๏น ๏ถ ๏ฉx ๏ซ 2 y๏น L๏ง ๏ช ๏บ๏ท = ๏ช ๏บ ๏ซ2x ๏ญ y ๏ป ๏จ ๏ซ y๏ป ๏ธ ๏ฌ ๏ฉ1 ๏น ๏ฉ0๏น ๏ผ Let S = ๏ญ X1 ๏ฝ ๏ช ๏บ , X 2 ๏ฝ ๏ช ๏บ ๏ฝ be the natural basis for โ2 and let ๏ซ0๏ป ๏ซ1 ๏ป ๏พ ๏ฎ ๏ฌ ๏ฉ ๏ญ1๏น ๏ฉ 2๏น ๏ผ T = ๏ญY1 ๏ฝ ๏ช ๏บ , Y2 ๏ฝ ๏ช ๏บ ๏ฝ be another basis for โ2 . Find the matrix of L with respect to S and ๏ซ2๏ป ๏ซ0๏ป ๏พ ๏ฎ ๏ฆ ๏ฉ1 ๏น ๏ถ T. Compute L ๏ง ๏ช ๏บ ๏ท using the matrix of L. ๏จ ๏ซ2๏ป ๏ธ Solution: ๏ฆ ๏ฉ1 ๏น ๏ถ ๏ฉ1 ๏น ๏ฉ ๏ญ1๏น ๏ฉ 2๏น L ๏ง ๏ช ๏บ ๏ท = ๏ช ๏บ ๏ฝ a1 ๏ช ๏บ ๏ซ a2 ๏ช ๏บ ๏ซ2๏ป ๏ซ2๏ป ๏ซ0๏ป ๏จ ๏ซ0 ๏ป ๏ธ 1 = ๏ญa1 ๏ซ 2a2 2 = 2a1 ๏ฉ1๏น The solution is a1 ๏ฝ 1 and a2 ๏ฝ 1, so ๏ L( X1 )๏T = ๏ช ๏บ ๏ซ1๏ป ๏ฆ ๏ฉ0 ๏น ๏ถ ๏ฉ2๏น ๏ฉ ๏ญ1๏น ๏ฉ 2๏น L ๏ง ๏ช ๏บ ๏ท = ๏ช ๏บ ๏ฝ a1 ๏ช ๏บ ๏ซ a2 ๏ช ๏บ ๏ซ ๏ญ1๏ป ๏ซ2๏ป ๏ซ0๏ป ๏จ ๏ซ1 ๏ป ๏ธ 2 = ๏ญa1 ๏ซ 2a2 −1 = 2a1 ๏ฉ ๏ญ1/ 2 ๏น The solution is a1 ๏ฝ ๏ญ1/ 2 and a2 ๏ฝ 3/ 4 , so ๏ L( X 2 )๏T = ๏ช ๏บ. ๏ซ 3/ 4 ๏ป Thus the matrix A of L with respect to S and T is ๏ฉ1 ๏ญ1/ 2 ๏น A๏ฝ๏ช ๏บ ๏ซ1 3 / 4 ๏ป ๏ฉ ๏ฆ ๏ฉ1 ๏น ๏ถ ๏น ๏ฉ ๏ฉ1 ๏น ๏น To compute for ๏ช L ๏ง ๏ช ๏บ ๏ท ๏บ using A, we first compute for ๏ช ๏ช ๏บ ๏บ . Since S is the natural ๏ช๏ซ ๏จ ๏ซ 2๏ป ๏ธ ๏บ๏ปT ๏ซ ๏ซ2๏ป ๏ป S basis for โ2 then 116 ๏ฉ ๏ฉ1 ๏น ๏น ๏ฉ1 ๏น ๏ช๏ช ๏บ๏บ = ๏ช ๏บ ๏ซ2๏ป ๏ซ ๏ซ2๏ป ๏ป S So, ๏ฉ ๏ฆ ๏ฉ1 ๏น ๏ถ ๏น ๏ฉ1 ๏ญ1/ 2 ๏น ๏ช L ๏ง ๏ช ๏บ ๏ท๏บ = ๏ช ๏บ ๏ซ1 3 / 4 ๏ป ๏ซ๏ช ๏จ ๏ซ 2๏ป ๏ธ ๏ป๏บT ๏ฉ1 ๏น ๏ฉ 0 ๏น ๏ช 2 ๏บ = ๏ช5 / 2 ๏บ ๏ซ ๏ป ๏ซ ๏ป ACTIVITY ๏ฌ ๏ฉ 1 ๏น ๏ฉ 0 ๏น ๏ฉ1 ๏น ๏ผ ๏ฏ ๏ฏ 1. Let S = ๏ญ ๏ช๏ช ๏ญ1๏บ๏บ , ๏ช๏ช 2 ๏บ๏บ , ๏ช๏ช 0 ๏บ๏บ ๏ฝ be a basis for โ3 . Find the coordinate vector of each of the ๏ฏ ๏ช 2 ๏บ ๏ช1 ๏บ ๏ช0๏บ ๏ฏ ๏ฎ๏ซ ๏ป ๏ซ ๏ป ๏ซ ๏ป ๏พ following vectors with respect to S. ๏ฉ1 ๏น (a) ๏ช๏ช 4 ๏บ๏บ ๏ช๏ซ 2 ๏บ๏ป ๏ฉ3๏น (b) ๏ช๏ช ๏ญ4 ๏บ๏บ ๏ช๏ซ 3 ๏บ๏ป ๏ฉ1๏น (c) ๏ช๏ช ๏ญ2 ๏บ๏บ ๏ช๏ซ 4 ๏บ๏ป ๏ฉx ๏ญ 2 y๏น ๏ฆ ๏ถ x ๏ฉ ๏น 2. Let L: โ2 → โ3 be defined by L ๏ง ๏ช ๏บ ๏ท ๏ฝ ๏ช๏ช 2 x ๏ซ y ๏บ๏บ . Let S and T be the natural bases for ๏จ ๏ซ y๏ป ๏ธ ๏ช x ๏ซ y ๏บ ๏ซ ๏ป ๏ฌ ๏ฉ 1 ๏น ๏ฉ0 ๏น ๏ผ โ and โ , respectively. Also, let S’ = ๏ญ๏ช ๏บ , ๏ช ๏บ ๏ฝ and T’ = ๏ฎ๏ซ๏ญ1๏ป ๏ซ1๏ป ๏พ 2 3 ๏ฌ ๏ฉ1 ๏น ๏ฉ 0 ๏น ๏ฉ 1 ๏น ๏ผ ๏ฏ๏ช ๏บ ๏ช ๏บ ๏ช ๏บ ๏ฏ ๏ญ ๏ช1 ๏บ , ๏ช1 ๏บ , ๏ช ๏ญ1๏บ ๏ฝ be bases for ๏ฏ ๏ช0 ๏บ ๏ช1 ๏บ ๏ช 1 ๏บ ๏ฏ ๏ฎ๏ซ ๏ป ๏ซ ๏ป ๏ซ ๏ป ๏พ โ2 and โ3 , respectively. Find the matrix of L with respect to (a) S and T (b) S’ and T’ 3. Let L: P1 → P3 be defined by L(p(t)) = t2p(t). Let S = { t, t + 1 } and T = { t3 , t2 – 1, t, t + 1 } be bases for P1 and P3, respectively. Find the matrix of L with respect to S and T. Compute [ L(–3t + 3) ]T using the matrix of L. 117 MODULE 6 EIGENVALUES AND EIGENVECTORS Introduction In this section we will discuss the concepts of eigenvalues, eigenvectors, and eigenspaces, algebraic and geometric multiplicity of an eigenvalue, Hamilton-Cayley theorem, similar matrices and diagonalizable matrices. Objectives At the end of this chapter, you are expected to be able to do the following: 1. Define eigenvalues and eigenvectors. 2. Compute for the eigenvalues and the associated eigenvectors of a matrix representing a linear transformation. 3. Determine the algebraic and geometric multiplicities of the eigenvalues. 3. Identify the properties of a diagonalizable matrix. 4. Determine the matrix of transition P so that ๐ −1 ๐ด๐ is a diagonal matrix, given a matrix A representing a linear transformation of a vector space into itself. 6.1 Characteristic Polynomial Definition 6.1.1: If A ๏ฝ [ a ij ] is an ๐ x ๐ matrix, the polynomial matrix ๐ฅ๐ผ − ๐ด = ๐ถ is called the characteristic matrix of ๐ด. The determinant of ๐ถ is called the characteristic polynomial of ๐ด. The equation det๐ถ = 0 is called the characteristic equation of ๐ด. 1 Example 1: Let ๐ด = [ 4 −3 ]. Then −2 ๐ฅ๐ผ2 − ๐ด = ๐ถ = [ = [ is called the characteristic matrix of ๐ด. ๐ฅ 0 0 1 ]−[ ๐ฅ 4 ๐ฅ−1 −4 3 ] ๐ฅ+2 −3 ] −2 118 The determinant of ๐ถ ๐ฅ−1 | −4 3 | = (๐ฅ − 1)(๐ฅ + 2) − (−12) ๐ฅ+2 = ๐ฅ 2 + ๐ฅ + 10 is called the characteristic polynomial of ๐ด and the equation ๐ฅ 2 + ๐ฅ + 10 = 0 is called the characteristic equation of ๐ด. 6.2 Hamilton-Cayley Theorem Let ๐(๐ฅ) = ๐0 ๐ฅ ๐ + ๐1 ๐ฅ ๐−1 + … + ๐๐−1 ๐ฅ + ๐๐ be a polynomial in ๐ฅ with real coefficients. If ๐ด is an ๐ x ๐ matrix then ๐(๐ด) is equal to the matrix ๐0 ๐ด๐ + ๐1 ๐ด๐−1 + … + ๐๐−1 ๐ด + ๐๐ ๐ผ๐ Note that we replace the constant term by ๐๐ ๐ผ๐ so that each term of ๐(๐ด) is a matrix. Theorem 6.2.1. (Hamilton-Cayley Theorem) If ๐ด is an ๐ x ๐ matrix and ๐(๐ฅ) is its characteristic polynomial, then ๐(๐ด) = 0. 2 3 Example 1: Let ๐ด = [ ]. The characteristic matrix of ๐ด is −1 4 ๐ฅ − 2 −3 ๐ถ=[ ] 1 ๐ฅ−4 Next we compute for the characteristic polynomial of ๐ด. ๐(๐ฅ) = det๐ถ ๐ฅ−2 ๐(๐ฅ) = | 1 −3 | ๐ฅ−4 = (๐ฅ − 2)(๐ฅ − 4) − (−3) = ๐ฅ 2 − 6๐ฅ + 11 It can be verified that the constant term is equal to the determinant of ๐ด. 119 Computing for ๐(๐ด) we have ๐(๐ด) = ๐ด2 − 6๐ด + 11๐ผ2 2 3 2 =[ ][ −1 4 −1 1 3 2 3 ] − 6[ ] + 11 [ 0 4 −1 4 1 18 12 18 11 = [ ]−[ ]+[ −6 13 −6 24 0 0 = [ 0 0 ] 1 0 ] 11 0 ] 0 Hence ๐ด satisfies its characteristic equation, that is ๐(๐ด) = 0. Since det๐ด = 11 ≠ 0 then ๐ด is nonsingular so ๐ด−1 exists. By the Hamilton-Cayley Theorem, we form the equation ๐ด2 − 6๐ด + 11๐ผ2 = 0 11๐ผ2 = 6๐ด − ๐ด2 ๐ผ๐ = ๐ผ๐ = 1 11 (6๐ด − ๐ด2 ) 1 11 (๐ด)(6๐ผ − ๐ด) Therefore, 1 ๐ด−1 = 11 (6๐ผ − ๐ด) 4 −3 ] 1 2 4/11 −3/11 = [ ] 1/11 2/11 1 = 11 [ 120 SAQ 6-1 Find the characteristic polynomial for the matrix 2 −2 3 [1 1 1] 1 3 −1 and a) Show by direct substitution that this matrix satisfies its characteristic equation. b) Find ๐ด−1 . ASAQ 6-1 The characteristic matrix of ๐ด is ๐ฅ−2 [ −1 −1 2 −3 ๐ฅ−1 −1 ] −3 ๐ฅ + 1 and its characteristic polynomial is ๐(๐ฅ) = (๐ฅ − 2)(๐ฅ − 1)(๐ฅ + 1) − 7 − [3(๐ฅ − 1) − 2(๐ฅ + 1) + 3(๐ฅ − 2)] = ๐ฅ 3 − 2๐ฅ 2 − 5๐ฅ + 6 Next we form the equation ๐(๐ด) = ๐ด3 − 2๐ด2 − 5๐ด + 6๐ผ3 = 0 (1) where 2 −2 3 2 −2 3 5 ๐ด2 = ๐ด โ ๐ด = [1 1 1 ] [1 1 1 ] = [4 1 3 −1 1 3 −1 4 5 3 1 2 −2 ๐ด3 = ๐ด2 โ ๐ด = [4 2 3] [1 1 4 −2 7 1 3 Substituting these to (1), we have 3 1 2 3] and −2 7 3 14 −4 1 ] = [13 3 −1 13 11 17 11] 3 121 14 ๐(๐ด) = [13 13 0 0 = [0 0 0 0 −4 17 2 −2 3 6 0 5 3 1 ] − 2 [ ] − 5 [ ] + [ 3 11 1 1 1 0 6 4 2 3 11 3 1 3 −1 0 0 4 −2 7 0 0] 6 0 0] 0 Thus ๐ด satisfies its characteristic equation. b. To solve for ๐ด−1 , we form the equation ๐ด3 − 2๐ด2 − 5๐ด + 6๐ผ3 = 0 6๐ผ3 = −๐ด3 + 2๐ด2 + 5๐ด 1 ๐ผ3 = 6 (−๐ด3 + 2๐ด2 + 5๐ด) 1 ๐ผ3 = 6 (๐ด)(−๐ด2 + 2๐ด + 5๐ผ3 ) Therefore 1 ๐ด−1 = 6 (−๐ด2 + 2๐ด + 5๐ผ3 ) 2 1 −5 −3 −1 = [[−4 −2 −3] + 2 [1 6 1 −4 2 −7 −2 3 5 0 0 ] + [ 1 1 0 5 0]] 3 −1 0 0 5 4 −7 5 = 6 [−2 5 −1] −2 8 −4 2/3 −7/6 5/6 −1/3 5/6 −1/6] =[ −1/3 4/3 −2/3 1 6.3 Eigenvalues, Eigenvectors, and Eigenspaces Definition 6.3.1: Let ๐ด be an ๐ x ๐ matrix. The real number ๐ is called an eigenvalue of ๐ด if there exists a nonzero vector ๐ฟ in โ๐ such that ๐ด๐ฟ = ๐๐ฟ (1) 122 Every nonzero vector ๐ฟ satisfying (1) is called an eigenvector of ๐ด associated with the eigenvalue ๐. Example 1: Let ๐ด = [ [ 2 −1 ]. Since −2 3 2 −1 1 1 ][ ] = 4[ ] −2 3 −2 −2 then 4 is an eigenvalue of ๐ด and [ 4. 1 ] is the eigenvector associated with the eigenvalue ๐ = −2 Theorem 6.3.1: The eigenvalues of ๐ด are the real roots of the characteristic polynomial of A. 1 0 Example 2. Let ๐ด = [−1 3 3 2 ๐ฅ−1 | 1 −3 0 0 ]. The characteristic polynomial of ๐ด is −2 0 ๐ฅ−3 −2 0 0 | = (๐ฅ − 1)(๐ฅ − 3)(๐ฅ + 2) ๐ฅ+2 By Theorem 6.3.1, the eigenvalues of ๐ด are the roots of the characteristic equation (๐ฅ − 1)(๐ฅ − 3)(๐ฅ + 2) = 0 Thus ๐1 = 1, ๐2 = 3, and ๐3 = −2 are the eigenvalues of ๐ด. To find the eigenvector associated with ๐1 = 1, we form the system (1๐ผ3 − ๐ด)๐ = 0: 0 0 0 ๐ฅ1 0 ๐ฅ [ 1 −2 0] [ 2 ] = [0] −3 −2 3 ๐ฅ3 0 6 A solution to this system is {(4 ๐, 8 ๐, ๐) : ๐ ∈ โ}. Thus if ๐ = 8 then ๐1 = [3] is the 8 eigenvector associated with ๐1 = 1. 3 3 123 To find the eigenvector associated with ๐2 = 3, we form the system (3๐ผ3 − ๐ด)๐ = 0: 2 0 0 ๐ฅ1 0 ๐ฅ [1 0 0] [ 2 ] = [0] −3 −2 5 ๐ฅ3 0 0 A solution to this system is {(0, 2 ๐, ๐) : ๐ ∈ โ}. Thus if ๐ = 2 then ๐2 = [5] is the 2 eigenvector associated with ๐2 = 3. 5 To find the eigenvector associated with ๐3 = −2, we form the system (−2๐ผ3 − ๐ด)๐ = 0: −3 0 0 ๐ฅ1 0 ๐ฅ [ 1 −5 0] [ 2 ] = [0] −3 −2 0 ๐ฅ3 0 0 A solution to this system is {(0, 0, ๐): ๐ ∈ โ}. Thus ๐2 = [0] is the eigenvector associated 1 with ๐3 = −2. ๏ฉ 2 1๏น Example 3: Let A ๏ฝ ๏ช ๏บ. ๏ซ ๏ญ1 3๏ป The characteristic polynomial of ๐ด is ๐ฅ−2 | 1 −1 | = (๐ฅ − 2)(๐ฅ − 3) − (−1) ๐ฅ−3 = ๐ฅ 2 − 5๐ฅ + 7 Since ๐ฅ 2 − 5๐ฅ + 7 = 0 has no real roots then A has no eigenvalues. Definition 6.3.2: Let L: V→ V be a linear transformation and let ๏ฌ be an eigenvalue of L. Denote by S( ๏ฌ ) the set of all eigenvectors associated with ๏ฌ , together with the zero vector. Then S( ๏ฌ ) is called the eigenspace of L associated with ๏ฌ . Definition 6.3.3: Let ๏ฌ be an eigenvalue of L. (a) The geometric multiplicity of ๏ฌ is the dimension of S( ๏ฌ ). (b) The algebraic multiplicity of ๏ฌ is its multiplicity as a root of the characteristic polynomial f( ๏ฌ ). 124 Example 4. Let L be a linear transformation represented by the matrix 2 0 0 ๐ด = [3 −1 0 ] 0 4 −1 Determine the geometric and algebraic multiplicities of its eigenvalues. Solution: The characteristic polynomial of A is ๐ฅ−2 0 0 ๐(๐ฅ) = | −3 ๐ฅ + 1 0 | 0 −4 ๐ฅ+1 = (๐ฅ − 2)(๐ฅ + 1)(๐ฅ + 1) = (๐ฅ − 2)(๐ฅ + 1)2 Thus the eigenvalues of A are ๐1 = 2 with algebraic multiplicity 1 and ๐2 = ๐3 = −1 with algebraic multiplicity 2. To find the geometric multiplicity of ๐1 = 2, we have to determine the dimension of the eigenspace S(2) associated with ๐1 = 2. Substituting 2 for x in the characteristic matrix of A we obtain 0 0 0 [−3 3 0] 0 −4 3 This matrix in reduced row echelon form is 1 [0 0 3 The solution is {(4 ๐, 3 4 0 −3/4 1 −3/4] 0 0 ๐, ๐) : ๐ ∈ โ}. Thus 3 S(2) = {(4 ๐, 3 4 ๐, ๐) : ๐ ∈ โ} = span {(3, 3, 4)} Hence ๐1 = 2 has geometric multiplicity 1. 125 Similarly, substituting −1 for x in the characteristic matrix we obtain −3 0 0 [−3 0 0] 0 −4 0 This matrix in reduced row echelon form is 1 0 [0 1 0 0 0 0] 0 Thus S(−1) = {(0, 0, ๐): ๐ ∈ โ} = span{(0, 0, 1)} Hence ๐2 = ๐3 = −1 has geometric multiplicity 1. Theorem 6.3.2. Let ๐ be an eigenvalue. Then the geometric multiplicity of ๐ does not exceed its algebraic multiplicity. Theorem 6.3.3. If the eigenvalues ๐1 , … , ๐๐ are all different and {๐1 , … , ๐2 } is a set of eigenvectors, ๐๐ corresponding to ๐๐ , then the set {๐1 , … , ๐2 } is linearly independent. Example 5. Let A be the matrix of Example 2. The three distinct eigenvalues of A are 6 0 0 ๐1 = 1, ๐2 = 3, and ๐3 = −2 and the corresponding eigenvectors are [3], [5], and [0]. It 8 2 1 6 0 0 can be verified that [3], [5], and [0] are linearly independent. 8 2 1 126 SAQ 6-2 3 2 Let ๐ด = [ 1 4 −2 −4 2 1 ]. −1 Determine the algebraic and geometric multiplicities of its eigenvalues. ASAQ 6-2 The characteristic polynomial of A is ๐ฅ−3 −2 −2 ๐(๐ฅ) = | −1 ๐ฅ − 4 −1 | 2 4 ๐ฅ+1 = ๐ฅ 3 − 6๐ฅ 2 + 11๐ฅ − 6 = (๐ฅ − 1)(๐ฅ − 2)(๐ฅ − 3) Thus the eigenvalues of A are 1, 2, and 3 all of which have algebraic multiplicity 1. To solve for the geometric multiplicity of ๐ = 1, we substitute 1 for x in the characteristic matrix obtaining the matrix −2 −2 −2 [−1 −3 −1] 2 4 2 This matrix in reduced row echelon form is 1 0 [0 1 0 0 Thus S(1) = {(−๐, 0, ๐): ๐ ∈ โ = span{(−1, 0, 1)} Hence ๐ = 1 has geometric multiplicity 1. 1 0] 0 127 To solve for the geometric multiplicity of ๐ = 2, we substitute 2 for x in the characteristic matrix obtaining the matrix −1 −2 −2 [−1 −2 −1] 2 4 3 This matrix in reduced row echelon form is equal to 1 2 [0 0 0 0 0 1] 0 Thus S(2) = {(−2๐, ๐, 0): ๐ ∈ โ = span{(−2, 1, 0)} Hence ๐ = 2 has geometric multiplicity 1. To solve for the geometric multiplicity of ๐ = 3, we substitute 3 for x in the characteristic matrix obtaining the matrix 0 −2 −2 [−1 −1 −1] 2 4 4 This matrix in reduced row echelon form is 1 0 [0 1 0 0 Thus S(3) = {(0, −๐, ๐): ๐ ∈ โ = span{(0, −1, 1)} Hence ๐ = 3 has geometric multiplicity 1. 0 1] 0 128 6.4 Diagonalization Definition 6.4.1: A matrix B is similar to a matrix A if there is a nonsingular matrix P such that B = P-1AP Example 1: Let ๐ด = [ ๐−1 ๐ด๐ we have 1 −1 −1 −1 1 ] and ๐ = [ ]. Then ๐−1 = [ 2 4 2 1 −2 ๐ต=[ = [ 1 ]. If we let ๐ต = −1 1 1 1 −1 −1 −1 ][ ][ ] −2 −1 2 4 2 1 3 0 ] 0 2 By Definition 6.4.1, B is similar to A. REMARKS: 1. A is similar to A. 2. If B is similar to A, then A is similar to B. 3. If A is similar to B and B is similar to C, then A is similar to C. Theorem 6.4.1. Similar matrices have the same characteristic polynomial. Theorem 6.4.2. Similar matrices have the same eigenvalues and eigenvectors. Example 2. Let A and B be the matrices of Example 1. The characteristic polynomial of A is ๐ฅ−1 1 ๐(๐ฅ) = | | = ๐ฅ 2 − 5๐ฅ + 6 −2 ๐ฅ − 4 and the characteristic polynomial of B is ๐ฅ−3 0 ๐(๐ฅ) = | | = ๐ฅ 2 − 5๐ฅ + 6 0 ๐ฅ−2 Thus A and B have the same characteristic polynomial. Consequently, A and B also have the same eigenvalues and eigenvectors. 129 Definition 6.4.2: We shall say that the matrix A is diagonalizable if it is similar to a diagonal matrix. In this case we also say that A can be diagonalized. Example 3. Let A be the matrix of Example 1. Since A is similar to a diagonal matrix then A is diagonalizable. Theorem 6.4.3: An ๐ x ๐ matrix A is diagonalizable if and only if it has ๐ linearly independent eigenvectors. In this case A is similar to a diagonal matrix D, with P-1AP = D, whose diagonal elements are the eigenvalues of A, while P is a matrix whose columns are ๐ linearly independent eigenvectors of A. Theorem 6.4.4: A matrix A is diagonalizable if all the roots of its characteristics polynomial are real and distinct. Example 4. Let ๐ด = [ 1 4 ]. The characteristic polynomial of A is 1 −2 ๐ฅ−1 ๐(๐ฅ) = | −1 −4 | = ๐ฅ 2 + ๐ฅ − 6 = (๐ฅ + 3)(๐ฅ − 2) ๐ฅ+2 Since the eigenvalues are real and distinct then A is diagonalizable. Example 5. Let ๐ด be the matrix of SAQ6-2. The distinct eigenvalues of A are 1, 2, and 3 −1 −2 hence A is diagonalizable. The linearly independent eigenvectors of A are [ 0 ] , [ 1 ] , 1 0 0 and [−1]. Thus by Theorem 6.4.3 1 −1 −2 0 1 2 2 ๐=[ 0 1 −1] and ๐−1 = [−1 −1 −1] (verify) 1 0 1 −1 −2 −1 Then ๐ท = ๐−1 ๐ด๐ 1 = [−1 −1 2 2 3 2 2 −1 −2 0 −1 −1] [ 1 4 1 ][ 0 1 −1] −2 −1 −2 −4 −1 1 0 1 1 2 2 −1 −2 0 = [−2 −2 −2] [ 0 1 −1] −3 −6 −3 1 0 1 130 1 0 = [0 2 0 0 0 0] 3 Note that D is a diagonal matrix whose diagonal elements are the eigenvalues of A. ACTIVITY ๏ฉ2 ๏ญ 2 3 ๏น 1. Let A ๏ฝ ๏ช๏ช0 3 ๏ญ 2๏บ๏บ . Find the characteristic polynomial, eigenvalues, and ๏ช๏ซ0 ๏ญ 1 2 ๏บ๏ป eigenvectors of A. ๏ฉ3 ๏ญ 2 1 ๏น 2. Let A ๏ฝ ๏ช๏ช0 2 0๏บ๏บ . Find a nonsingular matrix P such that P-1AP is diagonal. ๏ช๏ซ0 0 0๏บ๏ป 3. Let L be a linear transformation represented by the matrix ๏ฉ0 0 0 ๏น A ๏ฝ ๏ช๏ช0 1 0 ๏บ๏บ ๏ช๏ซ1 0 1 ๏บ๏ป Determine the geometric and algebraic multiplicities of its eigenvalues. 131 MODULE 7 INNER PRODUCT SPACES Introduction In this section we will discuss inner product, inner product spaces, orthonormal bases in โ๐ , the Gram-Schmidth orthogonalization process, diagonalization of symmetric matrix, and quadratic forms. Objectives After studying this module, you should be able to: 1. Explain the steps on the diagonalization of the matrix. 2. Enumerate the properties of a diagonizable symmetric matrix. 3. Perform diagonalization of a symmetric matrix by orthogonalization. 7.1 Inner Product in โ๐ Definition 7.1.1. The length or magnitude or norm of the vector ๐ = (๐ฅ1 , ๐ฅ2 , … , ๐ฅ๐ ) in โ๐ is โ๐โ = √๐ฅ1 2 + ๐ฅ2 2 + … + ๐ฅ๐ 2 The above formula is also use to define the distance from the point (๐ฅ1 , ๐ฅ2 , … , ๐ฅ๐ ) to the origin. Example 1. Let ๐ = (1, 2, −3) be a vector in โ3 . The length of ๐ is โ๐โ = √12 + 22 + (−3)2 = √14 Definition 7.1.2. If ๐ = (๐ฅ1 , ๐ฅ2 , … , ๐ฅ๐ ) and ๐ = (๐ฆ1 , ๐ฆ2 , … , ๐ฆ๐ ) are vectors in โ๐ , then their inner product (also called dot product) is given by ๐ โ ๐ = ๐ฅ1 ๐ฆ1 + ๐ฅ2 ๐ฆ2 + โฏ ๐ฅ๐ ๐ฆ๐ . Example 2. Let ๐ = (1, 0, 1) and ๐ = (2, 3, −1). Then their inner product is ๐ โ ๐ = 1 โ 2 + 0 โ 3 + 1(−1) = 1 132 Theorem 7.1.1. (Properties of Inner Product) If ๐, ๐ and ๐ are vectors in โ๐ and ๐ is a scalar, then: a. ๐ โ ๐ = โ๐โ2 ≥ 0, with equality if and only if ๐ = 0. b. ๐ โ ๐ = ๐ โ ๐ c. (๐ + ๐) โ ๐ = ๐ โ ๐ + ๐ โ ๐ d. (๐๐) โ ๐ = ๐ โ (๐๐) = ๐(๐ โ ๐) NOTE: A vector space V along with an inner product or scalar product operation satisfying all these properties is called an inner product space. Example 3. product. โ๐ is an inner product space since it satisfies all the properties of inner Definition 7.1.3. The angle between two nonzero vectors ๐ and ๐ is defined as the unique number ๐, 0 ≤ ๐ ≤ ๐, such that ๐๐๐ ๐ = ๐โ๐ โ๐โโ๐โ Example 4. Let ๐ = (0, 0, 1, 1) and ๐ = (1, 0, 1, 0). Then โ๐โ = √2, โ๐โ = √2 and ๐ โ ๐ = 1 Thus ๐๐๐ ๐ = 1 (√2)(√2) 1 ๐๐๐ ๐ = 2 1 ๐ = cos−1 (2) ๐ = 60๐ Definition 7.1.4. Two nonzero vectors ๐ and ๐ in โ๐ are said to be orthogonal if ๐ โ ๐ = 0. If one of the vectors is the zero vector, we agree to say that the vectors are orthogonal. They are said to be parallel if |๐ โ ๐| = โ๐โโ๐โ. They are in the same direction if ๐ โ ๐ = โ๐โโ๐โ. That is, they are orthogonal if ๐๐๐ ๐ = 0, parallel if ๐๐๐ ๐ = ±1 and in the same direction if ๐๐๐ ๐ = 1. 133 Example 5. Let ๐1 = (4, 2, 6, −8), ๐2 = (−2, 3, −1, −1) and ๐3 = (−2, −1, −3, 4). Then ๐1 โ ๐2 = 4(−2) + 2 โ 3 + 6(−1) + (−8)(−1) = 0, ๐2 โ ๐3 = (−2)(−2) + 3(−1) + (−1)(−3) + (−1)(4) = 0 and ๐1 โ ๐3 = 4(−2) + 2(−1) + 6(−3) + (−8)(4) = −60 This shows that ๐1 and ๐2 are orthogonal and ๐2 and ๐3 are also orthogonal. Moreover, โ๐1 โ = 2√30 and โ๐3 โ = √30 then |๐1 โ ๐3 | = โ๐1 โโ๐3 โ |−60| = (2√30)(√30) 60 = 60 This implies that ๐1 and ๐3 are parallel but not in the same direction. Definition 7.1.5. A unit vector ๐ in โ๐ is a vector of unit length. If ๐ is a nonzero vector, then the vector ๐= [ 1 โ๐โ ]๐ is a unit vector in the direction of ๐. Example 6. Consider the vector ๐ = (0, 4, 2, 3). Since โ๐โ = √29 then the vector ๐= 1 √29 (0, 4, 2, 3) = (0, Is a unit vector in the direction of ๐. 4 √29 , 2 √29 , 3 ) √29 134 72. Orthonormal Bases in โ๐ Definition 7.2.1. A set ๐ = {๐ฅ1 , ๐ฅ2 , … , ๐ฅ๐ } in โ๐ is called orthogonal if any two distinct vectors in S are orthogonal. An orthonormal set of vectors is an orthogonal set of unit vectors. Example 1. Let ๐1 = (1, 2, −1, 1), ๐2 = (0, −1, −2, 0) and ๐3 = (1, 0, 0, −1). Since ๐1 โ ๐2 = 0, ๐1 โ ๐3 = 0 and ๐2 โ ๐3 = 0 then the set {๐1, ๐2 , ๐3 } is orthogonal. The vectors ๐1 = ( 1 , 2 √7 √7 ,− 1 , 1 ) , ๐2 = (0, − √7 √7 1 √5 ,− 2 √5 , 0) and ๐3 = ( 1 √2 , 0, 0, − 1 ) √2 are unit vectors in the direction of ๐1 , ๐2 and ๐3 , respectively. Thus {๐1 , ๐2 , ๐3 } is an orthonormal set and span the same subspace as {๐1 , ๐2 , ๐3 }, that is, span{๐1 , ๐2 , ๐3 } = span{๐1 , ๐2 , ๐3 }. Notice that an orthonormal set is just a set of orthogonal vectors in which each vector has been normalized to unit length. Theorem 7.2.1. Let ๐ = {๐ฅ1 , ๐ฅ2 , … , ๐ฅ๐ } be an orthogonal set of nonzero vectors in โ๐ . Then S is linearly independent. Example 2. The vectors ๐1 , ๐2 and ๐3 of Example 1 are orthogonal hence they are linearly independent. Corollary 7.2.1. An orthonormal set of vectors in โ๐ is linearly independent. Example 3. The set of vectors {๐1 , ๐2 , ๐3 } of Example 1 is orthonormal hence it is linearly independent. Gram-Schmidt Orthogonalization Process Gram-Schmidth process is a procedure by which an orthonormal set of vectors is obtained from a linearly independent set of vectors in an inner product space. 135 Definition 7.2.2. (Gram-Schmidt Process) Let W be a nonzero subspace of โ๐ with basis ๐ = {๐1 , ๐2 , … , ๐๐ }. Then there exists an orthonormal basis ๐ = {๐1 , ๐2 , … , ๐๐ } for W. Steps in the Gram-Schmidth process: STEP 1. Let ๐1 = ๐1. STEP 2. Compute the vectors ๐2 , ๐3 , … , ๐๐ , by the formula ๐๐ โ๐๐ ๐๐ = ๐๐ − ∑๐−1 ๐=1 ( ๐ ๐ โ๐๐ ) ๐๐ for 2 ≤ ๐ ≤ ๐ The set of vectors ๐ ′ = {๐1 , ๐2 , … , ๐๐ } is an orthogonal set. STEP 3. Normalize each vector in ๐′ to obtain an orthonormal basis for W. Example 4. Use the Gram-Schmidth process to transform the basis {(1, 1, 1), (0, 1, 1), (1, 2, 3)} for โ3 into an orthonormal basis for โ3 . Solution: We apply the Gram-Schmidth process to obtain 3 vectors ๐1 , ๐2 , ๐3 which also span โ3 and orthogonal to each other. STEP 1. Let ๐1 = ๐1 = (1, 1, 1). STEP 2. We now compute for ๐2 and ๐3 : ๐2 โ ๐1 ๐2 = ๐2 − ( )๐ ๐1 โ ๐1 1 2 = (0, 1, 1) − ( ) (1, 1, 1) 3 2 1 1 = (− , , ) 3 3 3 and ๐3 โ ๐1 ๐3 โ ๐2 ๐3 = ๐3 − ( ) ๐1 − ( )๐ ๐1 โ ๐1 ๐2 โ ๐2 2 6 1 2 1 1 = (1, 2, 3) − ( ) (1, 1, 1) − ( ) (− , , ) 3 6/9 3 3 3 136 3 2 1 1 = (1, 2, 3) − (2, 2, 2) − (− , , ) 2 3 3 3 1 1 = (0, − , ) 2 2 2 1 1 1 1 The set of vectors ๐ ′ = {(1, 1, 1), (− 3 , 3 , 3) , (0, − 2 , 2)} is orthogonal. STEP 3. We normalize each vector found in STEP 2. Let ๐1 = ๐1 โ๐1 โ 1 ๐2 = ๐2 1 = โ๐2 โ √6/9 = = √3 3 ๐3 โ๐3 โ = = 1 , 1 , 1 √3 √3 √3 ) , (− 1 , 1 ) 2 1 1 2 1 1 √6 2 √2 , , 1 ) 1 1 √2/4 2 1 , √6 √6 √6 1 2 , √3 √3 √3 (− 3 , 3 , 3) (0, − 2 , 2) 1 1 (0, − 2 , 2) = (0, − Then ๐ = {( 1 (− 3 , 3 , 3) = (− ๐3 = (1, 1, 1) = ( 1 1 , 1 ) √2 √2 1 1 1 , ) , (0, − √2 , √2)} is an orthonormal basis for โ3 . √6 √6 √6 137 SAQ 7-1 Use the Gram-Schmidt process to find an orthonormal basis for the subspace W of โ4 with basis {(1, 1, -1, 0), (0, 2, 0, 1), (-1, 0, 0, 1)}. ASAQ 7-1 STEP 1. Let ๐1 = ๐1 = (1, 1, −1, 0) STEP 2. We now solve for ๐2 and ๐3 : ๐2 โ ๐1 ๐2 = ๐2 − ( )๐ ๐1 โ ๐1 1 2 = (0, 2, 0, 1) − (1, 1, −1, 0) 3 2 4 2 = (− , , , 1) 3 3 3 ๐3 โ ๐1 ๐3 โ ๐2 ๐3 = ๐3 − ( ) ๐1 − ( )๐ ๐1 โ ๐1 ๐2 โ ๐2 2 1 5/3 2 4 2 = (−1, 0, 0,1) − (− ) (1, 1, −1, 0) − (− , , , 1) 3 11/3 3 3 3 = (− 4 3 7 6 ,− ,− , ) 11 11 11 11 2 4 2 4 3 7 6 Thus ๐ ′ = {(1, 1, −1, 0), (− 3 , 3 , 3 , 1) , (− 11 , − 11 , − 11 , 11)} is an orthogonal basis for a subspace W of โ4 . 138 STEP 3. We normalize each vector in ๐′. Let ๐1 = ๐2 = ๐1 โ๐1 โ 1 = (1, 1, -1, 0) = ( √3 , 1 √3 √3 ,− 1 √3 ,0 ) ๐2 1 2 4 2 = (− 3 , 3 , 3 , 1) โ๐2 โ √33/9 = 3 2 4 2 ๐3 โ๐3 โ 1 = 11 1 1 3 3 4 , , 2 , 3 ) √33 √33 √33 √33 3 7 6 (− 11 , − 11 , − 11 , 11) 4 √110 = (− 2 4 √110/121 = (− 3 , 3 , 3 , 1) √33 = (− ๐3 = 1 3 7 6 (− 11 , − 11 , − 11 , 11) 4 √110 ,− 3 √110 1 3 ,− 7 , 6 √110 √110 ) 2 4 2 3 4 3 7 6 , , , ) , (− √110 , − √110 , − √110 , √110)} is 33 √33 √33 √33 Thus ๐ = {(√ , √ , − √ , 0 ) , (− √ an orthonormal basis for the subspace W of โ4 ACTIVITY 1. Which of the following are orthonormal sets of vectors? 1 1 1 1 1 a. ( , 0, − ) , ( , , ) , (0, 1, 0) √2 √2 √3 √3 √3 b. (0, 2, 2, 1), (−1, 1, 2, 2), (0, 1, −2, 1) 1 c. ( √6 ,− 2 √6 , 0, 1 ) , (− √6 1 √3 ,− 1 √3 , 0, − 1 √3 ) , (0, 0, 1, 0) 139 2. Use the Gram-Schmidt process to find an orthonormal basis for the subspace of โ4 with basis {(1, -1, 0, 1), (2, 0, 0, -1), (0, 0, 1, 0)}. 7.3 Diagonalization of Symmetric Matrix Theorem 7.3.1. All roots of the characteristic polynomial of a symmetric matrix are real numbers. −1 0 Example 1. Let ๐ด = [ 0 2 0 0 0 0]. The characteristic polynomial of A is 3 ๐ฅ+1 ๐(๐ฅ) = | 0 0 0 0 ๐ฅ−2 0 | 0 ๐ฅ−3 = (๐ฅ + 1)(๐ฅ − 2)(๐ฅ − 3) Clearly the roots of ๐(๐ฅ) are all real numbers. Corollary 7.3.1. If A is a symmetric matrix all of whose eigenvalues are distinct, then A is diagonalizable. Example 2. Let A be the matrix of Example 1. The eigenvalues ๐1 = −1, ๐2 = 2 and ๐3 = 3 are real and distinct. Hence A is diagonalizable. Theorem 7.3.2. If A is a symmetric matrix, then the eigenvectors that belong to distinct eigenvalues of A are orthogonal. Example 3. Let A be the matrix of Example 1. To find the associated eigenvectors, we form the system (๐๐ผ3 − ๐ด)๐ = 0 and solve for x. Thus 1 0 0 ๐1 = [0], ๐2 = [1] and X 3 = [0] 0 0 1 140 are the eigenvectors associated with ๐1 = −1, ๐2 = 2 and ๐3 = 3, respectively. It can be verified that {๐1 , ๐2 , ๐3 } is an orthogonal set of vectors in โ3 2. Definition 7.3.1. A nonsingular matrix A is called orthogonal if ๐ด−1 = ๐ด๐ . 1 0 0 Example 4. Let ๐ด = [0 1/√2 −1/√2]. 0 −1/√2 −1/√2 −1 Since ๐ด 1 = [0 0 0 0 1/√2 −1/√2] = ๐ด๐ then A is an orthogonal matrix. −1/√2 −1/√2 Theorem 7.3.3. The ๐ x ๐ matrix A is orthogonal if and only if the columns (and rows) of A form an orthonormal set of vectors in โ๐ . Example 5. Let A be the matrix of Example 4. The columns (and rows) of A are of unit length and are mutually orthogonal. Thus A is orthogonal. Theorem 7.3.4. If A is symmetric ๐ x ๐ matrix , then there exists an orthogonal matrix P such that ๐−1 ๐ด๐ = ๐๐ ๐ด๐ = ๐ท, a diagonal matrix, where the columns of P consist of the linearly independent eigenvectors of A and the diagonal elements of D are the eigenvalues of A associated with these eigenvectors. 0 −1 −1 Example 6. Let ๐ด = [−1 0 −1]. Find an orthogonal matrix P and a diagonal matrix D −1 −1 0 such that ๐ท = ๐−1 ๐ด๐ = ๐๐ ๐ด๐. Solution: The characteristic polynomial of A is ๐ฅ ๐(๐ฅ) = |1 1 1 1 ๐ฅ 1| = (๐ฅ − 1)2 (๐ฅ + 2) 1 ๐ฅ 141 The eigenvalues of A are ๐1 = 1, ๐2 = 1 and ๐3 = −2. To find the eigenvector associated with ๐1 = 1, we solve for the homogeneous system (๐1 ๐ผ3 − ๐ด)๐ = 0: 1 1 1 ๐ฅ1 0 [1 1 1] [๐ฅ2 ] = [0] 1 1 1 ๐ฅ3 0 (1) The solution to this system is (verify) −๐ − ๐ [ ๐ ] where ๐, ๐ ∈ โ ๐ Thus a basis for the solution space of (1) consists of the eigenvectors −1 −1 ๐1 = [ 1 ] and ๐2 = [ 0 ] 0 1 However, the two vectors are not orthogonal so we apply the Gram-Schmidt process to obtain an orthonormal basis for the solution space of (1). Let ๐1 = ๐1 = (−1, 1, 0). Then ๐2 โ ๐1 ๐2 = ๐2 − ( )๐ ๐1 โ ๐1 1 1 = (−1, 0, 1) − (−1, 1, 0) 2 1 1 = (− , − , 1) 2 2 Normalizing these two vectors we get ๐1 = ๐1 โ๐1 โ = 1 √2 (−1, 1, 0) 2 1 1 ๐ ๐2 = โ๐2 โ = (− 2 , − 2 , 1) √6 2 = (− 1 √6 ,− 1 , 2 ) √6 √6 142 Thus {๐1 , ๐2 } is an orthonormal basis of eigenvectors for the solution space of (1). Next we find the eigenvector associated with ๐3 = −2 by solving the homogeneous system (−2๐ผ3 − ๐ด)๐ = 0: −2 1 1 ๐ฅ1 0 [ 1 −2 1 ] [๐ฅ2 ] = [0] 1 1 −2 ๐ฅ3 0 (2) The solution to this system is (verify) ๐ [๐] where ๐ ∈ โ ๐ Thus a basis for the solution space of (2) consists of the eigenvector 1 ๐3 = [1] 1 Normalizing this we get ๐3 โ๐3 โ ๐3 = = 1 √3 =( 1 (1, 1, 1) , 1 , 1 ) √3 √3 √3 Note that ๐3 is orthogonal to both ๐1 and ๐2 thus {๐1 , ๐2 , ๐3 } is an orthonormal basis of โ3 consisting of eigenvectors of A. By Theorem 7.3.4, − ๐= 1 √2 1 √2 [ 0 1 1 √6 1 √3 1 √6 2 1 − − √6 1 − √2 1 and ๐ −1 = ๐๐ = − √6 √3 √3] 1 [ √3 1 √2 1 − √6 1 √3 0 2 √6 1 √ 3] 143 Hence ๐ท = ๐๐ ๐ด๐ 1 − √2 1 6 = − √ 1 [ √3 − 1 0 √2 1 − √6 1 √3 1 1 √2 1 √2 1 = − √6 − √6 2 0 2 [−1 √6 1 −1 √3] 2 0 2 √6 2 − √1 2 −1 −1 0 −1] −1 0 1 √2 [ − 1 √2 1 √2 [− √3 − √3 − √3] [ 0 0 1 1 √6 1 √3 1 √6 2 √3 1 − − √6 − √1 6 1 −√ 6 2 √6 1 √3 1 √3 1 √3] √3] 1 0 0 = [0 1 0 ] 0 0 −2 Note that D is a diagonal matrix whose main diagonal consists of the eigenvalues of A. SAQ 7-2 1 Let ๐ด = [0 0 0 0 3 −2]. Find an orthogonal matrix P and a −2 3 diagonal matrix D such that ๐ท = ๐−1 ๐ด๐ = ๐๐ ๐ด๐. 144 ASAQ 7-2 The characteristic polynomial of A is ๐ฅ−1 0 0 ๐(๐ฅ) = | 0 ๐ฅ−3 2 | 0 2 ๐ฅ−3 = (๐ฅ − 1)(๐ฅ − 3)(๐ฅ − 3) − 4(๐ฅ − 1) = (๐ฅ − 1)2 (๐ฅ − 5) The eigenvalues of A are ๐1 = 1, ๐2 = 1 and ๐3 = 5. To find the eigenvector associated with ๐1 = 1, we solve for the homogeneous system (๐1 ๐ผ3 − ๐ด)๐ = 0: 0 0 0 ๐ฅ1 0 [0 −2 2 ] [๐ฅ2 ] = [0] 0 2 −2 ๐ฅ3 0 (1) The solution to this system is (verify) ๐ [๐ ] where ๐, ๐ ∈ โ ๐ Thus a basis for the solution space of (1) consists of the eigenvectors 1 0 ๐1 = [0] and ๐2 = [1] 0 1 Note that ๐1 and ๐2 are already orthogonal. Next we normalize ๐2: 0 ๐2 ๐2 = = [1/√2] โ๐2 โ 1/√2 Thus {๐1 , ๐2 } is an orthonormal basis of eigenvectors for the solution space of (1). Next we find the eigenvector associated with ๐3 = 5 by solving the homogeneous system (5๐ผ3 − ๐ด)๐ = 0: 145 4 0 0 ๐ฅ1 0 [0 2 2] [๐ฅ2 ] = [0] 0 2 2 ๐ฅ3 0 (2) The solution to this system is (verify) 0 [−๐] ๐ ∈ โ ๐ Thus a basis for the solution space of (2) consists of the eigenvector 0 ๐3 = [−1] 1 Normalizing ๐3 we obtain 0 ๐3 ๐3 = = [−1/√2] โ๐3 โ 1/√2 Since ๐3 is orthogonal to both ๐1 and ๐2 then {๐1 , ๐2 , ๐3 } is an orthonormal basis of โ3 consisting of the eigenvectors of A. Thus 1 0 0 1 0 0 −1 ๐ ๐ = [0 1/√2 −1/√2] and ๐ = ๐ = [0 1/√2 1/√2] 0 1/√2 1/√2 0 −1/√2 1/√2 Hence 1 0 0 1 ๐ท = [0 1/√2 1/√2] [0 0 −1/√2 1/√2 0 1 = [0 0 0 0 1 1/√2 1/√2] [0 −5/√2 5/√2 0 1 0 0 = [0 1 0 ] 0 0 5 0 0 1 3 −2] [0 −2 3 0 0 0 1/√2 −1/√2] 1/√2 1/√2 0 0 1/√2 −1/√2] 1/√2 1/√2 146 ACTIVITY −3 0 1. Let ๐ด = [ 0 −2 −1 0 −1 0 ]. Find an orthogonal matrix P and a diagonal matrix D such −3 that ๐ท = ๐ −1 ๐ด๐ = ๐๐ ๐ด๐. REFERENCES: 1. Kolman. Bernard. Introductory Linear Algebra; 4th edition, New York: Mcmillan Publishing Company, 1988. 2. Lipschutz, Seymour. Linear Algebra; SI (Metric Edition), Singapore: McGraw-Hill International Book Company, 1981. 3. Johnson and Riess. Introduction to Linear Algebra; Phil: Addison-Wesley Publishing Company, Inc. 1981. 4. Perry, William L. Elementary Linear Algebra; USA: McGraw-Hill, Inc. 1988.