Inverse and Elementary Matrices 1. Inverse matrices Recall the economy problem: B" œ total production of coal B# œ total production of steel B$ œ total production of electricity Based on an economic model, have system of equations for B3 À B" Þ#B# Þ%B$ œ " (*) Þ"B" Þ*B# Þ#B$ œ Þ( Þ Þ"B" Þ#B# Þ*B$ œ #Þ* Ô B" × Recall: we defined \ œ B# à Õ B$ Ø Ô " Þ" E œ Õ Þ" Þ# Þ* Þ# Þ% × Þ# Þ* Ø Ô " × Þ( F œ Õ #Þ* Ø [known] [known] Then equation for unknown matrix \ is: E\ œ F [recall any system of equations can be written this way]. How to solve in general for \ ? Definition 1: An 8 ‚ 8 matrix is called the identity matrix M if it has 1's along the diagonal and zeroes elsewhere. Ô" Ö! Example 1: M œ % ‚ % identity matrix œ Ö ! Õ! ! " ! ! ! ! " ! !× !Ù Ù ! "Ø . [note size of the matrix is usually understood]. NOTE: 7 œ 8. Note: a b − ‘% À M#$ œ !à M$# œ !à M$$ œ ". Ô" Ö! Mb œ Ö ! Õ! ! " ! ! ! ! " ! That is, M78 œ ! unless ! ×Ô ," × Ô ," × ! ÙÖ ,# Ù Ö ,# Ù ÙÖ Ù œ Ö Ù œ bÞ ! ,$ ,$ ØÕ Ø Õ " ,% ,% Ø Similarly this holds for all sizes 8 of the identity. Proposition 1: If M is an 8 ‚ 8 identity matrix and a is an 8-vector, then M a œ aÞ Proof: Ô" Ö! Ö Ma œ Ö ! Ö ã Õ! ! " ! ã ! ! ! " ã ! á á ä á ! ×Ô +" × Ô +" × ! ÙÖ +# Ù Ö +# Ù ÙÖ Ù Ö Ù ! ÙÖ +$ Ù œ Ö +$ Ùß ÙÖ Ù Ö Ù ã ã ã " ØÕ +8 Ø Õ +8 Ø as seen directly by multiplication. Proposition 2: For any matrix E for which the product makes sense, M8 E œ EÞ Proof: ME œ MÒa" á a7 Ó œ ÒM a" á M a7 Ó œ Òa" á a7 Ó œ EÞ Definition 2: Given a square matrix E, a matrix F is called the inverse of E if EF œ M and FE œ MÞ A matrix for which an inverse exists is called invertible. Example 2: " E œ ” # # à "• " E" œ "$ ” # # Þ " • Theorem 2: If E has an inverse matrix F, then F is unique (i.e., there is no other inverse matrix). Proof: If F and G are both inverses of E then: F œ FM œ FÐEGÑ œ ÐFEÑG œ MG œ Gß so that any two inverses are the same, i.e., there is only one inverse. Theorem: Given square matrices E and F, if FE œ M , then EF œ M , i.e., F is automatically the inverse of E. (Proved later in this lecture - after invertible matrix theorem) Now: + Let's find inverse of E œ ” + ”- , : .• , B D " œ ” •” • . C A ! Get: +B ,C œ " -B .C œ ! B D Find \ œ ” C A• ! Þ "• s.t. +D ,A œ ! -D .A œ " Augmented matrix for the first pair: + ”- , . l l " !• , . l l ! "• Augmented matrix for second pair: + ”Solve in parallel: + ”- , . l l " ! ! "• Use row reduction: " ”- ,Î+ . l l Ä” Ä” " ! " ! ,Î+ " ! " l l œ” ! " Ä” • " ! "Î+ ! " ! l l ,Î+ Ð+. ,-ÑÎ+ "Î+ -ÎÐ+. ,-Ñ l l ! "• ! +ÎÐ+. ,-Ñ • "Î+ -,ÎÐ+Ð+. ,-ÑÑ -ÎÐ+. ,-Ñ ! " l "Î+ l -Î+ ,ÎÐ+. ,-Ñ +ÎÐ+. ,-Ñ • .ÎÐ+. ,-Ñ ,Ð+. ,-Ñ -ÎÐ+. ,-Ñ +Ð+. ,-Ñ • But note the entries in the first column on the right should be the solution for B and C: B .ÎÐ+. ,-Ñ ” C • œ ” -ÎÐ+. ,-Ñ •ß since the solution for B and C is obtained by manipulating the first column of the right hand side. And similarly the second column on the right should be D ,ÎÐ+. ,-Ñ ” A • œ ” +ÎÐ+. ,-Ñ • So the inverse is: B D .ÎÐ+. ,-Ñ E" œ F œ ” œ ” • C A -ÎÐ+. ,-Ñ œ Moral: to invert E œ ” + ”- , . l l " ! + - " . ” +. ,- ,ÎÐ+. ,-Ñ +ÎÐ+. ,-Ñ • , + • , ß row reduce .• ! " Ä” • " ! ! " l l B D œ ÒMlE" Ó (after row reduction) C A• that is, after row reduction the identity will be in the left block, and the desired inverse matrix appears in the right block. Thus: \E œ Mà can also check that E\ œ MÞ Check: + ”- , .• " +.-, . ” - , " œ ” • + ! ! Þ "• Thus we have a general formula for # ‚ # matrices. Note this is a general method Definition 3: A matrix which has an inverse is called nonsingular Theorem 3: If E and F are nonsingular then EF is also nonsingular and ÐEFÑ" œ F " E" Þ Pf: Note that ÐF " E" ÑÐEFÑ = ", so that ÐEFÑ" œ F " E" . Note that the other direction giving the identity can also be checked. [Note: this generalizes to larger products of matrices in the expected way; see corollary 1.1; you will do the proof in exercises.] [Note there are some other theorems in the book along same lines; proofs are simple] 2. Conclusion of the original matrix problem: Recall: We had B" Þ#B# Þ%B$ œ " Þ"B" Þ*B# Þ#B$ œ Þ( (*) Þ"B" Þ#B# Þ*B$ œ #Þ* Ô B" × Showed that if we defined \ œ B# à Õ B$ Ø Ô " Þ" E œ Õ Þ" Þ# Þ* Þ# Ô " × Þ( F œ Õ #Þ* Ø then to find \ we solve E\ œ Fß Þ% × Þ# Þ* Ø [known] [known], for the unknown matrix \ . How to do that? Suppose that matrix E has an inverse matrix E" : then E\ œ F Ê ÐE" EÑ\ œ E" F Ê \ œ E" F Thus we would have the unknown vector X as a product of two known matrices. How to find E" for a $ ‚ $ matrix? We'll find out later. Example (2 ‚ 2; a previous example revisited): Consider a metro area with constant population of 1 million. Area consists of city and suburbs. Analyze changing urban and suburban populations. Let G" and W" denote the two populations in the first year of our study, and in general G8 and W8 denote the populations in the 8>2 year of the study. Assume that each year 15% of people in city move to suburbs, and 10% of people in suburbs move to the city. Thus next year's city population will be .85 times this year's city population plus .10 times this year's suburban population: G8" œ Þ)& G8 Þ" W8 Similarly W8" œ Þ"& G8 .*! W8 Þ A) If the populations in year 2 were .5 and .5 million, what were they in year 1? Þ& œ Þ)& G" Þ" W" Þ& œ Þ"& G" .*! W" System of equations for G" and W" Þ Write as matrix: Þ)& Þ" G" Þ& ” Þ"& Þ* • ” W • = ” Þ& • " ðóóñóóò ï ï E \ F E" œ œ " Þ* ” Þ"& Þ* † Þ)& Þ" † Þ"& " Þ* ” Þ"& Þ(& Þ" "Þ# œ ” • Þ)& Þ# Þ" Þ Þ)& • Þ"$$$ "Þ"$$$ • Thus: "Þ# \ œ E" F œ ” Þ# Þ"$$$ Þ& Þ&$$% †” • œ ” • "Þ"$$$ Þ& Þ%''( • Which give original populations of the city and suburbs. 3. Elementary Matrices: Definition: An elementary matrix is a matrix obtained from the identity by a single elementary row operation. Ex 1: Elementary matrix of type I [interchange rows]: Ô! ! Õ" ! " ! "× ! !Ø Elementary matrix of type II [multiply row by a constant]: Ô" ! Õ! ! $ ! !× ! "Ø Elementary matrix of type III [add multiple of one row to another]: Ô" ! Õ$ !× ! "Ø ! " ! 4. Some theorems about elementary matrices: Note: now we will prove some theorems about elementary matrices; we will make them statements (most of which I will prove; will state when not proving them)Þ This is a story about elementary matrices we will be writing. Theorem 1: Let E be a matrix, and let F be the result of applying an ERO to E. Let I be the elementary matrix obtained from I by performing the same ERO. Then F œ IEÞ Proof is exercise. Ô" Example 1: Start with M œ ! Õ! to get the elementary matrix: Ô" Iœ ! Õ$ !× ! ß and add 3 † row 1 + row 3 "Ø ! " ! ! " ! !× ! Þ "Ø Now note that multiplying any matrix E by I performs the same operation on E: Ô" IE œ ! Õ$ ! " ! ! ×Ô # ! " ØÕ " # " " " !× Ô # # œ " $Ø Õ ) " " % !× # Þ $Ø Thus conclude: If have matrix E, and apply series of ERO's to E: E Ä E" Ä E# Ä E$ Ä á Ä E 8 Each ERO can be obtained by multiplying the previous matrix by an elementary matrix: E" œ I" E E# œ I# E " œ I # I " E E$ œ I$ E # œ I $ I # I " E ã (*) E8 œ I8 I8" á I# I" E Conclusion 1: If E8 is the result of applying a series of row operations to E, then E 8 can be expressed in terms of E as in (*) above. Note that every elementary row operation can be reversed by an elementary row operation of the same type. Since ERO's are equivalent to multiplying by elementary matrices, have parallel statement for elementary matrices: Theorem 2: Every elementary matrix has an inverse which is an elementary matrix of the same type. Proof: See book 5. More facts about matrices: Assume henceforth E is a square 8 ‚ 8 matrix. homogeneous system Î Ô B" × Ñ Ð ÖB Ù Ó : Ð E œ [E34 ]; \ œ Ö # Ù Ó ã Õ B% Ø Ò Ï Suppose we have E\ œ ! Þ Assume that only solution is \ œ ! (i.e. zero matrix for \ ). What can we conclude about E? Reduce E to reduced row echelon form: E Ä F Then F\ œ ! has the same solutions. But F has to have all non-zero rows, since if a row is 0 then there will be nontrivial solutions for \ (less equations than unknowns; see results from class last time). Therefore, F must be the identity. Conversely, if E is row equivalent to M , then E\ œ ! is equivalent to M\ œ !, so the only solution is \ œ !Þ We have just proved: Conclusion 2: E is row-equivalent to I, the identity matrix, 30 0 E\ œ ! has no nontrivial solutions. Example: " E œ ” 2 # % • " Ä ” ! # ! • System of equations: E\ œ ! Lots of solutions: B# œ B# B" œ #B# Þ Ä B" #B# œ ! !! œ ! B # Thus \ œ ” " • œ B# ” •ß B# arbitrary B# " solutions). (infinite number of Conclude: exist nontrivial solutions. Thus E is not row-eqivalent to M . [as can be checked directly] Again we assume here our matrices are square: 1. Nonsingular matrices are just products of elementary matrices Notice that if a matrix E is nonsingular (invertible), then if E\ œ !, then: E\ œ ! Ê E" E\ œ E" ! Ê M\ œ ! Ê \ œ !Þ Thus E\ œ ! has no nontrivial solutions. Thus E is row equivalent to I. Thus there exist elementary matrices I" ß á ß I5 such that: I5 I5" I5# á I# I" E œ M Ê E œ ÐI"" I#" á I5" ÑM œ I"" I#" á I5" Þ So E is a product of elementary matrices. Also, note that if E is a product of elementary matrices, then E is nonsingular since the product of nonsingular matrices is nonsingular. Thus Conclusion 3: E is a product of elementary matrices (note inverses of elementary matrices are also elementary) iff E is nonsingular, Now notice that: if E is nonsingular, then E œ I" I # á I 5 œ I " I # á I 5 M œ equivalent to MÞ a matrix row Can reverse the argument: if E row equivalent to M then E œ I" I# á I5 Mß so E is nonsingular. Conclusion 4: E 3s row equivalent to M iff E is nonsingular. Example 2: # E œ ” " " #• V" V # Í V" V # "Î# V# V" Í "Î#V# V" " ”" "Î#V" Í #V" " ”! "Î# $Î# • " ”! "Î# # • #Î$ V# Í $Î#V# " ”! "Î# " • ! . "• Thus E is equivalent to the identity, so by conclusion 4, we have that E is nonsingular. Thus E must be a product of elementary matrices. But note we can reverse the above process and go from I to E as shown above. # Thus E œ ” ! ! " • ” " " ! " • ” " ! ! " • ” $Î# ! "Î# " • ” " ! ! "• œ I" I # I $ I % M Thus since E is nonsingular it is indeed a product of elementary matrices. 2. Conclusions: We have shown the following are equivalent (if you follow through the arguments): 1. E is nonsingular 2. E\ œ ! has only trivial soln. 3. E is row equivalent to M 4. E\ œ F has a unique solution for every given column matrix F . 5. E is a product of elementary matrices. [Note that other equivalences are shown in section 2.2; these are the important ones for now] T <990 À We have showed 2, 3, are equivalent in Conclusion 2. 1 and 5 are equivalent by Conclusion 3. And 1 and 3 are equivalent by Conclusion 4. To show that % is equivalent to ", note that if 1 holds, then the equation E\ œ F implies E" E\ œ E" F which implies \ œ E" Fß which means \ has a unique solution, so that 4 holds. Conversely, if 4 holds then E\ œ F has a unique solution for each F. Setting F œ !, we get that E\ œ ! has only trivial solution, which shows E is nonsingular via condition 2 above. We thus have 2 Í 3 Í 1 Í 5, and that 1 Í 4. This shows that all statements are equivalent. 3. A useful Corollary: Corollary: Let E be a 8 ‚ 8 matrix. If the square matrix F satisfies FE œ M , Ð"Ñ EF œ M Ð#Ñ then automatically also i.e., F œ E" . Similarly, Ð#Ñ also implies Ð"ÑÞ Proof: Suppose Ð"Ñ holds. Then E\ œ ! Ê FE\ œ ! Ê \ œ ! Ê only trivial solution Ê E" exists. Also we have from Ð"Ñ: FEE" œ ME" Ê F œ E" Þ Now suppose Ð#Ñ holds. Then reversing the roles of E and F, it follows that (1) holdsÞ 4. Inverting a matrix: If E is invertible then E is row-equivalent to M , so I5 I5" á I" E œ Mß where the I5 are ERO which get M . Thus E" œ I5 I5" á I" Mß i.e. to get E" we need to perform the same elementary row operations on M that we perform on E to turn it into identity. Example 1: Find inverse of Eœ” # " " # • [note we already know how to invert this in another way] Inversion process: # ” " " # " Ä ” " " Ä ” ! " Ä ” ! " Ä ” ! ! "• l " l ! "Î# # l l "Î# ! ! "• "Î# $Î# l l "Î# "Î# ! "• "Î# " ! " l l l l "Î# "Î$ #Î$ "Î$ ! #Î$ • "Î$ #Î$ • So #Î$ E" œ ” "Î$ "Î$ Þ #Î$ • [Note that material below completes our coverage of material related to the inverse matrix theorem] 5. Inverses and linear transformations Recall if E is a nonsingular square matrix, and x is a vector, then we can define the function X ÐxÑ œ ExÞ Recall properties X Ðx yÑ œ EÐx yÑ œ Ex Ey œ X ÐxÑ X ÐyÑ X Ð- xÑ œ E- x œ -Ex œ -X ÐxÑÞ Thus X is a linear transformation. [recall X can be viewed as a rule for taking the vector x to Ex] We define the inverse X " of X as the operation which takes X x back to x. Thus X " ÐX ÐxÑÑ œ x [note this happens only if X is 1-1 and onto] Figure: E viewed as a linear transformation: Ex œ X ÐxÑÞ Note that X " inverts the operation. Note that E" ÐX xÑ œ E" ÐExÑ œ xÞ Thus it must be that X " ÐX xÑ œ E" ÐX xÑ so that X " corresponds to multiplication by E" . [this is stated as a theorem in the text]