Solutions 1. 1] Solution: Show "ab "a b from the field axioms. You need to show that "a b acts like the additive inverse of ab because then, it follows that it is the additive inverse. However, ab "a b a "a b 0b 0 and so "ab "a b. 2. 2] Solution: Show the maximum (minimum) of ax 2 bx c, a p 0, always occurs when x "b/2a . Then you get this equals ax 2 ba x ac . It equals 2 2 a x 2 ba x b 2 " b 2 ac 4a 4a 2 2 a x b 4ac "2 b 2a 4a b Thus a minimum occurs if a 0 and it occurs at x " 2a while if a 0, you get a maximum at the same value of x. 3. 3] Solution: Let ft t 2 4t for t u "2. Find a formula for f "1 y . Answer: Let y x 2 4x and solve for x f "1 y . "4 o 4y 16 2 Here y u "4 and you also need x u "2. Therefore, you need to use the sign. Hence x f "1 y y4 "2 4. 4] Solution: Suppose fx, y ax 2 by 2 2cxy and suppose a b 0 and ab " c 2 0. Show that under these conditions, the above function is nonnegative. This is a function of two variables. You can pick x, y to be any numbers and the above expression under the given conditions is always nonnegative. You just complete the square. Assume first that a 0. The function equals 2cxy c2y2 c2y2 by 2 a a2 " a2 cy 2 c 2 y 2 a x a " 2 by 2 a 2 cy 2 a x a by 2 " ca y 2 cy 2 1 2 a x a a y ab " c 2 u 0 by assumption. If a 0, then b 0 since their sum is positive. Then a similar argument works. a x2 2cxy a by 2 a x 2 You first complete the square on y and then check what is left over. 5. 5] Solution: Start with "b o b 2 " 4ca 2a and show directly that x given by either of the two numbers in the above formula satisfies the quadratic equation ax 2 bx c 0. x a "b b 2 "4ca 2a a 1 2a 2 b2 " 2 1 a b c" "b b 2 "4ca 2a 1 2a 2 c b b 2 " 4ac b "b b 2 "4ca 2a c 0 The other case is similar. 6. 6] Solution: Suppose there are two real solutions to ax 2 bx c 0, x 1 , x 2 . Show the maximum (minimum), depending on sign of a, of fx ax 2 bx c occurs at x 1 x 2 /2. When you complete the square, you obtain the polynomial equals 2 2 " b " 4ac a x b 2a 4a 2 You are also given that x 1 and x 2 make this expression equal to 0. Therefore, the two solutions to that equation are the two numbers given below. "b o b 2 " 4ac 2a It follows that if these are added and the result divided by 2 you obtain "2b 1 " 1 b 2a 2 2a which, from the above expression is the correct value to maximize or minimize the polynomial. 7. 7] Solution: For z equal to the following complex number, find z "1 . a. 2 5i 2 5 " 29 i 29 b. 9 " 9i 1 1 18 i 18 c. "5 2i 5 2 " 29 " 29 i d. 9 4i 9 4 " 97 i 97 8. 8] Solution: Let z 3 8i and let w 6 " 8i. Find zw, z w, z 2 , and w/z. " 72 i 82 24i, 9, "55 48i, " 46 73 73 9. 9] Solution: Give the complete solution to x 4 16 0. Graph these points in the complex plane. 1 i 2 , "1 " i 2 , "1 i 2 , 1 " i 2 y 4 2 -4 -2 2 4 x -2 -4 10. 10] Solution: Give the complete solution to x 3 64 0. Graph these points in the complex plane. "4, 2 " 2i 3 , 2i 3 2 y 4 2 -4 -2 2 4 x -2 -4 11. 11] Solution: If z is a complex number, show there exists F a complex number with |F| 1 and Fz |z|. If z 0, let F 1. If z p 0, let F |z|z 12. 12] Solution: De Moivre’s theorem says ¡rcos t i sin t ¢ n r n cos nt i sin nt for n a positive integer. Does this formula continue to hold for all integers, n, even negative integers? Explain. Yes, it holds for all integers. First of all, it clearly holds if n 0. Suppose now that n is a negative integer. Then "n 0 and so 1 1 "n ¡rcos t i sin t ¢ n r cos"nt i sin"nt ¡rcos t i sin t ¢ "n r n cosnt i sinnt rn cosnt " i sinnt cosnt " i sinnt cosnt i sinnt r n cosnt i sinnt because cosnt " i sinnt cosnt i sinnt 1. 13. 13] Solution: You already know formulas for cosx y and sinx y and these were used to prove De Moivre’s theorem. Now using De Moivre’s theorem, derive a formula for sin4x and one for cos4x . cos 4 x 4i cos 3 x sin x " 6 cos 2 x sin 2 x " 4i cos x sin 3 x sin 4 x The real part is cos4x and the imaginary part is sin4x . 14. 14] Solution: If z and w are two complex numbers and the polar form of z involves the angle 2 while the polar form of w involves the angle C, show that in the polar form for zw the angle involved is 2 C. Also, show that in the polar form of a complex number, z, r |z|. You have z |z|cos 2 i sin 2 and w |w|cos C i sin C . Then when you multiply these, you get |z||w|cos 2 i sin 2 cos C i sin C |z||w|cos 2 cos C " sin 2 sin C icos 2 sin C cos C sin 2 |z||w|cos2 C i sin2 C 15. 15] Solution: Factor x 3 64 as a product of linear factors. x 4 x 2i 3 " 2 x " 2i 3 " 2 16. 16] Solution: Write x 3 27 in the form x 3 x 2 ax b where x 2 ax b cannot be factored any more using only real numbers. x 3 x 2 " 3x 9 17. 17] Solution: Completely factor x 4 256 as a product of linear factors. x 2 2i 2 x 2 " 2i 2 x " 2 " 2i 2 x " 2 2i 2 18. 18] Solution: Completely factor x 4 81 as a product of of two quadratic polynomials each of which cannot be factored further without using complex numbers. x2 3 2 x 9 x2 " 3 2 x 9 19. 19] Solution: If z, w are complex numbers prove zw z w and then show by induction that z 1 Cz m z 1 Cz m . m m Also verify that ! k1 z k ! k1 z k . In words this says the conjugate of a product equals the product of the conjugates and the conjugate of a sum equals the sum of the conjugates. a ib c id ac " bd iad bc ac " bd " iad bc a " ib c " id ac " bd " iad bc which is the same thing. Thus it holds for a product of two complex numbers. Now suppose you have that it is true for the product of n complex numbers. Then z 1 Cz n1 z 1 Cz n z n1 and now, by induction this equals z 1 Cz n z n1 As to sums, this is even easier. n n n j1 j1 j1 !x j iy j ! x j i ! y j n n n n j1 j1 j1 j1 ! x j " i ! y j ! x j " iy j ! x j iy j . 20. 20] Solution: Suppose px a n x n a n"1 x n"1 C a 1 x a 0 where all the a k are real numbers. Suppose also that pz 0 for some z C. Show it follows that p z 0 also. You just use the above problem. If pz 0, then you have pz 0 a n z n a n"1 z n"1 C a 1 z a 0 a n z n a n"1 z n"1 C a 1 z a 0 a n z n a n"1 z n"1 C a 1 z a 0 a n z n a n"1 z n"1 C a 1 z a 0 p z 21. 21] Solution: Show that "2 5i and 9 " 10i are the only two zeros to the polynomial px x 2 "7 5i x 32 65i Thus the zeros do not necessarily come in conjugate pairs if the coefficients of the polynomial are not real. x " "2 5i x " 9 " 10i x 2 "7 5i x 32 65i 22. 22] Solution: I claim that 1 "1. Here is why. " 1 i2 "1 "1 "1 2 1 1. This is clearly a remarkable result but is there something wrong with it? If so, what is wrong? Something is wrong. There is no single "1 . 23. 23] Solution: De Moivre’s theorem is really a grand thing. I plan to use it now for rational exponents, not just integers. 1 1 1/4 cos 2= i sin 2= 1/4 cos=/2 i sin=/2 i. Therefore, squaring both sides it follows 1 "1 as in the previous problem. What does this tell you about De Moivre’s theorem? Is there a profound difference between raising numbers to integer powers and raising numbers to non integer powers? It doesn’t work. This is because there are four fourth roots of 1. 24. 24] Solution: If n is an integer, is it always true that cos 2 " i sin 2 n cosn2 " i sinn2 ? Explain. Yes, this is true. cos 2 " i sin 2 n cos"2 i sin"2 n cos"n2 i sin"n2 cosn2 " i sinn2 25. 25] Solution: Suppose you have any polynomial in cos 2 and sin 2. By this I mean an expression of the form ! m)0 ! n*0 a )* cos ) 2 sin * 2 where a )* C. Can this always be written in the form nm b cos +2 ! A"nm c A sin A2? Explain. ! mn +"nm + Yes it can. It follows from the identities for the sine and cosine of the sum and difference of angles that sin a sin b 1 cosa " b " cosa b 2 1 cos a cos b cosa b cosa " b 2 sin a cos b 1 sina b sina " b 2 Now cos 2 1 cos 2 0 sin 2 and sin 2 0 cos 2 1 sin 2. Suppose that whenever k t n, k cos k 2 ! a j cosj2 b j sinj2 j"k for some numbers a j , b j . Then n cos n1 2 ! a j cos2 cosj2 b j cos2 sinj2 j"n Now use the above identities to write all products as sums of sines and cosines of j " 1 2, j2, j 1 2. Then adjusting the constants, it follows n1 cos n1 2 ! a Uj cos2 cosj2 b Uj cos2 sinj2 j"n1 Now use the above formulas to get the thing you need. You can do something similar with sin n 2 and with products of the form cos ) 2 sin * 2. 26. 26] Solution: Suppose px a n x n a n"1 x n"1 C a 1 x a 0 is a polynomial and it has n zeros, z 1 , z 2 , C, z n listed according to multiplicity. (z is a root of multiplicity m if the polynomial fx x " z m divides px but x " z fx does not.) Show that px a n x " z 1 x " z 2 Cx " z n . px x " z 1 qx rx where rx is a nonzero constant or equal to 0. However, rz 1 0 and so rx 0. Now do to qx what was done to px and continue until the degree of the resulting qx equals 0. Then you have the above factorization. 27. 27] Solution: Give the roots of the following quadratic polynomials having real coefficients. a. x 2 " 6x 25, 3 4i, 3 " 4i b. 3x 2 " 18x 75, 3 4i, 3 " 4i c. 9x 2 " 24x 32, 43 43 i, 43 " 43 i d. 16x 2 " 24x 25, 34 i, 34 " i e. 16x 2 " 24x 25, 34 i, 34 " i 28. 28] Solution: Give the solutions to the following quadratic equations having possibly complex coefficients. a. x 2 " 6 " i x 39 " 3i 0, 3 5i, 3 " 6i b. x 2 " 6 " i x 39 " 3i 0, 3 5i, 3 " 6i c. 9x 2 " 18 " 3i x 21 " 3i 0, 1 i, 1 " 43 i d. 9x 2 21 ix 9 0, 34 i, " 43 i 4 Fn 1. 29] Solution: Find 89, 1, "6, "2 9"1, "5, 6, 8 . 63 "37 6 56 2. 30] Solution: Find parametric equations for the line through 5, 7, 0 and "4, "4, 3 . Recall that to find the parametric line through two points a, b, c and d, e, f , you find a direction vector of the form d " a, e " b, f " c and then the equation of the line is x, y, z a, b, c td " a, e " b, f " c where t R. You should check that this works in the sense that when t 0, you are at a, b, c and when t 1, the value is d, e, f . 5 7 0 t "9 "11 3 3. 31] Solution: Find parametric equations for the line through the point "7, 7, 0 with a direction vector 1, "5, "3 . Recall from calculus, the equation of a line through a, b, c having direction vector f, g, h is given by x, y, z a, b, c tf, g, h . t " 7 7 " 5t "3t 4. 32] Solution: Parametric equations of the line are x, y, z 2t " 6 3t 6 3t 3 . Find a direction vector for the line and a point on the line. Recall that to find a direction vector, you could subtract one point on the line from the other or else just recall that a parametric equation of a line is of the form Point on the line t direction vector and use this in the above. "6 6 3 is a point. 2 3 3 is a direction vector. 5. 33] Solution: "1 9 3 , 1 6 3 . Recall that to find the parametric line through two points a, b, c and d, e, f , you find a direction vector of the form d " a, e " b, f " c and then the equation of the line is x, y, z a, b, c td " a, e " b, f " c where t R. You should check that this works in the sense that when t 0, you are at a, b, c and when t 1, the value is d, e, f . Find parametric equations for the line through the two points 2t " 1 9 " 3t 3 6. 34] Solution: The equation of a line in two dimensions is written as y 2x " 2. Find parametric equations for this line. x t, y 2t " 2 and symmetric equations x 12 y 1. 7. 35] Solution: Find 53, "6, 0, 5 32, "6, 0, 9 . 21 "48 0 52 8. 36] Solution: Find parametric equations for the line through 2, 4, "1 and 1, 7, "6 . Recall that to find the parametric line through two points a, b, c and d, e, f , you find a direction vector of the form d " a, e " b, f " c and then the equation of the line is x, y, z a, b, c td " a, e " b, f " c where t R. You should check that this works in the sense that when t 0, you are at a, b, c and when t 1, the value is d, e, f . t 2 4 " 3t 5t " 1 9. 37] Solution: Find parametric equations for the line through the point "7, 2, "4 with a direction vector 1, 5, "6 . t " 7 5t 2 "6t " 4 10. 38] Solution: Parametric equations of the line are x, y, z 2t 3 6t 8 3t 6 . Find a direction vector for the line and a point on the line. 3 8 6 is a point. 2 6 3 is a direction vector. 11. 39] Solution: 0 4 5 , 2 "5 "5 . Recall that to find the parametric line through two points a, b, c and d, e, f , you find a direction vector of the form d " a, e " b, f " c and then the equation of the line is x, y, z a, b, c td " a, e " b, f " c where t R. You should check that this works in the sense that when t 0, you are at a, b, c and when t 1, the value is d, e, f . Find parametric equations for the line through the two points 2t 4 " 9t 5 " 10t 12. 40] Solution: Find the point on the line segment from from "5 5 4 to 1 "6 "5 "5 5 4 to 1 "6 "5 which is 1 5 of the way . " 15 " 19 " 16 5 5 13. 41] Solution: P P The point is P 1 P32 P 3 because this point is of the form P i 13 j 2 k 23 whenever i, j, k are distinct. The last expression is a point on the line between the midpoint of P j and P k and the point P i . This point is called the barycenter. 14. 42] Solution: The wind blows from the South at 50 kilometers per hour and an airplane which flies at 300 kilometers per hour in still air is heading East. What is the actual velocity of the airplane and where will it be after two hours. 300 50 and after two hours it is at . 600 100 15. 43] Solution: The wind blows from the West at 50 kilometers per hour and an airplane which flies at 500 kilometers per hour in still air is heading North East. What is the actual velocity of the airplane and where will it be after two hours. 250 2 50 250 2 it will be at 500 2 100 500 2 16. 44] Solution: The wind blows from the North at 20 kilometers per hour and an airplane which flies at 300 kilometers per hour in still air is supposed to go to the point whose coordinates are at 200 200 kilometers. In what direction should the airplane head? Let its direction be p, q . Then after much fuss, q need to solve the equations 449 1 30 1 30 ,p 1 30 449 " 1 30 . You 300p 300q " 20 p2 q2 1 where also p, q u 0. 17. 45] Solution: Three forces act on an object. Two are "7 "5 1 2 "1 "6 and Newtons. Find the third force if the object is not to move. 5 6 5 18. 46] Solution: Three forces act on an object. Two are 7 2 "1 third force if the total force on the object is to be and 1 "6 "5 "6 "6 "5 Newtons. Find the . "14 "2 1 19. 47] Solution: A river flows West at the rate of b miles per hour. A boat can move at the rate of 1 miles per hour. Find the smallest value of b such that it is not possible for the boat to proceed directly across the river. b 1. 20. 48] Solution: The wind blows from West to East at a speed of 20 miles per hour and an airplane which travels at 300 miles per hour in still air is heading North West. What is the velocity of the airplane relative to the ground? What is the component of this velocity in the direction North? The velocity is the sum of two vectors. 20 " 150 2 150 2 In the direction of North, the component is 150 2 . 21. 49] Solution: The wind blows from West to East at a speed of 60 miles per hour and an airplane can travel travels at 400 miles per hour in still air. How many degrees West of North should the airplane head in order to travel exactly North? Let a, b be a unit vector in the appropriate direction. Then you need 400a, b 601, 0 to have direction vector 0, 1 . Thus you need a400 60 0 and so a "60 400 and so b 1" "60 400 2 1 391 20 Thus the appropriate direction vector is "60 , 1 391 400 20 To find the angle you would need to find the cosine of the angle between 0, 1 and this vector. This is 1 391 20 and so the radian measure is 0. 150 57 Then this needs to be changed to degrees. 8. 626 9 22. 50] Solution: The wind blows from West to East at a speed of 60 miles per hour and an airplane which travels at 300 miles per hour in still air heading somewhat West of North so that, with the wind, it is flying due North. It uses 40. 0 gallons of gas every hour. It has to travel 200. 0 miles, how much gas will it use in flying to its destination? As above, the velocity of the plane is "60. 0 293. 94 and then there is the wind which gives a velocity vector of 60. 0 0 293. 94. How many hours? . Therefore, the component in the direction North is just 200. 0 0. 680 41 293. 94 How much gas? 40. 0 0. 680 41 27. 217 23. 51] Solution: An airplane is flying due north at 100. 0 miles per hour but it is not actually going due North because there is a wind which is pushing the airplane due east at 20. 0 miles per hour. After 1 hour, the plane starts flying 30 ( East of North. Assuming the plane starts at 0, 0 , where is it after 2 hours? Let North be the direction of the positive y axis and let East be the direction of the positive x axis. Where is it after one hour? 20. 0 100. 0 or in other words, in the direction of 1 2 , 3 2 . Now from here it heads 20 degrees east of North . Thus the new velocity vector is 70. 0 86. 603 Then after another hour, it will have gone this vector added to the first. Thus its position then is 90. 0 186. 6 24. 52] Solution: City A is located at the origin while city B is located at 300, 500 where distances are in miles. An airplane flies at 250 miles per hour in still air. This airplane wants to fly from city A to city B but the wind is blowing in the direction of the positive y axis at a speed of 50 miles per hour. Find a unit vector such that if the plane heads in this direction, it will end up at city B having flown the shortest possible distance. How long will it take to get there? Wind: 0, 50 . Direction it needs to travel: 3, 5 1 . Then you need 250a, b 0, 50 to have 34 this direction where a, b is an appropriate unit vector. Thus you need a2 b2 1 250b 50 5 250a 3 3 4 Thus a 5 , b 5 . Then the velocity of the plane relative to the ground is 150 250 . The speed of the plane relative to the ground is then 150 2 250 2 291. 55. It has to go a distance of 300 2 500 2 583. 10 miles. Therefore, it takes 583. 1 2 hours 291. 55 25. 53] Solution: A certain river is one half mile wide with a current flowing at 2. 0 miles per hour from East to West. A man takes a boat directly toward the opposite shore from the South bank of the river at a speed of 6. 0 miles per hour. How far down the river does he find himself when he has swam across? How far does he end up travelling? Water: "2. 0, 0 Boat: 0, 6. 0 . Resultant velocity: "2. 0 6. 0 Time taken to get across: 8. 333 3 10 "2 . How far down the river: "0. 166 67. Distance travelled: 0. 527 05. 26. 54] Solution: A certain river is one half mile wide with a current flowing at 2 miles per hour from East to West. A man can swim at 3 miles per hour in still water. In what direction should he swim in order to travel directly across the river? What would the answer to this problem be if the river flowed at 3 miles per hour and the man could swim only at the rate of 2 miles per hour? Man: 3a, b Water: "2, 0 . Then you need 3a 2 and so a 2/3 and hence b 5 /3. The 5 vector is then 23 , 3 . In the second case, he could not do it. You would need to have a unit vector a, b such that 2a 3. This is not possible, not even if you try real hard. 27. 55] Solution: Three forces are applied to a point which does not move. Two of the forces are 3. 0i 3. 0j 8. 0k Newtons and "7. 0i 8. 0j 3. 0k Newtons. Find the third force. x " 4 y 11 z 11 needs to equal the zero vector. You need 4 "11 "11 28. 56] Solution: The total force acting on an object is to be 1 i 4 j 3 k Newtons. A force of "8 i "2 j 8 k Newtons is being applied. What other force should be applied to achieve the desired total force? Need 1 4 3 x y z "8 "2 8 x y z Thus 9 6 "5 29. 57] Solution: A bird flies from its nest 8 km. in the direction 32 = north of east where it stops to rest on a tree. It then flies 5 km. in the direction due southeast and lands atop a telephone pole. Place an xy coordinate system so that the origin is the bird’s nest, and the positive x axis points east and the positive y axis points north. Find the displacement vector from the nest to the telephone pole. To the tree: To pole: 0 "8 " 52 2 " 52 2 Total displacement: " 52 2 " 52 2 " 8 30. 58] Solution: A car is stuck in the mud. There is a cable stretched tightly from this car to a tree which is 20 feet long. A person grasps the cable in the middle and pulls with a force of 150 pounds perpendicular to the stretched cable. The center of the cable moves 3 feet and remains still. What is the tension in the cable? The tension in the cable is the force exerted on this point by the part of the cable nearer the car as well as the force exerted on this point by the part of the cable nearer the tree. Assume the cable cannot be lengthened so the car moves slightly in moving the center of the cable. 3 2T cos ) 150 where cos ) 10 75 T cos) 250 31. 59] Solution: A car is stuck in the mud. There is a cable stretched tightly from this car to a tree which is 20 feet long. A person grasps the cable in the middle and pulls with a force of 80 pounds perpendicular to the stretched cable. The center of the cable moves 3 feet and remains still. What is the tension in the cable? The tension in the cable is the force exerted on this point by the part of the cable nearer the car as well as the force exerted on this point by the part of the cable nearer the tree. Assume the cable does lengthen and the car does not move at all. 3 2T cos ) 80 where cos ) 109 109 40 40 T cos) 3 109 32. 60] Solution: A box is placed on a strong plank of wood and then one end of the plank of wood is gradually raised. It is found that the box does not move till the angle of inclination of the wood equals 60 degrees at which angle the box begins to slide. What is the coefficient of static friction? You need to have 6wcos 60 wsin 60 and so tan 60 6. 33. 61] Solution: Recall that the open ball having center at a and radius r is given by Ba, r q £x : |x " a| r¤ Show that if y Ba, r , then there exists a positive number - such that By, - Ba, r . The symbol means that every point in By, - is also in Ba, r . In words, it states that By, - is contained in Ba, r . When you have done this, you will have shown that an open ball is open. This is a fantastically important observation although its major implications will not be explored very much in this book. You let - r " |y " a|. Then if z By, - , it follows from the triangle inequality that |z " a| t |z " y| |y " a| - |y " a| r " |y " a| |y " a| r Thus, z Ba, r . Dot Product 1. 62] Solution: Find the following dot products. a. 1, 4, 3, 10 2, "1, "1, 3 25 b. 2, 7, 5 "1, "3, 2 "13 c. 4, 2, 5, 6, 7 "1, 1, "1, 2, 0 5 2. 63] Solution: Use the geometric description of the dot product to verify the Cauchy Schwartz inequality and to show that equality occurs if and only if one of the vectors is a scalar multiple of the other. That formula said that a b |a||b|cos 2 where 2 is the included angle. Therefore, |a b| ||a||b|cos 2| |a||b||cos 2| t |a||b|. 3. 64] Solution: For u, v vectors in R 3 , define the product, u ' v q u 1 v 1 2u 2 v 2 3u 3 v 3 . Show the axioms for a dot product all hold for this funny product. Prove |u ' v| t u ' u 1/2 v ' v 1/2 . Hint: Do not try to do this with methods from trigonometry. These axioms are all obvious. Therefore, since the proof of the Cauchy Schwarz inequality holds for an arbitrary product satisfying those axioms, the Cauchy Schwarz inequality holds. 4. 65] Solution: Find the angle between the vectors 4i " j " k and i 8j 2k. 4 "1 "1 2 arccos 1 8 2 18 69 1 " 69 2 1 " 69 2 69 1. 741 9 69 5. 66] Solution: Find the angle between the vectors 4i " j " k and i 6j 2k. 4 "1 "1 2 arccos 18 41 2 " 123 1 6 2 2 " 123 2 41 1. 718 6 2 41 6. 67] Solution: Find proj u v where v 6. 0, 2. 0, "2 and u 1, 7. 0, 3 . 1,7.0,3 6. 0, 2. 0, "2 1,7.0,3 14 59 1,7.0,3 59 59 7. 68] Solution: 59 59 Find proj u v where v 3. 0, 3. 0, "2 and u 1, 7. 0, 3 . 1,7.0,3 1,7.0,3 1,7.0,3 3. 0, 3. 0, "2 18 59 59 59 59 59 8. 69] Solution: Find proj u v where v 2. 0, "3. 0, "2 and u 1, 5. 0, 3 . 1,5.0,3 2. 0, "3. 0, "2 1,5.0,3 " 19 35 1,5.0,3 35 35 35 35 9. 70] Solution: A boy drags a sled for 40 feet along the ground by pulling on a rope which is 45 degrees from the horizontal with a force of 10 pounds. How much work does this force do? 40 10 cos 14 = 17997. 10. 71] Solution: A dog drags a sled for 80 feet along the ground by pulling on a rope which is 30 degrees from the horizontal with a force of 20 pounds. How much work does this force do? 80 20 cos 16 = 47993. 11. 72] Solution: A boy drags a sled for 70 feet along the ground by pulling on a rope which is 20 degrees from the horizontal with a force of 60 pounds. How much work does this force do? 70 60 cos 19 = 83987. 12. 73] Solution: How much work in Newton meters does it take to slide a crate 80. 0 meters along a loading dock by pulling on it with a 400. 0 Newton force at an angle of 20. 0 degrees from the horizontal? 80. 0400. 0 cos0. 349 07 6. 399 10 5 13. 74] Solution: How much work in Newton meters does it take to slide a crate 40. 0 meters along a loading dock by pulling on it with a 100. 0 Newton force at an angle of 45. 0 degrees from the horizontal? 40. 0100. 0 cos0. 785 40 1. 799 7 10 5 14. 75] Solution: An object moves 11. 0 meters in the direction of j i. There are two forces acting on this object, F 1 i j 1. 0k, and F 2 "5i 7. 0 j "6k. Find the total work done on the object by the two forces. Hint: You can take the work done by the resultant of the two forces or you can add the work done by each force. Why? 11.0 1, 1, 0 "4, 8. 0, "5. 0 12 2 11. 04. 0 31. 113 2 15. 76] Solution: An object moves 10. 0 meters in the direction of j i. There are two forces acting on this object, F 1 i j 5. 0k, and F 2 "5i 1. 0 j "6k. Find the total work done on the object by the two forces. Hint: You can take the work done by the resultant of the two forces or you can add the work done by each force. Why? 10.0 1, 1, 0 "4, 2. 0, "1. 0 12 2 10. 0"2. 0 "14. 142 2 16. 77] Solution: An object moves 7. 0 meters in the direction of j i. There are two forces acting on this object, F 1 i j 5. 0k, and F 2 "5i 2. 0 j "6k. Find the total work done on the object by the two forces. Hint: You can take the work done by the resultant of the two forces or you can add the work done by each force. Why? 7.0 1, 1, 0 "4, 3. 0, "1. 0 12 2 7. 0"1. 0 "4. 949 7 2 17. 78] Solution: An object moves 6. 0 meters in the direction of j i " k. There are two forces acting on this object, F 1 i j 6. 0k, and F 2 "5i 6. 0 j "6k. Find the total work done on the object by the two forces. Hint: You can take the work done by the resultant of the two forces or you can add the work done by each force. Why? 6.0 1, 1, "1 "4, 7. 0, 0. 0 10. 392 3 18. 79] Solution: Does it make sense to speak of proj 0 v ? No. The 0 vector has no direction and the formula doesn’t make sense either. 19. 80] Solution: If F is a force and D is a vector, show proj D F |F|cos 2 u where u is the unit vector in the direction of D, u D/|D| and 2 is the included angle between the two vectors, F and D. |F|cos 2 is sometimes called the component of the force, F in the direction, D. D D 1 FD u |F||D| |D| proj D F F |D| cos 2 u |F|cos 2 u |D| |D| 20. 81] Solution: If a, b, and c are vectors. Show that b c µ b µ c µ where b µ b " proj a b . b c a a |a|2 b " b 2a a c " c 2a a |a| |a| bµ cµ b c µ b c " 21. 82] Solution: Show that a b 14 ¡|a b|2 " |a " b|2 ¢. 1 ¡|a b|2 " |a " b|2 ¢ 14 ¡a b a b " a " b a " b ¢ 4 14 a a " a a b b " b b 4a b a b. 22. 83] Solution: Prove from the axioms of the dot product the parallelogram identity, |a b|2 |a " b|2 2|a|2 2|b|2 . Explain what this says geometrically in terms of the diagonals and sides of a parallelogram. The left equals |a|2 |b|2 2a b |a|2 |b|2 " 2a b 2|a|2 2|b|2 This follows right away from the properties of the dot product. It says that the sum of the squares of the diagonals of a parallelogram equals the sum of the squares of the sides. 23. 84] Solution: Prove the Cauchy Schwarz inequality in R n as follows. For u, v vectors, note that u " proj v u u " proj v u u 0 Now simplify using the axioms of the dot product and then put in the formula for the projection. Show that you get equality in the Cauchy Schwarz inequality if and only if the above equals zero which means that u proj v u. What is the geometric significance of this condition? u " u 2v v |v| u " u 2v v |v| |u|2 " 2u v 2 1 2 u v 2 1 2 u 0 |v| |v| And so |u|2 |v|2 u u v 2 You get equality exactly when u proj v u uv2 v, in other words, when u is a multiple of v. |v| 24. 85] Solution: For C n , the dot product is often called the inner product. It was given by n x y x, y q ! xjyj j1 Why place the conjugate on the y j ? Why is it not possible to simply define this the usual way? It is because if you did, you would not necessarily have x, x u 0. For example, suppose x i, 1 i . Then if you did the usual thing you would get "1 1 i 2 " 1 2i which is not nonnegative. Also, you could have x, x 0 even though x p 0. For example, i, 1 . 25. 86] Solution: For the complex inner product, prove the parallelogram identity. |x y|2 |x " y|2 2|x|2 2|y|2 Start with the left. From axioms, |x y|2 |x " y|2 |x|2 |y|2 2 Rex, y |x|2 |y|2 " 2 Rex, y 2|x|2 2|y|2 26. 87] Solution: For the complex inner product, show that if Rex, y 0 for all y C n , then x 0. In fact, show that for all y, you have x, y 0. x, y C and so there exists a complex number 2 such that |2| 1 and 2x, y |x, y |. You x,y can take 2 |x,y | if x, y p 0. Of course if this equals zero, there is nothing to show. Then from the axioms 0 Rex, 2 y Re 2x, y Re|x, y | |x, y |. Letting y x, shows that x 0. 27. 88] Solution: For the complex inner product, show that for x, y Rex, y i Imx, y , it follows that Imx, y Rex, iy so that x, y Rex, y i Rex, iy which shows that the inner product is determined by the real part. Rex, iy i Imx, iy q x, iy "ix, y " i¡Rex, y i Imx, y ¢ and so, equating real parts, Rex, iy Imx, y Therefore, x, y Rex, y i Imx, y Rex, y i Rex, iy Cross Product 1. 89] Solution: Show that if a u 0 for all unit vectors, u, then a 0. If a p 0, then the condition says that |a u||a|sin 2 0 for all angles 2. Hence a 0 after all. 2. 90] Solution: Find the area of the triangle determined by the three points, 0, 1, "3 , 2, 2, 0 and "1, 1, "1 . 2 "7 1 is the cross product. Area is then 3 2 6 3. 91] Solution: Find the area of the triangle determined by the three points, 3, "2, 2 , 4, 2, 0 and "1, 1, 2 . 6 8 19 is the cross product. Area is then 1 2 461 4. 92] Solution: Find the area of the triangle determined by the three points, 3, 2, 0 , 4, 3, 1 and 2, 1, "1 . The area is 0. 5. 93] Solution: Find the area of the parallelogram determined by the two vectors, 3, 0, 2 , 0, "2, 0 . 4 0 "6 is the cross product. The area is then 2 13 6. 94] Solution: Find the area of the parallelogram determined by the two vectors, "2, 2, "3 , 1, 5, 0 . 15 "3 "12 is the cross product. The area is then 3 3 14 7. 95] Solution: Find the volume of the parallelepiped determined by the vectors, "2 "3 "3 , 4 2 1 , 1 1 "3 . "2 "3 "3 4 2 1 1 1 "3 "31 8. 96] Solution: Find the volume of the parallelepiped determined by the vectors, 3 "1 3 , 1 "3 0 , 1 1 "1 . 3 "1 3 1 "3 0 1 1 "1 20 9. 97] Solution: Suppose a, b, and c are three vectors whose components are all integers. Can you conclude the volume of the parallelepiped determined from these three vectors will always be an integer? Yes. It will involve the sum of product of integers and so it will be an integer. 10. 98] Solution: What does it mean geometrically if the box product of three vectors gives zero? It means that if you place them so that they all have their tails at the same point, the three will lie in the same plane. 11. 99] Solution: Using the notion of the box product yielding either plus or minus the volume of the parallelepiped determined by the given three vectors, show that a b c a b c In other words, the dot and the cross can be switched as long as the order of the vectors remains the same. Hint: There are two ways to do this, by the coordinate description of the dot and cross product and by geometric reasoning. It is better if you use geometric reasoning. In the picture, you have the same parallelepiped. The one on the left involves a b c which is equal to |a b||c|cos 2 which is the area of the base determined by a, b times the altitude which equals |c|cos 2. In the second picture, the altitude is measured from the base determined by c, b. This altitude is |a|cos ) and the area of the new base is |c d|. Hence the volume of the parallelepiped, determined in this other way is |b c||a|cos ) which equals a b c. Thus both a b c and a b c yield the same thing and it is the volume of the parallelepiped. If you switch two of the vectors, then the sign would change but in this case, both expressions would give "1 times the volume of the parallelepiped. 12. 100] Solution: Is a b c a b c? What is the meaning of a b c? Explain. Hint: Try i j j. i j j k j "i. However, i j j 0 and so the cross product is not associative. 13. 101] Solution: Discover a vector identity for u v w and one for u v w . u v w i / ijk u v j wk / ijk / jrs u r v s wk "/ jik / jrs u r v s wk "- ir - ks " - kr - is u r v s wk u k v i wk " u i v k wk u w v "v w u i Hence u v w u w v "v w u You can do the same thing for the other one or you could do the following. u v w "v w u "¡v u w "w u v¢ w u v "v u w 14. 102] Solution: Discover a vector identity for u v z w . It is easiest to use the identity just discovered, although you could do it directly with the permutation symbol and reduction identity. The expression equals u z w v " v z w u ¡u, z, w¢v "¡v, z, w¢u Here ¡u, z, w¢ denotes the box product. 15. 103] Solution: Simplify u v v w w z . Consider the cross product term. From the above, v w w z ¡v, w, z¢w "¡w, w, z¢v ¡v, w, z¢w Thus it reduces to u v ¡v, w, z¢w ¡v, w, z¢¡u, v, w¢ 16. 104] Solution: Simplify |u v|2 u v 2 " |u|2 |v|2 . |u v|2 / ijk u j v k / irs u r v s - jr - ks " - kr - js u r v s u j v k u j v k u j v k " u k v j u j v k |u|2 |v|2 " u v 2 It follows that the expression reduces to 0. You can also do the following. |u v|2 |u|2 |v|2 sin 2 2 |u|2 |v|2 1 " cos 2 2 |u|2 |v|2 " |u|2 |v|2 cos 2 2 |u|2 |v|2 " u v 2 which implies the expression equals 0. 17. 105] Solution: For u, v, w functions of t, show the product rules u v U u U v u v U u v U u U v u v U You can do this more elegantly by repeating the proof of the product rule given in calculus and using the properties of the two products. However, I will show it using the summation convention and permutation symbol. u v U U i q u v i / ijk u j v k U / ijk u Uj v k / ijk u k v Uk u U v u v U i and so u v U u U v u v U . Similar but easier reasoning shows the next one. 18. 106] Solution: If u is a function of t, and the magnitude |ut | is a constant, show from the above problem that the velocity u U is perpendicular to u. u u c where c is a constant. Differentiate both sides. uU u u uU 0 Hence u U u 0. For example, if you move about on the surface of a sphere, then your velocity will be perpendicular to the position vector from the center of the sphere. 19. 107] Solution: When you have a rotating rigid body with angular velocity vector (, then the velocity vector v q u U is given by v (u where u is a position vector. The acceleration is the derivative of the velocity. Show that if ( is a constant vector, then the acceleration vector a v U is given by the formula a ( ( u . Now simplify the expression. It turns out this is centripetal acceleration. a v U ( u U ( ( u Of course you can simplify the right side. It equals a ( u ( "( ( u Also, its magnitude is |(||( u| |(|2 |u|sin 2 and is perpendicular to (. 20. 108] Solution: Verify directly that the coordinate description of the cross product, a b has the property that it is perpendicular to both a and b. Then show by direct computation that this coordinate description satisfies |a b|2 |a|2 |b|2 " a b 2 |a|2 |b|2 1 " cos 2 2 where 2 is the angle included between the two vectors. Explain why |a b| has the correct magnitude. All that is missing is the material about the right hand rule. Verify directly from the coordinate description of the cross product that the right thing happens with regards to the vectors i, j, k. Next verify that the distributive law holds for the coordinate description of the cross product. This gives another way to approach the cross product. First define it in terms of coordinates and then get the geometric properties from this. However, this approach does not yield the right hand rule property very easily. From the coordinate description, a b a / ijk a j b k a i "/ jik a j b k a i "/ jik b k a i a j "a b a and so a b is perpendicular to a. Similarly, a b is perpendicular to b. The Problem above shows that |a b|2 |a|2 |b|2 1 " cos 2 2 |a|2 |b|2 sin 2 2 and so |a b| |a||b|sin 2, the area of the parallelogram determined by a, b. Only the right hand rule is a little problematic. However, you can see right away from the component definition that the right hand rule holds for each of the standard unit vectors. Thus i j k etc. i j k 1 0 0 0 1 0 k Systems Of Equations 1. 109] Solution: Do the three lines, 2y " x 2, 3x 4y 2, and "24x " 22y "8 have a common point of intersection? If so, find the point and if not, tell why they don’t have such a common point of intersection. 2y " x 2 3x 4y 2 , Solution is: x " 25 , y 4 5 2. 110] Solution: Do the three lines, 3x 3y 3, 2x " y 3, and 16y " 5x "12 have a common point of intersection? If so, find the point and if not, tell why they don’t have such a common point of intersection. 3x 3y 3 2x " y 3 , Solution is: x 4 3 , y " 13 3. 111] Solution: Do the three lines, 3x 3y 4, 2x y 2, and 2y " 5x "1 have a common point of intersection? If so, find the point and if not, tell why they don’t have such a common point of intersection. There is no point of intersection. 4. 112] Solution: Do the three planes, x 2z 0, y " 2x " 3z "1, y"x 0 have a common point of intersection? If so, find one and if not, tell why there is no such point. The point of intersection is x, y, z "2, "2, 1 . 5. 113] Solution: Do the three planes, x " 5z 10, y z 2, x y " 3z 10 have a common point of intersection? If so, find one and if not, tell why there is no such point. The point of intersection is x, y, z 0, 4, "2 . 6. 114] Solution: Do the three planes, x " z 3, 4x y " 3z 11, 5x y " 3z 15 have a common point of intersection? If so, find one and if not, tell why there is no such point. The point of intersection is x, y, z 4, "2, 1 . 7. 115] Solution: You have a system of k equations in two variables, k u 2. Explain the geometric significance of a. No solution. b. A unique solution. c. An infinite number of solutions. The lines don’t intersect, they intersect in a single point, they intersect in a whole line. 8. 116] Solution: Here is an augmented matrix in which ' denotes an arbitrary number and Q denotes a nonzero number. Determine whether the given augmented matrix is consistent. If consistent, is the solution unique? Q ' ' ' ' | ' 0 Q ' ' 0 | ' 0 0 Q ' ' | ' 0 0 0 0 Q | ' It appears that the solution exists but is not unique. 9. 117] Solution: Here is an augmented matrix in which ' denotes an arbitrary number and Q denotes a nonzero number. Determine whether the given augmented matrix is consistent. If consistent, is the solution unique? Q ' ' | ' 0 Q ' | ' 0 0 Q | ' It appears that there is a unique solution. 10. 118] Solution: Here is an augmented matrix in which ' denotes an arbitrary number and Q denotes a nonzero number. Determine whether the given augmented matrix is consistent. If consistent, is the solution unique? Q ' ' ' ' | ' 0 Q 0 ' 0 | ' 0 0 0 Q ' | ' 0 0 0 0 Q | ' It appears that there is a solution but it is not unique. 11. 119] Solution: Here is an augmented matrix in which ' denotes an arbitrary number and Q denotes a nonzero number. Determine whether the given augmented matrix is consistent. If consistent, is the solution unique? Q ' ' ' ' | ' 0 Q ' ' 0 | ' 0 0 0 0 Q | 0 0 0 0 0 ' | Q There might be a solution. If so, there are infinitely many. 12. 120] Solution: Suppose a system of equations has fewer equations than variables. Must such a system be consistent? If so, explain why and if not, give an example which is not consistent. No. Consider x y z 2 and x y z 1. 13. 121] Solution: If a system of equations has more equations than variables, can it have a solution? If so, give an example and if not, tell why not. These can have a solution. For example, x y 1, 2x 2y 2, 3x 3y 3 even has an infinite set of solutions. 14. 122] Solution: Find x such that the following augmented matrix is the augmented matrix of an inconsistent system. Then find values of x such that the augmented matrix pertains to a consistent system. 1 "2 x2 4 1 "2 2x 2 3 Inconsistent: x £1, "1¤ Consistent: x £1, "1¤. 15. 123] Solution: Find x such that the following augmented matrix is the augmented matrix of an inconsistent system. Then find values of x such that the augmented matrix pertains to a consistent system. 1 2 x2 1 2 2x 2 " 4 Inconsistent: x £2, "2¤ Consistent: x £2, "2¤. 16. 124] Solution: Find x such that the following augmented matrix is the augmented matrix of an inconsistent system. Then find values of x such that the augmented matrix pertains to a consistent system. 1 x 1 "1 2 x2 3 Consistent: x p 0, Inconsistent: x 0. 17. 125] Solution: Find x such that the following augmented matrix is the augmented matrix of an inconsistent system. Then find values of x such that the augmented matrix pertains to a consistent system. 1 3 1 x4 1 3 1 x4 1 3 1 2x 2 Consistent: x 2, Inconsistent: x p 2 18. 126] Solution: Find x such that the following augmented matrix is the augmented matrix of an inconsistent system. Then find values of x such that the augmented matrix pertains to a consistent system. 1 1 1 x2 6 1 1 1 x2 6 1 1 1 2x 2 5 Consistent: x £1, "1¤, Inconsistent: x £1, "1¤ 19. 127] Solution: Find x such that the following augmented matrix is the augmented matrix of an inconsistent system. Then find values of x such that the augmented matrix pertains to a consistent system. 1 3 4 x2 2 1 3 4 x2 2 1 3 4 2x 2 " 2 Consistent: x £"2, 2¤, Inconsistent: x £"2, 2¤ 20. 128] Solution: Choose h and k such that the augmented matrix shown has one solution. Then choose h and k such that the system has no solutions. Finally, choose h and k such that the system has infinitely many solutions. 1 h 2 . 1 2 k If 2 " h p 0 there is only one solution. If 2 " h 0 and k " 2 p 0 there is no solution. If 2 " h 0 and k " 2 p 0, then there are infinitely many. 21. 129] Solution: Choose h and k such that the augmented matrix shown has one solution. Then choose h and k such that the system has no solutions. Finally, choose h and k such that the system has infinitely many solutions. 4 h 2 . 28 56 k If 8 " h p 0 there is only one solution. If 8 " h 0 and k " 14 p 0 there is no solution. If 8 " h 0 and k " 14 p 0, then there are infinitely many. 22. 130] Solution: Give the complete solution to the system of equations whose augmented matrix is. x y z 2s 3t " 5 s t "1 2 1 "2 "3 "5 "2 4 3 6 5 10 , t, s R 23. 131] Solution: Give the complete solution to the system of equations whose augmented matrix is. 3 8 3 46 1 3 1 17 2 7 3 37 x1 x2 x3 4 5 "2 24. 132] Solution: Give the complete solution to the system of equations whose augmented matrix is. 3 "6 3 "18 1 "2 1 "6 2 "4 2 "12 x y z 2s " t " 6 s t , t, s R 25. 133] Solution: Give the complete solution to the system of equations whose augmented matrix is. "1 "4 "10 "27 1 3 8 22 "2 "5 "14 "38 Inconsistent. 26. 134] Solution: Give the complete solution to the system of equations whose augmented matrix is. 0 "1 0 1 3 1 "8 "1 "2 0 x1 x2 x3 4 8 0 "4 4 27. 135] Solution: Give the complete solution to the system of equations whose augmented matrix is. 0 "1 "1 "4 1 3 19 6 "1 "2 "5 "15 x y z 7 " 3t 4 " t t ,t R 28. 136] Solution: Give the complete solution to the system of equations whose augmented matrix is. 1 2 "2 15 1 3 "4 20 0 1 "2 x y z 5 " 2t 2t 5 t 5 ,t R 29. 137] Solution: Give the complete solution to the system of equations whose augmented matrix is. 0 "1 0 1 3 1 "6 "1 "2 0 x1 x2 x3 1 0 2 "1 "5 30. 138] Solution: Give the complete solution to the system of equations whose augmented matrix is. 1 2 1 5 1 3 1 8 0 1 1 7 x1 x2 x3 "5 3 4 31. 139] Solution: Give the complete solution to the system of equations. 3x 1 8x 2 3x 3 " 5x 4 " 11x 5 " 16 0 x 1 3x 2 x 3 " 2x 4 " 4x 5 " 6 0 2x 1 7x 2 3x 3 " 7x 4 " 8x 5 " 15 0 3x 1 7x 2 3x 3 " 4x 4 " 10x 5 " 15 0 32. 140] Solution: Give the complete solution to the system of equations whose augmented matrix is. 2 0 5 "8 15 1 0 3 "5 9 1 0 4 "7 12 "2 0 4 "10 12 x y z w "t s 2t 3 t , t, s R 33. 141] Solution: Give the complete solution to the system of equations whose augmented matrix is. "1 "4 "1 13 "25 1 1 "11 19 "2 "5 "1 15 "27 "5 "1 15 "51 6 3 , inconsistent. 34. 142] Solution: Give the complete solution to the system of equations whose augmented matrix is. 2 "4 5 15 14 1 "2 3 1 "2 4 12 10 9 8 6 "12 4 12 20 x y z w , t, s R 2t 2 t 2 " 3s s 35. 143] Solution: Give the complete solution to the system of equations whose augmented matrix is. 1 1 0 "1 4 1 3 1 "2 5 3 10 4 "5 17 "1 10 4 "3 x y z w 3 4 1 0 1 36. 144] Solution: Give the complete solution to the system of equations whose augmented matrix is. 1 3 2 "5 5 1 3 3 "8 8 0 0 1 "3 3 4 12 1 x y z w "3s " t " 1 s 3t 3 t 1 "1 , t, s R 37. 145] Solution: Give the complete solution to the system of equations. 4x 1 11x 2 4x 3 " 33x 4 " 29x 5 " 29 0 x 1 3x 2 x 3 " 9x 4 " 8x 5 " 8 0 3x 1 10x 2 4x 3 " 32x 4 " 23x 5 " 29 0 3x 1 10x 2 4x 3 " 32x 4 " 23x 5 " 29 0 x1 x2 x3 x4 x5 38. 146] 3s " 2t " 3 3s 3t 3 2t " 4s 2 t s , t, s R Solution: Give the complete solution to the system of equations. 6x 5 " 4x 2 " x 3 " x 1 6 0 x 1 3x 2 x 3 x 4 " 5x 5 " 5 0 2x 4 " 5x 2 " x 3 " 2x 1 5x 5 9 0 4x 4 " 5x 2 " x 3 " 4x 1 x 5 13 0 x1 x2 x3 x4 x5 t " 2s 2 s t 1 4s " 5t t s , t, s R 39. 147] Solution: Give the complete solution to the system of equations whose augmented matrix is. 3 8 3 "19 "19 10 1 3 1 "7 "7 4 2 7 3 "16 "15 10 1 7 3 "15 "13 12 x1 x2 x3 x4 x5 2s t " 2 2s 2t 2 "s t s , t, s R 40. 148] Solution: Give the complete solution to the system of equations whose augmented matrix is. "1 "4 "1 8 "6 1 "7 4 3 1 "2 "5 "1 10 "8 x y z w 4 "5 "1 t t 2 3t " 2 t ,t R 4 "8 41. 149] Solution: Four times the weight of Gaston is 150 pounds more than the weight of Ichabod. Four times the weight of Ichabod is 660 pounds less than seventeen times the weight of Gaston. Four times the weight of Gaston plus the weight of Siegfried equals 290 pounds. Brunhilde would balance all three of the others. Find the weights of the four sisters. 4g " I 150 4I " 17g "660 4g s 290 , Solution is : £g 60, I 90, b 200, s 50¤ gIs"b 0 42. 150] Solution: The steady state temperature, u in a plate solves Laplace’s equation, u 0. One way to approximate the solution which is often used is to divide the plate into a square mesh and require the temperature at each node to equal the average of the temperature at the four adjacent nodes. This procedure is justified by the mean value property of harmonic functions. In the following picture, the numbers represent the observed temperature at the indicated nodes. Your task is to find the temperature at the interior nodes, indicated by x, y, z, and w. One of the equations is z 1 0 w x . 4 10 You need 1 30 w x " y 0 4 20 1 30 0 z " w 0 4 y 1 y z 10 " x 0 4 20 1 w 0 10 " z 0 4 x , Solution is: ¡w 15, x 15, y 20, z 10¢. 43. 151] Solution: Consider the following diagram of four circuits. Those jagged places denote resistors and the numbers next to them give their resistance in ohms, written as (. The breaks in the lines having one short line and one long line denote a voltage source which causes the current to flow in the direction which goes from the longer of the two lines toward the shorter along the unbroken part of the circuit. The current in amps in the four circuits is denoted by I 1 , I 2 , I 3 , I 4 and it is understood that the motion is in the counter clockwise direction if I k ends up being negative, then it just means it moves in the clockwise direction. Then Kirchhoff’s law states that The sum of the resistance times the amps in the counter clockwise direction around a loop equals the sum of the voltage sources in the same direction around the loop. In the above diagram, the top left circuit should give the equation 2I 2 " 2I 1 5I 2 " 5I 3 3I 2 5 For the circuit on the lower left, you should have 4I 1 I 1 " I 4 2I 1 " 2I 2 "10 Write equations for each of the other two circuits and then give a solution to the resulting system of equations. You might use a computer algebra system to find the solution. It might be more convenient than doing it by hand. The other two equations are 6I 3 " 6I 4 I 3 I 3 5I 3 " 5I 2 "20 2I 4 3I 4 6I 4 " 6I 3 I 4 " I 1 0 Then the system is 2I 2 " 2I 1 5I 2 " 5I 3 3I 2 5 4I 1 I 1 " I 4 2I 1 " 2I 2 "10 6I 3 " 6I 4 I 3 I 3 5I 3 " 5I 2 "20 , Solution is: 2I 4 3I 4 6I 4 " 6I 3 I 4 " I 1 0 £¡I 1 "2. 010 7, I 2 "1. 269 9, I 3 "2. 735 5, I 4 "1. 535 3¢¤ 44. 152] Solution: Consider the following diagram of four circuits. Those jagged places denote resistors and the numbers next to them give their resistance in ohms, written as (. The breaks in the lines having one short line and one long line denote a voltage source which causes the current to flow in the direction which goes from the longer of the two lines toward the shorter along the unbroken part of the circuit. The current in amps in the four circuits is denoted by I 1 , I 2 , I 3 and it is understood that the motion is in the counter clockwise direction. If I k ends up being negative, then it just means the current flows in the clockwise direction. Then Kirchhoff’s law states that The sum of the resistance times the amps in the counter clockwise direction around a loop equals the sum of the voltage sources in the same direction around the loop. Find I 1 , I 2 , I 3 . You have 2I 1 5I 1 3I 1 " 5I 2 "10 I 2 3I 2 7I 2 5I 2 " 5I 1 "12 2I 3 4I 3 4I 3 I 3 " I 2 0 Simplifying this yields 10I 1 " 5I 2 "10 16I 2 " 5I 1 "12 11I 3 " I 2 0 Then you just find the solution. I 1 " 44 , I 2 " 34 , I 3 " 34 27 297 27 Thus all currents flow in the clockwise direction. Matrices 1. 153] Solution: Here are some matrices: "2 2 "3 A "3 1 2 2 "2 C 3 , "1 0 1 2 0 ,D 1 "3 2 2 ,B "2 ,E 2 1 . 0 Find if possible 3A, 3B " A, AC, CB, AE, EA. If it is not possible explain why. "6 6 "9 "9 3 6 , "7 4 0 "1 1 9 "4 4 2 , Not possible, , Not possible, Not possible. "10 6 7 2. 154] Solution: Here are some matrices: 2 A C 2 "3 "1 1 2 2 2 ,D 1 1 "3 ,B 2 2 , "1 "1 1 2 "1 2 1 ,E 2 . "1 Find if possible A, B " A, AC, CB, AE, EA. If it is not possible explain why. 2 2 "3 "1 1 2 , "5 0 0 "2 "1 5 "8 2 6 , Not possible, , Not possible, Not possible. "4 1 3 3. 155] Solution: Here are some matrices: A C 0 2 "2 1 1 2 0 2 1 1 ,B ,D "2 2 2 "1 2 1 1 2 2 1 ,E , 0 2 . Find if possible AB T , BA T " C, CA, E T A, E T E, EE T . If it is not possible explain why. 0 2 2 2 2 2 2 4. 156] , "2 2 0 , 4 , 1 , 0 0 0 4 0 4 "4 1 5 "3 Solution: Here are some matrices: "2 3 A 3 "2 1 "1 3 "2 C 3 ,D 1 3 ,B 3 2 "1 "3 1 "1 "3 2 , "2 ,E 1 . "3 Find if possible AB T , BA T " C, CA, E T A, E T E, EE T . If it is not possible explain why. 9 "4 , "5 "2 10 "9 "3 "3 6 "2 , "7 "3 7 11 "8 10 8 4 6 , 13 , 6 9 5. 157] Solution: Here are some matrices: 4 A 1 0 "3 1 2 1 4 C 2 1 0 ,B ,D 1 2 "1 "2 1 2 "2 2 1 ,E , 4 "2 . Find if possible AB T , BA T " C, CA, E T A, E T E, EE T . If it is not possible explain why. 1 "6 5 3 22 2 "4 , 0 1 "8 2 , 20 , , "8 5 8 5 3 2 16 "8 "8 4 6. 158] Solution: Suppose A and B are square matrices of the same size. Which of the following are correct? a. A " B 2 A 2 " 2AB B 2 b. AB 2 A 2 B 2 c. A B 2 A 2 2AB B 2 d. A B 2 A 2 AB BA B 2 e. A 2 B 2 AAB B f. A B 3 A 3 3A 2 B 3AB 2 B 3 g. A B A " B A 2 " B 2 Part d,e. 7. 159] Solution: Let A 2 2 1 1 Find all 2 2 matrices B such that AB 0. B x y "x "y 8. 160] Solution: Let A 2 3 "1 "3 what should k equal? . Is it possible to choose k such that AB BA? If so, 1 k 3k 6 7 Would need 2 3 ,B "5 "3k " 3 1 "3 2 " k 3 " 3k and this cannot happen. No such k. 9. 161] Solution: Let A 2 3 0 1 should k equal? You need 4 3k 6 0 2 3 ,B k 0 k . Is it possible to choose k such that AB BA? If so, what 4 9 0 k and so you must have k 1. 10. 162] Solution: Let A be an n n matrix. Show A equals the sum of a symmetric and a skew symmetric matrix. (M is skew symmetric if M "M T . M is symmetric if M T M.) Hint: Show that 12 A T A is symmetric and then consider using this as one of the matrices. T T A AA A"A . 2 2 11. 163] Solution: Show every skew symmetric matrix has all zeros down the main diagonal. The main diagonal consists of every entry of the matrix which is of the form a ii . It runs from the upper left down to the lower right. If A "A T , then a ii "a ii and so each a ii 0. 12. 164] Solution: Suppose M is a 3 3 skew symmetric matrix. Show there exists a vector ( such that for all u R3 Mu ( u Hint: Explain why, since M is skew symmetric it is of the form M 0 "F 3 F2 F3 0 "F 1 "F 2 F1 0 where the F i are numbers. Then consider F 1 i F 2 j F 3 k. Since M is skew symmetric, it is of the form mentioned above. Now it just remains to verify that i j k ( qF 1 i F 2 j F 3 k is of the right form. But F 1 F 2 F 3 u1 u2 u3 iF 2 u 3 " iF 3 u 2 " jF 1 u 3 jF 3 u 1 kF 1 u 2 " kF 2 u 1 . In terms of matrices, this is F2u3 " F3u2 . If you multiply by the matrix, you get "F 1 u 3 F 3 u 1 F1u2 " F2u1 0 "F 3 F2 u1 F3 0 "F 1 u2 "F 2 F1 0 u3 F2u3 " F3u2 F3u1 " F1u3 F1u2 " F2u1 which is the same thing. 13. 165] Solution: Using only the vector space axioms, show that 0 is unique. Suppose 0 U is another one. Then 0 U 0 U 0 0. 14. 166] Solution: Using only the vector space axioms, show that "A is unique. Suppose B also works. Then " A "A A B "A A B 0 B B 15. 167] Solution: Using only the vector space axioms, show that 0A 0. Here the 0 on the left is the scalar 0 and the 0 on the right is the zero for m n matrices. 0A 0 0 A 0A 0A. Now add "0A to both sides. Then 0 0A. 16. 168] Solution: Using only the vector space axioms, show that "1 A "A. A "1 A 1 "1 A 0A 0. Therefore, from the uniqueness of the additive inverse proved above, it follows that "A "1 A. 17. 169] Solution: Prove that for A, B m n matrices, and a, b scalars, it follows that aA bB T aA T bB T )A *B T )A *B ji )A ji *B ji ij )A T ij *B Tij )A T *B T ij 18. 170] Solution: Prove that I m A A where A is an m n matrix. I m A ij q ! j - ik A kj A ij . 19. 171] Solution: Show that if A "1 exists for an n n matrix, then it is unique. That is, if BA I and AB I, then B A "1 . A "1 A "1 I A "1 AB A "1 A B IB B. 20. 172] Solution: Give an example of matrices, A, B, C such that B p C, A p 0, and yet AB AC. An example, not the only one is 1 "1 1 1 "1 1 1 1 1 "1 2 2 "1 1 2 2 0 0 0 0 0 0 0 0 21. 173] Solution: Suppose AB AC and A is an invertible n n matrix. Does it follow that B C? Explain why or why not. What if A were a non invertible n n matrix? Yes. Multiply on the left by A "1 . This does not follow if A "1 does not exist. You could have 1 1 2 "2 1 1 "2 2 1 1 1 "1 1 1 "1 1 22. 174] Solution: Find your own examples: a. 2 2 matrices, A and B such that A p 0, B p 0 with AB p BA. i. 0 1 1 2 1 0 3 4 1 2 0 1 3 4 1 0 3 4 2 1 1 2 4 3 b. 2 2 matrices, A and B such that A p 0, B p 0, but AB 0. 0 0 0 0 1 "1 1 1 "1 1 1 1 i. 0 0 0 0 c. 2 2 matrices, A, D, and C such that A p 0, C p D, but AC AD. 1 "1 1 1 "1 1 1 1 i. 1 "1 2 2 "1 1 2 2 0 0 0 0 0 0 0 0 23. 175] Solution: Give an example of a matrix, A such that A 2 I and yet A p I and A p "I. 1 0 0 0 "1 0 0 0 1 24. 176] Solution: Let A 4 2 "4 "2 Find A "1 if it exists and if it does not, explain why. The row reduced echelon form of 4 2 "4 "2 1 is: 0 0 25. 177] Solution: Let A 0 "4 "1 "1 Find A "1 if it exists and if it does not, explain why. 0 "4 "1 "1 "1 1 4 " 14 "1 0 26. 178] Solution: Let A 1 2 3 1 "3 "2 Find A "1 if it exists and if it does not, explain why. and so there is no inverse. "1 3 1 "3 "2 2 3 1 3 "1 "1 27. 179] Solution: Let A "3 3 1 1 Find A "1 if it exists and if it does not, explain why. "3 3 1 "1 1 " 16 1 6 1 2 1 2 28. 180] Solution: Let A "2 "1 4 2 Find A "1 if it exists and if it does not, explain why. The row reduced echelon form of "2 "1 4 2 is: 1 1 2 and so there is no inverse. 0 0 29. 181] Solution: Let A be a 2 2 matrix which has an inverse. Say A a b c d . Find a formula for A "1 in terms of a, b, c, d. "1 a b c d d ad"bc c " ad"bc b " ad"bc a ad"bc 30. 182] Solution: Prove that if A "1 exists and Ax 0 then x 0. Multiply on both sides on the left by A "1 . Thus 0 A "1 0 A "1 Ax A "1 A x Ix x 31. 183] Solution: Let 3 5 12 A 1 2 5 1 2 6 Find A "1 if possible. If not possible, explain why. "1 3 5 12 1 2 5 1 2 6 2 "6 1 "1 6 "3 0 "1 1 32. 184] Solution: Let 1 0 "3 A 1 1 0 1 1 0 Find A "1 if possible. If not possible, explain why. There is no inverse. The row reduced echelon form is 1 0 "3 0 1 3 0 0 0 33. 185] Solution: Let 1 1 2 A 1 2 7 2 4 15 Find A "1 if possible. If not possible, explain why. "1 1 1 2 1 2 7 2 4 15 2 "7 3 "1 11 "5 0 "2 1 34. 186] Solution: Let 1 0 4 A 0 1 4 1 1 8 Find A "1 if possible. If not possible, explain why. There is no inverse. The row reduced echelon form is 1 0 4 0 1 4 0 0 0 35. 187] Solution: Let 1 0 "1 A 3 1 0 1 1 2 Find A "1 if possible. If not possible, explain why. There is no inverse. The row reduced echelon form is 1 0 "1 0 1 3 0 0 0 36. 188] Solution: Let "1 "3 8 1 "3 A 2 "3 "6 10 Find A "1 if possible. If not possible, explain why. "1 "1 "3 8 1 "3 2 "3 "6 10 2 "18 "7 "1 14 5 0 3 1 37. 189] Solution: "2x 1 " 3x 2 " 3x 3 2x 1 3x 3 Write 0 x 1 " 2x 2 2x 4 "2 "3 "3 0 A 2 0 3 0 0 0 0 0 1 "2 0 2 x1 in the form A x2 x3 x4 where A is an appropriate matrix. 38. 190] Solution: 4x 1 4x 2 " x 3 2x 1 x 3 Write 3x 3 x1 in the form A x1 x2 x4 x2 where A is an appropriate matrix. x3 x4 4 4 "1 0 A 2 0 1 0 0 0 3 0 1 1 0 1 39. 191] Solution: Show that if A is an n n invertible matrix and x is a n 1 matrix such that Ax b for b an n 1 matrix, then x A "1 b. Multiply on the left of both sides by A "1 . 40. 192] Solution: Using the inverse of the matrix, find the solution to the systems 2 "1 "4 x "1 1 3 y 1 0 0 z 3 2 "1 "4 x 1 "1 1 3 y 1 0 0 z 1 2 0 , , 2 "1 "4 x "1 1 3 y 1 0 0 z 2 "1 "4 x "1 1 3 y 1 0 0 z 1 Now give the solution in terms of l, m, and n to 2 "1 "4 x "1 1 3 y 1 0 0 z l m n . 2 1 0 3 "1 "2 . 3 0 , 5 0 "3 1 "2 , 1 , 10 9 "4 0 n 3l 4m " 2n For the general solution, it is n"m"l 41. 193] Solution: Using the inverse of the matrix, find the solution to the systems 1 "2 x "1 1 "5 y 1 0 4 z 3 0 1 "2 x 1 "1 1 "5 y 1 z 0 0 4 1 "2 x "1 1 "5 y 1 z 1 0 , 2 , 1 4 1 "2 x "1 1 "5 y 1 z 0 0 0 0 4 Now give the solution in terms of l, m, and n to 1 "2 x "1 1 "5 y 1 z 0 0 4 "13 , "1 1 22 , 0 4l " 4m " 3n 2m " l 2n m"ln m n 0 4 For the general solution, it is 4 9 1 l "9 "6 , . 2 1 0 3 "1 "2 . 42. 194] Solution: Using the inverse of the matrix, find the solution to the system "8 3 0 1 x "21 3 2 5 y "5 1 1 1 z "8 1 1 2 w l " 3m " 2n 8r l m n . r x 2l " 5m " 3n 13r y n " m 2r z 3l " 9m " 7n 25r w 43. 195] Solution: Using the inverse of the matrix, find the solution to the system l " 2m 7n " 5r m " 3n 2r 3m " l " 8n 6r 2m " l " 8n 6r 2 2 0 1 x 0 0 1 "1 y 2 "1 1 1 z 3 "1 1 2 w l m n . r x y z w 44. 196] Solution: "1 T Show that if A is an invertible n n matrix, then so is A T and A T A "1 . You just need to show that A "1 T acts like the inverse of A T because from uniqueness in the above problem, this will imply it is the inverse. But from properties of the transpose, A T A "1 T A "1 A T I T I T T A "1 A T AA "1 I T I Hence A "1 T A T "1 and this last matrix exists. 45. 197] Solution: Show that ABC "1 C "1 B "1 A "1 by verifying that ABC C "1 B "1 A "1 I and C "1 B "1 A "1 ABC I. From the above problem, BC "1 exists and equals BC "1 C "1 B "1 Therefore, ABC "1 exists and equals BC "1 A "1 C "1 B "1 A "1 . 46. 198] Solution: "1 2 If A is invertible, show A 2 A "1 . 2 A 2 A "1 AAA "1 A "1 AIA "1 AA "1 I 2 A "1 A 2 A "1 A "1 AA A "1 IA A "1 A I 47. 199] Solution: If A is invertible, show A "1 "1 A. "1 A "1 A AA "1 I and so by uniqueness, A "1 A. 48. 200] Solution: Let A and be a real m n matrix and let x R n and y R m . Show Ax, y R m x, A T y R n where , R k denotes the dot product in R k . In the notation above, Ax y x A T y. Use the definition of matrix multiplication to do this. Ax y ! k Ax k y k ! k ! i A ki x i y k ! i ! k A Tik x i y k x, A T y 49. 201] Solution: Use the fact that Ax y x A T y to verify directly that AB T B T A T without making any reference to subscripts. ABx, y Bx, A T y x, B T A T y ABx, y x, AB T y Since this is true for all x, it follows that, in particular, it holds for x B T A T y "AB T y and so from the axioms of the dot product, B T A T y "AB T y, B T A T y "AB T y 0 and so B T A T y "AB T y 0. However, this is true for all y and so B T A T "AB T 0. 50. 202] Solution: A matrix A is called a projection if A 2 A. Here is a matrix. 3 "6 "2 "1 4 1 6 "18 "5 Show that this is a projection. Show that a vector in the column space of a projection is left unchanged by multiplication by A. 3 "6 "2 "1 4 1 6 "18 "5 2 3 "6 "2 "1 4 1 6 "18 "5 So it is a projection. In general, a typical vector in the column space of A is Ax. Therefore, AAx A 2 x Ax Thus A leaves vectors in the column space unchanged. 51. 203] Solution: Suppose A is an n n matrix and for each j, n !|A ij | 1 i1 . k0 Show that the infinite series ! A converges in the sense that the ij th entry of the partial sums n converge for each ij. Hint: Let R q max j ! i1 |A ij |. Thus R 1. Show that A 2 ij t R 2 . Then generalize to show that A m ij t R m . Use this to show that the ij th entry of the partial sums is a Cauchy sequence. From calculus, these converge by completeness of the real or complex . numbers. Next show that I " A "1 ! k0 A k . The Leontief model in economics involves solving an equation for x of the form x Ax b, or I " A x b The vector Ax is called the intermediate demand and the vectors A k x have economic meaning as does the series which is also called the Neuman series. ! k A ik A kj t ! k |A ik ||A kj | t R ! k |A kj | t R 2 . Thus A 2 ij t R 2 . Suppose A m ij t R m for all ij. Then k !A m ik A kj A m1 ij t k !|A m ik ||A kj | t R m !|A kj | t R m1 k k It follows that for N M, N M ! Ak " k0 N ! A ij ! A kM1 ij . N k ij t kM1 ij k0 N k ! Ak ! A ij t k kM1 ! Rk kM RM 1"R and so the partial sums for each ij are a Cauchy sequence. Therefore, the partial sums converge. Also N I " A !A N k k0 and it was just shown that A N1 ! Ak I " A I " A N1 k0 v 0 as N v .. Therefore, passing to a limit, it follows that . I " A !A k0 . k ! Ak k0 I " A I Solutions 1. 1] Solution: Find the determinants of the following matrices. "5 "3 "4 0 "4 0 4 6 3 "4 2. 2] Solution: Find the determinants of the following matrices. "12 "10 "7 0 "3 0 9 10 4 "45 3. 3] Solution: Find the determinants of the following matrices. 10 18 9 "3 "6 "3 "11 "16 "10 12 4. 4] Solution: Find the determinants of the following matrices. 2 "1 1 "4 "1 "4 5 5 6 9 5. 5] Solution: Find the determinant of the following matrix. "2 0 2 "3 6 11 12 6 "2 "4 "4 "2 1 "4 "7 2 6 6. 6] Solution: Find the determinant of the following matrix. "5 "14 "16 "6 7 10 7 7 "6 "9 "9 "6 2 11 16 3 18 7. 7] Solution: Find the determinant of the following matrix. 20 31 34 17 "3 "6 "5 "3 5 7 7 5 "21 "31 "35 "18 "6 8. 8] Solution: Find the determinant of the following matrix. "6 "9 "12 "7 11 18 21 11 "5 "7 "7 "5 3 1 1 4 3 3 6 1 6 13 14 6 0 "3 "3 0 18 9. 9] Solution: Find the determinant of the following matrix. "8 "12 "16 "6 6 10. 10] Solution: Find the determinant of the following matrix. "6 "7 "10 "5 "3 "9 "11 "3 "2 0 0 "2 10 15 20 9 4 11. 11] Solution: Find the determinant of the following matrix. "5 "8 "5 "6 11 23 25 11 "3 "8 "8 "3 "4 "9 0 1 30 12. 12] Solution: Find the following determinant by expanding along the first row and then by expanding along the second column. "7 "14 2 8 13 2 "6 "4 "9 15 13. 13] Solution: Find the following determinant by expanding along the first row and then by expanding along the second column. 40 52 20 "20 "26 "10 "20 "26 "10 0 14. 14] Solution: Find the following determinant by expanding along the first row and then by expanding along the second column. 23 29 11 "8 "10 "3 "20 "24 "13 "30 15. 15] Solution: Find the following determinant by expanding along the first row and then by expanding along the second column. "34 "40 "20 17 20 10 20 23 13 0 16. 16] Solution: Find the following determinant by cofactor expansion. Pick whatever is easiest. "2 "7 "4 1 1 10 13 1 2 "1 "1 2 0 "1 "7 "3 "27 17. 17] Solution: Find the following determinant by cofactor expansion. Pick whatever is easiest. "18 "26 "38 "23 20 25 28 20 "11 "12 "12 "11 10 14 23 15 15 18. 18] Solution: Find the following determinant by cofactor expansion. Pick whatever is easiest. "23 "32 "40 "25 19 28 31 19 "11 "14 "14 "11 18 21 26 20 54 19. 19] Solution: Find the following determinant by cofactor expansion. Pick whatever is easiest. "14 "23 "24 "9 1 8 11 1 "1 "3 "3 "1 15 19 17 10 "30 20. 20] Solution: An operation is done to get from the first matrix to the second. Identify what was done and tell how it will affect the value of the determinant. a b c d a c , b d It does not change the determinant. This was just taking the transpose. 21. 21] Solution: Suppose u v w q r s 2 x y z Find ux det vy wz q 3x r 3y s 3z x y z The value is unchanged. It is still 2. 22. 22] Solution: Suppose that A, B are n n matrices and detA 4 while detB 2. Find detAB "1 . 2 23. 23] Solution: Suppose u v w "9 q r s x y z Find u det v w q 3x r 3y s 3z x y z The value is unchanged. It is still "9. 24. 24] Solution: Suppose that A is an n n matrix and A 2 A. Find the possible values for detA . You have detA 2 detA and so you have detA detA " 1 0 so detA 0 or 1. 25. 25] Solution: An operation is done to get from the first matrix to the second. Identify what was done and tell how it will affect the value of the determinant. a b c d , b a d c In this case the two columns were switched so the determinant of the second is "1 times the determinant of the first. 26. 26] Solution: Suppose u v w q r s "9 x y z Find u v w det x y z q r s It is 9 27. 27] Solution: Suppose u v w 10 q r s x y z Find u " 3x v " 3y w " 3z det x y z q r s It is "10 28. 28] Solution: An operation is done to get from the first matrix to the second. Identify what was done and tell how it will affect the value of the determinant. a b c d , a b ac bd It is unchanged. It just adds the top row to the bottom. 29. 29] Solution: Suppose that A, B are n n matrices and detA 5 while detB 3. Find detAB T . 15 30. 30] Solution: Suppose u v w q r s "7 x y z Find v u w " 3u det r q s " 3q y x z " 3x It is 7 31. 31] Solution: Suppose u v w 3 q r s x y z Find "3u v 3w w det "3q r 3w s "3x 3w y z It is 9 32. 32] Solution: An operation is done to get from the first matrix to the second. Identify what was done and tell how it will affect the value of the determinant. a b c d , a b 2c 2d The second row was multiplied by 2 so the determinant of the result is 2 times the original determinant. 33. 33] Solution: Suppose that A is a complex n n matrix and A m I. Find the possible values for detA . You have detA m 1 and so detA is an m th root of 1. 34. 34] Solution: Suppose u v w q r s "9 x y z Find u v 3u w det q r 3q s x y 3x z It is unchanged. Still "9. 35. 35] Solution: An operation is done to get from the first matrix to the second. Identify what was done and tell how it will affect the value of the determinant. a b c d c d , a b In this case two rows were switched and so the resulting determinant is "1 times the first. 36. 36] Solution: Suppose u v w 10 q r s x y z Find u"x v"y w"z det "q "r "s x y z It is 10 37. 37] Solution: Verify detAB detA detB , detA T detA , and that det is multilinear in the rows (columns) for 2 2 matrices. Here is the verification of the formula for products. det x y a b z w c d det ax cy bx dy cw az dw bz adwx " bcwx " adyz bcyz det x y det z w a b c d ad " bc wx " yz adwx " bcwx " adyz bcyz which is the same thing. 38. 38] Solution: ad ae af Find det bd be bf and explain why. cd ce cf a The column space is just the span of b c 39. 39] Solution: and so the determinant is 0. ' 1 1 Find det a b , det 1 1 1 a b c 2 2 2 a b c . Then conjecture a formula for 1 C 1 1 a0 a1 C an det B B B a n0 a n1 C a nn Finally prove your formula. This is called a Vandermonde determinant. Hint: The formula you should get is a j " a i 0tijtn which means to take all possible products of terms of the form a j " a i where i j. To verify this formula, consider pt q det 1 1 C 1 a0 a1 C t B B a n"1 0 a n0 a n"1 1 a n1 B C t n"1 C tn Explain why, using properties of determinants that pa j 0 for all j n. Therefore, explain why n"1 pt c t " a i i0 Now note that pa n is the thing which is wanted. Of course c is the coefficient of t n and you know what this is by induction. It equals the determinant of the upper left corner of the matrix whose determinant equals pt in the above formula. det 1 1 1 det b"a a b 1 1 a1 a2 a3 " a 21 a 2 a 21 a 3 a 1 a 22 " a 1 a 23 " a 22 a 3 a 2 a 23 a 21 a 22 a 23 a 3 " a 2 a 3 " a 1 a 2 " a 1 By induction, pt a j " a i t " a k 0tijtn"1 Hence n"1 k0 pa n n"1 a j " a i a n " a k k0 0tijtn"1 a j " a i 0tijtn 40. 40] Solution: ' Suppose a 0 a 1 a 2 C a n and numbers b 0 , b 1 , C, b n are numbers. Show that there exists a polynomial, pt of degree n such that pa i b i . This is known as Lagrange interpolation. Let pt ) 0 ) 1 x C ) n x n . Then you need to find the ) k such that for each a i , ) 0 ) 1 a i C ) n a ni b i In terms of matrices, you need to find the ) k such that 1 a 0 C a n0 )0 1 a 1 C a n1 )1 B B B 1 a n C a nn B b0 )n b1 B bn However, from the above problem, the matrix has an inverse and so there exists a unique solution to this problem. 41. 41] Solution: ' Suppose a 0 a 1 a 2 C a n and numbers b 0 , b 1 , C, b n are numbers. Show that there exists a unique polynomial, pt of degree n such that pa i b i . In fact, show that the following polynomial works. n t " a j ! b i j:jpia i " a j pt i0 j:jpi This obviously works. pa k b k j:jpk a k "a j b k . This is because for j:jpk a k "a j i p k, j:jpi a k " a j 0. Getting the solution to this is equivalent to finding ) k a solution to the system 1 a 0 C a n0 )0 1 a1 C a n1 )1 B B B B 1 a n C a nn )n b0 b1 B , bn the ) k being the coefficients of the desired polynomial. The above formula shows there exists a solution to this for every choice of the b k and therefore, the matrix must have an inverse. However, if this is so, then there is also at most one solution to the system. Hence the polynomial is unique. Of course by an earlier problem, the matrix is known to have an inverse because it is the transpose of the matrix of the one used in the Vandermonde determinant. 42. 42] Solution: An integration scheme is a formula 1 n n ! cifi X ;x xh ft dt i0 where X is in case ft 1, t, C, t n . Here f i is the value fx i for x x 0 x 1 C x n x h, these points being equally spaced. Find the integration scheme for n 1. You need to find the c i . Then give a technique based on this which will integrate every function which is linear on the subintervals determined by the partition a x 0 x 1 C x n b where x k " x k"1 h 0. You need c01 c11 ;x 1dt h c0x c1b ;x tdt 1 h2x h 2 xh xh Then it follows that c0 h , c1 h 2 2 and so the integration scheme is of the form h fx h fx h X ; xh ft dt 2 2 x Now suppose you have an interval ¡a, b¢. Apply the integration scheme on each interval of the form ¡a kh, a k 1 h¢ to get something which will integrate exactly all functions which are linear on each of these subintervals. Then ;a b n"1 ft dt ! ; xih xih n"1 ! h2 fa ih fa i 1 h ft dt i0 h 2 h 2 n"1 ! fa ih h2 i0 n"1 ! fa ih h2 i0 i0 n"1 ! fa i 1 h i0 n ! fa ih i1 h fa 2fa h 2fa 2h C 2fa n " 1 h fb 2 This is just the trapezoid rule for numerical integration. When you let n 2, you get Simpson’s rule and in fact the procedure integrates exactly polynomials of degree three. 43. 43] Solution: Let A be an r r matrix and suppose there are r " 1 rows (columns) such that all rows (columns) are linear combinations of these r " 1 rows (columns). Show detA 0. If the determinant is non zero, then this will be unchanged with row operations applied to the matrix. However, by assumption, you can obtain a row of zeros by doing row operations. Thus the determinant must have been zero after all. 44. 44] Solution: Show detaA a n detA where here A is an n n matrix and a is a scalar. detaA detaIA detaI detA a n detA . The matrix which has a down the main diagonal has determinant equal to a n . 45. 45] Solution: Is it true that detA B detA detB ? If this is so, explain why it is so and if it is not so, give a counter example. This is not true at all. Consider A 1 0 0 1 ,B "1 0 0 "1 . 46. 46] Solution: An n n matrix is called nilpotent if for some positive integer, k it follows A k 0. If A is a nilpotent matrix and k is the smallest possible integer such that A k 0, what are the possible values of detA ? It must be 0 because 0 det0 detA k detA k . 47. 47] Solution: A matrix is said to be orthogonal if A T A I. Thus the inverse of an orthogonal matrix is just its transpose. What are the possible values of detA if A is an orthogonal matrix? You would need detAA T detA detA T detA 2 1 and so detA 1, or "1. 48. 48] Solution: Fill in the missing entries to make the matrix orthogonal. A is orthogonal if A T A I. "1 2 1 6 12 6 1 2 1 6 12 6 0 6 3 " "1 2 1 6 12 6 1 2 _ _ _ 6 3 _ . 12 6 49. 49] Solution: Let A and B be two n n matrices. A L B (A is similar to B) means there exists an invertible matrix, S such that A S "1 BS. Show that if A L B, then B L A. Show also that A L A and that if A L B and B L C, then A L C. If A S "1 BS, then SAS "1 B and so if A L B, then B L A. It is obvious that A L A because you can let S I. Say A L B and B L C. Then A P "1 BP and B Q "1 CQ. Therefore, A P "1 Q "1 CQP QP "1 CQP and so A L C. 50. 50] Solution: Show that if A L B, then detA detB . To say A L B is to say there exists S such that A S "1 BS. If detA detB , does it follow that A L B? detA detS "1 BS detS "1 detB detS detB detS "1 S detB . The other direction does not work. You could consider 4 0 0 1 2 0 , 0 2 These are not similar. If you had "1 4 0 0 1 x y 2 0 x y z w 0 2 z w then you would have x y 4 0 z w 0 1 2 0 x y 0 2 z w 2x 2y and so 4x y 4z w 2z 2w but there is no solution to this unless x, y, z, w are all zero and this would violate the existence of "1 x y z w . 51. 51] Solution: Two n n matrices, A and B, are similar if B S "1 AS for some invertible n n matrix, S. Show that if two matrices are similar, they have the same characteristic polynomials. The characteristic polynomial of an n n matrix, M is the polynomial, det5I " M . Say M S "1 NS. Then det5I " M det5I " S "1 NS det5S "1 S " S "1 NS detS "1 5I " N S det5I " N 52. 52] Solution: Recall that the volume of a parallelepiped determined by the vectors a, b, c in R 3 is the absolute value of the box product a b c. Show that this box product is the same as the determinant aT det bT det a b c cT Also show that for a, b two vectors in R 2 the area of the parallelogram determined by these two vectors is the absolute value of the determinant det aT det bT a b It can be shown that the only reasonable way to define the volume of a parallelepiped determined by n vectors £a 1 , C, a n ¤ in R n is to take the absolute value of the determinant det a1 C an The formula in three dimensions follows right away because the box product is of the form i j k b1 b2 b3 a1 a2 a3 a1i a2j a3k c1 c2 c3 b1 b2 b3 c1 c2 c3 In two dimensions, you just use the formula for the area of the parallelogram determined by a 1 , a 2 , 0 and b 1 , b 2 , 0 . It reduces to the desired determinant. 53. 53] Solution: Suppose you have a box in three dimensions ¡a, b¢ ¡c, d¢ ¡e, f¢ R. Also suppose A is a 3 3 matrix. Show that the volume of AR is the absolute value of detA volume of R . Say 2 A y2 2 a 1 a 2 a 3 . Then if you have an ellipsoid determined by ax 2 b 2 cz 2 t 1, show that the volume of the ellipsoid is 43 =abc. Here each of a, b, c is positive. You can use the well known formula for the volume of a ball of radius 1 which is just 43 =. First note that R a, b, c R 0 where R 0 has the same length sides but the lower left corner is at 0, 0, 0 . The lengths of the sides of both boxes are b " a , d " c , f " e . Then a AR A b AR 0 c so the volume of AR is the same as the volume of AR 0 . Now let R 0 denote the box which has lower left corner at 0, 0, 0 and sides of length 1. Then from matrix multiplication, the parallelpiped determined by the columns of A is just AR 0 . Also, from geometric reasoning, if you have a parallelpiped and you simply scale the sides by multiplying by constants, the volume of the result is just the product of the constants times the volume of the unscaled parallelepiped. Now R 0 is just the scaled box R 0 the sides being multiplied by the above factors. You can verify this by drawing a picture and using a little trigonometry. Thus, letting VS denote the volume of S, VAR VAR 0 b " a d " c f " e VAR 0 b " a d " c f " e |detA | |detA |VR If you imagine stuffing various other sets U with non overlapping boxes, the same relation holds |detA |VU VAU a 0 0 Let U be the unit ball. Then let A 0 b 0 . It follows that the ellipsoid is just AU and so 0 0 c |detA |VU abc 4 = 3 54. 54] Solution: Tell whether the statement is true or false. a. If A is a 3 3 matrix with a zero determinant, then one column must be a multiple of some other column. 1 i. False. Consider 1 2 "1 5 4 0 3 3 b. If any two columns of a square matrix are equal, then the determinant of the matrix equals zero. i. True c. For A and B two n n matrices, detA B detA detB . i. False d. For A an n n matrix, det3A 3 detA i. False e. If A "1 exists then detA "1 detA "1 . i. True f. If B is obtained by multiplying a single row of A by 4 then detB 4 detA . i. True g. For A an n n matrix, det"A "1 n detA . i. True h. If A is a real n n matrix, then detA T A u 0. i. True i. Cramer’s rule is useful for finding solutions to systems of linear equations in which there is an infinite set of solutions. i. False j. If A k 0 for some positive integer, k, then detA 0. i. True k. If Ax 0 for some x p 0, then detA 0. i. True 55. 55] Solution: Use Cramer’s rule to find the solution to x " 3y 3 x " 2y 1 "3, "2 56. 56] Solution: Use Cramer’s rule to find the solution to x " 3y "3 4x " 11y "11 0, 1 57. 57] Solution: Use Cramer’s rule to find the solution to x " 4y 0 5x " 19y 1 4, 1 58. 58] Solution: Use Cramer’s rule to find the solution to "3x " 2y " 2z " 3 0 "4x " 3y " 3z " 3 0 2x 2y 3z 2 0 "3, 5, "2 59. 59] Solution: Use Cramer’s rule to find the solution to 6x 7y 7z 16 0 5x 6y 6z 14 0 3x 3y 4z 6 0 2, "4, 0 60. 60] Solution: Use Cramer’s rule to find the solution to 4x 5y 5z " 31 0 3x 4y 4z " 24 0 3x 3y 4z " 20 0 4, 4, "1 61. 61] Solution: Here is a matrix, 4 5 5 6 8 8 5 5 6 Determine whether the matrix has an inverse by finding whether the determinant is non zero. If the determinant is nonzero, find the inverse using the formula for the inverse which involves the cofactor matrix. "1 4 5 5 6 8 8 4 " 52 0 2 " 12 5 2 "1 "5 5 5 6 1 62. 62] Solution: Here is a matrix, 2 3 3 4 5 8 1 1 2 Determine whether the matrix has an inverse by finding whether the determinant is non zero. If the determinant is nonzero, find the inverse using the formula for the inverse which involves the cofactor matrix. "1 2 3 3 4 5 8 1 1 2 2 "3 9 0 1 "4 "1 1 "2 63. 63] Solution: Here is a matrix, 1 2 2 0 1 1 3 3 4 Determine whether the matrix has an inverse by finding whether the determinant is non zero. If the determinant is nonzero, find the inverse using the formula for the inverse which involves the cofactor matrix. "1 1 2 2 0 1 1 3 3 4 1 "2 3 "2 "1 "3 3 0 1 64. 64] Solution: Here is a matrix, "2 "1 "1 "3 "2 "2 1 1 2 Determine whether the matrix has an inverse by finding whether the determinant is non zero. If the determinant is nonzero, find the inverse using the formula for the inverse which involves the cofactor matrix. "2 "1 "1 "1 1 1 4 "3 "1 "1 1 "3 "2 "2 1 "2 2 0 1 65. 65] Solution: Here is a matrix, 1 2 2 0 "2 "2 5 5 6 Determine whether the matrix has an inverse by finding whether the determinant is non zero. If the determinant is nonzero, find the inverse using the formula for the inverse which involves the cofactor matrix. "1 1 2 2 0 "2 "2 5 5 6 1 1 0 5 2 "1 "5 " 52 1 66. 66] Solution: Here is a matrix, 3 4 4 2 3 3 2 2 3 Determine whether the matrix has an inverse by finding whether the determinant is non zero. If the determinant is nonzero, find the inverse using the formula for the inverse which involves the cofactor matrix. "1 3 "4 0 0 1 "1 "2 2 1 3 4 4 2 3 3 2 2 3 67. 67] Solution: Here is a matrix, e 2t 2e 2t cos t sin t " sin t cos t 4e 2t " cos t " sin t Determine whether the matrix has an inverse by finding whether the determinant is non zero. Then find its inverse if it has one. e 2t "1 cos t sin t 2e 2t " sin t cos t 1 5 4 5 4 5 4e 2t " cos t " sin t e "2t cos t sin t " 1 5 0 2 sin t 5 2 cos t 5 " sin t cos t e "2t 2 sin t " 15 cos t 5 " 25 cos t " 15 sin t 68. 68] Solution: Here is a matrix, et " sin 2t cos 2t e t "2 sin 2t "2 cos 2t e t "4 cos 2t 4 sin 2t Determine whether the matrix has an inverse by finding whether the determinant is non zero. Then find its inverse if it has one. et cos 2t " sin 2t e t "2 sin 2t "2 cos 2t e t "4 cos 2t 4 sin 2t "1 4 5 1 5 2 5 e "t cos 2t cos 2t " 2 5 1 5 1 5 0 sin 2t " 12 sin 2t sin 2t " 12 cos 2t 1 10 1 10 e "t sin 2t " cos 2t 1 cos 2t 5 1 sin 2t 5 69. 69] Solution: Here is a matrix, e t cosh t sinh t e t sinh t cosh t e t cosh t sinh t Does there exist a value of t for which this matrix fails to have an inverse? Explain. e t cosh t sinh t det e t sinh t cosh t e t cosh t sinh t 0 and so this matrix fails to have a nonzero determinant at any value of t. 70. 70] Here is a matrix, "1 t t2 0 "1 2t t 0 4 Determine whether the matrix has an inverse by finding whether the determinant is non zero. Then find its inverse if it has one. "1 Solution: "1 t2 t 0 "1 2t t 0 2 " 3t 344 "4 3t 3t4 3 3tt3 4 2 3tt3 4 " 3tt 34 4 3 2 3t 3t4 t 3t 3 4 t2 3t 3 4 1 3t 3 4 2 4 provided 3t 3 4 p 0. 71. 71] Solution: Here is a matrix, cos 2t e "2t e 5t 5e 5t "e "2t sin 2t "2e "2t cos 2t sin 2t "2e "2t cos 2t " sin 2t 8e "2t sin 2t 25e 5t 8cos 2t e "2t Determine whether the matrix has an inverse by finding whether the determinant is non zero. determinant is " 106 1 "4t e4 72. 72] Solution: Here is a matrix, 1 t e "2t 0 1 "2e "2t 0 0 4e "2t Determine whether the matrix has an inverse by finding whether the determinant is non zero. Then find its inverse if it has one. 1 t e "2t 0 1 "2e "2t 0 0 "1 1 "t " 12 t " 4e "2t 1 4 1 2 0 1 1 4 0 0 e 2t 73. 73] Solution: Here is a matrix, 1 0 0 "3e 0 cos 3t e "3t 0 "3t cos 3t sin 3t "3e "e "3t "3t sin 3t cos 3t " sin 3t Determine whether the matrix has an inverse by finding whether the determinant is non zero. Then find its inverse if it has one. "1 1 0 0 0 cos 3t e "3t "e "3t sin 3t 0 "3e "3t cos 3t sin 3t "3e "3t cos 3t " sin 3t 1 0 0 0 e cos 3t " sin 3t 3t " 0 "e 3t cos 3t sin 3t " sin3t 3e "3t cos2 3t3e "3t sin 2 3t cos3t 3e "3t cos2 3t3e "3t sin 2 3t 74. 74] Solution: Here is a matrix, et cos 4t e 2t "e 2t sin 4t et 2e 2t cos 4t " 2 sin 4t "2e 2t 2 cos 4t sin 4t e t "12cos 4t e 2t " 16e 2t sin 4t 12e 2t sin 4t " 16cos 4t e 2t Determine whether the matrix has an inverse by finding whether the determinant is non zero. determinant is "68e 5t 75. 75] Solution: Here is a matrix, 1 0 cos t e 0 0 "t sin t e "t 0 "e "t cos t sin t e "t cos t " sin t Determine whether the matrix has an inverse by finding whether the determinant is non zero. Then find its inverse if it has one. "1 1 0 0 1 0 cos t e "t sin t e "t 0 "e "t cos t sin t e "t cos t " sin t 0 0 0 e t cos t " sin t "e t sin t 0 e t cos t sin t cos t e t 76. 76] Solution: Here is a matrix, e "3t te "3t 1 "3e "3t "e "3t 3t " 1 0 9e "3t 3e "3t 3t " 2 0 Determine whether the matrix has an inverse by finding whether the determinant is non zero. Then find its inverse if it has one. e "3t te "3t "1 "3e "3t "e "3t 3t " 1 0 9e "3t 0 1 3e "3t 3t " 2 0 1 3 e 3t 3t " 2 1 9 e 3t 3t " 1 0 "e 3t " 13 e 3t 1 2 3 1 9 77. 77] Solution: Show that if detA p 0 for A an n n matrix, it follows that if Ax 0, then x 0. If detA p 0, then A "1 exists and so you could multiply on both sides on the left by A "1 and obtain that x 0. 78. 78] Solution: Suppose A, B are n n matrices and that AB I. Show that then BA I. Hint: You might do something like this: First explain why detA , detB are both nonzero. Then AB A A and then show BABA " I 0. From this use what is given to conclude ABA " I 0. You have 1 detA detB . Hence both A and B have inverses. Letting x be given, ABA " I x AB Ax " Ax Ax " Ax 0 and so it follows from the above problem that BA " I x 0. Since x is arbitrary, it follows that BA I. 79. 79] Solution: Suppose A is an upper triangular matrix. Show that A "1 exists if and only if all elements of the main diagonal are non zero. Is it true that A "1 will also be upper triangular? Explain. Is everything the same for lower triangular matrices? The given condition is what it takes for the determinant to be non zero. Recall that the determinant of an upper triangular matrix is just the product of the entries on the main diagonal. The inverse will also be upper triangular. This follows from the method of row operations applied to finding the inverse. 80. 80] Solution: If A, B, and C are each n n matrices and ABC is invertible, why are each of A, B, and C invertible. This is obvious because detABC detA detB detC and if this product is nonzero, then each determinant in the product is nonzero and so each of these matrices is invertible. 81. 81] Solution: Let Ft det at bt ct dt . Verify F U t det Now suppose a U t b U t ct dt det at bt c U t d U t . at bt ct Ft det . dt et ft gt ht it Use Laplace expansion and the first part to verify F U t a U t b U t c U t det det dt et ft gt ht it at det bt ct d U t e U t f U t gt at bt ct dt et ht it . ft g U t h U t i U t Conjecture a general result valid for n n matrices and explain why it will be true. Can a similar thing be done with the columns? Yes, a similar thing can be done with the columns because the determinant of the transpose is the same as the determinant of the matrix. 82. 82] Solution: Let Ly y n a n"1 x y n"1 C a 1 x y U a 0 x y where the a i are given continuous functions defined on a closed interval, a, b and y is some function which has n derivatives so it makes sense to write Ly. Suppose Ly k 0 for k 1, 2, C, n. The Wronskian of these functions, y i is defined as Wy 1 , C, y n x q det y 1 x C y n x y U1 x C y Un x B B y n"1 x 1 C y n"1 x n Show that for Wx Wy 1 , C, y n x to save space, U W x det y 1 x C y n x y U1 x C y Un x B B . n y n 1 x C y n x Now use the differential equation, Ly 0 which is satisfied by each of these functions, y i and properties of determinants presented above to verify that W U a n"1 x W 0. Give an explicit solution of this linear differential equation, Abel’s formula, and use your answer to verify that the Wronskian of these solutions to the equation, Ly 0 either vanishes identically on a, b or never. Hint: To solve the differential equation, let A U x a n"1 x and multiply both sides of the differential equation by e Ax and then argue the left side is the derivative of something. The last formula above follows because W U x equals the sum of determinants of matrices which have two equal rows except for the last one in the sum which is the displayed expression. Now let m i x " a n"1 x y n"1 C a 1 x y Ui a 0 x y i i Since each y i is a solution to Ly 0, it follows that y n i t m i t . Now from the properties of determinants, being linear in each row, W U x "a n"1 x det y 1 x C y n x y U1 x C y Un x B B y n"1 x 1 C y n"1 x n "a n"1 x W U x Now let A U x a n"1 x . Then d e Ax Wx 0 dx and so Wx Ce "Ax . Thus the Wronskian either vanishes for all x or for no x. 83. 83] Solution: Find the determinant of the following matrix. "2 i 1 4i 3i det "2 i 1 4i 3i 0 0 12 " 3i 84. 84] Solution: Find the determinant of the following matrix. 22 " 24i 32 " 34i 16 " 10i "13 14i "19 20i "9 6i "8 10i "12 14i "6 4i 8 85. 85] Solution: Find the determinant of the following matrix. 19 " 36i 21 " 51i 5 " 15i "5 21i "3 30i "1 9i "14 15i "18 21i "4 6i 0 Rank, Dimension, Subspaces 1. 86] Solution: Let £u 1 , C, u n ¤ be vectors in R n . The parallelepiped determined by these vectors Pu 1 , C, u n is defined as n Pu 1 , C, u n q ! t k u k : t k ¡0, 1¢ for all k . k1 Now let A be an n n matrix. Show that £Ax : x Pu 1 , C, u n ¤ is also a parallelepiped. n By definition, APu 1 , C, u n ! k1 t k Au k : t k ¡0, 1¢ for all k which equals PAu 1 , C, Au n which is a parallelepiped. 2. 87] Solution: Draw Pe 1 , e 2 where e 1 , e 2 are the standard basis vectors for R 2 . Thus e 1 1, 0 , e 2 0, 1 . Now suppose E 1 1 0 1 where E is the elementary matrix which takes the third row and adds to the first. Draw £Ex : x Pe 1 , e 2 ¤. In other words, draw the result of doing E to the vectors in Pe 1 , e 2 . Next draw the results of doing the other elementary matrices to Pe 1 , e 2 . For E the given elementary matrix, the result is as follows. It is called a shear. Note that it does not change the area. In caseE 1 0 , the given square Pe 1 , e 2 becomes a rectangle. If ) 0, the square is 0 ) stretched by a factor of ) in the y direction. If ) 0 the rectangle is reflected across the x axis and ) 0 in addition is stretched by a factor of |)| in the y direction. When E the result is 0 1 similar only it features changes in the x direction. These elementary matrices do change the area unless |)| 1 when the area is unchanged. The permutation 0 1 simply switches e 1 and 1 0 e 2 so the result appears to be a square just like the one you began with. 3. 88] Solution: Either draw or describe the result of doing elementary matrices to Pe 1 , e 2 , e 3 , the parallelepiped determined by the three listed vectors. Describe geometrically the fact proved in the chapter that every invertible matrix is a product of elementary matrices. It is the same sort of thing. The elementary matrix either switches the e i about or it produces a shear or a magnification in one direction. 4. 89] Solution: Explain why the elementary matrix which comes from the row operation which replaces a row with a constant times another row added to it does not change the volume of a box which has the sides parallel to the coordinate axes. It is because the change happens in only two dimensions where a rectangle becomes a parallelogram which has the same base and height as the rectangle. Thus the area of this face does not change and so the volume of the resulting sheared box is the same as the volume of the box. 5. 90] Solution: Determine which matrices are in row reduced echelon form. a. 1 2 0 0 1 7 i. This one is not. 1 0 0 0 b. 0 0 1 2 0 0 0 0 i. This one is. 1 1 0 0 0 5 c. 0 0 1 2 0 4 0 0 0 0 1 3 This one is. 6. 91] Solution: Row reduce the following matrix to obtain the row reduced echelon form. "1 "1 "2 "6 1 1 1 3 1 1 0 0 1 1 0 0 row echelon form: 0 0 1 3 0 0 0 0 7. 92] Solution: Row reduce the following matrix to obtain the row reduced echelon form. "1 1 2 1 1 "2 "7 "3 1 "3 "12 "5 1 0 3 1 row echelon form: 0 1 5 2 0 0 0 0 8. 93] Solution: Row reduce the following matrix to obtain the row reduced echelon form. 3 11 "24 11 1 0 row echelon form: 3 1 4 "9 4 1 3 "6 3 0 0 1 "3 1 0 0 0 0 9. 94] Solution: Row reduce the following matrix to obtain the row reduced echelon form. "2 "4 "1 "1 1 2 0 "1 row echelon form: 10. 95] Solution: 0 0 1 3 0 0 0 0 "1 1 2 0 1 2 "1 "4 Row reduce the following matrix to obtain the row reduced echelon form. 1 0 row echelon form: 0 "2 "9 27 "29 1 4 "12 13 1 3 "9 10 1 0 1 "3 3 0 0 0 0 11. 96] Solution: Row reduce the following matrix to obtain the row reduced echelon form. 3 "4 21 "5 1 0 row echelon form: 3 1 "1 6 "1 1 "2 9 "3 1 0 1 "3 2 0 0 0 0 12. 97] Solution: Row reduce the following matrix to obtain the row reduced echelon form. "1 1 "1 "5 1 "1 0 1 "1 "1 "1 2 1 "1 0 2 row echelon form: 0 0 1 3 0 0 0 0 13. 98] Solution: Find the rank of the following matrix. Also find a basis for the row and column spaces. 1 "4 1 "1 1 "4 2 1 1 "4 1 "1 1 "4 0 "3 row echelon form: 0 0 1 2 0 0 0 0 1 Rank is 2 Basis for column space is 1 , 2 1 0 0 1 2 , Basis for row space is 1 1 1 "4 0 "3 14. 99] Solution: Find the rank of the following matrix. Also find a basis for the row and column spaces. "2 4 1 "2 "2 2 7 1 "2 "2 1 3 "3 6 "3 "9 4 6 "5 "18 1 "2 "2 0 "1 row echelon form: "2 1 1 "3 0 0 0 1 4 0 0 0 0 0 0 0 0 0 0 Then a basis for the column space is "5 , 2 1 Rank 2 Basis for row space is "3 1 "2 "2 0 "1 , 0 0 0 1 4 15. 100] Solution: Find the rank of the following matrix. Also find a basis for the row and column spaces. 1 "2 0 "2 1 "2 1 0 1 "2 0 "2 1 "2 0 "2 row echelon form: 0 0 1 2 0 0 0 0 0 Rank is 2 Basis for column space is 1 , 1 0 0 0 1 2 , Basis for row space is 1 1 1 "2 0 "2 16. 101] Solution: Find the rank of the following matrix. Also find a basis for the row and column spaces. 3 9 "4 "14 1 3 "1 "4 1 3 "2 "6 1 3 0 "2 row echelon form: 0 0 1 2 0 0 0 0 "4 Rank is 2 Basis for column space is 3 "1 , "2 0 0 1 2 , 1 Basis for row space is 1 1 3 0 "2 17. 102] Solution: Find the rank of the following matrix. Also find a basis for the row and column spaces. 4 8 "16 "9 6 1 2 "4 "2 2 1 2 "4 "3 0 3 6 "12 "9 0 1 2 "4 0 6 row echelon form: 0 0 0 1 2 0 0 0 0 0 0 0 0 0 0 Then a basis for the column space is "9 4 1 1 "2 , Rank 2 Basis for row space is "3 1 2 "4 0 6 , 0 0 0 1 2 "9 3 18. 103] Solution: Find the rank of the following matrix. Also find a basis for the row and column spaces. "1 "1 1 1 1 "2 "5 "32 1 1 "3 "4 "24 "2 "2 1 1 0 0 "1 1 1 "2 0 0 0 1 7 0 0 0 0 0 1 , 48 Rank 3 Basis for column space is 5 "2 , "3 "5 Basis for row space is "4 6 1 1 0 0 1 8 33 1 0 0 1 0 "1 row echelon form: 6 5 8 , 0 0 1 0 "1 , 0 0 0 1 7 19. 104] Solution: Find the rank of the following matrix. Also find a basis for the row and column spaces. 2 "5 23 1 "2 10 1 "3 13 1 "3 13 1 0 row echelon form: 4 0 1 "3 0 0 0 0 0 0 Then a basis for the column space is "5 2 1 Basis for column space: "2 , 1 "3 1 1 0 4 Rank 2 Basis for row space: "3 0 1 "3 , 20. 105] Solution: Find the rank of the following matrix. Also find a basis for the row and column spaces. 1 2 "1 0 "4 1 2 0 0 "5 1 2 "1 1 2 0 0 0 0 0 1 2 0 0 "5 0 0 1 0 "1 row echelon form: "1 1 1 1 0 , 0 "1 6 0 0 0 0 0 Rank 3 Basis for column space is 0 0 , 1 0 1 2 0 0 "5 0 0 0 1 Basis for row space is 0 , 0 0 1 0 "1 , 0 0 0 1 6 21. 106] Solution: Find the rank of the following matrix. Also find a basis for the row and column spaces. "1 "5 "15 1 4 12 1 3 9 "2 "6 "18 1 0 0 row echelon form: 0 1 3 0 0 0 0 0 0 Then a basis for the column space is "1 "5 1 Basis for column space: 4 , 1 "2 1 0 0 , Rank 2 Basis for row space: 3 "6 0 1 3 22. 107] Solution: Find the rank of the following matrix. Also find a basis for the row and column spaces. 2 4 5 8 35 1 2 3 4 16 1 2 2 5 26 1 2 2 5 26 1 2 0 0 "3 0 0 1 0 "3 row echelon form: 2 1 1 1 5 , 3 2 7 0 0 0 0 0 Rank 3, Basis for column space is 8 4 , Basis for row space is 5 2 1 2 0 0 "3 0 0 0 1 5 0 0 1 0 "3 , , 0 0 0 1 7 23. 108] Solution: Find the rank of the following matrix. Also find a basis for the row and column spaces. 1 "1 "1 0 row echelon form: "2 2 2 "1 0 1 "1 "1 0 1 1 "1 "1 "1 3 "3 3 "9 3 3 1 0 0 0 1 "2 0 0 0 0 0 0 0 0 0 0 Then a basis for the column space is "2 "1 1 , 1 "3 0 Rank 2 Basis for row space is "1 3 1 "1 "1 0 1 0 0 0 1 "2 , 24. 109] Solution: Suppose A is an m n matrix. Explain why the rank of A is always no larger than minm, n . It is because you cannot have more than minm, n nonzero rows in the row reduced echelon form. 25. 110] Solution: 5 1 Let H denote span 1 14 , 4 of H and determine a basis. It is dimension 3 and a basis is 5 1 1 14 , 4 3 2 3 2 33 , 7 5 15 , 3 3 8 20 12 19 62 5 5 , 1 2 . Find the dimension 8 5 , 8 1 2 8 26. 111] Solution: 5 1 Let H denote span 1 , 4 determine a basis. It is dimension 3 and a basis is 5 1 1 19 , 4 27. 112] Solution: 4 3 12 5 , 1 2 8 4 3 12 , 13 10 40 , 1 2 8 . Find the dimension of H and "1 0 1 Let H denote span , 0 "1 5 0 "1 "4 dimension of H and determine a basis. It is dimension 3 and a basis is "1 0 1 , 0 "1 3 0 3 , "2 0 1 , 0 , 0 "2 "2 7 0 . Find the "5 0 1 , 0 "2 "2 28. 113] Solution: "7 3 Let H denote span 1 1 "2 , "3 8 , 11 "6 2 "8 27 "2 , . Find the dimension of H and "4 "8 22 determine a basis. It is dimension 2 and a basis is "7 3 1 1 "2 , "3 "6 2 29. 114] Solution: "2 Let H denote span 1 1 "3 and determine a basis. It is dimension 2 and a basis is "2 1 1 "9 , "3 30. 115] Solution: 4 3 "9 "9 , 4 3 "9 "33 , 15 12 "36 "2 , 1 1 "3 . Find the dimension of H 2 1 Let H denote span 2 27 , 1 of H and determine a basis. It is dimension 3 and a basis is 2 1 2 7 , 1 4 6 7 15 , 24 4 6 6 , 3 8 12 3 4 3 12 12 16 , 9 14 . Find the dimension 7 6 , 3 3 8 4 31. 116] Solution: 4 1 Let H denote span 1 , 3 1 0 , 3 3 , 3 4 0 9 12 2 6 2 . Find the dimension of H and determine a basis. It is dimension 3 and a basis is 4 1 1 3 , 3 1 0 0 12 , 3 4 12 32. 117] Solution: 1 Let H denote span 2 1 , 0 determine a basis. It is dimension 2 and a basis is 1 2 1 2 , 0 33. 118] Solution: 6 2 0 6 2 0 , 16 6 0 , 4 2 0 . Find the dimension of H and 34. 35. 36. 37. Let M £u u 1 , u 2 , u 3 , u 4 R 4 : sinu 1 1¤. Is M a subspace? Explain. Not a subspace. =2 , 0, 0, 0 is in it. 4 =2 , 0, 0, 0 is not. 119] Solution: Let M u u 1 , u 2 , u 3 , u 4 R 4 : u i u 0 for each i 1, 2, 3, 4 . Is M a subspace? Explain. Not a subspace. 1, 1, 1, 1 is in it. However, "1 1, 1, 1, 1 is not. 120] Solution: Let w R 4 and let M £u u 1 , u 2 , u 3 , u 4 R 4 : w u 0¤. Is M a subspace? Explain. Subspace 121] Solution: Let M £u u 1 , u 2 , u 3 , u 4 R 4 : |u 1 | t 4¤. Is M a subspace? Explain. This is not a subspace. 1, 1, 1, 1 is in it, but 201, 1, 1, 1 is clearly not. 122] Solution: Let w, w 1 be given vectors in R 4 and define M u u 1 , u 2 , u 3 , u 4 R 4 : w u 0 and w 1 u 0 . Is M a subspace? Explain. Sure. This is a subspace. It is obviously closed with respect to vector addition and scalar multiplication. 38. 123] Solution: Prove the following theorem: If A, B are n n matrices and if AB I, then BA I and B A "1 . Hint: First note that if AB I, then it must be the case that A is onto. Explain why this requires span columns of A F n . Now explain why, using the corollary that this requires A to be one to one. Next explain why ABA " I 0 and why the fact that A is one to one implies BA I. If AB I, then B must be one to one. Otherwise there exists x p 0 such that Bx 0. But then you would have x Ix ABx A0 0 In particular, the columns of B are linearly independent. Therefore, B is also onto. Also, BA " I Bx BAB x "Bx 0 Since B is onto, it follows that BA " I maps every vector to 0 and so this matrix is 0. Thus BA I. 39. 124] Solution: Let M £u u 1 , u 2 , u 3 , u 4 R 4 : u 3 u u 1 ¤. Is M a subspace? Explain. This is not a subspace. 0, 0, 1, 0 is in it. But 0, 0, "1, 0 is not. 40. 125] Solution: Suppose £x 1 , C, x k ¤ is a set of vectors from F n . Show that 0 is in spanx 1 , C, x k . ! ki1 0x k 0. 41. 126] Solution: Explain why if A is m n where m n, then A cannot be one to one. The columns cannot be a basis for F m because there are too many of them. Therefore, they are dependent. Hence there exists x p 0 such that Ax 0. 42. 127] Solution: Let M £u u 1 , u 2 , u 3 , u 4 R 4 : u 3 u 1 0¤. Is M a subspace? Explain Yes. This is a subspace. It is closed with respect to vector addition and scalar multiplication. 43. 128] Solution: Prove the following theorem: If A, B are n n matrices and if AB I, then BA I and B A "1 . Hint: First note that if AB I, then it must be the case that A is onto. Explain why this requires span columns of A F n . Now explain why, using the corollary that this requires A to be one to one. Next explain why ABA " I 0 and why the fact that A is one to one implies BA I. If AB I, then B must be one to one. Otherwise there exists x p 0 such that Bx 0. But then you would have x Ix ABx A0 0 In particular, the columns of B are linearly independent. Therefore, B is also onto. Also, BA " I Bx BAB x "Bx 0 Since B is onto, it follows that BA " I maps every vector to 0 and so this matrix is 0. Thus BA I. 44. 129] Solution: Study the definition of span. Explain what is meant by the span of a set of vectors. Include pictures. The span of vectors is a long flat thing. In three dimensions, it would be a plane which goes through the origin. 45. 130] Solution: Let £v 1 , C, v m ¤ be vectors. Prove that spanv 1 , C, v m is a subspace. m m Say a, b are scalars. Let ! k1 ) k v k and ! k1 * k v k are two things in spanv 1 , C, v m . Then m m k1 k1 a ! )kvk b ! *kvk m !a) k * k b v k spanv 1 , C, v m k1 46. 131] Solution: Let A be an m n matrix. Then NA q kerA q £x F n : Ax 0¤ Show that NA is a subspace of F n . Say Ax Ay 0. Then if a, b are scalars, Aax by aAx bAy a0 b0 0 47. 132] Solution: Let A be an m n matrix. Then the image of A is defined as ImA q AF n q £Ax : x F n ¤ Show that NA is a subspace of F m . Note that this is really the column space. It is done in the book but you should do it here without dealing with columns. Let a, b be scalars. Let Ax, Ay be two things in ImA . Then is it the case that aAx bAy is in ImA ? aAx bAy Aax by ImA 48. 133] Solution: Suppose you have a subspace V of F n . Is it the case that V equals NA kerA for some m n matrix? You know that V has a basis £v 1 , C, v m ¤, m t n. If m n, then V F n and you take A 0 nn . Otherwise, choose vectors u m1 , C, u n such that u k v j 0 and u j u i - ij . How do you do this? v '1 B v 'm is longer than it is tall. Hence there exists u 1 p 0 such that this matrix times u 1 to equal 0. Next, to get u 2 , do the same thing with v '1 B v 'm u '1 and continue till the matrix has the same height as length. Each time divide by the magnitude of the vector obtained. Consider the n " m n matrix. u 'm1 Aq B u 'n By construction, it sends each v k to 0 and so kerA spanv 1 , C, v m . Now suppose x m n i1 jm1 m n n i1 jm1 jm1 ! a i v i ! b ju j kerA Then 0 Ax ! a i Av i ! b jAu j ! b ju j and so each b j 0. Hence x spanv 1 , C, v m . 49. 134] Solution: Suppose you have a subspace V of F n . Is it the case that V equals ImA for some matrix A? Yes, this is obvious. You just form the matrix which has a basis for V as its columns. 50. 135] Solution: Consider the vectors of the form v u"v"w : u, v, w R . w Is this set of vectors a subspace of R 3 ? If so, explain why, give a basis for the subspace and find its dimension. It is of dimension 3 and a basis is 0 1 1 "1 , 0 0 0 , "1 1 51. 136] Solution: Consider the vectors of the form 2u v " 5w 4v " 2u 10w : u, v, w R . 2v " 4u 14w Is this set of vectors a subspace of R 3 ? If so, explain why, give a basis for the subspace and find its dimension. It is of dimension 2 and a basis is 2 "2 "4 1 , 4 2 52. 137] Solution: Consider the vectors of the form 3u v 5w 7u 3v 13w : u, v, w R . 15u 6v 27w Is this set of vectors a subspace of R 3 ? If so, explain why, give a basis for the subspace and find its dimension. It is of dimension 2 and a basis is 3 1 , 7 15 3 6 53. 138] Solution: Consider the vectors of the form 2u v w"v"u : u, v, w R . w " 2v " 2u Is this set of vectors a subspace of R 3 ? If so, explain why, give a basis for the subspace and find its dimension. It is of dimension 3 and a basis is 2 1 "1 , "2 "1 "2 0 , 1 1 54. 139] Solution: Consider the vectors of the form 4u 12v 15w u 3v 4w 4u 12v 12w : u, v, w R . 3u 9v 9w Is this set of vectors a subspace of R 4 ? If so, explain why, give a basis for the subspace and find its dimension. It is of dimension 2 and a basis is 4 1 4 3 15 , 4 12 9 55. 140] Solution: Consider the vectors of the form 3u v 5u v " w : u, v, w R . w " 2v " 7u Is this set of vectors a subspace of R 3 ? If so, explain why, give a basis for the subspace and find its dimension. It is of dimension 3 and a basis is 3 1 , 5 1 "7 "2 0 , "1 1 56. 141] Solution: Consider the vectors of the form 3u v " w 3u 2v w : u, v, w R . uvw Is this set of vectors a subspace of R 3 ? If so, explain why, give a basis for the subspace and find its dimension. It is of dimension 2 and a basis is 3 3 1 , 1 2 1 57. 142] Solution: Consider the vectors of the form 2u v 7w 2u " 4v 2w : u, v, w R . "12v " 12w Is this set of vectors a subspace of R 3 ? If so, explain why, give a basis for the subspace and find its dimension. It is of dimension 2 and a basis is 2 2 0 1 , "4 "12 58. 143] Solution: Consider the vectors of the form 3u v 6u 6v 2w : u, v, w R . 2u 4v 2w Is this set of vectors a subspace of R 3 ? If so, explain why, give a basis for the subspace and find its dimension. It is of dimension 3 and a basis is 3 6 1 , 0 , 6 2 4 2 2 59. 144] Solution: Consider the vectors of the form 3u v 5w 0 : u, v, w R . "6u " 4v " 14w Is this set of vectors a subspace of R 3 ? If so, explain why, give a basis for the subspace and find its dimension. It is of dimension 2 and a basis is 3 1 , 0 0 "6 "4 60. 145] Solution: Consider the vectors of the form 4u 4v 7w 0 4u 4v 4w : u, v, w R . 3u 3v 3w Is this set of vectors a subspace of R 4 ? If so, explain why, give a basis for the subspace and find its dimension. It is of dimension 2 and a basis is 4 0 4 3 7 , 0 4 3 61. 146] Solution: Consider the vectors of the form "w 2u 2v 4w 0 : u, v, w R . "u " v " w Is this set of vectors a subspace of R 4 ? If so, explain why, give a basis for the subspace and find its dimension. It is of dimension 2 and a basis is "1 0 2 0 , "1 4 0 "1 62. 147] Solution: Consider the vectors of the form 2u v w " 2v " 3u : u, v, w R . 2w " 6v " 8u Is this set of vectors a subspace of R 3 ? If so, explain why, give a basis for the subspace and find its dimension. It is of dimension 3 and a basis is 2 "3 "8 1 , "2 "6 0 , 1 2 63. 148] Solution: If you have 5 vectors in F 5 and the vectors are linearly independent, can it always be concluded they span F 5 ? Explain. Yes. If not, there would exist a vector not in the span. But then you could add in this vector and obtain a linearly independent set of vectors with more vectors than a basis. 64. 149] Solution: If you have 6 vectors in F 5 , is it possible they are linearly independent? Explain. No. If so, you would have a longer independent set than a spanning set. 65. 150] Solution: Suppose A is an m n matrix and £w 1 , C, w k ¤ is a linearly independent set of vectors in AF n F m . Now suppose Az i w i . Show £z 1 , C, z k ¤ is also independent. k Say ! i1 c i z i 0. Then you can do A to it. k ! c i Az i i1 k ! ci wi 0 i1 and so, by linear independence of the w i , it follows that each c i 0. 66. 151] Solution: Suppose V, W are subspaces of F n . Show V 9 W defined to be all vectors which are in both V and W is a subspace also. This is obvious. If x, y V 9 W, then for scalars ), *, the linear combination )x *y must be in both V and W since they are both subspaces. 67. 152] Solution: Suppose V and W both have dimension equal to 7 and they are subspaces of F 10 . What are the possibilities for the dimension of V 9 W? Hint: Remember that a linear independent set can be extended to form a basis. See the next problem. 68. 153] Solution: Suppose V has dimension p and W has dimension q and they are each contained in a subspace, U which has dimension equal to n where n maxp, q . What are the possibilities for the dimension of V 9 W? Hint: Remember that a linear independent set can be extended to form a basis. Let £x 1 , C, x k ¤ be a basis for V 9 W. Then there is a basis for V and W which are respectively £x 1 , C, x k , y k1 , C, y p ¤, £x 1 , C, x k , z k1 , C, z q ¤ It follows that you must have k p " k q " k t n and so you must have pq"n t k 69. 154] Solution: If b p 0, can the solution set of Ax b be a plane through the origin? Explain. No. It can’t. It does not contain 0. 70. 155] Solution: Suppose a system of equations has fewer equations than variables and you have found a solution to this system of equations. Is it possible that your solution is the only one? Explain. No. There must then be infinitely many solutions. If the system is Ax b, then there are infinitely many solutions to Ax 0 and so the solutions to Ax b are a particular solution to Ax b added to the solutions to Ax 0 of which there are infinitely many. 71. 156] Solution: Suppose a system of linear equations has a 2 4 augmented matrix and the last column is a pivot column. Could the system of linear equations be consistent? Explain. No. This would lead to 0 1. 72. 157] 73. 74. 75. 76. Solution: Suppose the coefficient matrix of a system of n equations with n variables has the property that every column is a pivot column. Does it follow that the system of equations must have a solution? If so, must the solution be unique? Explain. Yes. It has a unique solution. 158] Solution: Suppose there is a unique solution to a system of linear equations. What must be true of the pivot columns in the augmented matrix. The last one must not be a pivot column and the ones to the left must each be pivot columns. 159] Solution: State whether each of the following sets of data are possible for the matrix equation Ax b. If possible, describe the solution set. That is, tell whether there exists a unique solution no solution or infinitely many solutions. a. A is a 5 6 matrix, rankA 4 and rankA|b 4. Hint: This says b is in the span of four of the columns. Thus the columns are not independent. i. Infinite solution set. b. A is a 3 4 matrix, rankA 3 and rankA|b 2. i. This surely can’t happen. If you add in another column, the rank does not get smaller. c. A is a 4 2 matrix, rankA 4 and rankA|b 4. Hint: This says b is in the span of the columns and the columns must be independent. i. You can’t have the rank equal 4 if you only have two columns. d. A is a 5 5 matrix, rankA 4 and rankA|b 5. Hint: This says b is not in the span of the columns. i. In this case, there is no solution to the system of equations represented by the augmented matrix. e. A is a 4 2 matrix, rankA 2 and rankA|b 2. In this case, there is a unique solution since the columns of A are independent. 160] Solution: Suppose A is an m n matrix in which m t n. Suppose also that the rank of A equals m. Show that A maps F n onto F m . Hint: The vectors e 1 , C, e m occur as columns in the row reduced echelon form for A. This says that the columns of A have a subset of m vectors which are linearly independent. Therefore, this set of vectors is a basis for F m . It follows that the span of the columns is all of F m . Thus A is onto. 161] Solution: Suppose A is an m n matrix in which m u n. Suppose also that the rank of A equals n. Show that A is one to one. Hint: If not, there exists a vector, x such that Ax 0, and this implies at least one column of A is a linear combination of the others. Show this would require the column rank to be less than n. The columns are independent. Therefore, A is one to one. 77. 162] Solution: Explain why an n n matrix, A is both one to one and onto if and only if its rank is n. The rank is n is the same as saying the columns are independent which is the same as saying A is one to one which is the same as saying the columns are a basis. Thus the span of the columns of A is all of F n and so A is onto. If A is onto, then the columns must be linearly independent since otherwise the span of these columns would have dimension less than n and so the dimension of F n would be less than n. 78. 163] Solution: Suppose A is an m n matrix and B is an n p matrix. Show that dimkerAB t dimkerA dimkerB . Hint: Consider the subspace, BF p 9 kerA and suppose a basis for this subspace is £w 1 , C, w k ¤. Now suppose £u 1 , C, u r ¤ is a basis for kerB . Let £z 1 , C, z k ¤ be such that Bz i w i and argue that kerAB spanu 1 , C, u r , z 1 , C, z k . Here is how you do this. Suppose ABx 0. Then Bx kerA 9 BF p and so Bx ! i1 Bz i showing that k k x " ! z i kerB . i1 Consider BF 9 kerA and let a basis be £w 1 , C, w k ¤. Then each w i is of the form Bz i w i . Therefore, £z 1 , C, z k ¤ is linearly independent and ABz i 0. Now let £u 1 , C, u r ¤ be a basis for k kerB . If ABx 0, then Bx kerA 9 BF p and so Bx ! i1 c i Bz i which implies p k x " ! c i z i kerB i1 and so it is of the form k x " ! cizi i1 r ! d juj j1 It follows that if ABx 0 so that x kerAB , then x spanz 1 , C, z k , u 1 , C, u r . Therefore, dimkerAB t k r dimBF p 9 kerA dimkerB t dimkerA dimkerB 79. 164] Solution: Explain why Ax 0 always has a solution even when A "1 does not exist. a. Just let x 0. Then this solves the equation. b. What can you conclude about A if the solution is unique? i. You can conclude that the columns of A are linearly independent and so A "1 exists. c. What can you conclude about A if the solution is not unique? i. You can conclude that the columns of A are dependent and so A "1 does not exist. Thus A is not one to one and A is not onto. 80. 165] Solution: Suppose detA " 5I 0. Show there exists x p 0 such that A " 5I x 0. If detA " 5I 0 then A " 5I "1 does not exist and so the columns are not independent which means that for some x p 0, A " 5I x 0. 81. 166] Solution: Let A be an n n matrix and let x be a nonzero vector such that Ax 5x for some scalar 5. When this occurs, the vector x is called an eigenvector and the scalar, 5 is called an eigenvalue. It turns out that not every number is an eigenvalue. Only certain ones are. Why? Hint: Show that if Ax 5x, then A " 5I x 0. Explain why this shows that A " 5I is not one to one and not onto. Now argue detA " 5I 0. What sort of equation is this? How many solutions does it have? If A is n n, then the equation is a polynomial equation of degree n. 82. 167] Solution: Let m n and let A be an m n matrix. Show that A is not one to one. Hint: Consider the n n matrix, A 1 which is of the form A A1 q 0 where the 0 denotes an n " m n matrix of zeros. Thus det A 1 0 and so A 1 is not one to one. Now observe that A 1 x is the vector, Ax A1x 0 which equals zero if and only if Ax 0. Do this using the Fredholm alternative. Since A 1 is not one to one, it follows there exists x p 0 such that A 1 x 0. Hence Ax 0 although x p 0 so it follows that A is not one to one. From another point of view, if A were one to one, then kerA H R n and so by the Fredholm alternative, A T would be onto R n . However, A T has only m columns so this cannot take place. 83. 168] Solution: Let A be an m n real matrix and let b R m . Show there exists a solution, x to the system A T Ax A T b Next show that if x, x 1 are two solutions, then Ax Ax 1 . Hint: First show that A T A T A T A. Next show if x kerA T A , then Ax 0. Finally apply the Fredholm alternative. This will give existence of a solution. That A T A T A T A follows from the properties of the transpose. Therefore, ker A T A T H kerA T A H Suppose A T Ax 0. Then A T Ax, x Ax, Ax and so Ax 0. Therefore, A T b, x b, Ax b, 0 0 It follows that A T b ker A T A T H and so there exists a solution x to the equation A T Ax A T b by the Fredholm alternative. 84. 169] Solution: Show that in the context of the above problem that if x is the solution there, then |b " Ax| t |b " Ay| for every y. Thus Ax is the point of AR n which is closest to b of every point in AR n . |b "Ay|2 |b "Ax Ax "Ay|2 |b "Ax|2 |Ax " Ay|2 2b "Ax, Ax " y |b "Ax|2 |Ax " Ay|2 2A T b "A T Ax, x " y |b "Ax|2 |Ax " Ay|2 and so, Ax is closest to b out of all vectors Ay. 85. 170] Solution: 2 Let A be an n n matrix and consider the matrices I, A, A 2 , C, A n . Explain why there exist scalars, c i not all zero such that n2 ! c i A i 0. i1 Then argue there exists a polynomial, p5 of the form 5 m d m"1 5 m"1 C d 1 5 d 0 such that pA 0 and if q5 is another polynomial such that qA 0, then q5 is of the form p5 l5 for some polynomial, l5 . This extra special polynomial, p5 is called the minimal 2 polynomial. Hint: You might consider an n n matrix as a vector in F n . 2 The dimension of F n is n 2 . Therefore, there exist scalars c k such that n2 ! ckAk 0 k0 Let p5 be the monic polynomial having smallest degree such that pA 0. If qA 0 then from the Euclidean algorithm, q5 p5 l5 r5 where the degree of r5 is less than the degree of p5 or else r5 equals 0. However, if it is not zero, you could plug in A and obtain 0 qA 0 rA and this would contradict the definition of p5 as being the polynomial having smallest degree which sends A to 0. Hence q5 p5 l5 . Linear Transformations 1. 171] Solution: Show the map T : R n v R m defined by Tx Ax where A is an m n matrix and x is an m 1 column vector is a linear transformation. This is obvious from the properties of matrix multiplication. 2. 172] Solution: Find the matrix for the linear transformation which rotates every vector in R 2 through an angle of " 14 =. 1 2 " 12 2 2 1 2 1 2 2 2 3. 173] Solution: Find the matrix for the linear transformation which rotates every vector in R 2 through an angle of 5=/12. Hint: Note that 5=/12 2=/3 " =/4. 1 4 1 4 6 " 2 1 4 1 4 2 " 14 2 " 6 1 4 6 " 1 4 1 4 6 2 4. 174] Solution: Find the matrix for the linear transformation which rotates every vector in R 2 through an angle of =. "1 0 0 "1 5. 175] Solution: Find the matrix for the linear transformation which rotates every vector in R 2 through an angle of 3 =. 2 0 1 "1 0 6. 176] Solution: Show the map T : R n v R m defined by Tx Ax where A is an m n matrix and x is an m 1 column vector is a linear transformation. This is obvious from the properties of matrix multiplication. 7. 177] Solution: Find the matrix for the linear transformation which rotates every vector in R 2 through an angle of " 16 =. 1 2 1 2 3 " 12 1 2 3 8. 178] Solution: Find the matrix for the linear transformation which rotates every vector in R 2 through an angle of 4 =. 3 2 1 4 3 1 4 2 1 4 3 " 1 4 2 " 1 4 1 4 2 1 4 2 3 3 1 4 9. 179] Solution: Find the matrix for the linear transformation which rotates every vector in R 2 through an angle of 1 = and then reflects through the y axis. 6 "1 0 0 1 2 3 1 2 1 " 12 1 2 3 " 12 3 1 2 1 2 1 2 3 10. 180] Solution: Find the matrix for the linear transformation which rotates every vector in R 2 through an angle of 2 = and then reflects through the y axis. 3 1 0 0 "1 " 12 1 2 " 12 3 3 " 12 " 12 " 12 3 " 12 3 1 2 11. 181] Solution: Find the matrix for the linear transformation which reflects through the y axis and then rotates every vector in R 2 through an angle of 12 =. 0 "1 "1 0 1 0 0 1 0 "1 "1 0 12. 182] Solution: Find the matrix for the linear transformation which rotates every vector in R 2 through an angle of 2 = and then reflects through the y axis. 3 "1 0 0 1 " 12 1 2 3 " 12 3 " 12 1 2 1 2 1 2 3 3 " 12 13. 183] Solution: Find the matrix for the linear transformation which rotates every vector in R 2 through an angle of " 12 = and then reflects through the x axis. 1 0 0 0 "1 1 0 1 "1 0 1 0 14. 184] Solution: Find the matrix for the linear transformation which reflects through the x axis and then rotates every vector in R 2 through an angle of 34 =. " 12 2 " 12 2 1 2 1 " 12 2 0 "1 " 12 2 2 0 1 2 1 2 1 2 2 2 2 15. 185] Solution: Find the matrix for the linear transformation which reflects through the x axis and then rotates every vector in R 2 through an angle of 13 =. " 12 3 1 2 1 2 1 0 "1 1 2 3 0 1 2 1 2 1 2 3 " 12 3 16. 186] Solution: Recall the special vectors i, j, k such that i points in the direction of the positive x axis, j in the direction of the positive y axis and k in the direction of the positive z axis. The vectors i, j, k form a right handed system as shown in the following picture. Find the matrix of the linear transformation which rotates every vector in R 3 through an angle of " 16 = about the positive z axis when viewed from high on the positive z axis and then reflects through the xy plane. 1 0 0 1 2 0 1 0 " 12 0 0 "1 0 1 2 3 1 2 3 0 0 1 2 0 1 1 2 3 " 12 0 1 2 0 3 0 0 "1 17. 187] Solution: Recall the special vectors i, j, k such that i points in the direction of the positive x axis, j in the direction of the positive y axis and k in the direction of the positive z axis. The vectors i, j, k form a right handed system as shown in the following picture. Find the matrix of the linear transformation which rotates every vector in R 3 through an angle of 1 = about the positive x axis when viewed from the positive x axis and then reflects through the yz 3 plane. "1 0 0 1 0 0 0 1 0 0 1 2 " 12 3 0 0 1 0 1 2 3 1 2 "1 0 0 0 1 2 " 12 3 0 1 2 3 1 2 18. 188] Solution: Recall the special vectors i, j, k such that i points in the direction of the positive x axis, j in the direction of the positive y axis and k in the direction of the positive z axis. The vectors i, j, k form a right handed system as shown in the following picture. Find the matrix of the linear transformation which rotates every vector in R 3 through an angle of 1 = about the positive z axis when viewed from high on the positive z axis and then reflects 2 through the xy plane. 1 0 0 0 "1 0 0 1 0 1 0 0 0 0 "1 0 0 1 0 "1 0 1 0 0 0 0 "1 19. 189] Solution: Recall the special vectors i, j, k such that i points in the direction of the positive x axis, j in the direction of the positive y axis and k in the direction of the positive z axis. The vectors i, j, k form a right handed system as shown in the following picture. Find the matrix of the linear transformation which rotates every vector in R 3 through an angle of 3 = about the positive z axis when viewed from high on the positive z axis and then reflects 2 through the xy plane. 1 0 0 0 0 1 0 "1 0 0 0 0 "1 0 1 0 0 0 1 1 0 "1 0 0 0 0 "1 20. 190] Solution: Recall the special vectors i, j, k such that i points in the direction of the positive x axis, j in the direction of the positive y axis and k in the direction of the positive z axis. The vectors i, j, k form a right handed system as shown in the following picture. Find the matrix of the linear transformation which rotates every vector in R 3 through an angle of 1 = about the positive z axis when viewed from high on the positive z axis and then reflects 2 through the yz plane. "1 0 0 0 "1 0 0 1 0 1 0 0 0 0 1 0 0 1 0 1 0 1 0 0 0 0 1 21. 191] Solution: Recall the special vectors i, j, k such that i points in the direction of the positive x axis, j in the direction of the positive y axis and k in the direction of the positive z axis. The vectors i, j, k form a right handed system as shown in the following picture. Find the matrix of the linear transformation which reflects every vector in R 3 through the xy plane and then rotates every vector in R 3 through an angle of 14 = about the positive y axis when viewed from the positive y axis and then reflects through the xz plane. 1 0 0 1 2 0 "1 0 0 0 1 2 0 0 1 2 1 " 12 2 0 2 0 1 2 1 0 0 0 1 0 0 0 "1 2 0 " 12 2 0 "1 0 " 12 2 0 " 12 2 1 2 2 22. 192] Solution: Recall the special vectors i, j, k such that i points in the direction of the positive x axis, j in the direction of the positive y axis and k in the direction of the positive z axis. The vectors i, j, k form a right handed system as shown in the following picture. Find the matrix of the linear transformation which rotates every vector in R 3 through an angle of 1 = about the positive y axis when viewed from the positive y axis and then reflects through the xz 2 plane. 1 0 0 0 0 1 0 "1 0 0 1 0 0 "1 0 0 0 1 0 0 1 0 "1 0 "1 0 0 23. 193] Solution: Recall the special vectors i, j, k such that i points in the direction of the positive x axis, j in the direction of the positive y axis and k in the direction of the positive z axis. The vectors i, j, k form a right handed system as shown in the following picture. Find the matrix of the linear transformation which reflects every vector in R 3 through the xz plane and then rotates every vector through an angle of 14 = about the positive z axis when viewed from high on the positive z axis. 1 2 2 " 12 2 0 1 2 2 1 2 0 2 0 1 0 0 0 0 "1 0 1 0 0 1 2 2 1 2 2 " 12 2 0 1 1 2 0 2 0 0 1 24. 194] Solution: T Find the matrix for proj u v where u Columns are 1 11 1 11 3 11 1 11 1 11 3 11 . 2 2 6 3 11 3 11 9 11 1 11 1 11 3 11 , concatenate: 1 11 1 11 3 11 3 11 3 11 9 11 25. 195] Solution: T Find the matrix for proj u v where u Columns are 0 0 0 0 4 5 2 5 2 5 1 5 0 0 2 1 . 0 0 0 4 5 2 5 2 5 1 5 0 , concatenate: 0 26. 196] Solution: Find the matrix for proj u v where u Columns are 2 7 1 7 " 37 1 7 1 14 3 " 14 "2 "1 3 T . " 37 3 " 14 , concatenate: 9 14 2 7 1 7 " 37 " 37 1 7 1 14 3 " 14 3 " 14 9 14 27. 197] Solution: Find the matrix for proj u v where u Columns are 4 9 " 29 " 49 2 "1 "2 " 29 " 49 1 9 2 9 2 9 4 9 T . , concatenate: 4 9 " 29 " 49 " 29 " 49 1 9 2 9 2 9 4 9 28. 198] Solution: Let M be the matrix for proj u v , u R n , n 1. Explain why detM 0 The columns of M are all multiples of a single vector so they are linearly dependent. 29. 199] Solution: Show that the function T u defined by T u v q v " proj u v is also a linear transformation. av bw u u |u|2 v u w u av " a u bw "b u 2 |u| |u|2 aT u v bT u w T u av bw av bw " 30. 200] Solution: Show that v " proj u v , u 0 and conclude every vector in R n can be written as the sum of two vectors, one which is perpendicular and one which is parallel to the given vector. v " proj u v , u v, u " vu |u|2 v, u " v, u 0. u, u Therefore, v v " proj u v proj u v . The first is perpendicular to u and the second is a multiple of u so it is parallel to u. 31. 201] Solution: Here are some descriptions of functions mapping R n to R n . a. T multiplies the j th component of x by a nonzero number b. b. T replaces the i th component of x with b times the j th component added to the i th component. c. T switches two components. Show these functions are linear and describe their matrices. Each of these is an elementary matrix. The first is the elementary matrix which multiplies the j th diagonal entry of the identity matrix by b. The second is the elementary matrix which takes b times the j th row and adds to the i th row and the third is just the elementary matrix which switches the i th and the j th rows where the two components are in the i th and j th positions. 32. 202] Solution: Let u a, b be a unit vector in R 2 . Find the matrix which reflects all vectors across this vector. Hint: You might want to notice that a, b cos 2, sin 2 for some 2. First rotate through "2. Next reflect through the x axis which is easy. Finally rotate through 2. a. cos2 " sin2 1 sin2 0 "1 cos2 0 cos 2 2 " sin 2 2 2 cos 2 sin 2 2 cos 2 sin 2 sin 2 2 " cos 2 2 cos"2 " sin"2 sin"2 cos"2 Now to write in terms of a, b , note that a/ a 2 b 2 cos 2, b/ a 2 b 2 sin 2. Now plug this in to the above. The result is a 2 "b 2 a 2 b 2 2 ab a 2 b 2 2 a 2ab b 2 b 2 "a 2 a 2 b 2 1 a2 b2 Since this is a unit vector, a 2 b 2 1 and so you get a2 " b2 2ab 2ab b " a2 2 a2 " b2 2ab 2ab b " a2 2 33. 203] Solution: Let u be a unit vector. Show the linear transformation of the matrix I " 2uu T preserves all distances and satisfies T I " 2uu T I " 2uu T I. This matrix is called a Householder reflection. More generally, any matrix Q which satisfies Q T Q QQ T is called an orthogonal matrix. Show the linear transformation determined by an orthogonal matrix always preserves the length of a vector in R n . Hint: First show that for any matrix A, Ax, y x, A T y T I " 2uu T I " 2uu T I " 2uu T I " 2uu T 1 ¥ I " 2uu T " 2uu T 4u u T u u T I Now, why does this matrix preserve distance? For short, call it Q and note that Q T Q I. Then |x|2 Q T Qx, x Qx, Qx |Qx|2 and so Q preserves distances. 34. 204] Solution: Suppose |x| |y| for x, y R n . The problem is to find an orthogonal transformation Q Q T Q I which has the property that Qx y and Qy x. Show x"y Q q I"2 " y T 2 x |x " y| does what is desired. From the above problem this preserves distances and Q T Q. Now do it to x. x"y Qx " y x " y " 2 x " y, x " y y " x |x " y|2 x"y Qx y x y " 2 " y T x y 2 x |x " y| x"y 2 " |y|2 x y xy"2 2 |x| |x " y| and so Qx " Qy y " x Qx Qy x y Hence, adding these yields Qx y and then subtracting them gives Qy x. 35. 205] Solution: Let a be a fixed vector. The function T a defined by T a v a v has the effect of translating all vectors by adding a. Show this is not a linear transformation. Explain why it is not possible to realize T a in R 3 by multiplying by a 3 3 matrix. Linear transformations take 0 to 0. Also T a u v p T a u T a v. 36. 206] Solution: In spite of the above problem, we can represent both translations and rotations by matrix multiplication at the expense of using higher dimensions. This is done by the homogeneous coordinates. I will illustrate in R 3 where most interest in this is found. For each vector v v 1 , v 2 , v 3 T , consider the vector in R 4 v 1 , v 2 , v 3 , 1 T . What happens when you do 1 0 0 a1 v1 0 1 0 a2 v2 0 0 1 a3 v3 0 0 0 1 1 ? Describe how to consider both rotations and translations all at once by forming appropriate 4 4 matrices. That product above is of the form a1 v1 a2 v2 a3 v3 1 If you just discard the one at the bottom, you have found a v. Then to do both rotations and translations, you would look at matrices of the form R 0 0 1 where R is a rotation matrix and for translation by a, you use I a 0 1 To obtain a rotation followed by a translation by a, you would just multiply these two matrices. to get If you did this to the vector x 1 I a R 0 0 1 0 1 , you would get R a 0 1 Rx a 1 . Now discard the 1 and you have what you want. 37. 207] Solution: You want to add 3 "1 "2 to every point in R 3 and then rotate about the z axis counter clockwise through an angle of 1 4 2 " 12 2 0 0 1 2 1 2 1 2 2 2 =. Find what happens to the point 1 0 0 3 1 0 0 0 1 0 "1 3 0 0 1 0 0 0 1 "2 4 0 0 0 1 0 0 0 1 1 1 3 4 . 2 3 2 and so the answer is 2 1 2 . 3 2 2 38. 208] Solution: You want to add clockwise through an angle of " 12 to every point in R 3 and then rotate about the z axis counter "1 "1 "3 2 3 " 12 3 0 0 =. Find what happens to the point 1 0 0 "1 3 3 1 2 . "1 " 12 0 0 0 1 0 "1 1 0 0 1 0 0 0 1 "3 2 0 0 0 1 0 0 0 1 1 1 "2 to every point in R 3 and then rotate about the z axis counter 1 2 3 1 3 "1 and so the answer is 1 "1 . 3 "1 39. 209] Solution: You want to add clockwise through an angle of 1 2 =. Find what happens to the point 0 "1 0 0 1 0 0 1 3 "3 1 0 0 0 0 1 0 1 2 4 0 0 1 0 0 0 1 "2 3 0 0 0 1 0 0 0 1 1 1 1 3 2 3 . "3 and so the answer is 4 1 40. 210] Solution: You want to rotate about the z axis counter clockwise through an angle of 13 = and then add 3 3 "2 to every point in R 3 . Find what happens to the point 1 2 1 . . 1 0 0 3 0 1 0 3 " 12 3 0 0 1 2 1 2 3 1 2 0 0 2 0 0 1 "2 0 0 1 0 1 0 0 0 0 0 0 1 1 1 " 3 7 2 1 2 1 3 4 "1 . 1 41. 211] Solution: You want to rotate about the z axis counter clockwise through an angle of 54 = and then add . "3 0 "1 to every point in R 3 . Find what happens to the point 2 1 4 1 0 0 "3 " 12 2 0 1 0 " 12 2 0 1 2 " 12 2 0 0 2 2 0 0 1 0 0 1 "1 0 0 1 0 4 0 0 0 0 0 0 1 1 1 " 12 2 " 3 " 32 2 3 1 42. 212] Solution: Give the general solution to the following system of equations wx"z 0 x"wy 0 x"wy 0 by considering it as finding the solution to a system of the form Ax 0 Now, using what you just did, find the general solution to the system wx"z3 0 x"wy2 0 x"wy2 0 The solution to the first system of equations is x y z "1 1 "1 s 1 w 2 t , s, t R 0 0 1 The general solution to the second system is x y z w "1 1 s "1 1 0 t 2 0 1 "3 1 0 0 . 43. 213] Solution: Give the general solution to the following system of equations x " 3w 3z 0 x " 4w y 7z 0 x " 4w y 7z 0 by considering it as finding the solution to a system of the form Ax 0 Now, using what you just did, find the general solution to the system x " 3w 3z 2 0 x " 4w y 7z 1 0 x " 4w y 7z 1 0 The solution to the first system of equations is "3 x y z "4 s 1 w 3 1 t , s, t R 0 0 1 The general solution to the second system is "3 x y z w s "4 1 0 "2 3 t 1 0 1 1 0 0 44. 214] Solution: Give the general solution to the following system of equations x " w " 2z 0 x " 2w y 2z 0 x " 2w y 2z 0 by considering it as finding the solution to a system of the form Ax 0 Now, using what you just did, find the general solution to the system x " w " 2z 1 0 x " 2w y 2z " 2 0 x " 2w y 2z " 2 0 The solution to the first system of equations is x y z 2 "4 s 1 w 1 1 t , s, t R 0 0 1 The general solution to the second system is x 2 y "4 s z 1 w "1 1 t 0 1 3 0 0 1 0 45. 215] Solution: Give the general solution to the following system of equations 3w x " z 0 8w x y 3z 0 8w x y 3z 0 by considering it as finding the solution to a system of the form Ax 0 Now, using what you just did, find the general solution to the system 3w x " z " 2 0 8w x y 3z " 4 0 8w x y 3z " 4 0 The solution to the first system of equations is x y z "3 1 s "4 1 w t "5 0 0 , s, t R 1 The general solution to the second system is x y z w 46. 216] Solution: "3 1 s "4 1 0 t "5 0 1 2 2 0 0 Give the general solution to the following system of equations 3w x " 2z 0 8w x y " 4z 0 8w x y " 4z 0 by considering it as finding the solution to a system of the form Ax 0 Now, using what you just did, find the general solution to the system 3w x " 2z " 1 0 8w x y " 4z " 2 0 8w x y " 4z " 2 0 The solution to the first system of equations is x y z "3 2 2 s 1 w t "5 0 0 , s, t R 1 The general solution to the second system is x y z w "3 2 s 2 1 0 t "5 0 1 1 1 0 0 47. 217] Solution: Give the general solution to the following system of equations 3w x " z 0 5w x y 0 5w x y 0 by considering it as finding the solution to a system of the form Ax 0 Now, using what you just did, find the general solution to the system 3w x " z " 1 0 5w x y " 3 0 5w x y " 3 0 The solution to the first system of equations is x y z "3 1 "1 s 1 w "2 t 0 0 , s, t R 1 The general solution to the second system is x y z "3 1 s w "1 1 1 "2 t 0 0 1 2 0 0 48. 218] Solution: Find kerA NA where 2 1 3 "3 6 3 2 7 "4 10 A 2 1 3 "3 6 3 1 2 "5 8 x1 1 2 "2 x2 "5 "1 "2 x3 s 1 t r 0 0 x4 0 1 0 x5 0 0 1 49. 219] Solution: Find kerA NA where 2 1 11 "2 A 6 3 2 19 "2 10 2 1 11 "2 6 3 1 14 "4 8 x1 "3 2 "2 x2 "5 "2 "2 x3 s t 1 r 0 0 x4 0 1 0 x5 0 0 1 50. 220] Solution: Find kerA NA where 2 1 5 A 9 0 3 2 7 15 1 2 1 5 9 1 3 1 8 12 1 x1 "3 "3 x2 1 "3 x3 s t 1 0 x4 1 1 x5 0 0 51. 221] Solution: Find kerA NA where A 2 1 0 10 0 3 2 1 17 1 2 1 0 10 1 3 1 "1 13 1 x1 1 "3 x2 "2 "4 x3 52. 222] Solution: Find kerA NA where s 1 t 0 x4 1 1 x5 0 0 2 1 "1 "3 0 3 2 A 1 "3 1 2 1 "1 "3 1 3 1 "4 "6 1 x1 3 3 x2 "5 "3 x3 s 1 t 0 x4 1 1 x5 0 0 53. 223] Solution: Find kerA NA where "5 7 3 2 15 "7 9 7 2 1 A 8 2 1 8 "5 3 1 9 "8 12 x1 "1 3 "5 x2 "6 "1 3 x3 s 1 t r 0 0 x4 0 1 0 x5 0 0 1 54. 224] Solution: Find kerA NA where 2 1 10 "2 A 0 3 2 17 "1 "1 2 1 10 "2 0 3 1 13 "5 1 x1 "3 3 "1 x2 "4 "4 2 x3 s 1 t 0 r 0 x4 0 1 0 x5 0 0 1 55. 225] Solution: Suppose Ax b has a solution. Explain why the solution is unique precisely when Ax 0 has only the trivial (zero) solution. If not, then there would be a infintely many solutions to Ax 0 and each of these added to a solution to Ax b would be a solution to Ax b. 56. 226] Solution: Show that if A is an m n matrix, then kerA is a subspace. If x, y kerA then Aax by aAx bAy a0 b0 0 and so kerA is closed under linear combinations. Hence it is a subspace. 57. 227] Solution: Give an example of a 2 3 matrix which maps R 3 onto R 2 . Is it possible that there could exist an example of such a matrix which is one to one? There is no one to one such matrix because there must be a free variable. Also, to be one to one, the columns would have to be an independent set and this is not possible. Factorizations 1. 228] Solution: Find an LU factorization for the matrix 5 1 2 2 5 3 7 3 "5 3 9 1 10 10 29 12 and use to compute the determinant of the matrix. "10 2. 229] Solution: Find an LU factorization for the matrix 1 1 2 1 1 "3 "1 "1 "2 "1 "3 "1 "4 "1 "4 2 1 0 0 0 1 1 2 1 1 "3 1 0 0 0 2 5 1 2 "3 1 1 0 0 0 "3 1 "3 2 0 0 "3 1 4 5 1 0 10 9 8 "4 3. 230] Solution: Find an LU factorization for the matrix 1 2 "3 "5 "2 5 4 5 20 7 "5 "5 1 0 0 0 5 1 2 "3 "1 1 0 0 0 "1 7 1 4 "3 1 0 0 0 "1 1 "1 4 0 0 0 4 7 1 4. 231] Solution: Find an LU factorization for the matrix "14 "14 19 18 5 1 2 1 1 5 4 7 2 3 "15 0 15 1 0 0 0 5 1 2 1 1 1 1 0 0 0 3 5 1 2 "3 1 1 0 0 0 1 1 "3 3 0 0 0 "3 1 2 5 1 0 "1 "4 9 21 "7 7 5. 232] Solution: Find an LU factorization for the matrix 4 1 2 3 "8 "4 2 "5 "4 "7 14 1 "8 "12 14 4 and use to compute the determinant of the matrix. "16 6. 233] Solution: Find an LU factorization for the matrix 3 1 2 "3 6 2 9 "5 18 6 "1 "20 0 1 0 0 0 3 1 2 "3 2 1 0 0 0 0 5 1 6 "3 1 0 0 0 2 1 0 0 0 0 6 4 5 1 7. 234] Solution: Find an LU factorization for the matrix 0 30 15 4 1 2 "3 "4 "4 2 4 4 10 "11 "5 "12 "21 1 0 0 0 4 1 2 "3 "1 1 0 0 0 "3 4 1 1 "3 1 0 0 0 "1 1 "3 6 0 0 0 1 4 1 14 20 8. 235] Solution: Find an LU factorization for the matrix 4 1 2 3 "4 "1 5 "2 "12 "3 14 "5 0 1 0 0 0 4 1 2 3 "1 1 0 0 0 0 7 1 "3 3 1 0 0 0 "1 1 0 0 0 "3 7 7 1 0 0 42 11 9. 236] Solution: Is there only one LU factorization for a given matrix? Hint: Consider the equation 0 1 0 1 1 0 0 1 1 1 0 0 . Sometimes there is more than one LU factorization as is the case in this example. The above equation clearly gives an LU factorization. However, it appears that 0 1 0 1 1 0 0 1 0 1 0 1 also. Therefore, this is another. 10. 237] Solution: Computer algebra systems are very good at finding PLU factorizations. Describe how to use a PLU factorization to compute the determinant of a matrix. You let m be the number of switches used to form P and then the determinant is just "1 m times the product of the diagonal entries of U. This is because if A PLU, then detA detP detL detU "1 m 1 above product of diagonal entries. 11. 238] Solution: 1 2 1 Find a PLU factorization of . 1 2 2 2 1 1 1 2 1 1 2 2 2 1 1 1 0 0 1 0 0 1 2 1 0 0 1 2 1 0 0 "3 "1 0 1 0 1 0 1 0 0 1 12. 239] Solution: 1 2 1 1 2 2 Find a PLU factorization of . 2 4 1 3 2 1 1 2 1 1 2 2 2 4 1 3 2 1 1 0 0 0 1 0 0 0 1 2 1 0 0 0 1 3 1 0 0 0 "4 "2 0 0 1 0 2 0 1 0 0 0 "1 0 1 0 0 1 0 "1 1 0 0 0 Here are steps for doing this. First use the top row to zero out the entries in the first column which are below the top row. This yields 1 2 1 0 0 1 0 0 "1 0 "4 "2 Obviously there will be no way to obtain an LU factorization because a switch of rows must be done. Switch the second and last row in the original matrix. This will yield one which does have an LU factorization. 1 0 0 0 1 2 1 0 0 0 1 1 2 2 0 0 1 0 2 4 1 0 1 0 0 3 2 1 1 2 1 3 2 1 2 4 1 1 2 2 1 0 0 0 1 2 1 3 1 0 0 0 "4 "2 2 0 1 0 0 0 "1 1 0 "1 1 0 0 0 Then the original matrix is 1 2 1 1 2 2 2 4 1 3 2 1 1 0 0 0 1 0 0 0 1 2 0 0 0 1 3 1 0 0 0 "4 "2 0 0 1 0 2 0 1 0 0 0 "1 0 1 0 0 1 0 "1 1 0 0 0 13. 240] Solution: 1 2 1 2 4 1 Find a PLU factorization of and use it to solve the systems 1 0 2 2 2 1 1 2 1 a. 2 4 1 1 0 2 2 2 1 1 2 1 2 4 1 1 0 2 2 2 1 z 1 0 0 0 1 2 1 0 0 1 0 1 1 0 0 0 "2 1 0 1 0 0 2 0 1 0 0 0 "1 0 0 0 1 2 1 2 1 0 0 0 1 x y 1 0 0 0 2 1 1 a. First solve 1 0 0 0 1 0 0 0 t 0 0 1 0 1 1 0 0 u 0 1 0 0 2 0 1 0 v 0 0 0 1 2 1 2 1 w 1 t 2 0 1 0 u 1 1 0 0 v 2 1 2 1 w 1 2 1 1 Thus t 1, v 0, u 0, w "1. Next solve 1 2 1 0 "2 1 0 0 "1 0 0 0 1 x y z 1 1 This is not too hard because it is the same as solving 1 0 0 0 2 0 0 "1 1 which clearly has no solution. Thus the original problem has no solution either. Note that the augmented matrix has the following row reduced echelon form. 1 2 1 1 2 4 1 2 1 0 2 1 1 0 0 0 0 1 0 0 , row echelon form: 0 0 1 0 2 2 1 1 0 0 0 1 i. Thus there is no solution. 1 2 1 b. 2 4 1 1 0 2 2 2 1 a x b y c z d There might not be any solution from part a. Thus suppose there is. First solve t c. u v w 1 0 0 0 t 2 0 1 0 u 1 1 0 0 v 2 1 2 1 w a b c d a c"a . Next solve b " 2a 3a " 2b " c d 1 2 1 0 "2 1 0 0 "1 0 0 0 a x c"a y b " 2a z 3a " 2b " c d Note that there will be no solution unless 3a " 2b " c d 0, but if this condition holds, then the solution to the problem is 2b " 4a c x y z To check this, note that 3 2 a" 1 2 b" 2a " b 1 2 c 1 2 1 a 2b " 4a c 2 4 1 3 2 1 0 2 a" 1 2 b" 1 2 b c c 2a " b 2 2 1 2b " 3a c where the bottom entry on the right equals d if there is any solution. 14. 241] Solution: Find a QR factorization for the matrix 1 1 1 0 0 1 0 1 1 0 1 "1 1 1 1 0 0 1 0 1 1 2 2 0 1 0 1 "1 1 2 2 1 6 1 3 " 16 2 3 2 3 2 3 1 3 " 13 " 13 3 2 3 0 3 0 2 1 1 2 1 2 2 2 2 3 0 0 0 " 12 2 1 2 2 3 0 15. 242] Solution: Find a QR factorization for the matrix 1 3 "2 1 1 1 2 3 "2 1 1 0 1 11 3 11 1 11 1 2 11 11 11 13 66 5 " 66 1 33 2 11 0 2 " 16 2 2 11 " 16 2 2 11 2 3 2 11 4 " 11 11 0 6 11 0 2 11 0 16. 243] Solution: Find a QR factorization for the matrix 1 2 1 0 1 1 1 0 2 1 2 1 2 1 0 1 1 1 0 2 17. 244] 2 0 1 2 2 1 3 1 3 " 13 3 " 16 6 3 1 3 3 1 6 2 2 2 3 0 3 6 0 0 3 2 2 0 1 2 2 3 6 11 2 11 11 2 11 2 Solution: Find a QR factorization for the matrix 1 1 1 0 1 0 1 0 2 1 2 1 1 1 0 1 0 1 0 2 1 3 0 1 2 1 6 2 2 6 2 3 " 16 6 " 13 3 1 3 1 3 1 2 2 3 0 3 0 1 2 2 2 3 0 3 2 " 16 1 3 2 6 3 18. 245] Solution: If you had a QR factorization, A QR, describe how you could use it to solve the equation Ax b. This is not usually the way people solve this equation. However, the QR factorization is of great importance in certain other problems, especially in finding eigenvalues and eigenvectors. You would have QRx b and so then you would have Rx Q T b. Now R is upper triangular and so the solution of this problem is fairly simple. 19. 246] Solution: Starting with an independent set of n vectors in R n £v 1 , C, v n ¤, form the matrix and QR factorization v1 C vn q1 C qn R where Q q q 1 C q n is the orthogonal matrix. Thus Q T Q I and the columns of Q form an orthonormal set of vectors. Show that for each k t n, span q 1 C q k span v 1 C v k . From the equation, v 1 r 11 q 1 , v 2 r 12 q 1 r 22 q 2 and so forth. Also R "1 must exist because the determinant of the left side is nonzero, and from the usual process for finding it, this matrix is also upper triangular. Thus v1 C vn R "1 q1 C qn and now a similar thing can be said. Thus in any linear combination of the v i one can replace v k with a suitable linear combination of the q j and in any linear combination of the q i one can replace q k with a suitable linear combination of the v j . Hence the desired conclusion is obtained. Linear Programming 1. 247] Solution: Maximize and minimize z x 1 " 2x 2 x 3 subject to the constraints x 1 x 2 x 3 t 10, x 1 x 2 x 3 u 2, and x 1 2x 2 x 3 t 7 if possible. All variables are nonnegative. The constraints lead to the augmented matrix 1 1 1 1 0 0 10 1 1 1 0 "1 0 2 1 2 1 0 7 0 1 The obvious solution is not feasible. Do a row operation. 1 1 1 1 0 0 10 0 0 0 1 1 0 8 0 1 0 "1 0 1 "3 An obvious solution is still not feasible. Do another operation couple of row operations. 1 1 1 0 "1 0 2 0 0 0 1 1 0 8 0 1 0 0 1 1 5 At this point, you can spot an obvious feasible solution. Now assemble the simplex tableau. 1 1 1 0 "1 0 0 2 0 0 0 1 1 0 0 8 0 1 0 0 1 1 0 5 "1 2 "1 0 0 0 1 0 Now preserve the simple columns. 1 1 1 0 "1 0 0 2 0 0 0 1 1 0 0 8 0 1 0 0 1 1 0 5 0 2 0 0 "1 0 1 0 First lets work on minimizing this. There is a 2. The ratios are then 5, 2 so the pivot is the 1 on the top of the second column. The next tableau is 1 1 1 0 "1 0 0 2 0 0 0 1 1 0 0 8 "1 0 "1 0 2 1 0 3 "2 0 "2 0 1 0 1 "4 There is a 1 on the bottom. The ratios of interest for that column are 3/2, 8, and so the pivot is the 2 in that column. Then the next tableau is 0 1 0 1 2 " 12 0 7 2 13 2 "1 0 "1 0 2 1 0 3 1 2 1 2 " 1 2 1 2 1 0 0 " 3 2 0 0 3 2 0 0 " 1 " 11 2 1 2 Now you stop because there are no more positive numbers to the left of 1 on the bottom row. The minimum is "11/2 and it occurs when x 1 x 3 x 6 0 and x 2 7/2, x 4 13/2, x 6 "11/2. Next consider maximization. The simplex tableau was 1 1 1 0 "1 0 0 2 0 0 0 1 1 0 0 8 0 1 0 0 1 1 0 5 "1 2 "1 0 0 0 1 0 This time you work on getting rid of the negative entries. Consider the "1 in the first column. There is only one ratio to consider so 1 is the pivot. 1 1 1 0 "1 0 0 2 0 0 0 1 1 0 0 8 0 1 0 0 1 1 0 5 0 3 0 0 "1 0 1 2 There remains a "1. The ratios are 5 and 8 so the next pivot is the 1 in the third row and column 5. 1 2 1 0 0 1 0 7 0 "1 0 1 0 "1 0 3 0 1 0 0 1 1 0 5 0 4 0 0 0 1 1 7 Then no more negatives remain so the maximum is 7 and it occurs when x 1 7, x 2 0, x 3 0, x 4 3, x 5 5, x 6 0. 2. 248] Solution: Maximize and minimize the following if possible. All variables are nonnegative. a. z x 1 " 2x 2 subject to the constraints x 1 x 2 x 3 t 10, x 1 x 2 x 3 u 1, and x 1 2x 2 x 3 t 7 i. an augmented matrix for the constraints is 1 1 1 1 0 0 10 1 1 1 0 "1 0 1 1 2 1 0 0 1 7 The obvious solution is not feasible. Do some row operations. 0 0 0 1 1 1 1 1 0 "1 0 1 "1 0 "1 0 0 9 2 1 5 Now the obvious solution is feasible. Then include the objective function. 0 0 0 1 1 0 0 9 1 1 1 0 "1 0 0 1 "1 0 "1 0 2 1 0 5 "1 2 0 0 0 0 1 0 1 First preserve the simple columns. 0 0 0 1 0 0 9 1 1 1 0 "1 0 0 1 "1 0 "1 0 2 1 0 "3 0 "2 0 2 0 1 "2 5 Lets try to maximize first. Begin with the first column. The only pivot is the 1. Use it. 0 0 0 1 1 0 0 9 1 1 1 0 "1 0 0 1 0 1 0 0 1 1 0 6 0 3 1 0 "1 0 1 1 There is still a -1 on the bottom row to the left of the 1. The ratios are 9 and 6 so the new pivot is the 1 on the third row. 0 "1 0 1 0 "1 0 3 1 2 1 0 0 1 0 7 0 1 0 0 1 1 0 6 0 4 1 0 0 1 1 7 Then the maximum is 7 when x 1 7 and x 1 , x 3 0. Next consider the minimum. 0 0 0 1 1 0 0 9 1 1 1 0 "1 0 0 1 "1 0 "1 0 2 1 0 "3 0 "2 0 2 0 1 "2 5 There is a positive 2 in the bottom row left of 1. The pivot in that column is the 2. 1 0 " 12 0 0 0 1 2 0 13 2 7 2 "1 0 "1 0 2 1 0 5 1 2 1 2 1 2 1 2 0 1 "2 0 "1 0 0 "1 1 "7 The minimum is "7 and it happens when x 1 0, x 2 7/2, x 3 0. b. z x 1 " 2x 2 " 3x 3 subject to the constraints x 1 x 2 x 3 t 8, x 1 x 2 3x 3 u 1, and x1 x2 x3 t 7 i. This time, lets use artificial variables to find an initial simplex tableau. Thus you add in an artificial variable and then do a minimization procedure. 1 1 1 1 0 0 0 0 8 1 1 3 0 "1 0 1 0 1 1 1 1 0 0 1 0 0 7 0 0 0 0 0 0 "1 1 0 First preserve the seventh column as a simple column by a row operation. 1 1 1 1 0 0 0 0 8 1 1 3 0 "1 0 1 0 1 1 1 1 0 0 1 0 0 7 1 1 3 0 "1 0 0 1 1 Now use the third column. 0 " 13 0 2 3 2 3 1 1 3 0 "1 0 2 3 2 3 0 1 3 0 1 1 23 3 0 1 0 20 3 0 "1 1 0 0 0 1 3 1 " 0 0 0 0 1 3 It follows that a basic solution is feasible if 2 3 2 3 1 3 0 23 3 1 1 3 0 "1 0 1 2 3 2 3 20 3 0 1 1 3 0 0 1 Now assemble the simplex tableau 2 3 2 3 0 0 23 3 1 1 3 0 "1 0 0 1 2 3 2 3 0 1 1 3 0 0 1 3 1 0 20 3 "1 2 3 0 0 0 1 0 Preserve the simple columns by doing row operations. 1 3 2 3 2 3 1 1 3 0 "1 0 0 2 3 2 3 0 1 0 0 23 3 1 0 0 1 3 20 3 1 0 "2 1 0 0 1 0 1 "1 Lets do minimization first. Work with the second column. 0 0 "2 1 1 1 0 0 "2 0 1 1 0 "3 0 "3 0 2 0 1 "2 3 1 0 0 7 0 "1 0 0 1 6 Recall how you have to pick the pivot correctly. There is still a positive number in the bottom row left of the 1. Work with that column. 0 0 0 1 0 "1 0 1 1 1 1 0 0 1 0 7 0 0 "2 0 1 1 0 6 "3 0 1 0 0 "2 1 "14 There is still a positive number to the left of 1 on the bottom row. 0 0 0 1 0 "1 0 1 1 1 1 0 0 1 0 7 2 2 0 0 1 3 0 20 "4 "1 0 0 0 "3 1 "21 It follows that the minimum is "21 and it occurs when x 1 x 2 0, x 3 7. Next consider the maximum. The simplex tableau was 2 3 2 3 0 0 23 3 1 1 3 0 "1 0 0 1 2 3 2 3 20 3 0 1 1 3 0 0 1 3 1 0 "2 1 0 0 1 0 1 "1 0 0 "2 1 1 0 0 7 Use the first column. 1 1 3 0 "1 0 0 1 0 0 "2 0 0 3 6 1 1 0 6 0 "1 0 1 1 There is still a negative on the bottom row to the left of 1. 0 0 0 1 0 "1 0 1 1 1 1 0 0 1 0 7 0 0 "2 0 1 1 0 6 0 3 1 1 7 4 0 0 There are no more negatives on the bottom row left of 1 so stop. The maximum is 7 and it occurs when x 1 7, x 2 0, x 3 0. c. z 2x 1 x 2 subject to the constraints x 1 " x 2 x 3 t 10, x 1 x 2 x 3 u 1, and x 1 2x 2 x 3 t 7. i. The augmented matrix for the constraints is 1 "1 1 1 0 0 10 1 1 1 0 "1 0 1 1 2 1 0 7 0 1 The basic solution is not feasible because of that "1. Lets do a row operation to change this. I used the 1 in the second column as a pivot and zeroed out what was above and below it. Now it seems that the basic solution is feasible. 2 0 2 1 "1 0 11 1 1 1 0 "1 0 "1 0 "1 0 2 1 1 5 Assemble the simplex tableau. 2 0 2 1 "1 0 0 11 1 1 1 0 "1 0 0 "1 0 "1 0 2 1 0 5 0 0 0 1 0 "2 "1 0 1 Then do a row operation to preserve the simple columns. 2 0 2 1 "1 0 0 11 1 1 1 0 "1 0 0 "1 0 "1 0 "1 0 1 2 1 1 0 5 0 "1 0 1 1 Lets do minimization first. Work with the third column because there is a positive entry on the bottom. 0 "2 0 1 1 1 1 0 "1 0 0 1 0 1 0 0 1 1 0 6 "2 "1 0 0 0 0 1 0 1 0 0 9 It follows that the minimum is 0 and it occurs when x 1 x 2 0, x 3 1. Now lets maximize. 2 0 2 1 "1 0 0 11 1 1 1 0 "1 0 0 "1 0 "1 0 "1 0 1 2 1 1 0 5 0 "1 0 1 1 Lets begin with the first column. 0 "2 0 1 1 0 0 9 1 1 1 0 "1 0 0 1 0 1 0 0 0 1 2 0 "2 0 1 2 1 1 0 6 There is still a "2 to the left of 1 in the bottom row. 0 "3 0 1 0 "1 0 3 1 2 1 0 0 1 0 7 0 1 0 0 1 1 0 6 0 3 2 0 0 2 1 14 There are no negatives left so the maximum is 14 and it happens when x 1 7, x 2 x 3 0. d. z x 1 2x 2 subject to the constraints x 1 " x 2 x 3 t 10, x 1 x 2 x 3 u 1, and x 1 2x 2 x 3 t 7. i. The augmented matrix for the constraints is 1 "1 1 1 0 0 10 1 1 1 0 "1 0 1 1 2 1 0 7 0 1 Of course the obvious or basic solution is not feasible. Do a row operation involving a pivot in the second row to try and fix this. 2 0 2 1 "1 0 11 1 1 1 0 "1 0 "1 0 "1 0 2 1 1 5 Now all is well. Begin to assemble the simplex tableau. 2 0 2 1 "1 0 0 11 1 1 1 0 "1 0 0 "1 0 "1 0 2 1 0 5 0 0 0 1 0 "1 "2 0 Do a row operation to preserve the simple columns. 1 2 0 2 1 "1 0 0 11 1 1 1 0 "1 0 0 "1 0 "1 0 1 0 2 2 1 1 0 5 0 "2 0 1 2 Next lets maximize. There is only one negative number in the bottom left of 1. 0 27 2 7 2 "1 0 "1 0 2 1 0 5 0 7 3 2 1 2 3 2 1 2 0 1 0 1 2 1 2 1 0 0 0 1 0 0 0 1 1 Thus the maximum is 7 and it happens when x 2 7/2, x 3 x 1 0. Next lets find the minimum. 2 0 2 1 "1 0 0 11 1 1 1 0 "1 0 0 "1 0 "1 0 1 0 2 2 1 1 0 5 0 "2 0 1 2 Start with the column which has a 2. 0 "2 0 1 1 1 1 0 "1 0 0 1 0 1 0 0 1 1 0 6 "1 "2 0 0 0 0 1 0 1 0 0 9 There are no more positive numbers so the minimum is 0 when x 1 x 2 0, x 3 1. 3. 249] Solution: Consider contradictory constraints, x 1 x 2 u 12 and x 1 2x 2 t 5. You know these two contradict but show they contradict using the simplex algorithm. You can do this by using artificial variables, x 5 . Thus 1 1 "1 0 1 0 12 1 2 0 1 0 0 5 0 0 0 0 "1 1 0 Do a row operation to preserve the simple columns. 1 1 "1 0 1 0 12 1 2 0 1 0 0 5 1 1 "1 0 0 1 12 Next start with the 1 in the first column. 0 "1 "1 "1 1 0 7 1 2 0 1 0 0 5 0 "1 "1 "1 0 1 7 Thus the minimum value of z x 5 is 7 but, for there to be a feasible solution, you would need to have this minimum value be 0. 4. 250] Solution: Find a solution to the following inequalities for x, y u 0 if it is possible to do so. If it is not possible, prove it is not possible. a. 6x 3y u 4 8x 4y t 5 i. Use an artificial variable. Let x 1 x, x 2 y and slack variables x 3 , x 4 with artificial variable x 5 . Then minimize x 5 as described earlier. 6 3 "1 0 1 0 4 8 4 0 1 0 0 5 0 0 0 0 "1 1 0 Keep the simple columns. 6 3 "1 0 1 0 4 8 4 0 1 0 0 5 6 3 "1 0 0 1 4 Now proceed to minimize. 0 0 "1 " 34 1 0 8 4 0 1 1 4 0 0 5 0 0 "1 " 34 0 1 1 4 It appears that the minimum value of x 5 is 1/4 and so there is no solution to these inequalities with x 1 , x 2 u 0. 6x 1 4x 3 t 11 b. 5x 1 4x 2 4x 3 u 8 6x 1 6x 2 5x 3 t 11 i. The augmented matrix is 6 0 4 1 0 0 11 5 4 4 0 "1 0 8 6 6 5 0 0 1 11 It is not clear whether there is a solution which has all variables nonnegative. However, if you do a row operation using 5 in the first column as a pivot, you get 0 " 24 " 45 1 5 6 5 0 7 5 5 4 4 0 "1 0 8 0 6 5 1 5 0 6 5 1 7 5 and so a solution is x 1 8/5, x 2 x 3 0. 6x 1 4x 3 t 11 c. 5x 1 4x 2 4x 3 u 9 6x 1 6x 2 5x 3 t 9 i. The augmented matrix is 6 0 4 1 0 0 11 5 4 4 0 "1 0 9 6 6 5 0 0 1 9 Lets include an artificial variable and seek to minimize x 7 . 6 0 4 1 0 0 0 0 11 5 4 4 0 "1 0 1 0 9 6 6 5 0 0 1 0 0 9 0 0 0 0 0 0 "1 1 0 0 0 0 0 11 Preserving the simple columns, 6 0 4 1 5 4 4 0 "1 0 1 0 9 6 6 5 0 1 0 0 9 5 4 4 0 "1 0 0 1 9 0 Use the first column. 0 "6 "1 1 0 "1 " 6 6 1 6 5 0 "1 0 0 2 0 "1 " 56 1 0 0 0 1 3 2 0 0 9 0 "1 " 16 0 "1 " 56 0 1 3 2 It appears that the minimum value for x 7 is 3/2 and so there will be no solution to these inequalities for which all the variables are nonnegative. x1 " x2 x3 t 2 d. x 1 2x 2 u 4 3x 1 2x 3 t 7 i. The augmented matrix is 1 "1 1 1 0 0 2 1 2 0 0 "1 0 4 3 0 2 0 0 1 7 Lets add in an artificial variable and set things up to minimize this artificial variable. 1 "1 1 1 0 0 0 0 2 1 2 0 0 "1 0 1 0 4 3 0 2 0 0 1 0 0 7 0 0 0 0 0 0 "1 1 0 1 "1 1 1 0 0 0 0 2 Then 1 2 0 0 "1 0 1 0 4 3 0 2 0 1 2 0 0 "1 0 0 1 4 0 1 0 0 7 Work with second column. 0 1 1 " 12 0 1 2 0 4 1 2 0 0 "1 0 1 0 4 3 0 2 0 0 1 0 0 7 0 0 0 0 0 0 "1 1 0 3 2 It appears the minimum value of x 7 is 0 and so this will mean there is a solution when x 2 2, x 3 0, x 1 0. 5x 1 " 2x 2 4x 3 t 1 e. 6x 1 " 3x 2 5x 3 u 2 5x 1 " 2x 2 4x 3 t 5 i. The augmented matrix is 5 "2 4 1 0 0 1 6 "3 5 0 "1 0 2 5 "2 4 0 0 1 5 lets introduce an artificial variable x 7 and then minimize x 7 . 5 "2 4 1 0 0 0 1 6 "3 5 0 "1 0 1 0 2 5 "2 4 0 0 1 0 0 5 0 0 0 "1 1 0 0 0 0 Then, preserving the simple columns, 0 5 "2 4 1 0 0 0 0 1 6 "3 5 0 "1 0 1 0 2 5 "2 4 0 0 1 0 0 5 6 "3 5 0 "1 0 0 1 2 work with the first column. 5 "2 4 0 " 3 5 1 5 " 0 0 "1 0 1 5 " "1 0 0 1 0 0 " 3 5 1 0 6 5 6 5 0 0 0 1 "1 0 1 0 4 5 1 0 0 4 4 5 There is still a positive entry. "2 4 5 " 1 4 0 " 1 0 " 1 2 5 4 0 "1 0 0 0 0 0 1 "1 0 1 0 0 3 4 1 0 0 4 " 14 " 12 0 " 54 "1 0 0 1 3 4 It appears that there is no solution to this system of inequalities because the minimum value of x 7 is not 0. 5. 251] Solution: Minimize z x 1 x 2 subject to x 1 x 2 u 2, x 1 3x 2 t 20, x 1 x 2 t 18. Change to a maximization problem and solve as follows: Let y i M " x i . Formulate in terms of y 1 , y 2 . You could find the maximum of 2M " x 1 " x 2 for the given constraints and this would happen when x 1 x 2 is as small as possible. Thus you would maximize y 1 y 2 subject to the constraints M " y1 M " y2 u 2 M " y 1 3M " y 2 t 20 M " y 1 M " y 2 t 18 To simplify, this would be 2M " 2 u y 1 y 2 4M " 20 t y 1 3y 2 2M " 18 t y 1 y 2 You could simply regard M as large enough that y i u 0 and use the techniques just developed. The augmented matrix for the constraints is then 0 0 2M " 2 1 3 0 "1 0 4M " 20 1 1 0 "1 2M " 18 1 1 1 0 Here M is large. Use the 3 as a pivot to zero out above and below it. 2 3 1 3 0 1 1 3 0 "1 2 3 M 14 3 4M " 20 0 "1 1 3 0 0 2 3 0 2 3 M" 34 3 Then it is still the case that the basic solution is not feasible. Lets use the bottom row and pick the 2/3 as a pivot. 0 0 1 0 1 16 0 3 0 " 32 3 2 3M " 3 2 3 1 3 0 0 "1 2 3 M" 34 3 Now it appears that the basic solution is feasible provided M is large. Then assemble the simplex tableau. 0 0 1 0 1 0 16 0 3 0 " 32 3 2 0 3M " 3 2 3 0 0 1 3 "1 "1 0 0 "1 0 0 2 3 1 M" 34 3 0 Do row operations to preserve the simple columns. 0 0 1 0 1 0 16 0 3 0 3 2 0 3M " 3 0 0 " 32 1 3 "1 0 2 3 0 0 0 0 "1 1 2M " 18 2 3 M" 34 3 There is a negative number to the left of the 1 on the bottom row and we want to maximize so work with this column. Assume M is very large. Then the pivot should be the top entry in this column. 0 0 1 0 3 " 0 1 0 16 0 0 3M " 27 0 0 2 3 0 1 0 1 " 32 1 3 0 0 1 0 2 3 3 2 M 14 3 2M " 2 It follows that the maximum of y 1 y 2 is 2M " 2 and it happens when y 1 M 7, y 2 M " 9, y 3 0. Thus the minimum of x 1 x 2 is M " y 1 M " y 2 M " M 7 M " M " 9 2 Solutions 1 Solution: State the eigenvalue problem from a geometric perspective. It means that Ax is parallel to x. 2 Solution: If A is an n × n matrix and c is a nonzero constant, compare the eigenvalues of A and cA. Say Ax = λx. Then cAx = cλx and so the eigenvalues of cA are just cλ where λ is an eigenvalue of A. 3 Solution: If A is an invertible n × n matrix, compare the eigenvalues of A and A −1 . More generally, for m an arbitrary integer, compare the eigenvalues of A and A m . A m x = λ m x for any integer. In the case of −1, A −1 λx = AA −1 x = x so A −1 x = λ −1 x. Thus the eigenvalues of A −1 are just λ −1 where λ is an eigenvalue of A. 4 Solution: State the eigenvalue problem from an algebraic perspective. It means Ax =λx where x ≠ 0. 5 Solution: If A is the matrix of a linear transformation which rotates all vectors in R 2 through 60 ∘ , explain why A cannot have any real eigenvalues. Is there an angle such that rotation through this angle would have a real eigenvalue. What eigenvalues would be obtainable in this way? If it did have λ ∈ R as an eigenvalue, then there would exist a vector x such that Ax = λx for λ a real number. Therefore, Ax and x would need to be parallel. However, this doesn’t happen because A rotates the vectors. You could let θ = 180 ∘ or θ = 0 ∘ . Then the eigenvalues would be 1 or −1. 6 Solution: Let A, B be invertible n × n matrices which commute. That is, AB = BA. Suppose x is an eigenvector of B. Show that then Ax must also be an eigenvector for B. BAx = ABx = Aλx = λAx. Here is is assumed that Bx = λx. 7 Solution: Suppose A is an n × n matrix and it satisfies A m = A for some m a positive integer larger than 1. Show that if λ is an eigenvalue of A then |λ| equals either 0 or 1. Let x be the eigenvector. Then A m x = λ m x, A m x = Ax = λx and so λm = λ Hence if λ ≠ 0, then λ m−1 = 1 and so |λ| = 1. 8 Solution: Show that if Ax = λx and Ay = λy, then whenever a, b are scalars, Aax + by = λax + by. Does this imply that ax + by is an eigenvector? Explain. The formula is obvious from properties of matrix multiplications. However, this vector might not be an eigenvector because it might equal 0 and Eigenvectors are never 0 9 Solution: Is it possible for a nonzero matrix to have only 0 as an eigenvalue? Sure. 0 1 works. 0 0 10 Solution: Find the eigenvalues and eigenvectors of the matrix −17 17 28 −37 −32 −50 59 16 53 One eigenvalue is 5. Determine whether the matrix is defective. −17 17 28 −37 −32 −50 59 16 , eigenvectors: 53 1 3 − 23 2 ↔ −3, ↔5 −4 7 1 11 Solution: Find the eigenvalues and eigenvectors of the matrix 5 1 0 0 5 0 −3 −91 2 One eigenvalue is 2. Determine whether the matrix is defective. 5 1 0 0 5 0 −3 −91 2 defective. 12 −1 , eigenvectors: 0 1 0 ↔ 5, 0 1 ↔ 2 The matrix is Solution: Find the eigenvalues and eigenvectors of the matrix −23 −66 −88 −4 −13 −16 8 24 31 One eigenvalue is −3. Determine whether the matrix is defective. −23 −66 −88 −4 −13 −16 8 24 , eigenvectors: 31 −4 − 114 −3 , 0 ↔ −1, 1 1 ↔ −3 − 12 0 1 The matrix is not defective. 13 Solution: Find the eigenvalues and eigenvectors of the matrix −12 45 15 −2 9 −8 24 11 2 One eigenvalue is 2. Determine whether the matrix is defective. −12 45 15 −2 9 −8 24 11 1 , eigenvectors: 2 15 8 1 4 3 , 0 1 1 ↔ 3, 0 ↔2 1 The matrix is not defective. 14 Solution: Find the eigenvalues and eigenvectors of the matrix −11 −26 −8 6 14 4 −4 −17 −10 One eigenvalue is 1. Determine whether the matrix is defective. −11 −26 −8 6 14 −4 −17 −10 defective 4 , eigenvectors: 4 3 − 23 1 3 2 ↔ −4, −1 1 ↔1 15 Solution: Find the eigenvalues and eigenvectors of the matrix 41 12 78 8 −3 12 −20 −4 −37 One eigenvalue is −3. Determine whether the matrix is defective. 41 12 78 8 −3 12 − 53 − 32 ↔ −1, − 23 , eigenvectors: −20 −4 −37 −1 ↔ −3, 1 1 −2 ↔5 − 12 1 It is not defective. 16 Solution: Find the eigenvalues and eigenvectors of the matrix 53 20 160 8 −1 22 −18 −6 −54 One eigenvalue is −3. Determine whether the matrix is defective. 53 20 160 8 −1 22 − 83 − 52 ↔ −2, − 23 , eigenvectors: −18 −6 −54 1 1 −3 − 12 ↔3 1 It is not defective. 17 Solution: Find the eigenvalues and eigenvectors of the matrix 50 −1 114 −60 −10 −23 12 22 −26 51 One eigenvalue is −2. Determine whether the matrix is defective. ↔ −3, 11 5 − 25 114 −60 50 −10 −23 12 22 −26 51 , eigenvectors: ↔ 2, 1 9 4 − 12 ↔ −2, 1 2 ↔1 − 13 1 It is not defective. 18 Solution: Find the eigenvalues and eigenvectors of the matrix −2 −24 −12 0 6 4 −19 −14 2 One eigenvalue is −2. Determine whether the matrix is defective. 6 5 − 25 −2 −24 −12 0 2 6 4 , eigenvectors: −19 −14 ↔ −4, 1 5 4 − 12 ↔ −2 1 19 Solution: Find the eigenvalues and eigenvectors of the matrix −15 −56 4 15 0 0 −12 −48 −1 One eigenvalue is 1. Determine whether the matrix is defective. −15 −56 4 15 0 0 −4 0 , eigenvectors: 0 −12 −48 −1 , 1 1 0 The matrix is not defective. 20 Solution: Find the eigenvalues and eigenvectors of the matrix −7 −361 8 0 1 0 −4 −181 5 One eigenvalue is −3. Determine whether the matrix is defective. ↔ −1, 7 6 − 13 1 ↔1 −7 −361 8 0 1 1 , eigenvectors: 0 2 ↔ 1, 0 −4 −181 5 defective. ↔ −3 The matrix is 0 1 1 21 Solution: Find the eigenvalues and eigenvectors of the matrix −9 −6 −12 −4 1 −4 12 6 15 One eigenvalue is 3. Determine whether the matrix is defective. − 45 −9 −6 −12 −4 1 −4 12 6 15 − 34 ↔ 3, − 25 , eigenvectors: 1 ↔ 3, − 12 1 −1 ↔1 − 13 1 It is not defective. 22 Solution: Find the eigenvalues and eigenvectors of the matrix −13 −716 −48 0 −1 0 4 239 15 One eigenvalue is 3. Determine whether the matrix is defective. −13 −716 −48 0 −1 0 −4 , eigenvectors: −3 ↔ −1, 0 4 239 15 matrix is defective. 1 0 1 23 Solution: Find the eigenvalues and eigenvectors of the matrix −8 −20 5 5 27 −5 5 85 −13 ↔ 3 The One eigenvalue is 2. Determine whether the matrix is defective. Hint: This one has some complex eigenvalues. −8 −20 5 5 27 −5 5 85 −13 2 29 5 29 − + 5 29 2 29 , eigenvectors: i i ↔ 2 + 5i, 2 29 5 29 + − 1 5 29 2 29 i ↔ 2 − 5i, i 1 6 1 6 ↔2 1 1 24 Solution: Let A be a real 3 × 3 matrix which has a complex eigenvalue of the form a + ib where b ≠ 0. Could A be defective? Explain. Either give a proof or an example. The characteristic polynomial is of degree three and it has real coefficients. Therefore, there is a real root and two distinct complex roots. It follows that A cannot be defective because it has three distinct eigenvalues. 25 Solution: Find the eigenvalues and eigenvectors of the matrix −8 3 −14 21 −5 7 −66 92 −24 One eigenvalue is 2. Determine whether the matrix is defective. Hint: This one has some complex eigenvalues. −8 3 −14 21 −5 7 , eigenvectors: −66 92 −24 1 − 10 − 1 5 − 1 5 1 10 − 101 + i i ↔ 1 + 4i, 1 5 + 1 1 5 1 10 i 1 ↔ 1 − 4i, i 1 ↔2 1 1 26 Solution: Find the eigenvalues and eigenvectors of the matrix 7 −15 6 −3 1 0 −18 24 −8 One eigenvalue is −2. Determine whether the matrix is defective. Hint: This one has some complex eigenvalues. 7 −15 6 −3 1 0 −18 24 −8 1 − 10 − 3 10 − , eigenvectors: − 101 + 3 i 10 1 i 10 ↔ 1 + 3i, 3 10 1 + 3 i 10 1 i 10 1 ↔ 1 − 3i, ↔ −2 1 1 1 27 Solution: Let T be the linear transformation which reflects all vectors in R 3 through the xy plane. Find a matrix for T and then obtain its eigenvalues and eigenvectors. A= 1 0 0 0 1 0 0 0 −1 1 0 0 0 1 0 0 , eigenvectors: 1 ↔ −1, 0 0 0 −1 0 1 0 , 0 1 ↔1 0 28 Solution: Let T be the linear transformation which reflects vectors about the x axis. Find a matrix for T and then find its eigenvalues and eigenvectors. The matrix of T is 1 0 0 −1 1 0 0 −1 , eigenvectors: . 0 1 1 ↔ −1, 0 ↔1 29 Solution: Let A be the 2 × 2 matrix of the linear transformation which rotates all vectors in R 2 through an angle of θ. For which values of θ does A have a real eigenvalue? When you think of this geometrically, it is clear that the only two values of θ are 0 and π or these added to integer multiples of 2π. 30 Solution: Find the eigenvalues and eigenvectors of the matrix 0 −10 3 −12 32 −7 −61 175 −39 One eigenvalue is −1. Determine whether the matrix is defective. Hint: This one has some complex eigenvalues. 0 −10 3 −12 32 −7 , eigenvectors: −61 175 −39 1 − 26 − 5 26 − − 261 + 5 i 26 1 i 26 ↔ −3 + 5i, 5 26 1 + 5 i 26 1 i 26 ↔ −3 − 5i, 1 3 1 3 ↔ −1, 1 1 31 Solution: Let T be the linear transformation which rotates all vectors in R 2 counterclockwise through an angle of π/2. Find a matrix of T and then find eigenvalues and eigenvectors. A= 0 −1 1 0 0 −1 1 −i , eigenvectors: 0 ↔ −i, 1 i ↔i 1 32 Solution: Suppose A is a 3 × 3 matrix and the following information is available. 0 A −1 0 =2 −1 −1 −4 ,A −1 −3 A 1 1 =1 1 1 1 1 −3 = −3 −3 −4 −3 1 Find A −5 4 You have given that 0 1 −3 −1 1 −4 −1 1 −3 and so it follows that −1 0 A 1 −3 −1 1 −4 −1 1 −3 = 2 0 0 0 1 0 0 0 −3 −12 12 1 A= −1 −15 17 −1 −12 14 Therefore, the desired answer is given by 109 142 115 33 Solution: Suppose A is a 3 × 3 matrix and the following information is available. −2 A −3 −2 =1 −3 −3 −6 ,A −3 −5 A 1 1 = −3 1 1 1 1 −5 = −1 −5 −6 −5 1 Find A −3 4 You have given that −2 1 −5 −3 1 −6 −1 −2 1 −5 A −3 1 −5 1 = −3 1 −6 −3 1 −5 and so it follows that −11 10 −2 A= −12 9 0 −12 10 −1 Therefore, the desired answer is given by −49 −39 −46 34 Solution: 0 0 0 −3 0 0 −1 0 Suppose A is a 3 × 3 matrix and the following information is available. −1 A A −2 −1 =2 −2 −2 −2 −2 −2 −5 1 =2 ,A 1 =0 1 1 1 1 −5 −4 −4 3 Find A −4 3 You have given that −1 1 −2 −2 1 −5 −1 −1 1 −2 −2 1 −5 A −2 1 −4 −2 1 −4 2 0 0 = 0 0 0 0 0 2 and so it follows that −2 0 2 A= −4 2 2 −4 0 4 Therefore, the desired answer is given by 0 −14 0 35 Solution: Here is a symmetric matrix. 13 6 − 16 − 13 − 16 − 13 13 6 1 3 1 3 8 3 Find its largest eigenvalue and an eigenvector associated with this eigenvalue. This eigenvector gives the direction of the principal stretch. The largest eigenvalue is 3. The eigenvectors and eigenvalues are −1 1 1 0 ↔ 2, 1 2 1 ↔ 3, −1 1 ↔2 36 Solution: Here is a symmetric matrix. 24 11 − 112 − 116 − 112 − 116 24 11 6 11 6 11 40 11 Find its largest eigenvalue and an eigenvector associated with this eigenvalue. This eigenvector gives the direction of the principal stretch. The largest eigenvalue is 4. The eigenvectors and eigenvalues are 1 1 3 ↔ 2, 1 ↔ 2, −3 0 ↔4 −1 −3 2 37 Solution: Here is a symmetric matrix. 13 6 5 6 2 3 5 6 13 6 − 23 2 3 − 23 11 3 Find its largest eigenvalue and an eigenvector associated with this eigenvalue. This eigenvector gives the direction of the principal stretch. The largest eigenvalue is 4. The eigenvectors and eigenvalues are −4 1 1 ↔ 3, 1 ↔ 1, 4 0 ↔4 −1 2 4 38 Solution: Here is a symmetric matrix. 25 6 − 76 5 3 − 76 25 6 − 53 5 3 − 53 11 3 Find its largest eigenvalue and an eigenvector associated with this eigenvalue. This eigenvector gives the direction of the principal stretch. The largest eigenvalue is 7. The eigenvectors and eigenvalues are −1 1 1 0 ↔ 3, 1 2 1 ↔ 2, −1 1 ↔7 39 Solution: Here is a symmetric matrix. 25 6 − 76 5 3 − 76 25 6 − 53 5 3 − 53 11 3 Find its largest eigenvalue and an eigenvector associated with this eigenvalue. This eigenvector gives the direction of the principal stretch. The largest eigenvalue is 7. The eigenvectors and eigenvalues are −1 1 1 0 ↔ 3, 1 2 1 ↔ 2, ↔7 −1 1 40 Solution: Here is a Markov (migration matrix) for three locations 1 2 1 2 0 1 8 5 8 1 4 1 4 1 4 1 2 The total number of individuals in the migration process is 320 After a long time, how many are in each location? 80 160 80 41 Solution: Here is a Markov (migration matrix) for three locations 1 10 2 5 1 2 1 2 1 6 1 3 2 9 2 3 1 9 The total number of individuals in the migration process is 949 After a long time, how many are in each location? 280 372 297 42 Solution: Here is a Markov (migration matrix) for three locations 7 10 1 5 1 10 1 14 1 2 3 7 1 6 1 2 1 3 The total number of individuals in the migration process is 736 After a long time, how many are in each location? 200 308 228 43 Solution: The following table describes the transition probabilities between the states rainy, partly cloudy and sunny. The symbol p.c. indicates partly cloudy. Thus if it starts off p.c. it ends up sunny the next day with probability 15 . If it starts off sunny, it ends up sunny the next day with probability and so forth. 2 5 rains sunny p.c. rains sunny p.c. 1 5 1 5 3 5 1 5 2 5 2 5 6 11 1 11 4 11 Given this information, what are the probabilities that a given day is rainy, sunny, or partly cloudy? This is a vector in the direction 190 100 This vector is given by 242 5 14 25 133 121 266 44 Solution: The following table describes the transition probabilities between the states rainy, partly cloudy and sunny. The symbol p.c. indicates partly cloudy. Thus if it starts off p.c. it ends up sunny the next day with probability 15 . If it starts off sunny, it ends up sunny the next day with probability and so forth. 1 3 rains sunny p.c. 1 10 1 5 7 10 rains sunny p.c. 1 3 1 3 1 3 3 10 2 5 3 10 Given this information, what are the probabilities that a given day is rainy, sunny, or partly cloudy? This is a vector in the direction 100 126 This vector is given by 160 50 193 63 193 80 193 45 Solution: You own a trailer rental company in a large city and you have four locations, one in the South East, one in the North East, one in the North West, and one in the South West. Denote these locations by SE,NE,NW, and SW respectively. Suppose that the following table is observed to take place. SE NE NW SW SE NE NW SW 2 9 1 3 2 9 2 9 1 10 7 10 1 10 1 10 1 10 1 5 3 5 1 10 1 5 1 10 1 5 1 2 In this table, the probability that a trailor starting at NE ends in NW is 1/10, the probability that a trailor starting at SW ends in NW is 1/5, and so forth. Approximately how many will you have in each location after a long time if the total number of trailors is 463? 63 184 126 90 46 Solution: The following table describes the transition probabilities between the states rainy, partly cloudy and sunny. The symbol p.c. indicates partly cloudy. Thus if it starts off p.c. it ends up sunny the next day with probability and so forth. 1 5 . If it starts off sunny, it ends up sunny the next day with probability 1 5 rains sunny p.c. 1 10 1 5 7 10 rains sunny p.c. 2 5 1 5 2 5 3 8 3 8 1 4 Given this information, what are the probabilities that a given day is rainy, sunny, or partly cloudy? This is a vector in the direction 180 165 This vector is given by 256 180 601 165 601 256 601 47 Solution: You own a trailer rental company in a large city and you have four locations, one in the South East, one in the North East, one in the North West, and one in the South West. Denote these locations by SE,NE,NW, and SW respectively. Suppose that the following table is observed to take place. SE NE NW SW SE NE NW SW 1 3 1 3 1 6 1 6 1/4 1/10 1/5 1/4 1/5 1/10 1/4 3/5 1/5 1/4 1/10 1/2 In this table, the probability that a trailor starting at NE ends in NW is 1/10, the probability that a trailor starting at SW ends in NW is 1/5, and so forth. Approximately how many will you have in each location after a long time if the total number of trailors is 1030. 210 220 350 250 48 Solution: In the city of Nabal, there are three political persuasions, republicans (R), democrats (D), and neither one (N). The following table shows the transition probabilities between the political parties, the top row being the initial political party and the side row being the political affiliation the following year. R D N R D N 1 10 1 10 4 5 1 5 1 5 3 5 3 13 7 13 3 13 Find the probabilities that a person will be identified with the various political persuasions. Which party will end up being most important? This is a vector in the direction 190 330 which has the entries add to 1. This vector is given by 455 38 195 22 65 7 15 49 Solution: The University of Poohbah offers three degree programs, scouting education (SE), dance appreciation (DA), and engineering (E). It has been determined that the probabilities of transferring from one program to another are as in the following table. SE DA E SE .8 .1 .3 DA .1 .7 .5 E .2 .2 .1 where the number indicates the probability of transferring from the top program to the program on the left. Thus the probability of going from DA to E is .2. Find the probability that a student is enrolled in the various programs. 8 10 −1 1 10 1 10 1 1 1 1 10 7 10 3 10 1 2 −1 2 10 2 10 14 5 13 5 0 , row echelon form: 0 −1 0 z = z 32 5 z z 5 14 32 5 5 13 5 32 5 32 = 7 16 13 32 5 32 0. 437 5 = 0. 406 25 0. 156 25 1 0 − 145 0 0 1 − 135 0 0 0 0 0 50 Solution: Here is a Markov (migration matrix) for three locations 3 5 1 10 3 10 1 8 3 4 1 8 1 12 7 12 1 3 The total number of individuals in the migration process is 406 After a long time, how many are in each location? 90 232 84 51 Solution: You own a trailer rental company in a large city and you have four locations, one in the South East, one in the North East, one in the North West, and one in the South West. Denote these locations by SE,NE,NW, and SW respectively. Suppose that the following table is observed to take place. SE NE NW SW SE NE NW SW 1 2 1 10 1 5 1 5 1 10 7 10 1 10 1 10 1 10 1 5 3 5 1 10 1 5 1 10 1 5 1 2 In this table, the probability that a trailor starting at NE ends in NW is 1/10, the probability that a trailor starting at SW ends in NW is 1/5, and so forth. Approximately how many will you have in each location after a long time if the total number of trailors is 350? 70 112 98 70 52 Solution: Here is a Markov (migration matrix) for three locations 2 5 1 2 1 10 1 8 1 2 3 8 1 2 3 14 2 7 The total number of individuals in the migration process is 1000 After a long time, how many are in each location? 310 424 266 53 Solution: Let A be the n × n, n > 1, matrix of the linear transformation which comes from the projection v → proj w v. Show that A cannot be invertible. Also show that A has an eigenvalue equal to 1 and that for λ an eigenvalue, |λ| ≤ 1. Obviously A cannot be onto because the range of A has dimension 1 and the dimension of this space should be 3 if the matrix is onto. Therefore, A cannot be invertible. Its row reduced echelon form cannot be I since if it were, A would be onto. Aw = w so it has an eigenvalue equal to 1. Now suppose Ax = λx. Thus, from the Cauchy Schwarz inequality, |x, w| |x||w| |w| ≥ |w| = |λ||x| |x| = 2 |w|2 |w| and so |λ| ≤ 1. 54 Solution: Consider the dynamical system in which A is an n × n matrix, xn + 1 = Axn, x0 = x 0 Suppose the eigen pairs are v i , λ i and suppose that v 1 , v 2 , ⋯, v n is a basis for R n . If n x0 = ∑ civi, i=1 show that the solution to the dynamical system is of the form n xn = ∑ c i λ ni v i i=1 n The solution is A x 0 and this reduces to n xn = ∑ ciA i=1 n n vi = ∑ c i λ ni v i i=1 55 Solution: Consider the recurrance relation x n+2 = x n+1 + x n where x 1 and x 0 are given positive numbers. 1+ 5 Show that if lim n→∞ xxn+1 exists, then it must equal 2 . n You have x n+2 = 1 + x n x n+1 x n+1 Assuming this limit exists, and equals r, it must satisfy r = 1 + 1r and so r 2 − r − 1 = 0, Solution is: 12 5 + 12 , 12 − 12 5 . Since the solution is obviously positive, it must be the first. 56 Solution: Consider the dynamical system xn + 1 yn + 1 .8 = .8 xn −. 8 . 8 yn −i eigenvalues and eigenvectors are 0. 8 + 0. 8i , 0. 8 − 0. 8i . Show the i . Find a formula 1 1 for the solutiont to the dynamical system for given initial condition x 0 , y 0 T . Show that the magnitude of xn, yn T must diverge provided the initial condition is not zero. Next graph the vector field for .8 .8 x −. 8 . 8 y − x y Note that this vector field seems to indicate a conclusion different than what you just obtained. Therefore, in this context of discreet dynamical systems the consideration of such a picture is not all that reliable. It is straight forward to verify that these are the eigenvectors and eigenvalues. Thus the solution to the dynamical system is of the form a −i 1 0. 8 + 0. 8i n + b i 1 0. 8 − 0. 8i n Now |0. 8 − 0. 8i| = 1. 131 4 = |0. 8 + 0. 8i| and so the magnitude of any solution having nonzero initial conditions is unbounded as n → ∞. However, here is the graph of the vector field. It does not appear to indicate that the solutions are getting larger in magnitude from one point to the next and in fact, seems to indicate that they get smaller. 57 Solution: Let v be a unit vector and let A = I − 2vv T . Show that A has an eigenvalue equal to −1. Lets see what it does to v. I − 2vv T v = v − 2vv T v = v − 2v =−1v Yes. It has an eigenvector and the eigenvalue is −1. 58 Solution: Let M be an n × n matrix and suppose x 1 , ⋯, x n are n eigenvectors which form a linearly independent set. Form the matrix S by making the columns these vectors. Show that S −1 exists and that S −1 MS is a diagonal matrix (one having zeros everywhere except on the main diagonal) having the eigenvalues of M on the main diagonal. When this can be done the matrix is diagonalizable. Since the vectors are linearly independent, the matrix S has an inverse. Denoting this inverse by w T1 S −1 = ⋮ w Tn it follows by definition that w Ti x j = δ ij . Therefore, w T1 S −1 MS = S −1 Mx 1 , ⋯, Mx n = ⋮ λ 1 x 1 , ⋯, λ n x n w Tn λ1 = 0 ⋱ λn 0 59 Solution: Show that a matrix, M is diagonalizable if and only if it has a basis of eigenvectors. Hint: The first part is done earlier. It only remains to show that if the matrix can be diagonalized by some matrix, S giving D = S −1 MS for D a diagonal matrix, then it has a basis of eigenvectors. Try using the columns of the matrix S. The formula says that MS = SD th Letting x k denote the k column of S, it follows from the way we multiply matrices that Mx k = λ k x k th where λ k is the k diagonal entry on D. 60 Solution: Suppose A is an n × n matrix which is diagonally dominant. This means |a ii | > ∑|a ij |. j≠i Show that A −1 must exist. The diagonally dominant condition implies that none of the Gerschgorin disks contain 0. Therefore, 0 is not an eigenvalue. Hence A is one to one, hence invertible. 61 Solution: Let M be an n × n matrix. Then define the adjoint of M, denoted by M ∗ to be the transpose of the conjugate of M. For example, ∗ 2 i 1+i 3 = 2 1−i −i 3 . A matrix, M, is self adjoint if M ∗ = M. Show the eigenvalues of a self adjoint matrix are all real. If the self adjoint matrix has all real entries, it is called symmetric. First note that AB ∗ = B ∗ A ∗ . Say Mx = λx, x ≠ 0. Then λ |x|2 = λ x ∗ x = λx ∗ x =Mx ∗ x = x ∗ M ∗ x = x ∗ Mx = x ∗ λx = λ|x|2 Hence λ = λ . 62 Solution: Suppose A is an n × n matrix consisting entirely of real entries but a + ib is a complex eigenvalue having the eigenvector, x + iy. Here x and y are real vectors. Show that then a − ib is also an eigenvalue with the eigenvector, x − iy. Hint: You should remember that the conjugate of a product of complex numbers equals the product of the conjugates. Here a + ib is a complex number whose conjugate equals a − ib. Ax = a + ibx. Now take conjugates of both sides. Since A is real, A x = a − ib x 63 Solution: Recall an n × n matrix is said to be symmetric if it has all real entries and if A = A T . Show the eigenvectors and eigenvalues of a real symmetric matrix are real. These matrices are self adjoint by definition, so this follows from the above problems. 64 Solution: Recall an n × n matrix is said to be skew symmetric if it has all real entries and if A = −A T . Show that any nonzero eigenvalues must be of the form ib where i 2 = −1. In words, the eigenvalues are either 0 or pure imaginary. Suppose A is skew symmetric. Then what about iA? iA ∗ = −iA ∗ = −iA T = iA and so iA is self adjoint. Hence it has all real eigenvalues. iAx = λx where λ is real. Thus Ax = −iλx and therefore, the eigenvalues of A are all of the form iλ where λ is real. 65 Solution: A discreet dynamical system is of the form xk + 1 = Axk, x0 = x 0 where A is an n × n matrix and xk is a vector in R n . Show first that xk = A k x 0 for all k ≥ 1. If A is nondefective so that it has a basis of eigenvectors, v 1 , ⋯, v n where Av j = λ j v j you can write the initial condition x 0 in a unique way as a linear combination of these eigenvectors. Thus n x0 = ∑ a jv j j=1 Now explain why n xk = n ∑ a jA v j = ∑ a jλ kj v j k j=1 j=1 which gives a formula for xk, the solution of the dynamical system. The first formula is obvious from induction. Thus n Akx0 = ∑ a jAk v j = j=1 n ∑ a jλ kj v j j=1 because if Av = λv, then if A k x = λ k x, do A to both sides. Thus A k+1 x = λ k+1 x. 66 Solution: Suppose A is an n × n matrix and let v be an eigenvector such that Av = λv. Also suppose the characteristic polynomial of A is detλI − A = λ n + a n−1 λ n−1 + ⋯ + a 1 λ + a 0 Explain why A n + a n−1 A n−1 + ⋯ + a 1 A + a 0 Iv = 0 If A is nondefective, give a very easy proof of the Cayley Hamilton theorem based on this. Recall this theorem says A satisfies its characteristic equation, A n + a n−1 A n−1 + ⋯ + a 1 A + a 0 I = 0. A n + a n−1 A n−1 + ⋯ + a 1 A + a 0 Iv = λ n + a n−1 λ n−1 + ⋯ + a 1 λ + a 0 v = 0 If A is nondefective, then it has a basis of eigenvectors v 1 , ⋯, v n and so a typical vector w is of the form n w = ∑ i=1 c i v i and so, letting the characteristic polynomial be denoted as Pλ, n PAw = PA ∑ c i v i = i=1 n ∑ c i PAv i = ∑ c i 0 = 0 i=1 i Since w is arbitrary, this shows that PA = 0. 67 Solution: Suppose an n × n nondefective matrix A has only 1 and −1 as eigenvalues. Find A 12 . From the above formula, A 12 x 0 = Thus A 12 n n j=1 j=1 ∑ a jλ 12j v j = ∑ a jv j = x 0 = I. 68 Solution: Suppose the characteristic polynomial of an n × n matrix A is 1 − λ n . Find A mn where m is an integer. Hint: Note first that A is nondefective. Why? The eigenvalues are distinct because they are the n th roots of 1. Hence from the above formula, if x is a given vector with n x= ∑ a jv j j=1 then n A nm x = A nm ∑ a j v j = j=1 so A nm n ∑ a jA nm v j = j=1 n ∑ a jv j = x j=1 = I. 69 Solution: Sometimes sequences come in terms of a recursion formula. An example is the Fibonacci sequence. x 0 = 1 = x 1 , x n+1 = x n + x n−1 Show this can be considered as a discreet dynamical system as follows. x n+1 = xn 1 1 xn 1 0 x n−1 x1 , 1 = x0 1 Now find a formula for x n . What are the eigenvalues and eigenvectors of this matrix? 1 1 1 0 , eigenvectors: 1 2 − 1 2 5 1 2 ↔ 1 − 1 5, 2 2 1 5 + 1 2 ↔ 1 5 + 1 2 2 1 Now also 1 − 1 5 2 10 1 2 − 1 2 5 1 + 1 5 + 1 10 2 Therefore, the solution is of the form x n+1 xn = 1 2 5 + 1 1 2 = 1 1 1 − 1 5 2 10 + 1 5 + 1 10 2 1 2 n 1 − 1 5 2 2 − 1 2 5 1 n 1 5 + 1 2 2 1 2 5 + 1 2 1 In particular, xn = 1 − 1 5 2 10 1 − 1 5 2 2 n + 1 5 + 1 10 2 1 5 + 1 2 2 n 70 Solution: Let A be an n × n matrix having characteristic polynomial detλI − A = λ n + a n−1 λ n−1 + ⋯ + a 1 λ + a 0 Show that a 0 = −1 n detA. The characteristic polynomial equals detλI − A. To get the constant term, you plug in λ = 0 and obtain det−A = −1 n detA. 71 Solution: Here is a matrix A= 53 8 − 454 25 8 − 214 A= 19 9 − 209 2 3 − 13 Find lim n→∞ A n . 10 5 −18 −9 72 Solution: Here is a matrix Find lim n→∞ A n . 6 3 −10 −5 73 Solution: Find the solution to the initial value problem ′ x = y x0 xt yt x 6 y 5 3 = y0 0 −1 3 12e 2t − 9e 2 2t 3 = 27e 2 2t − 24e 2t 3 74 Solution: Find the solution to the initial value problem ′ x = y x0 = y0 xt yt − 1 e 1 = e 3 2t 2 3 2t 2 −13 −5 x 30 12 y 7 −2 x 12 −3 y 2 2 6e 2 2t − 8 5 18e 2 2t − 16 5 75 Solution: Find the solution to the initial value problem ′ x y x0 y0 xt yt = = = 3 2 7e 3t − 4e t 14e 3t − 12e t Matrices And Inner Product 76 Solution: Here are some matrices. Label according to whether they are symmetric, skew symmetric, or orthogonal. If the matrix is orthogonal, determine whether it is proper or improper. a. 1 0 0 1 2 0 1 2 0 − 1 2 1 2 This one is orthogonal and is a proper transformation. b. 1 2 −3 2 1 4 −3 4 7 This is symmetric. 0 −2 −3 c. 2 0 −4 3 4 0 This one is skew symmetric. 77 Solution: Here are some matrices. Label according to whether they are symmetric, skew symmetric, or orthogonal. If the matrix is orthogonal, determine whether it is proper or improper. a. 1 0 0 1 2 0 1 2 0 − 1 2 1 2 This one is orthogonal and is a proper transformation. b. 1 2 −3 2 1 4 −3 4 7 This is symmetric. 0 −2 −3 c. 2 0 −4 3 4 0 This one is skew symmetric. 78 Solution: Show that every real matrix may be written as the sum of a skew symmetric and a symmetric matrix in a unique way. Hint: If A is an n × n matrix, show that B ≡ 12 A − A T is skew symmetric. T T A = A+A + A−A . There is only one way to do it. If A = S + W where S = S T , W = −W T , then 2 2 you also have A T = S − W and so 2S = A + A T while 2W = A − A T . 79 Solution: Let x be a vector in R n and consider the matrix, I − 2xx T |x|2 . Show this matrix is both symmetric and orthogonal. T I− 2xx T |x|2 = I− 2xx T |x|2 so it is symmetric. T I − 2xx2 |x| T I − 2xx2 |x| T = I + 4 xx xx |x|4 T T T − 4 xx2 |x| T = I + 4 xx4 − 4 xx2 = I |x| |x| because x T x 1 |x|2 = 1. 80 Solution: For U an orthogonal matrix, explain why ||Ux|| = ||x|| for any vector, x. Next explain why if U is an n × n matrix with the property that ||Ux|| = ||x|| for all vectors, x, then U must be orthogonal. Thus the orthogonal matrices are exactly those which preserve distance. First, suppose U is orthogonal. ‖Ux‖ 2 = Ux, Ux = U T Ux, x = Ix, x = ‖x‖ 2 Next suppose distance is preserved by U. Then Ux + y, Ux + y = ‖Ux‖ 2 + ‖Uy‖ 2 + 2Ux, Uy = ‖x‖ 2 + ‖y‖ 2 + 2U T Ux, y But since U preserves distances, it is also the case that Ux + y, Ux + y = ‖x‖ 2 + ‖y‖ 2 + 2x, y Hence x, y = U T Ux, y and so U T U − Ix, y = 0 Since y is arbitrary, it follows that U T U − I = 0. Thus U is orthogonal. 81 Solution: A quadratic form in three variables is an expression of the form a 1 x 2 + a 2 y 2 + a 3 z 2 + a 4 xy + a 5 xz + a 6 yz. Show that every such quadratic form may be written as x x y z A y z where A is a symmetric matrix. a1 x y z a 4 /2 a 4 /2 a 5 /2 a2 a 5 /2 a 6 /2 x a 6 /2 y a3 z 82 Solution: Given a quadratic form in three variables, x, y, and z, show there exists an orthogonal matrix U and variables x ′ , y ′ , z ′ such that x′ x y z =U y′ z′ with the property that in terms of the new variables, the quadratic form is λ 1 x ′ 2 + λ 2 y ′ 2 + λ 3 z ′ 2 where the numbers, λ 1 , λ 2 , and λ 3 are the eigenvalues of the matrix, A in the above problem. The quadratic form may be written as x T Ax where A = A T . By the theorem about diagonalizing a symmetric matrix, there exists an orthogonal matrix U such that U T AU = D, A = UDU T Then the quadratic form is x T UDU T x = U T x T DU T x where D is a diagonal matrix having the real eigenvalues of A down the main diagonal. Now simply let x ′ ≡ U T x. 83 Solution: If A is a symmetric invertible matrix, is it always the case that A −1 must be symmetric also? How about A k for k a positive integer? Explain. If A is symmetric, then A = U T DU for some D a diagonal matrix in which all the diagonal entries are non zero. Hence A −1 = U −1 D −1 U −T . Now U −1 U −T = U T U −1 = I −1 = I and so A −1 = QD −1 Q T , where Q is orthogonal. Is this thing on the right symmetric? Take its transpose. This is QD −1 Q T which is the same thing, so it appears that a symmetric matrix must have symmetric inverse. Now consider raising it to a power. Ak = UTDkU and the right side is clearly symmetric. 84 Solution: If A, B are symmetric matrices, does it follow that AB is also symmetric? 2 1 Let A = 1 3 AB = ,B = 2 1 1 1 1 3 1 0 1 1 1 0 = 3 2 4 1 which is not symmetric. 85 Solution: Suppose A, B are symmetric and AB = BA. Does it follow that AB is symmetric? AB T = B T A T = BA = AB so the answer in this case is yes. 86 Solution: Here are some matrices. What can you say about the eigenvalues of these matrices just by looking at them? 0 0 a. 0 0 0 −1 0 1 0 The eigenvalues are all within 1 of 0, and pure imaginary or zero. b. 1 2 −3 2 1 4 −3 4 7 The eigenvalues are in D1, 6 ∪ D7, 7. They are also real. 0 −2 −3 c. 2 0 −4 3 4 0 The eigenvalues are all imaginary or 0. They are no farther than 7 from 0. 1 2 3 d. 0 2 3 0 0 2 The eigenvalues are 1,2,2. 87 Solution: c 0 Find the eigenvalues and eigenvectors of the matrix 0 0 0 −b 0 b 0 . Here b, c are real numbers. c 0 0 0 0 −b 0 b , eigenvectors: 0 1 0 ↔ c, 0 0 ↔ −ib, −i 0 ↔ ib i 1 1 88 Solution: c 0 Find the eigenvalues and eigenvectors of the matrix 0 0 a −b 0 b . Here a, b, c are real a numbers. c 0 0 0 a −b 0 b , eigenvectors: a 0 0 ↔ a − ib, −i 1 1 ↔ a + ib, i ↔c 0 1 0 89 Solution: Determine which of the following sets of vectors are orthonormal sets. Justify your answer. a. 1, 1, 1, −1 i. This one is not orthonormal. 1 2 b. , −1 2 , 1, 0 i. This one is not orthonormal. c. 13 , 23 , 23 , −2 , −1 , 23 , 23 , −2 , 3 3 3 1 3 T 1/3 i. 2/3 2/3 −2/3 −1/3 2/3 1/3 2/3 2/3 −2/3 −1/3 2/3 2/3 −2/3 1/3 2/3 orthonormal set of vectors. −2/3 1/3 1 0 0 = 0 1 0 0 0 1 90 Solution: Find the eigenvalues and an orthonormal basis of eigenvectors for A. so this is an A= − 43 − 13 − 13 − 13 7 6 − 176 − 176 − 1 3 . 7 6 Hint: Two eigenvalues are −2, −1. 1 3 3 1 3 3 1 3 3 − 13 2 3 −2, 1 6 2 3 1 6 2 3 0 −1, 1 2 − 12 4 2 2 91 Solution: Find the eigenvalues and an orthonormal basis of eigenvectors for A. A= 61 22 − 35 22 23 11 − 35 22 23 11 7 11 − 18 11 − 223 7 11 . Hint: Two eigenvalues are −3, 0. 1 6 1 6 − 13 6 6 −3, 6 1 66 7 66 2 33 − 113 11 6 11 6 11 0, 1 11 − 111 6 11 4 11 11 92 Solution: Find the eigenvalues and an orthonormal basis of eigenvectors for A. A= 1 −2 −2 −2 1 2 − 32 − 32 −2 . 1 2 Hint: Two eigenvalues are −3, 3. 1 3 1 3 1 3 − 13 2 3 3 3 3 −3, 1 6 1 6 0 3, 2 3 1 2 2 2 − 12 2 2 3 93 Solution: Find the eigenvalues and an orthonormal basis of eigenvectors for A. A= − 32 0 7 2 0 −5 0 7 2 0 − 32 . Hint: Two eigenvalues are 2, −5. − 32 0 7 2 0 −5 0 7 2 0 − 32 1 2 1 2 − 16 6 2 ↔ 2, 0 , eigenvectors: 2 1 3 1 6 − 13 3 ↔ −5, 6 ↔ −5 − 13 3 1 3 6 3 94 Solution: Find the eigenvalues and an orthonormal basis of eigenvectors for A. 20 11 − 116 9 11 A= − 116 9 11 − 116 20 11 15 11 − 116 . Hint: One eigenvalue is 1. 1 − 12 2 p 2 +2 p p 2 +2 1 p 2 +2 1, 1, 0 1 2 2 3 22 − 111 2 11 3 22 2 11 3 2 11 95 Solution: Find the eigenvalues and an orthonormal basis of eigenvectors for A. 1 0 0 A= . 0 1 0 0 0 1 Hint: Two eigenvalues are 1, 1. 1 2 1 0 0 0 1 0 0 0 1 , eigenvectors: ↔ 1, 0 1 2 − 16 6 2 2 1 3 6 1 6 6 − 13 3 ↔ 1, 96 Solution: Find the eigenvalues and an orthonormal basis of eigenvectors for A. − 13 3 1 3 3 ↔1 3 0 0 A= . 0 3 0 0 0 3 Hint: One eigenvalue is 3. 1 3 38 − 12 2 p 2 +2 p 3, p 2 +2 1 2 1 p 2 +2 3, 0 2 38 3 38 2 3 − 381 2 38 2 38 97 Solution: Find the eigenvalues and an orthonormal basis of eigenvectors for A. 3 1 1 A= . 1 3 1 1 1 3 Hint: One eigenvalue is 2. 1 − 12 2 p 2 +2 p 2, p 2 +2 1 2 1 p 2 +2 2, 0 2 1 3 1 3 3 1 3 3 5 3 98 Solution: Find the eigenvalues and an orthonormal basis of eigenvectors for A. 3 2 A= 0 1 2 . 0 1 0 1 2 0 3 2 Hint: Two eigenvalues are 2, 1. 3 2 0 1 2 0 1 0 1 2 0 3 2 1 2 , eigenvectors: ↔ 2, 0 1 2 − 16 6 2 2 1 3 1 6 6 − 13 3 ↔ 1, 6 99 Solution: Find the eigenvalues and an orthonormal basis of eigenvectors for A. − 13 3 1 3 3 ↔1 A= 19 6 7 6 − 23 7 6 19 6 − 23 − 23 − 23 . 11 3 Hint: Two eigenvalues are 3, 2. 1 6 6 1 6 6 1 3 6 − 12 2 3, 1 2 2 2, 1 3 3 1 3 3 5 − 13 3 0 100 Solution: Find the eigenvalues and an orthonormal basis of eigenvectors for A. A= 1 3 1 3 3 1 3 3 8 3 2 3 − 13 3, 3, 0 1 2 2 3 8 3 . − 16 6 − 12 2 3 − 13 2 3 5 3 2 3 1 3 1 6 − 16 6 2 101 Solution: Find the eigenvalues and an orthonormal basis of eigenvectors for A. A= 13 6 2 3 − 76 2 3 8 3 − 23 − 76 − 23 . 13 6 Hint: Two eigenvalues are 1, 2. 13 6 2 3 − 76 1 2 − 76 − 23 − 16 6 ↔ 1, 2 , eigenvectors: 13 6 2 0 1 2 2 3 8 3 − 23 1 3 6 1 6 6 − 13 3 ↔ 2, − 13 3 1 3 ↔4 3 102 Solution: Find the eigenvalues and an orthonormal basis of eigenvectors for A. A= 137 54 5 27 − 25 54 − 25 54 5 27 79 27 5 27 5 27 137 54 . Hint: One eigenvalue is 3. 1 p p 2 +2 5 18 − 19 5 18 − 12 2 p 2 +2 3, 1 2 1 p 2 +2 3, 0 2 2 3 2 2 3 2 3 103 Solution: Find the eigenvalues and an orthonormal basis of eigenvectors for A. A= 1 3 1 3 1 3 − 12 2 3 3 3 2, 2, 0 1 2 2 5 2 −1 1 2 −1 4 −1 1 2 −1 5 2 1 6 − 13 1 6 . 6 6 5 6 104 Solution: Show that if A is a real symmetric matrix and λ and μ are two different eigenvalues, then if x is an eigenvector for λ and y is an eigenvector for μ, then x ⋅ y = 0. Also all eigenvalues are real. First CD T =D T C T A is real λ is eigenvalue = x T A x = x T Ax = xT λ x = λ xT x λx T x = Ax T x and so λ = λ . This shows that all eigenvalues are real. It follows all the eigenvectors can be assumed real. Why? Because A is real. Ax = λx, A x = λ x , so x + x is an eigenvector. Hence it can be assumed all eigenvectors are real. Now let x, y, μ and λ be given as above. λx ⋅ y = λx ⋅ y = Ax ⋅ y = x ⋅ Ay = x ⋅μy = μx ⋅ y = μx ⋅ y and so λ − μx ⋅ y = 0. Since λ ≠ μ, it follows x ⋅ y = 0. 105 Solution: Suppose U is an orthogonal n × n matrix. Explain why rankU = n. You could observe that detUU T = detU 2 = 1 so detU ≠ 0. 106 Solution: Show that the eigenvalues and eigenvectors of a real matrix occur in conjugate pairs. This follows from the observation that ifAx = λx, then A x = λ x 107 Solution: If a real matrix, A has all real eigenvalues, does it follow that A must be symmetric. If so, explain why and if not, give an example to the contrary. 1 3 Certainly not. 0 2 108 Solution: Suppose A is a 3 × 3 symmetric matrix and you have found two eigenvectors which form an orthonormal set. Explain why their cross product is also an eigenvector. There exists another eigenvector such that, with these two you have an orthonormal basis. Let the third eigenvector be u and the two given ones v 1 , v 2 . Then u is perpendicular to the two given vectors and so it has either the same or opposite direction to the cross product. Hence u =kv 1 × v 2 for some scalar k. 109 Solution: Show that if u 1 , ⋯, u n is an orthonormal set of vectors in F n , then it is a basis. Hint: It was shown earlier that this is a linearly independent set. If you wish, replace F n with R n . Do this version if you do not know the dot product for vectors in C n . The vectors are linearly independent as shown earlier. If they do not span all of F n , then there is a vector v ∉ spanu 1 , ⋯, u n . But then u 1 , ⋯, u n , v would be a linearly independent set which has more vectors than a spanning set, e 1 , ⋯, e n and this is a contradiction. 110 Solution: Fill in the missing entries to make the matrix orthogonal. −1 2 −1 6 1 3 −1 2 −1 6 1 3 1 2 −1 6 a 1 2 −1 6 a 0 6 3 b 0 6 3 b −1 2 −1 6 1 3 1 2 _ _ _ 6 3 _ T . = 3a − 1 3 1 3 3a − 1 3 1 a2 + 1 3 1 3 3b − ab − 1 3 3b − 1 3 ab − 2 3 1 3 1 3 1 3 2 3 b2 + Must have a = 1/ 3 , b = 1/ 3 −1 2 −1 6 1 2 0 T 1 3 −1 2 −1 6 1 3 −1 6 1/ 3 1 2 −1 6 1/ 3 6 3 1/ 3 0 6 3 1/ 3 1 0 0 = 0 1 0 0 0 1 111 Solution: Fill in the missing entries to make the matrix orthogonal. 2 3 2 2 2 3 _ _ _ 0 _ 1 6 2 . T 2 3 2 2 2 3 − 2 2 − 13 0 2 3 2 2 a 2 3 − 2 2 a b − 13 0 b 1 6 2 1 6 1 = a= 1 6 1 6 2a − 2b − a2 + 1 18 2 9 ,b = 4 3 2 2 3 2 2 1 6 2 3 − 2 2 − 13 0 1 3 2 2a − ab − 1 18 1 6 2 2b − 1 6 ab − 17 18 2 9 2 9 2 9 1 9 b2 + T 2 3 2 2 3 2 2 3 − 2 2 1 3 2 4 3 2 − 13 0 4 3 2 2 1 1 6 2 1 0 0 = 112 Solution: Fill in the missing entries to make the matrix orthogonal. 0 1 0 0 0 1 − 1 3 − 1 3 Try 2 5 2 3 0 2 3 1 5 = 2 3 0 _ _ c 1 3 d 2 3 0 2 3 1 5 4 15 5 _ _ c d 4 15 5 cd + 2 9 4 15 5c− 8 45 cd + 2 9 d2 + 4 9 4 15 5d + 4 9 8 45 4 15 Would require that c = − 2 5 2 3 0 2 3 1 5 5d + 2 3 5 2 4 9 1 −5 3 5 ,d = − 3 5 1 3 −5 3 5 2 3 0 2 3 1 5 4 15 5 5 T 2 5 41 45 5c− . 4 15 c2 + 4 15 1 3 − 2 5 2 5 T 2 3 5 1 0 0 −5 3 5 4 15 = 0 1 0 so this seems to have 0 0 1 5 worked. 113 Solution: Find the eigenvalues and an orthonormal basis of eigenvectors for A. Diagonalize A by finding an orthogonal matrix, U and a diagonal matrix D such that U T AU = D. 6 A= 0 0 0 −1 1 0 −1 1 . T 1 0 0 0 − 12 1 2 0 2 2 1 2 1 2 6 0 1 0 2 0 −1 1 0 2 0 −1 0 1 0 − 12 1 2 0 1 2 1 2 2 2 6 2 2 = 0 0 0 −2 0 0 0 0 114 Solution: Find the eigenvalues and an orthonormal basis of eigenvectors for A. Diagonalize A by finding an orthogonal matrix, U and a diagonal matrix D such that U T AU = D. A= 13 1 4 1 13 4 4 4 10 . Hint: Two eigenvalues are 12 and 18. 13 1 4 1 13 4 4 4 10 , eigenvectors: − 16 6 − 16 6 1 3 − 12 2 ↔ 6, 1 2 2 3 ↔ 12, 2 0 1 3 1 3 3 1 3 3 3 ↔ 18. The matrix U has these as its columns. 115 Solution: Find the eigenvalues and an orthonormal basis of eigenvectors for A. Diagonalize A by finding an orthogonal matrix, U and a diagonal matrix D such that U T AU = D. A= 3 0 0 0 3 2 1 2 0 1 2 3 2 − 1 2 0 0 3 2 1 2 0 1 2 3 2 . , eigenvectors: 0 1 2 3 0 0 2 ↔ 1, 2 1 1 2 2 1 2 2 ↔ 2, 0 ↔ 3. These vectors are 0 the columns of the matrix U. 116 Solution: Find the eigenvalues and an orthonormal basis of eigenvectors for A. Diagonalize A by finding an orthogonal matrix, U and a diagonal matrix D such that U T AU = D. 17 −7 −4 A= −7 17 −4 −4 −4 14 Hint: Two eigenvalues are 18 and 24. 17 −7 −4 −7 17 −4 −4 −4 14 , eigenvectors: . 1 3 3 1 3 1 3 3 − 16 6 ↔ 6, − 12 2 ↔ 18, − 16 6 1 3 3 1 2 2 3 ↔ 24 . The 2 0 matrix U has these as its columns. 117 Solution: Find the eigenvalues and an orthonormal basis of eigenvectors for A. Diagonalize A by finding an orthogonal matrix, U and a diagonal matrix D such that U T AU = D. 3 A= 0 0 0 −1 1 0 −1 1 . T 1 0 0 0 − 12 1 2 0 2 2 1 2 1 2 3 0 1 0 2 0 −1 1 0 2 0 −1 0 1 0 − 12 1 2 0 1 2 1 2 2 2 3 = 2 0 0 −2 0 0 2 0 0 0 118 Solution: Find the eigenvalues and an orthonormal basis of eigenvectors for A. Diagonalize A by finding an orthogonal matrix, U and a diagonal matrix D such that U T AU = D. 1 A= 0 0 0 −4 1 0 −4 1 . T 1 0 0 0 − 12 2 0 1 2 2 1 2 1 2 0 0 1 2 0 −4 1 0 − 12 2 2 0 −4 0 1 1 0 1 2 0 1 2 1 2 2 1 = 2 2 0 0 0 −5 0 0 −3 0 119 Solution: Find the eigenvalues and an orthonormal basis of eigenvectors for A. Diagonalize A by finding an orthogonal matrix, U and a diagonal matrix D such that U T AU = D. 3 0 0 A= . 0 2 1 0 1 2 T 1 0 0 0 − 12 2 0 1 2 2 1 2 1 2 3 0 0 1 0 2 0 2 1 0 − 12 2 2 0 1 2 0 1 2 0 2 1 2 1 2 3 0 0 2 2 = 0 1 0 0 0 3 120 Solution: Explain why a matrix, A is symmetric if and only if there exists an orthogonal matrix, U such that A = U T DU for D a diagonal matrix. If A is given by the formula, then A T = U T D T U = U T DU = A Next suppose A = A T . Then by the theorems on symmetric matrices, there exists an orthogonal matrix U such that UAU T = D for D diagonal. Hence A = U T DU 121 Solution: The proof of a theorem concluded with the following observation. If −ta + t 2 b ≥ 0 for all t ∈ R and b ≥ 0, then a = 0. Why is this so? If a ≠ 0, then the derivative of the function ft = −ta + t 2 b is non zero at t = 0. However, this requires that ft < 0 for values of t near 0. 122 Solution: Using Schur’s theorem, show that whenever A is an n × n matrix, detA equals the product of the eigenvalues of A. There exists U unitary such that A = U ∗ TU such that T is uppser triangular. Thus A and T are similar. Hence they have the same determinant. Therefore, detA = detT, but detT equals the product of the entries on the main diagonal which are the eigenvalues of A. 123 Solution: In the proof of a theorem, the following argument was used. If x ⋅ w = 0 for all w ∈ R n , then x = 0. Why is this so? Let w = x. 124 Solution: Suppose A is a 3 × 2 matrix. Is it possible that A T is one to one? What does this say about A being onto? Prove your answer. A T is a 2 × 3 and so it is not one to one by an earlier corollary. Not every column can be a pivot column. 125 Solution: Find the least squares solution to the system 6y − 3x = 3 x + 3y = 2 x + 4y = 1 29 − 275 11 25 126 Solution: Find the least squares solution to the system 3y − 3x = 3 2x + 3y = 2 x + 6y = 1 − 31 83 34 83 127 Solution: Find the least squares solution to the system 2x + y = 3 x + 3y = 1 x + 4y = 1 118 75 4 − 25 128 Solution: You are doing experiments and have obtained the ordered pairs, 0, 1, 1, 2, 2, 3. 5, 3, 4 Find m and b such that y = mx + b approximates these four points as well as possible. Now do the same thing for y = ax 2 + bx + c, finding a, b, and c to give the best approximation. Here is the second of the two. You really want the following system. 0a + 0b + c = 1 a+b+c = 2 4a + 2b + c = 3. 5 9a + 3b + c = 4 T 0 0 1 0 0 1 1 1 1 1 1 1 4 2 1 4 2 1 9 3 1 9 3 1 98 36 14 = 36 14 6 14 4 6 98 36 14 a 36 14 6 b 14 4 c 1 1 1 1 98 36 14 a 52. 0 36 14 6 b 14 4 c 6 6 1 0 1 4 9 = 0 1 2 3 = 52. 0 2 = 3. 5 10. 5 4 −0. 125 , Solution is: 21. 0 1. 425 10. 5 0. 925 y -4 21. 0 5 -2 2 4 -5 x 129 Solution: You are doing experiments and have obtained the ordered pairs, 0, 2. 0, 1, 1. 0, 2, 2. 0, 3, 2. 0 Find m and b such that y = mx + b approximates these four points as well as possible. Now do the same thing for y = ax 2 + bx + c, finding a, b, and c to give the best approximation. Here is the second of the two. You really want the following system. 0a + 0b + c = 2. 0 a + b + c = 1. 0 4a + 2b + c = 2. 0 9a + 3b + c = 2. 0 T 0 0 1 0 0 1 1 1 1 1 1 1 4 2 1 4 2 1 9 3 1 9 3 1 98 36 14 = 36 14 6 14 4 6 98 36 14 a 36 14 6 b 14 4 c 1 1 1 1 98 36 14 a 27. 0 36 14 6 b 14 4 c 6 6 0 1 4 9 = = 0 1 2 3 11. 0 7. 0 2. 0 1. 0 2. 0 27. 0 = 7. 0 2. 0 0. 25 , Solution is: 11. 0 −0. 65 1. 85 0. 25x 2 − 0. 65x + 1. 85 y 10 5 -4 -2 0 2 4 x 130 Solution: You are doing experiments and have obtained the ordered pairs, 0, 7. 0, 1, −3. 0, 2, 1. 0, 3, 1. 0 Find m and b such that y = mx + b approximates these four points as well as possible. Now do the same thing for y = ax 2 + bx + c, finding a, b, and c to give the best approximation. Here is the second of the two. You really want the following system. 0a + 0b + c = 7. 0 a + b + c = −3. 0 4a + 2b + c = 1. 0 9a + 3b + c = 1. 0 T 0 0 1 0 0 1 1 1 1 1 1 1 4 2 1 4 2 1 9 3 1 9 3 1 98 36 14 = 36 14 6 14 4 6 98 36 14 a 36 14 6 b 14 4 c 1 1 1 1 98 36 14 a 10. 0 36 14 6 b 14 4 c 6 6 7. 0 0 1 4 9 = = 10. 0 −3. 0 0 1 2 3 = 1. 0 6. 0 1. 0 2. 5 −8. 9 , Solution is: 2. 0 6. 0 6. 1 2. 5x 2 − 8. 9x + 6. 1 y 100 50 -4 131 -2 0 2 2. 0 4 x Solution: Suppose you have several ordered triples, x i , y i , z i . Describe how to find a polynomial, z = a + bx + cy + dxy + ex 2 + fy 2 for example giving the best fit to the given ordered triples. Is there any reason you have to use a polynomial? Would similar approaches work for other combinations of functions just as well? You do something similar to the above. You write an equation which would be satisfied if each x i , y i , z i were a real solution. Then you obtain the least squares solution. 132 Solution: The two level surfaces, 2x + 3y − z + w = 0 and 3x − y + z + 2w = 0 intersect in a subspace of R 4 , find a basis for this subspace. Next find an orthonormal basis for this subspace. Using row reduced echelon form, it is easy to obtain −2z − 7w w + 5z 11z 11w Thus a basis is −2 −7 5 1 , 11 0 0 11 Then an orthonormal basis is 46 − 3135 6 209 1 − 15 6 1 6 6 11 30 6 1 1254 , 6 209 1 − 330 6 209 5 209 0 6 209 133 Solution: −3 Find an orthonormal basis for the span of the vectors 1 0 3 − 10 10 1 10 10 0 134 Solution: , 1 110 110 3 110 110 1 11 10 11 − 111 11 , − 113 11 1 11 11 1 , 1 4 0 , 0 1 . The set, V ≡ x, y, z : 2x + 3y − z = 0 is a subspace of R 3 . Find an orthonormal basis for this subspace. The subspace is of the form x y 2x + 3y 1 and a basis is 0 , 0 Therefore, an orthonormal basis is 1 2 3 1 5 − 353 5 14 5 , 0 2 5 5 1 14 5 14 3 70 5 14 135 Solution: Find an orthonormal basis for the span of the vectors 3, −4, 0, 7, −1, 0, 1, 7, 1. 3/5 4/5 −4/5 , 0 , 3/5 0 0 0 1 136 Solution: Find an orthonormal basis for the span of the vectors 3, 0, −4, 5, 0, 10, −7, 1, 1 3 5 0 − 4 5 , 0 , 0 4 5 3 5 1 0 137 Solution: Using the Gram Schmidt process, find an orthonormal basis for the span of the vectors, 1, 2, 1, 0, 2, −1, 3, 1, and 1, 0, 0, 1. To do this, have the computer use the QR factorization. 1 2 1 2 −1 0 1 3 0 0 1 1 = 1 6 1 3 6 1 6 0 2 3 6 1 6 − 29 6 5 18 2 3 1 9 2 3 2 3 5 111 1 333 3 37 3 37 17 − 333 3 37 22 333 3 37 7 111 2 − 111 111 111 − 371 111 7 − 111 111 1 2 6 3 2 2 3 5 18 0 0 1 9 0 0 0 ⋅ 1 6 6 6 2 3 3 37 0 Then a solution is 1 6 1 3 6 1 6 6 1 6 6 2 3 − 29 2 3 , 5 18 1 9 0 , 2 3 2 3 5 111 3 37 1 333 17 − 333 22 333 3 37 3 37 3 37 138 Solution: 3 Find an orthonormal basis for the span of the vectors 1 0 3 10 1 10 1 − 410 410 10 3 410 , 10 2 41 0 , 410 10 41 2 41 − 416 41 1 41 41 1 , 1 4 0 , 0 . 1 41 139 Solution: Find an orthonormal basis for the span of the vectors 3, 0, −4, 11, 0, 2, 1, 1, 7 3 11 1 0 0 −4 2 − 4 5 4 5 , 4 5 0 − 7 3 5 0 = 1 3 5 0 3 5 4 5 5 −5 0 1 0 10 5 3 5 0 1 0 0 5 0 0 , 1 0 140 Solution: Let A, B be a m × n matrices. Define an inner product on the set of m × n matrices by A, B F ≡ traceAB ∗ . Show this is an inner product satisfying all the inner product axioms. Recall for M an n × n matrix, n traceM ≡ ∑ i=1 M ii . The resulting norm, ||⋅||F is called the Frobenius norm and it can be used to measure the distance between two matrices. It satisfies the properties of an inner product. Note that traceAB ∗ = ∑ ∑ A ik B ik = ∑ ∑ A ik B ik = traceBa k i k i so A, B F = B, A F The product is obviously linear in the first argument. If A, A F = 0, then ∑ ∑ A ik A ik = ∑|A ik |2 = 0 k i i,k 141 Solution: Let A be an m × n matrix. Show ||A||2F ≡ A, A F = ∑ σ 2j j where the σ j are the singular values of A. From the singular value decomposition, σ 0 U ∗ AV = 0 0 σ 0 ,A=U V∗ 0 0 Then σ 0 traceAa = trace U 0 0 σ2 0 = trace U 0 0 V∗ V U∗ σ 0 0 0 = trace U∗ σ2 0 0 0 142 Solution: Here is a matrix A 5 0 0 0 3 10 2 10 − 101 2 10 0 3 10 2 10 − 101 2 10 Find a singular value decomposition. 1 A= 0 0 0 1 2 1 2 0 5 0 0 1 2 − 12 2 0 2 0 0 1 2 0 0 0 0 2 143 Solution: Here is a matrix A 2 0 3 10 1 10 0 10 − 101 10 10 3 10 10 = ∑ σ 2j j 3 + 1 3 5 2 − 2 3 1 3 5 2 3 15 22 1 11 11 − 3 11 − 1 11 3 11 + 11 + 2 3 11 − 1 33 3 15 22 2 11 1 33 1 33 2 3 11 2 3 11 2 3 11 − 1 33 1 11 3 11 − 1 11 3 11 1 11 5 22 3 11 − 2 11 − 3 11 + 2 33 5 22 1 33 2 11 2 3 11 2 3 11 Find a singular value decomposition. A= 1 2 2 1 6 1 2 2 − 16 2 3 1 3 0 2 3 2 3 − 13 3 1 3 1 3 5 0 0 1 2 2 3 0 2 0 1 2 2 − 223 2 11 3 0 0 1 0 3 22 1 11 2 11 2 11 − 111 11 1 11 3 11 11 11 144 Solution: Here is a matrix. 7 26 − 653 10 13 10 13 3 26 21 65 2 5 13 2 5 13 Find a singular value decomposition. 2 13 13 3 13 3 13 2 13 − 13 13 13 4 0 1 10 10 3 10 0 3 3 10 10 − 101 2 5 2 5 145 Solution: The trace of an n × n matrix M is defined as ∑ i M ii . In other words it is the sum of the entries on the main diagonal. If A, B are n × n matrices, show traceAB = traceBA. Now explain why if A = S −1 BS it follows traceA = traceB. Hint: For the first part, write these in terms of components of the matrices and it just falls out. traceAB = ∑ i ∑ k A ik B ki , traceBA = ∑ i ∑ k B ik A ki . These give the same thing. Now traceA = traceS −1 BS = traceBSS −1 = traceB. 146 Solution: Using the above problem and Schur’s theorem, show that the trace of an n × n matrix equals the sum of the eigenvalues. A = UTU ∗ for some unitary U. Then the trace of A equals the trace of T. However, the trace of T is just the sum of the eigenvalues of A. 147 Solution: If A is a general n × n matrix having possibly repeated eigenvalues, show there is a sequence A k of n × n matrices having distinct eigenvalues which has the property that the ij th entry of A k converges to the ij th entry of A for all ij. Hint: Use Schur’s theorem. A = U ∗ TU where T is upper triangular and U is unitary. Change the diagonal entries of T slightly so that the resulting upper triangular matrix T k has all distinct diagonal entries and T k → T in the sense that the ij th entry of T k converges to the ij th entry of T. Then let A k = U ∗ T k U. It follows that A k → A in the sense that corresponding entries converge. 148 Solution: Prove the Cayley Hamilton theorem as follows. First suppose A has a basis of eigenvectors v k nk=1 , Av k = λ k v k . Let pλ be the characteristic polynomial. Show pAv k = pλ k v k = 0. Then since v k is a basis, it follows pAx = 0 for all x and so pA = 0. Next in the general case, use the above problem to obtain a sequence A k of matrices whose entries converge to the entries of A such that A k has n distinct eigenvalues. Explain why A k has a basis of eigenvectors. Therefore, from the first part and for p k λ the characteristic polynomial for A k , it follows p k A k = 0. Now explain why and the sense in which lim p k A k = pA. k→∞ First say A has a basis of eigenvectors v 1 , ⋯, v n . A k v j = λ kj v j . Then it follows that n pAv k = pλ k v k . Hence if x is any vector, let x = ∑ k=1 x k v k and it follows that n pAx = pA ∑ xkvk n = k=1 = ∑ x k pAv k k=1 n n k=1 k=1 ∑ x k pλ k v k = ∑ x k 0v k = 0 Hence pA = 0. Now drop the assumption that A is nondefective. From the above, there exists a sequence A k which is non defective which converges to A and also p k λ → pλ uniformly on bounded intervals because these characteristic polynomials are defined in terms of determinants of the corresponding matrix. Thus the coefficients of p k converge to the corresponding coefficients of p. See the above construction of the A k . It is probably easiest to use the Frobinius norm for the last part. ‖p k A k − pA‖ F ≤ ‖p k A k − pA k ‖ F + ‖pA k − pA‖ F The first term converges to 0 because the convergence of A k to A implies all entries of A k lie in a bounded set. The second term converges to 0 also because the entries of A k converge to the corresponding entries of A. Also, p k A k = 0 from the first part. 149 Solution: Show that the Moore Penrose inverse A + satisfies the following conditions. AA + A = A, A + AA + = A + , A + A, AA + are Hermitian. Next show that if A 0 satisfies the above conditions, then it must be the Moore Penrose inverse and that if A is an n × n invertible matrix, then A −1 satisfies the above conditions. Thus the Moore Penrose inverse generalizes the usual notion of inverse but does not contradict it. Hint: Let U ∗ AV = Σ ≡ σ 0 0 0 and suppose V+ A 0 U = P Q R S where P is the same size as σ. Now use the conditions to identify P = σ, Q = 0 etc. First recall what the Moore Penrose invers was in terms of the singular value decomposition. It equals V σ −1 0 0 σ 0 U ∗ where A = U 0 V ∗ with U, V unitary and of the right size. 0 0 Therefore, σ 0 AA + A = U 0 0 σ 0 =U σ −1 0 V∗ V 0 U∗U 0 σ 0 V∗ 0 0 V∗ = A 0 0 Next A + AA + = V =V σ −1 0 0 0 σ −1 0 0 σ 0 U∗U V∗ V 0 0 σ −1 0 0 0 U∗ U∗ = A+ 0 Next, A+A = V σ −1 0 0 σ 0 U∗U 0 I 0 V∗ = V 0 0 0 0 V∗ which is clearly Hermitian. + AA = U σ 0 ∗ V V 0 0 σ −1 0 0 0 I 0 U∗ = U 0 0 U∗ which is Hermitian. Next suppose A 0 satisfies the above Penrose conditions. Then A0 = V P Q R S U∗, A = U σ 0 V∗ 0 0 and it is necessary to identify P, Q, R, S. Here P is the same size as σ. A0A = V =V P Q R S σP ∗ σR ∗ 0 σ 0 U∗U 0 0 V∗ = V Pσ 0 Rσ 0 V∗ V∗ 0 and so Rσ = 0 which implies R = 0. AA 0 = U =U σ 0 0 0 P∗ σ 0 ∗ Q σ 0 V∗ V U∗ P Q 0 S U∗ = U σP σQ 0 0 U∗ Hence Q ∗ σ = 0 so Q = 0. σ 0 A 0 AA 0 = V P 0 =V P 0 σ 0 P 0 0 S 0 0 0 S =V PσP 0 U∗ = A0 = V U∗U 0 S 0 0 P 0 V∗ V 0 0 0 S U∗ U∗ P 0 U∗ 0 S and so S = 0. Finally, AA 0 A = U =U =U σ 0 V∗ V 0 0 P 0 0 0 σ 0 P 0 σ 0 0 0 0 0 0 0 σPσ 0 0 0 V∗ = U σ 0 U∗U σ 0 0 0 V∗ V∗ V∗ 0 0 Then σPσ = σ and so P = σ −1 . Hence A 0 = A + . 150 Solution: Find the least squares solution to 1 1 1 1 x y 1 1+ a = b c Next suppose is so small that all 2 terms are ignored by the computer but the terms of order are not ignored. Show the least squares equations in this case reduce to 3+ x 3 + 3 + 2 y 3 = a+b+c a + b + 1 + c . Find the solution to this and compare the y values of the two solutions. Show that one of these is −2 times the other. This illustrates a problem with the technique for finding least squares solutions presented as the solutions to aAx = ay. One way of dealing with this problem is to use the QR factorization. This is illustrated in the next problem. It turns out that this helps alleviate some of the round off difficulties of the above. The least squares problem is 1 1 1 1 1 +1 1 1 1 1 1 1+ x y = 1 1 1 1 1 +1 a b c which reduces to +3 3 x + 3 2 + 2 + 3 a+b+c = y a + b + c + 1 and now, since the stupid computer doesn’t see 2 , what it sees is the following. +3 x + 3 2 + 3 y 3 = a+b+c a + b + c + 1 This yields the solution 1 3 2 2 a + b + c − + 3a + b − 2c 1 3 2 3a + 3b − 6c So what is the real least squares solution? 1 6 2 2 2 a + b + c + + 3a + b − 2c − 61 2 3a + 3b − 6c The two y values are very different one is −2 times the other. 151 Solution: Show that the equations aAx = ay can be written as R ∗ Rx = R ∗ Q ∗ y where R is upper triangular and R ∗ is lower triangular. Explain how to solve this system efficiently. Hint: You first find Rx and then you find x which will not be hard because R is upper triangular. A = QR and so a = R ∗ Q ∗ . Hence the least squares problem reduces to R ∗ Q ∗ QRx = R ∗ Q ∗ y R ∗ Rx =R ∗ Q ∗ y Then you can solve this by first solving R ∗ z = R ∗ Q ∗ y. Next, after finding z, you would solve for x in Rx = z. Numerical Methods For Linear Systems 152 Solution: Solve the system 4 1 1 x 1 5 2 y 0 2 6 z 2 = 1 3 using the Gauss Seidel method and the Jacobi method. Check your answer by also solving it using row operations. 4 1 1 x 1 5 2 y 0 2 6 z 2 = 0. 39 , Solution is: 1 3 −0. 09 0. 53 153 Solution: Solve the system 6 1 1 x 1 13 2 y 0 2 z 5 1 = 2 3 using the Gauss Seidel method and the Jacobi method. Check your answer by also solving it using row operations. 6 1 1 x 1 13 2 y 0 z 2 5 6. 060 6 × 10 −2 1 = 2 , Solution is: 6. 060 6 × 10 −2 3 0. 575 76 154 Solution: Solve the system 10 1 1 x 1 8 2 y 0 2 6 z 1 = 2 3 using the Gauss Seidel method and the Jacobi method. Check your answer by also solving it using row operations. 10 1 1 x 1 8 2 y 0 2 6 z 4. 128 4 × 10 −2 1 = 2 , Solution is: 0. 130 73 3 0. 456 42 155 Solution: Solve the system 12 1 1 x 1 8 2 y 0 2 7 z 1 = 2 3 using the Gauss Seidel method and the Jacobi method. Check your answer by also solving it using row operations. 12 1 1 x 1 8 2 y 0 2 7 z 3. 877 2 × 10 −2 1 = , Solution is: 2 0. 148 63 3 0. 386 11 156 Solution: Solve the system 13 1 1 x 1 13 2 y 0 2 z 5 1 = 2 3 using the Gauss Seidel method and the Jacobi method. Check your answer by also solving it using row operations. 13 1 1 x 1 13 2 y 0 2 z 5 2. 784 8 × 10 −2 1 = , Solution is: 2 3 6. 329 1 × 10 −2 0. 574 68 157 Solution: If you are considering a system of the form Ax = b and A −1 does not exist, will either the Gauss Seidel or Jacobi methods work? Explain. What does this indicate about using either of these methods for finding eigenvectors for a given eigenvalue? When this method works, there is a unique solution and so these methods are no good for the eigenvalue problem. Numerical Methods For Eigenvalues 158 Solution: Using the power method, find the eigenvalue correct to one decimal place having largest 0 −4 −4 absolute value for the matrix A = along with an eigenvector associated with 7 10 5 −2 0 6 this eigenvalue. 20 0 −4 −4 7 10 5 1 −2 0 6 1 −1729 412 493 480 034 304 1 = 1729 447 677 852 123 136 1729 348 721 805 623 296 −1729 412 493 480 034 304 1 1729 447 677 852 123 136 −0. 999 98 = 1729 447 677 852 123 136 1. 0 1729 348 721 805 623 296 0. 999 94 −1 It appears that an eigenvector is . Does it work? 1 1 0 −4 −4 −1 7 10 5 1 −2 0 6 1 −8 = −1 =8 8 8 . 1 1 159 Solution: Using the power method, find the eigenvalue correct to one decimal place having largest 15 6 1 absolute value for the matrix A = along with an eigenvector associated with this −5 2 1 1 2 7 eigenvalue. 20 15 6 1 1 −5 2 1 1 1 1 2 7 1. 150 0 × 10 22 = −5. 748 9 × 10 21 1. 729 4 × 10 18 1. 150 0 × 10 22 1 1. 150 0×10 22 −5. 748 9 × 10 21 1. 729 4 × 10 1. 0 = −0. 499 9 1. 503 8 × 10 −4 18 1 −. 5 This is pretty close to . How well does this work? 0 15 6 1 1 −5 2 1 −. 5 1 2 7 0 12. 0 = 1 = 12 −6. 0 0 −. 5 0 160 Solution: Using the power method, find the eigenvalue correct to one decimal place having largest 10 4 2 absolute value for the matrix A = along with an eigenvector associated with −3 2 −1 0 this eigenvalue. 0 4 20 10 4 2 1 −3 2 −1 1 0 1 0 4 3. 458 8 × 10 18 = −1. 729 4 × 10 18 1. 099 5 × 10 12 3. 458 8 × 10 18 1 3. 458 8×10 18 −1. 729 4 × 10 1. 0 = 18 −0. 5 3. 178 8 × 10 −7 1. 099 5 × 10 12 1 −. 5 This looks a lot like . How well does this work? 0 10 4 2 1 −3 2 −1 0 0 8. 0 = −. 5 4 1 =8 −4. 0 0 −. 5 0 0 161 Solution: Using the power method, find the eigenvalue correct to one decimal place having largest 15 14 −3 absolute value for the matrix A = along with an eigenvector associated −13 −18 9 5 −1 10 with this eigenvalue. 20 14 −3 −13 −18 9 1 −1 1 15 5 10 −6. 044 6 × 10 23 1 = 1. 208 9 × 10 24 −6. 044 6 × 10 23 −6. 044 6 × 10 23 1 1. 208 9×10 24 1. 208 9 × 10 24 −0. 500 01 = 1. 0 −0. 500 01 −6. 044 6 × 10 23 −. 5 Looks like 1 −. 5 14 −3 −. 5 −13 −18 9 1 −1 −. 5 15 5 10 −. 5 8. 0 = −16. 0 = −16 8. 0 162 Solution: Find the eigenvalues and eigenvectors for the matrix 1 −. 5 1 2 3 2 2 1 3 1 4 First of all, note that the matrix is symmetric and so all the eigenvalues are real. You can get fairly close to them by just graphing. 1 2 3 , characteristic polynomial: X 3 − 7X 2 + 15 2 2 1 3 1 4 X − 7X 2 + 15 3 y 10 -2 -10 -20 -30 2 4 6 x You can see that the eigenvalues are close to −1. 5, 1. 8, 6. 5. Use the shifted inverse power method. −1 1 2 3 1 0 0 + 1. 5 2 2 1 = 0 1 0 3 1 4 0 0 1 −1. 939 4 −2. 060 6 −1. 939 4 1. 151 5 0. 848 48 1 −2. 060 6 0. 848 48 1. 151 5 1 −1. 939 4 1. 151 5 0. 848 48 −2. 060 6 0. 848 48 1. 151 5 4. 432 0 × 10 14 1 4. 432 0 × 10 14 −2. 021 3 × 10 14 −1. 939 4 −2. 060 6 20 4. 424 2 1 4. 432 0×10 14 4. 424 2 = −2. 021 3 × 10 14 −2. 111 0 × 10 14 1. 0 = −2. 111 0 × 10 14 −0. 456 07 Thus we would expect −0. 476 31 1. 0 −0. 456 07 to be close to an eigenvector for A. Then −0. 476 31 1 2 3 1. 0 2 2 1 −0. 456 07 3 1 4 −0. 476 31 −1. 341 1 = 0. 611 55 and so the eigenvalue should be close to 0. 638 69 −1. 341 1. −1. 341 1 1. 0 Check: −1. 341 1 −0. 456 07 −0. 476 31 = 0. 611 64 0. 638 78 so it works very well, to three decimal places. Next find the one close to 1. 8 −1 1 2 3 −0. 303 03 −0. 757 58 0. 757 58 1 0 0 − 1. 8 2 2 1 3 1 4 = 0 1 0 0 0 1 −0. 303 03 −0. 757 58 0. 757 58 1 0. 757 58 −2. 251 1 1 1 −5. 129 7×10 17 8. 122 6 × 10 = 2 2 1 −1. 583 4 3 1 4 1. 0 1. 261 1 × 10 17 = 8. 122 6 × 10 17 −5. 129 7 × 10 17 −1. 583 4 −5. 129 7 × 10 17 −0. 245 84 −2. 251 1 3. 679 7 −0. 245 84 17 1 2 3 0. 757 58 1 3. 679 7 1. 261 1 × 10 17 3. 679 7 20 −0. 757 58 −5. 822 5 3. 679 7 −0. 757 58 −5. 822 5 This ought to be close. 1. 0 −0. 412 64 = −2. 658 5 1. 679 1 An eigenvalue ought to be clost to 1. 679 1. −0. 245 84 Check: 1. 679 1 −0. 412 79 = −1. 583 4 −2. 658 7 1. 0 It worked well. One could iterate some 1. 679 1 more if desired. Next find the one close to 6. 5 −1 1 2 3 2 2 1 1 0 0 − 6. 5 3 1 4 1. 673 5 1. 306 1 2. 530 6 = 0 1 0 1. 306 1 0. 775 51 1. 877 6 0 0 1 2. 530 6 1. 877 6 3. 387 8 20 1. 673 5 1. 306 1 2. 530 6 1 1. 306 1 0. 775 51 1. 877 6 1 2. 530 6 1. 877 6 3. 387 8 1 5. 758 6 × 10 15 1 8. 067 9×10 15 4. 201 0 × 10 15 0. 713 77 2 2 1 0. 520 71 3 1 4 1. 0 = 4. 201 0 × 10 15 8. 067 9 × 10 15 0. 713 77 = 8. 067 9 × 10 15 1 2 3 5. 758 6 × 10 15 0. 520 71 1. 0 4. 755 2 = 3. 469 0 6. 662 eigenvalue is about 6. 662 and an 0. 713 77 eigenvector ought to be about . 0. 520 71 1. 0 0. 713 77 Check: 6. 662 4. 755 1 = 0. 520 71 , very close to the matrix times this 3. 469 0 1. 0 6. 662 eigenvector. −0. 245 84 0. 713 77 6. 662, , 0. 520 71 1. 0 −1. 583 4 1. 679 1, 1. 0 , −1. 341 1, −0. 456 07 −0. 476 31 1. 0 163 Solution: 3 2 1 Find the eigenvalues and eigenvectors of the matrix A = 2 1 3 numerically. In this 1 3 2 case the exact eigenvalues are ± 3 , 6. Compare with the exact answers. Graphing the characteristic equation, you see that the eigenvectors are close to 1. 7, −1. 7, 6. You see this from graphing. In fact 6 is an eigenvalue and then you find the eigenvector the usual way. 1 . You need to find the others. 1 1 −1 3 2 1 1 0 0 − 1. 7 2 1 3 = 0 1 0 1 3 2 19. 471 19. 471 0 0 1 1. 289 6 4. 016 9 1 −14. 165 4. 016 9 10. 381 1 −1. 216 9 × 10 25 3. 260 8 × 10 4. 016 9 −14. 165 4. 016 9 10. 381 −1. 216 9 × 10 25 = 3. 260 8 × 10 24 8. 908 7 × 10 24 1. 0 1 −1. 216 9×10 25 8. 908 7 × 10 24 732. 1. 289 6 1 −5. 074 0 24 −5. 074 0 20 −5. 074 0 −14. 165 −5. 074 0 −14. 165 = −0. 267 96 −0. 732 08 3 2 1 1. 0 2 1 3 −0. 267 96 1 3 2 −0. 732 08 1. 732 = −0. 464 2 −1. 268 so the eigenvalue ought to be about 1. 1. 0 1. 732 = −0. 267 96 Check: 1. 732 −0. 464 11 −0. 732 08 −1. 268 0 −1 3 2 1 −1. 168 8 5. 194 8 −3. 896 1 1 0 0 + 1. 7 2 1 3 = 0 1 0 1 3 2 0 0 1 −19. 351 14. 286 −3. 896 1 14. 286 −10. 260 20 −1. 168 8 5. 194 8 −3. 896 1 −4. 066 7 × 10 23 1 5. 194 8 −19. 351 14. 286 1 −3. 896 1 14. 286 −10. 260 1 −4. 066 7 × 10 23 = 1. 517 7 × 10 24 −1. 111 1 × 10 24 −0. 267 95 1 1. 517 7×10 24 1. 517 7 × 10 24 −1. 111 1 × 10 5. 194 8 = 1. 0 −0. 732 09 24 3 2 1 −0. 267 95 2 1 3 1. 0 1 3 2 −0. 732 09 0. 464 06 = −1. 732 2 1. 267 9 It appears the eigenvalue is −1. 732 2 or close to it. How well does it work? −0. 267 95 −1. 732 2 0. 464 14 = 1. 0 −0. 732 09 −1. 732 2 of course 3 = 1. 732 1 1. 268 1 164 Solution: 3 2 1 Find the eigenvalues and eigenvectors of the matrix A = 2 5 3 numerically. The exact 1 3 2 eigenvalues are 2, 4 + 15 , 4 − 15 . Compare your numerical results with the exact values. Is it much fun to compute the exact eigenvectors? This time, I shall just raise to a high power and take a QR factorization. The first column of Q will then be close to an eigenvector unless the first column of A is an eigenvector. 20 1. 448 7 × 10 17 2. 713 5 × 10 17 1. 632 8 × 10 17 3 2 1 2 5 3 1 3 2 0. 415 99 = 2. 713 5 × 10 17 5. 082 3 × 10 17 3. 058 1 × 10 17 1. 632 8 × 10 17 3. 058 1 × 10 17 1. 840 1 × 10 17 0. 806 58 0. 779 18 −7. 803 9 × 10 0. 468 86 −0. 585 95 0. 419 97 −2 −0. 621 92 0. 660 94 = 3. 482 5 × 10 17 6. 522 6 × 10 17 3. 924 8 × 10 17 0 1. 539 × 10 13 1. 324 1 × 10 13 0 0 1. 399 9 × 10 12 T 0. 415 99 0. 806 58 0. 419 97 3 2 1 0. 415 99 0. 779 18 −7. 803 9 × 10 −2 −0. 621 92 2 5 3 0. 779 18 −7. 803 9 × 10 −2 −0. 621 92 −0. 585 95 1 3 2 0. 468 86 0. 468 86 0. 660 94 0. 806 58 −0. 585 95 8. 768 9 × 10 −5 3. 295 6 × 10 −5 7. 873 0 8. 768 9 × 10 −5 1. 746 2 0. 641 05 −5 0. 641 05 0. 380 82 3. 295 6 × 10 Now it appears that one eigenvalue is close to 7. 873 0 and the others are the eigenvalues for the 1. 746 2 0. 641 05 lower right corner. , eigenvalues: 2. 000 0, 0. 127 02. Note that 2 is an 0. 641 05 0. 380 82 −1 3 2 1 eigenvalue and you will see this if you try to take 2 5 3 1 0 0 −2 1 3 2 1 2 1 0 1 0 0 0 1 0 0 1 −1 0 , row echelon form: 2 3 3 0 3 0 1 0 1 3 0 0 0 0 0 0 −3 an eigenvector is 1 1 Check: 3 2 1 −3 2 5 3 1 1 3 2 1 −6 = 2 2 next, look for the one close to . 127 −1 3 2 1 2 5 3 1 0 0 −. 127 1 3 2 526. 16 526. 16 = 0 1 0 0 0 1 −3087. 5 4664. 4 10 18133. −27395. 1 4664. 4 −27395. 41388. 1 −3087. 5 18133. −27395. 4664. 4 −27395. 41388. 2. 133 9 × 10 46 1 −3087. 5 −3087. 5 4664. 4 = −1. 253 3 × 10 47 1. 893 4 × 10 47 . 0. 419 97 0. 660 94 2. 133 9 × 10 46 1 1. 893 4×10 47 0. 112 7 = −1. 253 3 × 10 47 1. 893 4 × 10 3 2 1 0. 112 7 2 5 3 −0. 661 93 1 3 2 1. 0 1. 0 0. 014 24 = −0. 084 25 0. 126 91 1. 430 3 × 10 −2 0. 112 7 0. 126 91 −0. 661 93 47 −0. 661 93 = −8. 400 6 × 10 −2 1. 0 so this worked well. 0. 126 91 Next look for the other one. −1 3 2 1 2 5 3 1 0 0 − 7. 87 = 0 1 0 1 3 2 0 0 1 57. 869 108. 7 65. 414 1 108. 7 203. 45 122. 50 1 65. 414 122. 50 73. 578 1 1 4. 156 9×10 50 4. 156 9 × 10 = 2. 501 4 × 10 50 0. 533 88 2 5 3 1. 0 1 3 2 0. 601 75 1. 0 0. 601 75 108. 7 203. 45 122. 50 2. 219 3 × 10 50 = 4. 156 9 × 10 50 2. 501 4 × 10 50 1. 0 0. 601 75 4. 203 4 = 7. 873 so the eigenvalue is about 7. 873 4. 737 4 0. 533 88 Check: 7. 873 65. 414 0. 533 88 50 3 2 1 108. 7 65. 414 122. 50 73. 578 20 2. 219 3 × 10 50 57. 869 4. 203 2 = 7. 873 . Worked well. 4. 737 6 4 + 15 = 7. 873 0 165 Solution: 0 2 1 Find the eigenvalues and eigenvectors of the matrix A = 2 5 3 numerically. We don’t 1 3 2 know the exact eigenvalues in this case. Check your answers by multiplying your numerically computed eigenvectors by the matrix. The characteristic polynomial is x 3 − 7x 2 − 4x + 1. x 3 − 7x 2 − 4x + 1 y -2 2 -20 4 6 x -40 -60 The eigenvalues are close to −. 7, . 4, 7. 5. Now we find these more closely and also the eigenvectors. −1 0 2 1 1 0 0 +. 7 2 5 3 = 0 1 0 1 3 2 88. 889 −11. 111 88. 889 −32. 963 3. 703 7 −11. 111 3. 703 7 0. 370 37 0 0 1 20 −236. 67 88. 889 88. 889 −32. 963 3. 703 7 1 −11. 111 3. 703 7 0. 370 37 1 −11. 111 2. 585 6 × 10 48 1 2. 585 6 × 10 48 −9. 693 9 × 10 −236. 67 = −9. 693 9 × 10 47 1. 193 1 × 10 47 1. 0 1 2. 585 6×10 48 47 = −0. 374 92 4. 614 4 × 10 −2 1. 193 1 × 10 47 −0. 703 70 0 2 1 1. 0 2 5 3 −0. 374 92 1 3 2 4. 614 4 × 10 −2 = 0. 263 83 −3. 247 2 × 10 −2 λ 1 = −0. 703 70 −0. 703 7 1. 0 Check: −0. 703 70 = −0. 374 92 4. 614 4 × 10 0. 263 83 −2 −3. 247 2 × 10 −2 −1 0 2 1 −0. 990 34 −0. 120 77 0. 845 41 1 0 0 −. 4 2 5 3 = 0 1 0 1 3 2 0 0 1 −0. 990 34 −0. 120 77 0. 845 41 1 0. 845 41 −3. 526 6 1 −3. 045 5 × 10 12 5. 824 3 × 10 12 0. 845 41 −3. 526 6 −0. 242 19 1 5. 824 3×10 12 = −0. 522 90 1. 0 1. 932 4 −1. 410 6 × 10 12 1 1. 932 4 −1. 410 6 × 10 12 1. 932 4 20 −0. 120 77 −0. 990 34 1. 932 4 −0. 120 77 −0. 990 34 = −3. 045 5 × 10 12 5. 824 3 × 10 12 0 2 1 −0. 242 19 2 5 3 −0. 522 90 1 3 2 1. 0 −0. 045 8 = −0. 098 88 0. 189 11 λ 2 = 0. 189 11 −4. 580 1 × 10 −2 −0. 242 19 = −0. 522 90 check: 0. 189 11 −9. 888 6 × 10 −2 1. 0 worked well. 0. 189 11 −1 0 2 1 1 0 0 − 7. 5 2 5 3 1 3 2 5. 428 6 16. 0 9. 714 3 = 0 1 0 0 0 1 16. 0 46. 0 1 9. 714 3 28. 0 16. 857 6. 878 8 × 10 36 4. 174 9 × 10 36 = 6. 878 8 × 10 36 4. 174 9 × 10 36 1 2. 386 4 × 10 36 28. 0 2. 386 4 × 10 36 1 28. 0 46. 0 9. 714 3 28. 0 16. 857 20 5. 428 6 16. 0 9. 714 3 16. 0 0. 346 92 1 6. 878 8×10 36 = 1. 0 0. 606 92 0 2 1 0. 346 92 2. 606 9 2 5 3 1. 0 1 3 2 0. 606 92 4. 560 8 0. 346 92 2. 607 0 = 7. 514 6 λ 3 = 7. 514 6 Check: 7. 514 6 1. 0 0. 606 92 = 7. 514 6 4. 560 8 166 Solution: 0 2 1 Find the eigenvalues and eigenvectors of the matrix A = 2 0 3 numerically. We don’t 1 3 2 know the exact eigenvalues in this case. Check your answers by multiplying your numerically computed eigenvectors by the matrix. 0 2 1. 0 2 0 3 1 3 2 0. 379 2 , eigenvectors: 0. 584 81 0. 717 08 ↔ 4. 975 4, 0. 814 41 0. 439 25 ↔ −0. 300 56, 0. 156 94 ↔ −2. 674 9 −0. 795 85 −0. 558 66 0. 416 76 167 Solution: 3 2 3 Consider the matrix A = 2 1 4 and the vector 1, 1, 1 T . Estimate the distance 3 4 0 between the Rayleigh quotient determined by this vector and some eigenvalue of A. Recall that there was a formula for this. Let 1 1 1 q= 3 2 3 1 2 1 4 1 3 4 0 1 = 22 3 3 Then there is an eigenvalue λ such that 3 2 3 λ − 22 3 ≤ 1 2 1 4 1 3 4 0 1 1 − 22 3 1 1 = 1 2 3 3 168 Solution: 1 2 1 Consider the matrix A = 2 1 4 and the vector 1, 1, 1 T . Estimate the distance 1 4 5 between the Rayleigh quotient determined by this vector and some eigenvalue of A. 1 1 1 q= 1 2 1 1 2 1 4 1 1 4 5 1 3 =7 Then |λ − 7| ≤ 1 2 1 1 2 1 4 1 1 4 5 1 3 1 −7 1 1 = 6 169 Solution: Using Gerschgorin’s theorem, find upper and lower bounds for the eigenvalues of 3 2 3 A= . 2 6 4 3 4 −3 From the bottom line, a lower bound is −10 From the second line, an upper bound is 12. 170 Solution: The QR algorithm works very well on general matrices. Try the QR algorithm on the following matrix which happens to have some complex eigenvalues. A= 1 2 3 1 2 −1 −1 −1 1 Use the QR algorithm to get approximate eigenvalues and then use the shifted inverse power method on one of these to get an approximate eigenvector for one of the complex eigenvalues. The real parts of the eigenvalues are larger than −4. I will consider the matrix which comes from adding 5I to the given matrix. Thus, consider 1 2 1 2 1 0 0 0 1 2 1 2 6 = 2 0 15 2 1 2 1 7 −1 −1 −1 6 1 2 −1 2 − 12 2 − 12 2 3 1 2 1 2 2 2 − 12 2 0 2 6 2 5 2 2 0 6 2 3 1 1 7 −1 0 −1 −1 6 0 2 1 2 11 2 2 = 6. 0 = 0 1 2 − 12 0 2 2 1 2 1 2 2 2 −0. 707 11 3. 535 5 1. 414 2 7. 5 0. 5 0 0. 5 5. 5 This is the upper Hessenberg matrix which goes with A. 35 −0. 707 11 3. 535 5 6. 0 1. 414 2 7. 5 0. 5 0 0. 5 5. 5 = 3. 582 3 × 10 29 4. 290 6 × 10 29 6. 739 6 × 10 29 6. 190 5 × 10 30 7. 422 7 × 10 30 1. 165 9 × 10 31 1. 409 7 × 10 30 1. 690 3 × 10 30 2. 655 3 × 10 30 This has a QR factorization with Q = 5. 633 4 × 10 −2 −0. 998 39 7. 287 3 × 10 −3 0. 973 49 5. 330 5 × 10 −2 −0. 222 43 0. 221 68 1. 962 4 × 10 −2 0. 974 92 Then −0. 707 11 3. 535 5 6. 0 Q T 1. 414 2 7. 5 0. 5 0 0. 5 5. 5 −1. 281 6 0. 230 07 7. 695 7 1. 310 2 × 10 Q= −4 5. 898 3 −3. 601 3 −1. 781 5 × 10 −5 0. 310 73 5. 406 1 It is now clear that the eigenvalues are approximately those of 5. 898 3 −3. 601 3 0. 310 73 5. 406 1 and 7. 695 7. That is, 5. 652 2 + 1. 028 8i, 5. 652 2 − 1. 028 8i, 7. 695 7 Subtracting 5 gives the approximate eigenvalues for the original matrix, 0. 652 2 + 1. 028 8i, 0. 652 2 − 1. 028 8i, 2. 695 7 Lets find the eigenvector which goes with the first of the above. −1 1 2 3 1 2 −1 −1 −1 1 1 0 0 − 0. 652 2 + 1. 028 8i 0 1 0 0 0 1 = −4251. 5 − 8395. 3i −16391. +3972. 6i −26508. +5426. 7i 1411. 0 + 4647. 5i 8687. 2 − 554. 36i 13959. −389. 59i 2431. 6 − 3583. 0i −5253. 6 − 5712. 0i −8094. 1 − 9459. 8i −4251. 5 − 8395. 3i −16391. +3972. 6i −26508. +5426. 7i 1411. 0 + 4647. 5i 8687. 2 − 554. 36i 13959. −389. 59i 2431. 6 − 3583. 0i −5253. 6 − 5712. 0i −8094. 1 − 9459. 8i 12 1 = 1 1 −3. 942 8 × 10 51 + 2. 747 1 × 10 51 i 2. 249 7 × 10 51 − 1. 044 0 × 10 51 i −1. 984 8 × 10 51 − 9. 746 7 × 10 50 i Then the first approximation will be −3. 942 8 × 10 51 + 2. 747 1 × 10 51 i 1 −3. 942 8 × 10 51 + 2. 747 1 × 10 51 i 2. 249 7 × 10 51 − 1. 044 0 × 10 51 i −1. 984 8 × 10 51 − 9. 746 7 × 10 50 i 1. 0 = −0. 508 31 − 8. 937 5 × 10 −2 i 0. 222 94 + 0. 402 53i −4251. 5 − 8395. 3i −16391. +3972. 6i −26508. +5426. 7i 1411. 0 + 4647. 5i 8687. 2 − 554. 36i 13959. −389. 59i 2431. 6 − 3583. 0i −5253. 6 − 5712. 0i −8094. 1 − 9459. 8i −3658. 8 − 18410. 0i 1. 0 −2 −0. 508 31 − 8. 937 5 × 10 i = 214. 5 + 9684. 9i 6594. 9 − 5577. 1i 0. 222 94 + 0. 402 53i The next approximate eigenvector is −3658. 8 − 18410. 0i 214. 5 + 9684. 9i 6594. 9 − 5577. 1i 1 −3658. 8 − 18410. 0i 1. 0 = −0. 508 31 − 8. 936 9 × 10 −2 i 0. 222 94 + 0. 402 53i ⋅ This didn’t change by much. In fact, it didn’t change at all. Thus the approximate eigenvalue is obtained by solving 1 = −3658. 8 − 18410. 0i λ − 0. 652 2 + 1. 028 8i Thus λ = 0. 652 19 + 1. 028 9i and the approximate eigenvector is 1. 0 −0. 508 31 − 8. 936 9 × 10 −2 i 0. 222 94 + 0. 402 53i Lets check it. 1 2 3 0. 652 2 + 1. 028 9i 1. 0 −2 2 −1 −0. 508 31 − 8. 936 9 × 10 i −1 −1 1 0. 222 94 + 0. 402 53i 1 = −0. 268 75 + 0. 491 90i 0. 652 19 + 1. 028 9i 1. 0 0. 652 19 + 1. 028 9i −0. 239 56 − 0. 581 27i = −2 −0. 508 31 − 8. 936 9 × 10 i 0. 222 94 + 0. 402 53i −0. 239 56 − 0. 581 29i −0. 268 76 + 0. 491 91i It worked very well. Thus you know the other eigenvalue for the other complex eigenvalue is 1. 0 −0. 508 31 + 8. 936 9 × 10 −2 i 0. 222 94 − 0. 402 53i 171 Solution: Use the QR algorithm to approximate the eigenvalues of the symmetric matrix 1 2 3 2 −8 1 3 20 1 2 3 2 −8 1 3 1 0 = 1 0 1. 250 3 × 10 17 −6. 367 1 × 10 17 3. 099 4 × 10 16 −6. 367 1 × 10 17 3. 242 5 × 10 18 −1. 578 3 × 10 17 3. 099 4 × 10 −1. 578 3 × 10 7. 683 2 × 10 16 0. 192 47 0. 835 59 −0. 980 14 0. 189 19 5. 939 7 × 10 −2 4. 771 2 × 10 −2 0. 515 74 0. 514 53 −0. 855 41 17 15 = 6. 496 1 × 10 17 −3. 308 2 × 10 18 1. 610 3 × 10 17 0 1. 826 7 × 10 13 1. 076 7 × 10 12 0 0 3. 180 1 × 10 11 T 0. 192 47 0. 835 59 −0. 980 14 0. 189 19 5. 939 7 × 10 −2 4. 771 2 × 10 −2 −7. 008 4 × 10 −6 3 0. 192 47 2 −8 1 −0. 980 14 3 3. 825 −0. 845 44 −0. 845 44 −2. 383 7 20 −0. 845 44 = −0. 845 44 −2. 383 7 1 0 4. 771 2 × 10 −0. 132 54 0. 991 18 7. 906 × 10 11 −1. 057 2 × 10 11 −1. 057 2 × 10 11 1. 422 7 × 10 10 0. 132 54 −0. 845 44 3. 825 −0. 132 54 0. 991 18 0. 991 18 −0. 845 44 −2. 383 7 3. 938 1 −9. 737 8 × 10 −5 −9. 737 8 × 10 −5 −2. 496 8 −0. 132 54 0. 991 18 Now the eigenvalues are close to −8. 441 4, 3. 938 1, −2. 496 8. To find eigenvectors, −1 2 3 1 0 0 + 8. 441 4 2 −8 1 3 1 = 0 1 0 0 0 0 1 −4375. 9 22285. 22285. −1. 134 9 × 10 −1084. 8 5524. 2 −1084. 8 5 5524. 2 −268. 77 20 −4375. 9 22285. −1084. 8 22285. −1. 134 9 × 10 5 5524. 2 1 −1084. 8 5524. 2 −268. 77 1 −9. 894 1 × 10 2 3 2 −8 1 3 1 0 = −0. 196 36 1 2. 032 6×10 101 2. 032 6 × 10 101 −3. 991 3 × 10 100 1 −3. 991 3 × 10 100 1 0. 515 74 = 0. 132 54 = 1. 0 −4. 867 7 × 10 −2 99 −0. 196 36 1. 0 −4. 867 7 × 10 −2 1. 657 6 = −8. 441 4 0. 410 92 2. 032 6 × 10 101 −9. 894 1 × 10 99 0. 514 53 0. 189 19 5. 939 7 × 1 −2 8. 919 8 × 10 7 0 T 1 0. 835 59 7. 976 4 × 10 11 −1. 066 7 × 10 11 0. 132 54 0. 991 18 2 4. 960 7 × 10 −5 −7. 008 4 × 10 −6 4. 960 7 × 10 −5 0. 991 18 1 −0. 855 41 0. 515 74 −8. 441 4 3. 825 0. 514 53 = −0. 855 41 −0. 196 36 8. 441 4 −1. 657 6 = 1. 0 −4. 867 7 × 10 , 8. 441 4 −2 −0. 410 9 −1 1 2 3 1 0 0 − 3. 938 1 2 −8 1 3 1 14217. 3360. 4 11683. = 0 1 0 0 0 0 1 11683. 2761. 6 9601. 3 8 14217. 3360. 4 11683. 1 3360. 4 794. 21 2761. 6 1 11683. 2761. 6 9601. 3 1 1. 600 8 × 10 35 1. 315 5 × 10 35 1. 0 2 0. 236 37 35 0. 821 78 3 1. 0 2 −8 1 0. 236 37 3 0. 821 78 1 = 1 1. 600 8×10 35 3. 783 8 × 10 34 1 0 1. 0 3. 938 1 = 0. 930 82 3. 236 4 3. 938 1 3. 938 1 = 0. 236 37 0. 821 78 0. 930 85 3. 236 3 λ 2 = 3. 938 1 1 2 −1 3 1 0 0 + 2. 496 8 2 −8 1 3 1 0 = 0 1 0 0 0 1 8 5680. 5 768. 27 −7133. 0 768. 27 103. 74 −964. 66 1 −7133. 0 −964. 66 8957. 4 1 −1. 400 5 × 10 31 = 1. 300 4 × 10 32 −0. 796 37 2 −8 1 −0. 107 70 3 1 0 −0. 107 70 1. 0 1. 988 2 = 768. 27 −7133. 0 768. 27 103. 74 −964. 66 −7133. 0 −964. 66 8957. 4 −1. 400 5 × 10 31 1. 300 4 × 10 32 1. 0 3 2 = −0. 796 37 1 1. 300 4×10 32 5680. 5 −1. 035 6 × 10 32 1 −1. 035 6 × 10 32 1 3. 783 8 × 10 34 := 1. 600 8 × 10 35 1. 315 5 × 10 3360. 4 794. 21 2761. 6 0. 268 86 −2. 496 8 −0. 796 37 1. 988 4 −2. 496 8 = −0. 107 70 0. 268 91 −2. 496 8 1. 0 −0. 196 36 1. 0 3. 938 2, , 0. 236 38 8. 441 4, , 1. 0 −4. 867 7 × 10 −2 0. 821 82 −0. 796 37 −2. 496 8, −0. 107 70 1. 0 172 Solution: 3 Try to find the eigenvalues of the matrix 3 1 −2 −2 −1 using the QR algorithm. It has 0 1 0 eigenvalues 1, i, −i. You will see the algorithm won’t work well. To see that the method will not work well, consider the powers of this matrix. 2 3 3 1 3 = −2 −2 −1 0 1 0 3 3 1 4 0 −2 −3 0 −2 −2 −1 3 = −2 −2 −1 0 1 0 3 3 1 1 1 −1 0 0 1 −2 −3 0 4 −2 −2 −1 0 1 0 1 0 0 = 0 1 0 0 0 1 Thus the powers of the matrix will repeat forever. Therefore, you cannot expect things to get small and be able to find the eigenvalues by looking at those of a block upper triangular matrix which has very small entries below the block diagonal. Recall that, in terms of size of the entries, you can at least theoretically consider the QR factorization of A k and then look at Q ∗ AQ however, not much can happen given the fact that there are only finitely many A k . The problem here is that there is no gap between the size of the eigenvalues. The matrix has eigenvalues equal to i, −i and 1. To produce a situation in which this unfortunate situation will not occur, simply add a multiple of the identity to the matrix and use the modified one. This is always a reasonable idea. You can identify an interval, using Gerschgorin’s theorem which will contain all the eigenvalues of the matrix. Then when you add a large enough multiple of the identity, the result will have all positive real parts for the eigenvalues. If you do this to this matrix, the problem goes away. The idea is to produce gaps between the absolute values of the eigenvalues. Fields 173 Solution: Prove the Euclidean algorithm: If m, n are positive integers, then there exist integers q, r ≥ 0 such that r < m and n = qm + r Hint: You might try considering S ≡ n − km : k ∈ N and n − km < 0 and picking the smallest integer in S or something like this. The hint is a good suggestion. Pick the first thing in S. By the Archimedean property, S ≠ ∅. That is km > n for all k sufficiently large. Call this first thing q + 1. Thus n − q + 1m < 0 but n − qm ≥ 0. Then n − qm < m and so 0 ≤ r ≡ n − qm < m. 174 Solution: The greatest common divisor of two positive integers m, n, denoted as q is a positive number which divides both m and n and if p is any other positive number which divides both m, n, then p divides q. Recall what it means for p to divide q. It means that q = pk for some integer k. Show that the greatest common divisor of m, n is the smallest positive integer in the set S S ≡ xm + yn : x, y ∈ Z and xm + yn > 0 Two positive integers are called relatively prime if their greatest common divisor is 1. First note that either m or −m is in S so S is a nonempty set of positive integers. By well ordering, there is a smallest element of S, called p = x 0 m + y 0 n. Either p divides m or it does not. If p does not divide m, then by the above problem, m = pq + r where 0 < r < p. Thus m = x 0 m + y 0 nq + r and so, solving for r, r = m1 − x 0 + −y 0 qn ∈ S. However, this is a contradiction because p was the smallest element of S. Thus p|m. Similarly p|n. Now suppose q divides both m and n. Then m = qx and n = qy for integers, x and y. Therefore, p = mx 0 + ny 0 = x 0 qx + y 0 qy = qx 0 x + y 0 y showing q|p. Therefore, p = m, n. 175 Solution: A positive integer larger than 1 is called a prime number if the only positive numbers which divide it are 1 and itself. Thus 2,3,5,7, etc. are prime numbers. If m is a positive integer and p does not divide m where p is a prime number, show that p and m are relatively prime. Suppose r is the greatest common divisor of p and m. Then if r ≠ 1, it must equal p because it must divide p. Hence there exist integers x, y such that p = xp + ym which requires that p must divide m which is assumed not to happen. Hence r = 1 and so the two numbers are relatively prime. 176 Solution: There are lots of fields. This will give an example of a finite field. Let Z denote the set of integers. Thus Z = ⋯, −3, −2, −1, 0, 1, 2, 3, ⋯. Also let p be a prime number. We will say that two integers, a, b are equivalent and write a ∼ b if a − b is divisible by p. Thus they are equivalent if a − b = px for some integer x. First show that a ∼ a. Next show that if a ∼ b then b ∼ a. Finally show that if a ∼ b and b ∼ c then a ∼ c. For a an integer, denote by a the set of all integers which is equivalent to a, the equivalence class of a. Show first that is suffices to consider only a for a = 0, 1, 2, ⋯, p − 1 and that for 0 ≤ a < b ≤ p − 1, a ≠ b. That is, a = r where r ∈ 0, 1, 2, ⋯, p − 1. Thus there are exactly p of these equivalence classes. Hint: Recall the Euclidean algorithm. For a > 0, a = mp + r where r < p. Next define the following operations. a + b ≡ a + b ab ≡ ab Show these operations are well defined. That is, if a = a ′ and b = b ′ , then a + b = a ′ + b ′ with a similar conclusion holding for multiplication. Thus for addition you need to verify a + b = a ′ + b ′ and for multiplication you need to verify ab = a ′ b ′ . For example, if p = 5 you have 3 = 8 and 2 = 7. Is 2 × 3 = 8 × 7? Is 2 + 3 = 8 + 7? Clearly so in this example because when you subtract, the result is divisible by 5. So why is this so in general? Now verify that 0, 1, ⋯, p − 1 with these operations is a Field. This is called the integers modulo a prime and is written Z p . Since there are infinitely many primes p, it follows there are infinitely many of these finite fields. Hint: Most of the axioms are easy once you have shown the operations are well defined. The only two which are tricky are the ones which give the existence of the additive inverse and the multiplicative inverse. Of these, the first is not hard. −x = −x. Since p is prime, there exist integers x, y such that 1 = px + ky and so 1 − ky = px which says 1 ∼ ky and so 1 = ky. Now you finish the argument. What is the multiplicative identity in this collection of equivalence classes? The only substantive issue is why Z p is a field. Let x ∈ Z p where x ≠ 0. Thus x is not a multiple of p. Then from the above problem, x and p are relatively prime. Hence from another of the above problems, there exist integers a, b such that 1 = ap + bx Then 1 − bx = ap = 0 and it follows that bx = 1 −1 so b = x . Abstract Vector Spaces 177 Solution: Let M = u = u 1 , u 2 , u 3 , u 4 ∈ R 4 : |u 1 | ≤ 4. Is M a subspace? Explain. No. 1, 0, 0, 0 ∈ M but 101, 0, 0, 0 ∉ M. 178 Solution: Let M = u = u 1 , u 2 , u 3 , u 4 ∈ R 4 : sinu 1 = 1. Is M a subspace? Explain. Absolutely not. Let u = π2 , 0, 0, 0. Then 2u ∉ M although u ∈ M. 179 Solution: If you have 5 vectors in F 5 and the vectors are linearly independent, can it always be concluded they span F 5 ? Explain. If not, you could add in a vector not in their span and obtain 6 vectors which are linearly independent. This cannot occur thanks to the exchange theorem. 180 Solution: If you have 6 vectors in F 5 , is it possible they are linearly independent? Explain. No because there exists a spanning set of 5 vectors. Note that it doesn’t matter what F is here. 181 Solution: Show in any vector space, 0 is unique. Say 0 ′ is another one. Then 0 = 0 + 0 ′ = 0 ′ . This happens from the definition of what it means for a vector to be an additive identity. 182 Solution: ↑In any vector space, show that if x + y = 0, then y = −x. Say x + y = 0. Then −x = −x +x + y = −x + x + y = 0 + y = y. If it acts like the additive inverse, it is the additive inverse. 183 Solution: ↑Show that in any vector space, 0x = 0. That is, the scalar 0 times the vector x gives the vector 0. 0x = 0 + 0x = 0x + 0x. Now add −0x to both sides to conclude that 0x = 0. 184 Solution: ↑Show that in any vector space, −1x = −x. Lets show −1x acts like the additive inverse. x +−1x =1 + −1x = 0x = 0 from one of the above problems. Hence −1x = − x. 185 Solution: Let X be a vector space and suppose x 1 , ⋯, x k is a set of vectors from X. Show that 0 is in spanx 1 , ⋯, x k . Pick any vector, x i . Then 0x i is in the span, but this was shown to be 0 above. 186 Solution: Let X consist of the real valued functions which are defined on an interval a, b. For f, g ∈ X, f + g is the name of the function which satisfies f + gx = fx + gx and for α a real number, αfx ≡ αfx. Show this is a vector space with field of scalars equal to R. Also explain why it cannot possibly be finite dimensional. It doesn’t matter where the functions have their values provided it is some real vector space. The axioms of a vector space are all routine because they hold for a vector space. The only thing maybe not completely obvious is the assertions about the things which are supposed to exist. 0 would be the zero function which sends everything to 0. This is an additive identity. Now if f is a function, −fx ≡ −fx. Then f + −fx ≡ fx + −fx ≡ fx + −fx = 0 Hence f + −f = 0. For each x ∈ a, b, let f x x = 1 and f x y = 0 if y ≠ x. Then these vectors are obviously linearly independent. 187 Solution: Let S be a nonempty set and let V denote the set of all functions which are defined on S and have values in W a vector space having field of scalars F. Also define vector addition according to the usual rule, f + gs ≡ fs + gs and scalar multiplication by αfs ≡ αfs. Show that V is a vector space with field of scalars F. This is no different than the above. You simply understand that the vector space has the same field of scalars as the W. 188 Solution: Verify that any field F is a vector space with field of scalars F. However, show that R is a vector space with field of scalars Q. A field also has multiplication. However, you can consider the elements of the field as vectors and then it satisfies all the vector space axioms. When you multiply a number (vector) in R by a scalar in Q you get something in R. All the axioms for a vector space are now obvious. For example, if α ∈ Q and x, y ∈ R, αx + y = αx + αy from the distributive law on R. 189 Solution: Let F be a field and consider functions defined on 1, 2, ⋯, n having values in F. Explain how, if V is the set of all such functions, V can be considered as F n . Simply let fi be the i th component of a vector x ∈ F n . Thus a typical thing in F n is f1, ⋯, fn. 190 Solution: Let V be the set of all functions defined on N ≡1, 2, ⋯ having values in a field F such that vector addition and scalar multiplication are defined by f + gs ≡ fs + gs and αfs ≡ αfs respectively, for f, g ∈ V and α ∈ F. Explain how this is a vector space and show that for e i given by e i k ≡ 1 if i = k 0 if i ≠ k , the vectors e k ∞k=1 are linearly independent. n Say for some n, ∑ k=1 c k e k = 0, the zero function. Then pick i, n 0= ∑ c k e k i = c i e i i = c i k=1 Since i was arbitrary, this shows these vectors are linearly independent. 191 Solution: Suppose you have smooth functions y 1 , y 2 , ⋯, y n (all derivatives exist) defined on an interval a, b. Then the Wronskian of these functions is the determinant Wy 1 , ⋯, y n x = det y 1 x ⋯ y n x y ′1 x ⋯ y ′n x ⋮ ⋮ y n−1 x 1 ⋯ y n−1 x n Show that if Wy 1 , ⋯, y n x ≠ 0 for some x, then the functions are linearly independent. Say n ∑ ckyk = 0 k=1 Then taking derivatives you have n ∑ c k y jk = 0, j = 0, 1, 2⋯, n − 1 k=1 This must hold when each equation is evaluated at x where you can pick the x at which the above determinant is nonzero. Therefore, this is a system of n equations in n variables, the c i and the coefficient matrix is invertible. Therefore, each c i = 0. 192 Solution: Give an example of two functions, y 1 , y 2 defined on −1, 1 such that Wy 1 , y 2 x = 0 for all x ∈ −1, 1 and yet y 1 , y 2 is linearly independent. Let y 1 x = x 2 and let y 2 x = x|x|. Suppose c 1 y 1 + c 2 y 2 = 0 Then you could consider evaluating at 1 and get c1 + c2 = 0 and then at −1 and get c1 − c2 = 0 then it follows that c 1 = c 2 = 0. Thus, these are linearly independent. However, if x > 0, W= x2 x2 2x 2x =0 and if x < 0 W= x 2 −x 2 2x −2x Also, W = 0 if x = 0 because at this point, W = 0 0 0 0 =0 = 0. 193 Solution: Let the vectors be polynomials of degree no more than 3. Show that with the usual definitions of scalar multiplication and addition wherein, for px a polynomial, αpx = αpx and for p, q polynomials p + qx ≡ px + qx, this is a vector space. This is just a subspace of the vector space of functions because it is closed with respect to vector addition and scalar multiplication. Hence this is a vector space. 194 Solution: In the previous problem show that a basis for the vector space is 1, x, x 2 , x 3 . Using one of the above problems, this is really easy if you take the Wronskian of these functions. 1 x x2 det x3 0 1 2x 3x 2 0 0 2 6x 0 0 0 6 ≠ 0. 195 Solution: In the previous problem show that a basis for the vector space is 1, x, x 2 , x 3 . a. x + 1, x 3 + x 2 + 2x, x 2 + x, x 3 + x 2 + x i. Lets look at the Wronskian of these functions. If it is nonzero somewhere, then these are linearly independent and are therefore a basis because there are four of them. 1 + x x 3 + x 2 + 2x x 2 + x det 1 x3 + x2 + x 3x 2 + 2x + 2 2x + 1 3x 2 + 2x + 1 0 6x + 2 2 6x + 2 0 6 0 6 Lets plug in x = 0 and see what happens. You only need to have it be nonzero at one point. 1 0 0 0 det 1 2 1 1 0 2 2 2 = 12 ≠ 0 0 6 0 6 so these are linearly independent. b. x + 1, x 2 + x, 2x 3 + x 2 , 2x 3 − x 2 − 3x + 1 i. Suppose c 1 x 3 + 1 + c 2 x 2 + x + c 3 2x 3 + x 2 + c 4 2x 3 − x 2 − 3x + 1 = 0 Then combine the terms according to power of x. c 1 + 2c 3 + 2c 4 x 3 + c 2 + c 3 − c 4 x 2 + c 2 − 3c 4 x + c 1 + c 4 = 0 Is there a non zero solution to the system 3 c 1 + 2c 3 + 2c 4 = 0 c2 + c3 − c4 = 0 c 2 − 3c 4 = 0 , Solution is: c 1 = 0, c 2 = 0, c 3 = 0, c 4 = 0 Therefore, these are c1 + c4 = 0 linearly independent. 196 Solution: Let V be the space of polynomials of degree three or less. Consider a collection of polynomials in this vector space which are of the form a i x 3 + b i x 2 + c i x + d i , i = 1, 2, 3, 4 Show that this collection of polynomials is linearly independent on an interval a, b if and only if a1 b1 c1 d1 a2 b2 c2 d2 a3 b3 c3 d3 a4 b4 c4 d4 is an invertible matrix. Let p i x denote the i th of these polynomials. Suppose ∑ i C i p i x = 0. Then collecting terms according to the exponent of x, you need to have C1a1 + C2a2 + C3a3 + C4a4 = 0 C1b1 + C2b2 + C3b3 + C4b4 = 0 C1c1 + C2c2 + C3c3 + C4c4 = 0 C1d1 + C2d2 + C3d3 + C4d4 = 0 The matrix of coefficients is just the transpose of the above matrix. There exists a non trivial solution if and only if the determinant of this matrix equals 0. 197 Solution: Let the field of scalars be Q, the rational numbers and let the vectors be of the form a + b 2 where a, b are rational numbers. Show that this collection of vectors is a vector space with field of scalars Q and give a basis for this vector space. This is obvious because when you add two of these you get one and when you multiply one of these by a scalar, you get another one. A basis is 1, 2 . By definition, the span of these gives the collection of vectors. Are they independent? Say a + b 2 = 0 where a, b are rational numbers. If a ≠ 0, then b 2 = −a which can’t happen since a is rational. If b ≠ 0, then −a = b 2 which again can’t happen because on the left is a rational number and on the right is an irrational. Hence both a, b = 0 and so this is a basis. 198 Solution: Suppose V is a finite dimensional vector space. Based on the exchange theorem above, it was shown that any two bases have the same number of vectors in them. Give a different proof of this fact using the earlier material in the book. Hint: Suppose x 1 , ⋯, x n and y 1 , ⋯, y m are two bases with m < n. Then define φ : F n → V, ψ : F m → V by n φa ≡ ∑ akxk, m ψb ≡ k=1 ∑ b jyj j=1 −1 Consider the linear transformation, ψ ∘ φ. Argue it is a one to one and onto mapping from F n to F m . Now consider a matrix of this linear transformation and its row reduced echelon form. Since both of these are bases, both ψ and φ are one to one and onto. Therefore, the given linear transformation is also one to one and onto. It has a matrix A which has more columns than rows such that multiplication by A has the same effect as doing ψ −1 ∘ φ. However, the augmented matrix A|0 has free variables because there are more columns than rows for A. Hence A cannot be one to one. Therefore, ψ −1 ∘ φ also must fail to be one to one. This is a contradiction. 199 Solution: This and the following problems will present most of a differential equations course. To begin with, consider the scalar initial value problem y ′ = ay, yt 0 = y 0 When a is real, show the unique solution to this problem is y = y 0 e at−t 0 . Next suppose y ′ = a + iby, yt 0 = y 0 where yt = ut + ivt. Show there exists a unique solution and it is yt = y 0 e at−t 0 cos bt − t 0 + i sin bt − t 0 ≡ e a+ibt−t 0 y 0 . Next show that for a real or complex there exists a unique solution to the initial value problem y ′ = ay + f, yt 0 = y 0 and it is given by yt = e at−t 0 y 0 + e at ∫ e −as fsds. t t0 Consider the first part. First you can verify that yt = y 0 e at−t 0 works using elementary calculus. Why is it the only one which does so? Suppose y 1 t is a solution. Then yt − y 1 t = zt solves the initial value problem z ′ t = azt, zt 0 = 0. Thus z ′ − az = 0. Multiply by e −at on both sides. By the chain rule, e −at z ′ − az = d e −at zt = 0, dt −at and so there is a constant C such that e zt = C. Since zt 0 = 0, this constant is 0, and so zt = 0. Next consider the second part involving the complex stuff. You can verify that y 0 e at−t 0 cos bt − t 0 + i sin bt − t 0 = yt does indeed solve the initial value problem from using elementary calculus. Now suppose you have another solution y 1 t. Then let zt = yt − y 1 t. It solves z ′ = a + ibz, zt 0 = 0 Now zt = ut + ivt. It is also clear that z t ≡ ut − ivt solves the equation z ′ = a − ib z , z t 0 = 0 Thus from the product rule which you can easily verify from the usual proof of the product rule to be also true for complex valued functions, d |zt|2 = d z z = z ′ z + z z ′ = a + ibz z + za − ib z dt dt = 2a|z|2 , |z|t 0 = 0 Therefore, from the first part, |z| = 0 and so y = y 1 . ′ Note that this implies that e ia+ibt = a + ibe ia+ibt where e ia+ibt is given above. Now consider the last part. Sove y ′ = ay + f, yt 0 = y 0 . y ′ − ay = f, yt 0 = y 0 Multiply both sides by e −at−t 0 d e −at−t 0 y = e −at−t 0 ft dt Now integrate from t 0 to t. Then e −at−t 0 yt − y 0 = ∫t t e −as−t 0 fsds 0 Hence yt = e at−t 0 y 0 + e at−t 0 ∫ e −as−t 0 fsds t t0 = e at−t 0 y 0 + e at ∫ e −as fsds t t0 200 Solution: Now consider A an n × n matrix. By Schur’s theorem there exists unitary Q such that Q −1 AQ = T where T is upper triangular. Now consider the first order initial value problem x ′ = Ax, xt 0 = x 0 . Show there exists a unique solution to this first order system. Hint: Let y = Q −1 x and so the system becomes y ′ = Ty, yt 0 = Q −1 x 0 Now letting y =y 1 , ⋯, y n T , the bottom equation becomes y ′n = t nn y n , y n t 0 = Q −1 x 0 n . * Then use the solution you get in this to get the solution to the initial value problem which occurs one level up, namely y ′n−1 = t n−1n−1 y n−1 + t n−1n y n , y n−1 t 0 = Q −1 x 0 n−1 Continue doing this to obtain a unique solution to *. There isn’t much to do. Just see that you understand it. In the above problem, equations of the above form are shown to have a unique solution and there is even a formula given. Thus there is a unique solution to y ′ = Ty, yt 0 = Q −1 x 0 Then let y = Q −1 x and plug this in. Thus Q −1 x ′ = TQ −1 x, Q −1 xt 0 = Q −1 x 0 Then x ′ = QTQ −1 x = Ax, xt 0 = x 0 . The solution is unique because of uniqueness of the solution for y. You can go backwards, starting with x and defining y ≡ Q −1 x and then there is only one thing y can be and so x is also unique. 201 Solution: Now suppose Φt is an n × n matrix of the form Φt = x 1 t ⋯ x n t where x ′k t = Ax k t. Explain why Φ ′ t = AΦt if and only if Φt is given in the form of ♠. Also explain why if c ∈ F n , yt ≡ Φtc solves the equation y ′ t = Ayt. Suppose Φ ′ t = AΦt where Φt is as described. This happens if and only if Φ ′ t ≡ = =A x ′1 t ⋯ x ′n t x 1 t ⋯ x n t Ax 1 t ⋯ Ax n t from the way we multiply matrices. Which happens if and only if x ′k = Ax k Say Φ ′ t = AΦt. Then consider yt = ∑ c k x k t. k ♠ Then y ′ t = ∑ c k x ′k t = ∑ c k Ax k t = A ∑ c k x k t k k = AΦtc = Ayt k 202 Solution: In the above problem, consider the question whether all solutions to x ′ = Ax ♣ n are obtained in the form Φtc for some choice of c ∈ F . In other words, is the general solution to this equation Φtc for c ∈ F n ? Prove the following theorem using linear algebra. Theorem: Suppose Φt is an n × n matrix which satisfies Φ ′ t = AΦt. Then the general solution to ♣ is Φtc if and only if Φt −1 exists for some t. Furthermore, if Φ ′ t = AΦt, then either Φt −1 exists for all t or Φt −1 never exists for any t. (detΦt is called the Wronskian and this theorem is sometimes called the Wronskian alternative.) Suppose first that Φtc is the general solution. Why does Φt −1 necessarily exist? If for some t 0 , Φt 0 fails to have an inverse, then there exists c ∉ Φt 0 R n . However, there exists a unique a solution to the system x ′ = Ax, xt 0 = c from one of the above problems. It follows that xt ≠ Φtv for any choice of v because for any v, c = xt 0 ≠ Φt 0 v Thus Φtc for arbitrary choice of c fails to deliver the general solution as assumed. Next suppose Φt 0 −1 exists for some t 0 and let xt be any solution to the system of differential equations. Then let yt = ΦtΦt 0 −1 xt 0 . It follows y ′ = Ay and x ′ = Ax and also that xt 0 = yt 0 so by the uniqueness part of the above problem, x = y. Hence Φtc for arbitrary c is the general solution. Consider det Φt if it equals zero for any t 0 then Φtc does not deliver the general solutions. If det Φt is nonzero for any t then Φtc delivers the general solution. Either Φtc does deliver the general solution or it does not. Therefore, you cannot have two different values of t, t 1 , t 2 such that det Φt 1 = 0 but det Φt 2 ≠ 0. In other words det Φt either vanishes for all t or it vanishes for no t. Thus the inverse of Φt either exists for all t or for no t. 203 Solution: Let Φ ′ t = AΦt. Then Φt is called a fundamental matrix if Φt −1 exists for all t. Show there exists a unique solution to the equation x ′ = Ax + f, xt 0 = x 0 * and it is given by the formula xt = ΦtΦt 0 −1 x 0 + Φt ∫ Φs −1 fsds t t0 Now these few problems have done virtually everything of significance in an entire undergraduate differential equations course, illustrating the superiority of linear algebra. The above formula is called the variation of constants formula. It is easy to see that x given by the formula does solve the initial value problem. This is just calculus. x ′ t = Φ ′ tΦt 0 −1 x 0 + Φ ′ t ∫ Φs −1 fsds + ΦtΦt −1 ft t t0 = A ΦtΦt 0 −1 x 0 + Φt ∫ Φs −1 fsds t + ft t0 = Axt + ft As for the initial condition, xt 0 = Φt 0 Φt 0 −1 x 0 + Φt 0 ∫ Φs −1 fsds = x 0 t0 t0 It is also easy to see that this must be the solution. If you have two solutions, then let ut = x 1 t − x 2 t and observe that u ′ t = Aut, ut 0 = 0. From the above problem, there is at most one solution to this initial value problem and it is u = 0. 204 Solution: Show there exists a special Φ such that Φ ′ t = AΦt, Φ0 = I, and suppose Φt −1 exists for all t. Show using uniqueness that Φ−t = Φt −1 and that for all t, s ∈ R Φt + s = ΦtΦs Explain why with this special Φ, the solution to the differential equation x ′ = Ax + f, xt 0 = x 0 can be written as xt = Φt − t 0 x 0 + ∫ Φt − sfsds. t t0 Hint: Let Φt be such that the j th column is x j t where x ′j = Ax j , x j 0 = e j . Use uniqueness as required. You simply let Φt = x 1 t ⋯ x n t where x ′k t = Ax k t, x k 0 = e k Then det Φ0 = 1 and so det Φt ≠ 0 for any t. Furthermore, there is only one solution to Φ ′ t = AΦt along with the initial condition Φ0 = I and it is the one just described. This follows from the uniqueness of the intial value problem x ′ = Ax, x0 = x 0 discussed earlier. Now it is interesting to note that A and Φt commute. To see this, consider yt ≡ AΦtc − ΦtAc When t = 0 y0 = Ac − Ac = 0. What about its derivative? y ′ t = AΦ ′ tc − Φ ′ tAc = A 2 Φtc − AΦtAc = AAΦtc − ΦtAc = Ayt By uniqueness, it follows that yt = 0. Thus these commute as claimed, since c is arbitrary. Now from the product rule (the usual product rule holds when the functions are matrices by the same proof given in calculus. ) ΦtΦ−t ′ = Φ ′ tΦ−t − ΦtΦ ′ −t = AΦtΦ−t − ΦtAΦ−t = ΦtAΦ−t − ΦtAΦ−t = 0 Hence ΦtΦ−t must be a constant matrix since the derivative of each component of this product equals 0. However, this constant can only equal I because when t = 0 the product is I. Therefore, Φt −1 = Φ−t. Next consider the claim that Φt + s = ΦtΦs. Fix s and let t vary. Φ ′ t + s − Φ ′ tΦs = AΦt + s − AΦtΦs = AΦt + s − ΦtΦs Now Φ0 + s − Φ0Φs = Φs − Φs = 0 and so, by uniqueness, t → Φt + s − ΦtΦs equals 0. The rest follows from the variation of constants formula derived earlier. xt = ΦtΦ0 −1 x 0 + Φt ∫ Φs −1 fsds t 0 = Φtx 0 + Φt ∫ Φ−sfs t 0 = Φtx 0 + ∫ ΦtΦ−sfs t 0 t = Φtx 0 + ∫ Φt − sfs 0 The reason you can take Φt inside the integral is that this is constant with respect to the variable of integration. The j th entry of the product is Φt ∫ Φ−sfs t 0 = j ∑ Φt jk ∫ 0 ∑ Φ−s ki f i sds t k = i ∑ ∫ 0 Φt jk Φ−s ki f i ds t k,i = ∑ ∫ 0 ΦtΦ−s ji f i ds t i t = ∫ 0 ∑ΦtΦ−s ji f i ds = ∫ 0 ΦtΦ−sfsds i t j You could also simply verify directly that this formula works much as was done earlier. 205 Solution: ∗ Using the Lindemann Weierstrass theorem show that if σ is an algebraic number sin σ, cos σ, ln σ, and e are all transcendental. Hint: Observe, that ee −1 + −1e 0 = 0, 1e lnσ + −1σe 0 = 0, 1 e iσ − 1 e −iσ + −1 sinσe 0 = 0. 2i 2i The hint gives it away. Consider the claim about ln σ. The equation shown does hold from the definition of ln σ. However, if ln σ were algebraic, then e lnσ , e 0 would be linearly dependent with field of scalars equal to the algebraic numbers, contrary to the Lindemann Weierstrass theorem. The other instances are similar. In the case of cos σ, you could use the identity 1 e iσ + 1 e −iσ − e 0 cos σ = 0 2 2 iσ −iσ 0 contradicting independence of e , e , e . Inner Product Spaces 206 Solution: Let V be the continuous complex valued functions defined on a finite closed interval I. Define an inner product as follows. 〈f, g〉 ≡ ∫I fxgxpxdx where px some function which is strictly positive on the closed interval I. It is understood in writing this that ∫I fx + igxdx ≡ ∫I fxdx + i ∫I gxdx Then with this convention, the usual calculus theorems hold about evaluating integrals using the fundamental theorem of calculus and so forth. You simply apply these theorems to the real and imaginary parts of a complex valued function. Show V is an inner product space with the given inner product. All of the axioms of the inner product are obvious except one, the one which says that if 〈f, f〉 = 0 then f = 0. This one depends on continuity of the functions. Suppose then that it is not true. In other words, 〈f, f〉 = 0 and yet f ≠ 0. Then for some x ∈ I, fx ≠ 0. By continuity, there exists δ > 0 such that if y ∈ I ∩ x − δ, x + δ ≡ I δ , then |fy − fx| < |fx|/2 It follows that for y ∈ I δ , |fy| > |fx| − |fx/2| = |fx|/2. Hence 〈f, f〉 ≥ ∫I |fy|2 pxdy ≥ |fx|2 /2 length of I δ minp > 0, δ a contradiction. Note that min p > 0 because p is a continuous function defined on a closed and bounded interval and so it achieves its minimum by the extreme value theorem of calculus. 207 Solution: Let V be the polynomials of degree at most n which are defined on a closed interval I and let x 0 , x 1 , ⋯, x n be n + 1 distinct points in I. Then define n 〈f, g〉 ≡ ∑ fx k gx k k=0 Show that V is an inner product space with respect to the given inner product. All of the axioms of the inner product are obvious for this one also, except for the one which says that if 〈f, f〉 = 0, then f = 0. Suppose than that 〈f, f〉 = 0. Then f equals 0 at n + 1 points of the interval and yet f is a polynomial of degree at most n. Therefore, this would be a contradiction unless f is identically equal to 0 for all x. This is because a polynomial of degree n has at most n zeros. 208 Solution: Let V be any complex vector space and let v 1 , ⋯, v n be a basis. Decree that 〈v i , v j 〉 = δ ij . Then define n n j=1 k=1 n ∑ cjv j, ∑ d k v k ≡ ∑ c jd k 〈v j, v k 〉 = ∑ c k d k j,k k=1 This makes the complex vector space into an inner product space. n A generic vector of V will be denoted by w = ∑ i=1 wi v i . Now 〈w, u〉 ≡ n n k=1 k=1 ∑ wk u k = ∑ wk u k ≡ 〈u, w〉 Letting a, b be scalars, 〈aw + bz, u〉 = ∑awk + bz k u k = a ∑ wk u k + b ∑ z k u k k k k ≡ a〈w, u〉 + b〈z, u〉 〈w, w〉 ≡ n n k=1 k=1 ∑ wk wk = ∑|wk |2 ≥ 0 The only way this can equal zero is to have each wk = 0. Since v k is a basis, this happens if and only if w = 0. Thus the inner product by decree is an inner product. 209 Solution: Let V consist of sequences a = a k ∞k=1 , a k ∈ C, with the property that ∞ ∑|a k |2 < ∞ k=1 and the inner product is then defined as ∞ 〈a, b〉 ≡ ∑ akbk k=1 The last example is obviously an inner product as noted earlier except for needing to verify that the inner product makes sense. However, from the Cauchy Schwarz inequality, n ∑ akbk 1/2 n ≤ k=1 ∑|a k | 2 k=1 ∑|a k | k=1 ∑|b k | 2 k=1 1/2 ∞ ≤ 1/2 n 2 1/2 ∞ ∑|b k | 2 <∞ k=1 210 Solution: In each of the examples of the above problems, write the Cauchy Schwarz inequality. In the first example, the Cauchy Schwarz inequality says ∫I fxgxpxdx 1/2 ∫I |fx|2 pxdx ≤ ∫I |gx|2 pxdx 1/2 In the next example, the Cauchy Schwarz inequality says n ∑ fx k gx k 1/2 n ≤ k=0 ∑|fx k | 1/2 n ∑|gx k | 2 k=0 2 k=0 In the third example, the Cauchy Schwarz inequality says n ∑ u k wk 1/2 n ≤ k=1 ∑|u k | 2 k=1 1/2 n ∑|wk | 2 k=1 where u = ∑ k u k v k and w = ∑ k wk v k . In the last example, ∞ ∑ akbk k=1 1/2 ∞ ≤ ∑|a k | 2 k=1 1/2 ∞ ∑|b k | 2 k=1 211 Solution: Verify that in an inner product space, |αz| = |α||z| and that |z| ≥ 0 and equals zero if and only if z = 0. |αz|2 ≡ 〈αz, αz〉 = α〈z, αz〉 = α〈αz, z〉 = α α 〈z, z〉 ≡ |α|2 |z|2 and so |α||z| = |αz| It is clear |z| ≥ 0 because |z| = 〈z, z〉 . Suppose |z| = 0. Then 〈z, z〉 = 0 and so, by the axioms of the inner product, z = 0. 212 Solution: Consider the Cauchy Schwarz inequality. Show that it still holds under the assumptions 〈u, v〉 = 〈v, u〉, 〈au + bv, z〉 = a〈u, z〉 + b〈v, z〉, and 〈u, u〉 ≥ 0. Thus it is not necessary to say that 〈u, u〉 = 0 only if u = 0. It is enough to simply state that 〈u, u〉 ≥ 0. This was all that was used in the proof. Look at the proof carefully and you will see this is the case. 213 Solution: Consider the integers modulo a prime, Z p . This is a field of scalars. Now let the vector space be Z p n where n ≥ p. Define now 1/2 n 〈z, w〉 ≡ ∑ z i wi i=1 Does this satisfy the axioms of an inner product? Does the Cauchy Schwarz inequality hold for this 〈 〉? Does the Cauchy Schwarz inequality even make any sense? It might be the case that 〈z, z〉 = 0 and yet z ≠ 0. Just let z =z 1 , ⋯, z n where exactly p of the z i equal 1 but the remaining are equal to 0. Then 〈z, z〉 would reduce to 0 in the integers mod p. Another problem is the failure to have an order on Z p . Consider first Z 2 . Is 1 positive or negative? If it is positive, then 1 + 1 would need to be positive. But 1 + 1 = 0 in this case. If 1 is negative, then −1 is positive, but −1 is equal to 1. Thus 1 would be both positive and negative. You can consider the general case where p > 2 also. Simply take a ≠ 1. If a is positive, then consider a, a 2 , a 3 ⋯. These would all have to be positive. However, eventually a repeat will take place. Thus a n = a m m < n, and so a m a k − 1 = 0 where k = n − m. Since a m ≠ 0, it follows that a k = 1 for a suitable k. It follows that the sequence of powers of a must include each of 1, 2, ⋯, p − 1 and all these would therefore, be positive. However, 1 + p − 1 = 0 contradicting the assertion that Z p can be ordered. So what would you mean by saying 〈z, z〉 ≥ 0? The Cauchy Schwarz inequality would not even apply. 214 Solution: If you only know that 〈u, u〉 ≥ 0 along with the other axioms of the inner product and if you define |z| the same way, which properties of a norm are lost? You lose the one which says that if |z| = 0 then z = 0. However, all the others are retained. 215 Solution: In an inner product space, an open ball is the set Bx, r ≡ y : |y − x| < r. If z ∈ Bx, r, show there exists δ > 0 such that Bz, δ ⊆ Bx, r. In words, this says that an open ball is open. Hint: This depends on the triangle inequality. Let δ = r − |z − x|. Then if y ∈ Bz, δ, |y − x| ≤ |y − z| + |z − x| < δ + |z − x| = r − |z − x| + |z − x| = r and so Bz, δ ⊆ Bx, r. 216 Solution: Let V be the real inner product space consisting of continuous functions defined on −1, 1 with the inner product given by ∫ −1 fxgxdx 1 Show that 1, x, x 2 are linearly independent and find an orthonormal basis for the span of these vectors. Lets first find the Grammian. 2 G= 1 1 ∫ −1 xdx 1 ∫ −1 xdx ∫ −1 x 2 dx 1 ∫ −1 x 2 1 ∫ −1 x 3 1 ∫ −1 x 2 1 ∫ −1 x 3 = 1 ∫ −1 x 4 dx Then the inverse is G −1 = 9 8 0 − 158 0 3 2 0 − 158 0 45 8 2 0 2 3 0 2 3 0 2 3 0 2 5 a b c Now let R = 0 d h . Solve 0 0 f T a b c a b c 0 d h 0 d h 0 0 f 0 0 f 9 8 0 − 158 0 3 2 0 − 158 0 45 8 = Thus 9 8 0 − 158 0 3 2 0 − 158 0 45 8 a 2 + b 2 + c 2 bd + ch cf bd + ch = d 2 + h 2 hf cf f2 hf Hence a solution is f = 3 2 5 , h = 0, c = − 1 2 5 , d = 1 2 3 , b = 0, a = 1 2 4 4 2 2 and so the orthonormal basis is 1 2 1 x x2 2 0 1 2 0 0 1 2 2 1 2 2 3x − 14 2 5 2 3 0 3 4 2 5 x2 − 0 3 4 2 5 1 4 2 5 217 Solution: A regular Sturm Liouville problem involves the differential equation for an unknown function of x which is denoted here by y, ′ pxy ′ + λqx + rxy = 0, x ∈ a, b and it is assumed that pt, qt > 0 for any t along with boundary conditions, C 1 ya + C 2 y ′ a = 0 C 3 yb + C 4 y ′ b = 0 where C 21 + C 22 > 0, and C 23 + C 24 > 0. There is an immense theory connected to these important problems. The constant λ is called an eigenvalue. Show that if y is a solution to the above problem corresponding to λ = λ 1 and if z is a solution corresponding to λ = λ 2 ≠ λ 1 , then ∫ a qxyxzxdx = 0. b Let y go with λ and z go with μ. zpxy ′ ′ + λqx + rxyz = 0 ′ ypxz ′ + μqx + rxzy = 0 Subtract. zpxy ′ ′ − ypxz ′ ′ + λ − μqxyz = 0 Now integrate from a to b. First note that zpxy ′ ′ − ypxz ′ ′ = d pxy ′ z − pxz ′ y dx and so what you get is pby ′ bzb − pbz ′ byb − pay ′ aza − paz ′ aya + λ − μ ∫ qxyxzxdx = 0 b a Look at the stuff on the top line. From the assumptions on the boundary conditions, C 1 ya + C 2 y ′ a = 0 C 1 za + C 2 z ′ a = 0 and so yaz ′ a − y ′ aza = 0 Similarly, ybz ′ b − y ′ bzb = 0 Hence, that stuff on the top line equals zero and so the orthogonality condition holds. 218 Solution: Using the above problem or standard techniques of calculus, show that 2 sinnx π ∞ n=1 are orthonormal with respect to the inner product 〈f, g〉 = π ∫ 0 fxgxdx Hint: If you want to use the above problem, show that sinnx is a solution to the boundary value problem y ′′ + n 2 y = 0, y0 = yπ = 0 It is obvious that sinnx solves this boundary value problem. Those boundary conditions in the above problem are implied by these. In fact, 1 sin 0 + 0 cos0 = 0 1 sin π + 0 cos π = 0 Then it follows that π ∫ 0 sinnx sinmxdx = 0 unless n = m. π Now also ∫ sin 2 nxdx = 0 219 Solution: 1 2 π and so 2 π sinnx are orthonormal. Find S 5 fx where fx = x on −π, π. Then graph both S 5 fx and fx if you have access to a system which will do a good job of it. 5 Recall that this is the partial sum of the Fourier series. ∑ k=−5 b k e ikx where bk = 1 2π −1 k i k π ∫ −π xe −ikx dx = Thus 5 S 5 fx = ∑ k=−5 −1 k i e ikx = k 5 = ∑ −1k k+1 ∑ −1 k k icoskx + i sinkx k=−5 5 sinkx = k=−5 ∑ 2−1 k k+1 sinkx k=1 y -3 5 -2 2 -1 1 2 3 x -2 220 Solution: Find S 5 fx where fx = |x| on −π, π. Then graph both S 5 fx and fx if you have access to a system which will do a good job of it. a. π b k = 1 ∫ |x|e −ikx dx = − 22 2π −π k π if k is odd and 0 if k is even. However, b 0 = 12 π Thus the sum desired is −1 S 5 fx = 5 ∑ − k12 π 1 + −1 k+1 coskx + ∑ − k12 π 1 + −1 k+1 coskx k=−5 k=1 + 1π 2 since the sin terms cancel due to the fact that sin is odd. The above equals 2 1π−∑ 4 cos2k + 1x 2 + 1 2 π 2k k=0 y3 2 1 -3 -2 -1 0 1 2 3 x 221 Solution: Find S 5 fx where fx = x 2 on −π, π. Then graph both S 5 fx and fx if you have access to a system which will do a good job of it. bk = 1 2π b0 = 1 2π π ∫ −π x 2 e −kix dx = π ∫ −π x 2 dx = 2 −1 k k2 1 π2 3 Then the series is 5 S 5 fx = ∑ k=1 −1 2 −1 k e ikx + ∑ 2 −1 k e ikx + π 2 3 k2 k2 k=−5 5 = ∑ k42 −1 k coskx + k=1 ∑ 5 4 k=1 k 2 −1 coskx + k π2 3 π2 3 y 10 8 6 4 2 -3 -2 -1 0 1 2 3 x 222 Solution: Let V be the set of real polynomials defined on 0, 1 which have degree at most 2. Make this into a real inner product space by defining 〈f, g〉 ≡ f0g0 + f1/2g1/2 + f1g1 Find an orthonormal basis and explain why this is an inner product. It was described more generally why this was an inner product in an earlier problem. To find an orthonormal basis, consider a basis 1, x, x 2 . Then the Grammian is 3 0 + 1/2 + 1 0 + 1/4 + 1 0 + 1/2 + 1 0 + 1/4 + 1 0 + 1/21/4 + 1 0 + 1/4 + 1 0 + 1/21/4 + 1 3 2 5 4 9 8 3 = 3 2 5 4 0 + 1/4 2 + 1 5 4 9 8 17 16 Now G −1 equals 1 −3 2 −3 26 −24 2 −24 24 a b c Let R = 0 d h 0 0 f T a b c a b c 0 d h 0 d h 0 0 f 0 0 f a 2 + b 2 + c 2 bd + ch cf = bd + ch d 2 + h 2 hf cf hf f2 So you need to solve the following . a 2 + b 2 + c 2 bd + ch cf bd + ch d +h cf hf 2 2 hf f2 f = 2 6 , h = −2 6 , c = 1 6 , d = 6 So the orthonormal basis is 1 3 3 2x− 1 2 223 Solution: Consider R n with the following definition. −3 2 −3 26 −24 2 −24 24 2 ,b = − 1 2 ,a = 1 3 2 3 3 − 12 2 1 3 1 x x2 = 1 1 6 6 0 2 −2 6 0 0 2 6 2 2 6 x2 − 2 6 x + 1 6 6 n 〈x, y〉 ≡ ∑ xiyii i=1 Does this define an inner product? If so, explain why and state the Cauchy Schwarz inequality in terms of sums. It obviously defines an inner product because it satisfies all the axioms of one. The Cauchy Schwarz inequality says n ∑ xiyii 1/2 n ∑ ≤ i=1 x 2i i i=1 1/2 n ∑ y 2i i i=1 This is a fairly surprising inequality it seems to me. 224 Solution: For f a piecewise continuous function, n S n fx = 1 2π π ∑ e ikx ∫ −π fye −iky dy . k=−n Show this can be written in the form S n fx = π ∫ −π fyD n x − ydy where n D n t = 1 2π ∑ e ikt k=−n This is called the Dirichlet kernel. Show that sinn + 1/2t D n t = 1 2π sint/2 For V the vector space of piecewise continuous functions, define S n : V → V by S n fx = π ∫ −π fyD n x − ydy. Show that S n is a linear transformation. (In fact, S n f is not just piecewise continuous but infinitely π differentiable. Why?) Explain why ∫ D n tdt = 1. Hint: To obtain the formula, do the −π following. n e it/2 D n t = 1 2π ∑ e ik+1/2t e i−t/2 D n t = 1 2π ∑ e ik−1/2t k=−n n k=−n Change the variable of summation in the bottom sum and then subtract and solve for D n t. n S n fx = 1 2π ∑ e ikx k=−n π ∫ −π fye −iky dy = 1 2π Let D n t be as described. Then the above equals π ∫ −π D n x − yfydy. π n ∫ −π fy ∑ e ikx−y dy. k=−n So now you want to find a formula for D n t. The hint is really good. n e it/2 D n t = 1 2π ∑ e ik+1/2t k=−n n e i−t/2 D n t = 1 2π n−1 ∑ e ik−1/2t = 1 2π k=−n ∑ e ik+1/2t k=−n+1 D n te it/2 − e −it/2 = 1 e in+1/2t − e −in+1/2t 2π D n t2i sint/2 = 1 2i sin n + 1 t 2 2π 1 sintn + 2 D n t = 1 2π sin 12 t You know that t → D n t is periodic of period 2π. Therefore, if fy = 1, S n fx = π π ∫ −π D n x − ydy = ∫ −π D n tdt However, it follows directly from computation that S n fx = 1. 225 Solution: Let V be an inner product space and let U be a finite dimensional subspace with an orthonormal basis u i ni=1 . If y ∈ V, show n |y|2 ≥ ∑|〈y, u k 〉|2 k=1 Now suppose that u k ∞k=1 is an orthonormal set of vectors of V. Explain why lim〈y, u k 〉 = 0. k→∞ When applied to functions, this is a special case of the Riemann Lebesgue lemma. n y − ∑〈y, u k 〉u k , w =0 k=1 for all w ∈ spanu i ni=1 . Therefore, n n k=1 k=1 |y| = y − ∑〈y, u k 〉u k + ∑〈y, u k 〉u k 2 2 Now if 〈u, v〉 = 0, then you can see right away from the definition that |u + v|2 = |u|2 + |v|2 n n Applying this to u = y − ∑ k=1 〈y, u k 〉u k , v = ∑ k=1 〈y, u k 〉u k , the above equals n = y − ∑〈y, u k 〉u k 2 k=1 n = y − ∑〈y, u k 〉u k k=1 n + ∑〈y, u k 〉u k 2 k=1 2 n + ∑|〈y, u k 〉|2 , k=1 the last step following because of similar reasoning to the above and the assumption that the u k ∞ are orthonormal. It follows the sum ∑ k=1 |〈y, u k 〉|2 converges and so lim k→∞ 〈y, u k 〉 = 0 because if a series converges, then the k th term must converge to 0. 226 Solution: Let f be any piecewise continuous function which is bounded on −π, π. Show, using the above problem, that π lim n→∞ π ∫ −π ft sinntdt = lim ∫ ft cosntdt = 0 n→∞ −π Let the inner product space consist of piecewise continuous bounded functions with the inner product defined by 〈f, g〉 ≡ π ∫ −π fxgxdx Then, from the above problem and the fact shown earlier that 1 2π e ikx form an orthonormal k∈Z set of vectors in this inner product space, it follows that lim 〈f, e inx 〉 = 0 n→∞ without loss of generality, assume that f has real values. Then the above limit reduces to having both the real and imaginary parts converge to 0. This implies the thing which was desired. Note also that if α ∈ −1, 1, then π lim n→∞ π ∫ −π ft sinn + αtdt = lim ∫ ftsinnt cos α + cosnt sin αdt = 0 n→∞ −π 227 Solution: ∗ Let f be a function which is defined on −π, π. The 2π periodic extension is given by the formula fx + 2π = fx. In the rest of this problem, f will refer to this 2π periodic extension. Assume that f is piecewise continuous, bounded, and also that the following limits exist fx + y − fx + fx − y − fx + , lim lim y y y→0+ y→0+ Here it is assumed that fx + ≡ lim fx + h, fx − ≡ lim fx − h h→0+ h→0+ both exist at every point. The above conditions rule out functions where the slope taken from either side becomes infinite. Justify the following assertions and eventually conclude that under these very reasonable conditions lim S fx = fx + + fx −/2 n→∞ n the mid point of the jump. In words, the Fourier series converges to the midpoint of the jump of the function. S n fx = S n fx − fx + + fx − 2 = π ∫ −π fx − yD n ydy π ∫ −π fx − y − fx + + fx − D n ydy 2 = π π ∫ 0 fx − yD n ydy + ∫ 0 fx + yD n ydy π − ∫ fx + + fx −D n ydy 0 ≤ π ∫ 0 fx − y − fx −D n ydy + π ∫ 0 fx + y − fx +D n ydy Now apply some trig. identities and use the result of the above problems to conclude that both of these terms must converge to 0. From the definition of D n , the top formula holds. Now observe that D n is an even function. Therefore, the formula equals S n fx = = π 0 π π ∫ 0 fx − yD n ydy + ∫ −π fx − yD n ydy ∫ 0 fx − yD n ydy + ∫ 0 fx + yD n ydy fx + y + fx − y 2D n ydy 2 π π Now note that ∫ 2D n y = 1 because ∫ D n ydy = 1 and D n is even. Therefore, = 0 π ∫0 −π fx + + fx − S n fx − 2 π fx + y − fx + + fx − y − fx − = ∫ 2D n ydy 2 0 From the formula for D n y given earlier, this is dominated by an expression of the form π fx + y − fx + + fx − y − fx − C ∫ sinn + 1/2ydy siny/2 0 for a suitable constant C. The above is equal to π fx + y − fx + + fx − y − fx − y C ∫ sinn + 1/2ydy y 0 siny/2 y and the expression siny/2 equals a bounded continuous function on 0, π except at 0 where it is undefined. This follows from elementary calculus. Therefore, changing the function at this single point does not change the integral and so we can consider this as a continuous bounded function defined on 0, π. Also, from the assumptions on f, fx + y − fx + + fx − y − fx − y→ y is equal to a piecewise continuous function on 0, π except at the point 0. Therefore, the above integral converges to 0 by the previous problem. This shows that the Fourier series generally tries to converge to the midpoint of the jump. 228 Solution: Using the Fourier series obtained above for fx = x and the result that the Fourier series converges to the midpoint of the jump obtained above to find an interesting formula by examining where the Fourier series converges when x = π/2. Of course you can get many other interesting formulas in the same way. Hint: You should get n S n fx = ∑ 2−1 k k=1 k+1 sinkx Let x = π/2. Then you must have π = lim n→∞ 2 = lim n→∞ n 2−1 k+1 sin k π k 2 ∑ k=1 n n = lim n→∞ ∑ 2k2− 1 sin k=1 2k − 1 π 2 ∑ 2k2− 1 −1 k+1 . k=1 Then, dividing by 2 you get π = lim n→∞ 4 n k+1 ∑ −1 2k − 1 k=1 You could also consider the Fourier series for fx = x 2 and get n lim n→∞ ∑ k42 −1 k coskx + k=1 π2 = x2 3 because the periodic extension of this function is continuous. Let x = 0 n lim n→∞ ∑ k42 −1 k + k=1 π2 = 0 3 and so π 2 = lim n→∞ 3 n ∑ k=1 4 −1 k+1 ≡ k2 ∞ ∑ k42 −1 k+1 k=1 This is one of those calculus problems where you show it converges absolutely by the comparison test with a p series. However, here is what it converges to. 229 Solution: Let V be an inner product space and let K be a convex subset of V. This means that if x, z ∈ K, then the line segment x + tz − x = 1 − tx + tz is contained in K for all t ∈ 0, 1. Note that every subspace is a convex set. Let y ∈ V and let x ∈ K. Show that x is the closest point to y out of all points in K if and only if for all w ∈ K, Re〈y − x, w − x〉 ≤ 0. n In R , a picture of the above situation where x is the closest point to y is as follows. The condition of the above variational inequality is that the angle θ shown in the picture is larger than 90 degrees. Recall the geometric description of the dot product in terms of an included angle. Consider for t ∈ 0, 1 the following. |y − x + tw − x|2 where w ∈ K and x ∈ K. It equals ft = |y − x|2 + t 2 |w − x|2 − 2t Re〈y − x, w − x〉 Suppose x is the point of K which is closest to y. Then f ′ 0 ≥ 0. However, f ′ 0 = −2 Re〈y − x, w − x〉. Therefore, if x is closest to y, Re〈y − x, w − x〉 ≤ 0. Next suppose this condition holds. Then you have |y − x + tw − x|2 ≥ |y − x|2 + t 2 |w − x|2 ≥ |y − x|2 By convexity of K, a generic point of K is of the form x + tw − x for w ∈ K. Hence x is the closest point. 230 Solution: Show that in any inner product space the parallelogram identity holds. |x + y|2 + |x − y|2 = 2|x|2 + 2|y|2 Next show that in a real inner product space, the polarization identity holds. 〈x, y〉 = 1 |x + y|2 − |x − y|2 . 4 This follows right away from the axioms of the inner product. |x + y|2 + |x − y|2 = |x|2 + |y|2 + 2 Re〈x, y〉 + |x|2 + |y|2 − 2 Re〈x, y〉 = 2|x|2 + 2|y|2 Of course the same reasoning yields 1 |x + y|2 − |x − y|2 = 1 |x|2 + |y|2 + 2〈x, y〉 − |x|2 + |y|2 − 2〈x, y〉 4 4 = 〈x, y〉 231 Solution: ∗ This problem is for those who know about Cauchy sequences and completeness of F p for F either R or C, and about closed sets. Suppose K is a closed nonempty convex subset of a finite dimensional subspace U of an inner product space V. Let y ∈ V. Then show there exists a unique point x ∈ K which is closest to y. Hint: Let λ = inf|y − z| : z ∈ K Let x n be a minimizing sequence, |y − x n | → λ. Use the parallelogram identity in the above problem to show that x n is a Cauchy sequence. Now let u k pk=1 be an orthonormal basis for U. Say p xn = ∑ c nk u k k=1 Verify that for c n ≡ c n1 , ⋯, c np ∈ F p |x n − x m | = |c n − c m |F p . Now use completeness of F p and the assumption that K is closed to get the existence of x ∈ K such that |x − y| = λ. The hint is pretty good. Let x k be a minimizing sequence. The connection between x k and c k ∈ F k is obvious because the u k are orthonormal. That is, |x n − x m | = |c n − c m |F p . Use the parallelogram identity. y − x k − y − x m 2 2 + y − x k + y − x m 2 2 =2 y − xk 2 2 +2 y − xm 2 Hence 1 |x − x |2 = 1 |y − x |2 + 1 |y − x |2 − y − x k + x m 2 k k m 4 m 2 2 2 2 2 ≤ 1 |y − x k | + 1 |y − x m | − λ 2 2 2 Now the right hand side converges to 0 since x k is a minimizing sequence. Therefore, x k is a Cauchy sequence in U. Hence the sequence of component vectors c k is a Cauchy sequence in F n and so it converges thanks to completeness of F. It follows that x k also must converge to some x. Then since K is closed, it follows that x ∈ K. Hence λ = |x − y|. 232 Solution: ∗ Let K be a closed nonempty convex subset of a finite dimensional subspace U of a real inner product space V. (It is true for complex ones also.) For x ∈ V, denote by Px the unique closest point to x in K. Verify that P is Lipschitz continuous with Lipschitz constant 1, |Px − Py| ≤ |x − y|. Hint: Use the above problem which involves characterizing the closest point with a variational inequality. From the problem, 〈Px − Py, y − Py〉 ≤ 0 〈Py − Px, x − Px〉 ≤ 0 Thus 〈Px − Py, x − Px〉 ≥ 0 Hence 〈Px − Py, x − Px〉 − 〈Px − Py, y − Py〉 ≥ 0 and so 〈Px − Py, x − y − Px − Py〉 ≥ 0 |x − y||Px − Py| ≥ 〈Px − Py, Px − Py〉 = |Px − Py|2 233 Solution: ∗ This problem is for people who know about compactness. It is an analysis problem. If you have only had the usual undergraduate calculus course, don’t waste your time with this problem. Suppose V is a finite dimensional normed linear space. Recall this means that there exists a norm ‖⋅‖ defined on V as described above, ‖v‖ ≥ 0 equals 0 if and only if v = 0 ‖v + u‖ ≤ ‖u‖ + ‖v‖, ‖αv‖ = |α|‖v‖. Let |⋅| denote the norm which is defined by |v|2 ≡ n n j=1 j=1 ∑|v j |2 where v = ∑ v ju j u j a basis, the norm which comes from the inner product by decree. Show |⋅| and ‖⋅‖ are equivalent. That is, there exist constants δ, Δ > 0 such that for all x ∈ V, δ|x| ≤ ‖x‖ ≤ Δ|x|. Next explain why every two norms on a finite dimensional vector space must be equivalent in the above sense. Let u k nk=1 be a basis for V and if x ∈ V, let x i be the components of x relative to this basis. Thus the x i are defined according to ∑ xi ui = x i Then since u i is an orthonormal basis by decree, it follows that ∑|x i |2 . |x|2 = i Now letting x k be a sequence of vectors of V let x k denote the sequence of component vectors in F n . One direction is easy, saying that ‖x‖ ≤ Δ|x|. If this is not so, then there exists a sequence of vectors x k such that ‖x k ‖ > k|x k | dividing both sides by ‖x k ‖ it can be assumed that 1 > k|x k | = |x k |. Hence x k → 0 in F k . But from the triangle inequality, n ‖x k ‖ ≤ ∑ x ki ‖u i ‖ i=1 = 0, this is a contradiction to each ‖x k ‖ = 1. It follows that there Therefore, since exists Δ such that for all x, ‖x‖ ≤ Δ|x| lim k→∞ x ki Now consider the other direction. If it is not true, then there exists a sequence x k such that 1 |x k | > ‖x k ‖ k Dividing both sides by |x k |, it can be assumed that |x k | = |x k | = 1. Hence, by compactness of the closed unit ball in F n , there exists a further subsequence, still denoted by k such that x k → a ∈ F n and it also follows that |a|F n = 1. Also the above inequality implies lim k→∞ ‖x k ‖ = 0. Therefore, n n j=1 j=1 xk = 0 ∑ x kj u j = lim ∑ a ju j = lim k→∞ k→∞ which is a contradiction to the u j being linearly independent. Therefore, there exists δ > 0 such that for all x, δ|x| ≤ ‖x‖. Now if you have any other norm on this finite dimensional vector space, say |||⋅|||, then from what was just shown, there exist scalars δ i and Δ i all positive, such that δ 1 |x| ≤ ‖x‖ ≤ Δ 1 |x| δ 2 |x| ≤ |||x||| ≤ Δ 2 |x| It follows that Δ Δ Δ Δ Δ |||x||| ≤ 2 ‖x‖ ≤ 2 1 |x| ≤ 2 1 |||x|||. δ1 δ1 δ1δ2 Hence δ1 Δ |||x||| ≤ ‖x‖ ≤ 1 |||x|||. Δ2 δ2 In other words, any two norms on a finite dimensional vector space are equivalent norms. What this means is that every consideration which depends on analysis or topology is exactly the same for any two norms. What might change are geometric properties of the norms. Linear Transformations 234 Solution: Find the matrix for the linear transformation which rotates every vector in R 2 through an angle of − 14 π. 1 2 − 12 2 2 1 2 1 2 2 2 235 Solution: Find the matrix for the linear transformation which rotates every vector in R 2 through an angle of 34 π. − 12 2 − 12 2 1 2 2 − 12 2 236 Solution: Find the matrix for the linear transformation which rotates every vector in R 2 through an angle of 43 π. − 12 − 12 3 1 2 3 − 12 237 Solution: Find the matrix for the linear transformation which rotates every vector in R 2 through an angle of 14 π. and then reflects through the y axis. 1 0 0 −1 1 2 1 2 2 − 12 2 2 1 2 2 = 1 2 − 12 2 − 12 2 2 − 12 2 238 Solution: Find the matrix for the linear transformation which rotates every vector in R 2 through an angle of 5π/12. Hint: Note that 5π/12 = 2π/3 − π/4. cos 2π − sin 2π 3 3 cos −π − sin −π 4 4 sin 2π 3 sin −π 4 cos 2π 3 1 4 2 3 − 1 4 2 − 14 2 3 − 1 4 2 3 + 1 4 2 2 3 − 1 4 = cos −π 4 1 4 1 4 2 2 239 Solution: Find the matrix for the linear transformation which rotates every vector in R 2 through an angle of 13 π. and then reflects through the y axis. −1 0 0 − 12 3 1 2 1 2 1 1 2 3 − 12 = 1 2 1 2 3 3 1 2 240 Solution: Find the matrix for the linear transformation which rotates every vector in R 2 through an angle of 16 π. 1 2 3 1 2 − 12 1 2 3 241 Solution: Find the matrix for the linear transformation which rotates every vector in R 2 through an angle of 76 π. − 12 3 1 2 − 12 − 12 3 242 Solution: Find the matrix for the linear transformation which rotates every vector in R 2 through an angle of π/12. Hint: Note that π/12 = π/3 − π/4. − 12 3 1 2 1 2 3 1 2 1 2 − 1 2 2 2 − 12 2 1 2 2 = 1 4 2 3 + 1 4 1 4 2 3 − 1 4 2 − 14 2 3 − 2 1 4 2 − 1 4 1 4 2 2 3 243 Solution: Let V be a finite dimensional inner product space, the field of scalars equal to either R or C. Verify that f given by fv ≡ 〈v, z〉 is in LV, F, the space of linear transformations from V to the field F. Next suppose f is an arbitrary element of LV, F. Show the following. a. If f = 0, the zero mapping, then fv = 〈v, 0〉 for all v ∈ V. This is obvious from the properties of the inner product. 〈v, 0〉 = 〈v, 0 + 0〉 = 〈v, 0〉 + 〈v, 0〉 so 〈v, 0〉 = 0. b. If f ≠ 0 then there exists z ≠ 0 satisfying 〈u, z〉 = 0 for all u ∈ kerf. kerf is a subspace and so there exists z 1 ∉ kerf. Then there exists a closest point of kerf to z 1 called x. Then let z = z 1 − x. By the theorems on minimization, 〈u, z〉 = 0 for all u ∈ kerf. c. Explain why fyz − fzy ∈ kerf. ffyz − fzy = fyfz − fzfy = 0. d. Use the above to show that there exists w such that fy = 〈y, w〉 for all y ∈ V. 0 = 〈fyz − fzy, z〉 = fy|z|2 − fz〈y, z〉 and so fz fy = y, 2 z |z| so w = fz |z|2 z appears to work. e. Show there is at most one such w. If w 1 , w 2 both work, then for every y, 0 = fy − fy = 〈y, w 1 〉 − 〈y, w 2 〉 = 〈y, w 1 − w 2 〉 In particular, this is true for y = w 1 − w 2 and so w 1 = w 2 . You have now proved the Riesz representation theorem which states that every f ∈ LV, F is of the form fy = 〈y, w〉 for a unique w ∈ V. 244 Solution: ↑Let A ∈ LV, W where V, W are two finite dimensional inner product spaces, both having field of scalars equal to F which is either R or C. Let f ∈ LV, F be given by fy ≡ 〈Ay, z〉 where 〈 〉 now refers to the inner product in W. Use the above problem to verify that there exists a unique w ∈ V such that fy = 〈y, w〉, the inner product here being the one on V. Let A ∗ z ≡ w. Show that A ∗ ∈ LW, V and by construction, 〈Ay, z〉 = 〈y, A ∗ z〉. In the case that V = F n and W = F m and A consists of multiplication on the left by an m × n matrix, give a description of A ∗ . It is required to show that A ∗ is linear. 〈y, A ∗ αz + βw〉 ≡ 〈Ay, αz + βw〉 = α 〈Ay, z〉 + β 〈Ay, w〉 ≡ α 〈y, A ∗ z〉 + β 〈y, A ∗ w〉 = 〈y, αA ∗ z〉 + 〈y, βA ∗ w〉 = 〈y, αA ∗ z + βA ∗ w〉 Since y is arbitrary, this shows that A ∗ is linear. In case A is an m × n matrix as described, A ∗ = A T 245 Solution: Let A be the linear transformation defined on the vector space of smooth functions (Those which have all derivatives) given by Af = D 2 + 2D + 1f. Find kerA. First solve D + 1z = 0. This is easy to do and gives zt = C 1 e −t . Now solve D + 1y = C 1 e −t . To do this, you multiply by an integrating factor. Thus d e t y = C 1 dt Take an antiderivative and obtain ety = C1t + C2 Thus the desired kernel consists of all functions y which are of the form yt = C 1 te t + C 2 e t where C i is a constant. 246 Solution: Let A be the linear transformation defined on the vector space of smooth functions (Those which have all derivatives) given by Af = D 2 + 5D + 4f. Find kerA. Note that you could first find kerD + 4 where D is the differentiation operator and then consider kerD + 1D + 4 = kerA and consider Sylvester’s theorem. In this case, the two operators D + 1 and D + 4 commute and are each one to one on the kernel of the other. Also, it is obvious that kerD + a consists of functions of the form Ce −at . Therefore, kerD + 1D + 4 consists of functions of the form y = C 1 e −t + C 2 e −4t where C 1 , C 2 are arbitrary constants. In other words, a basis for kerD + 1D + 4 is e −t , e −4t . 247 Solution: Let A be the linear transformation defined on the vector space of smooth functions (Those which have all derivatives) given by Af = D 2 + 6D + 8f. Find kerA. Note that you could first find kerD + 2 where D is the differentiation operator and then consider kerD + 2D + 4 = kerA and consider Sylvester’s theorem. In this case, the two operators D + 2 and D + 4 commute and are each one to one on the kernel of the other. Also, it is obvious that kera + D consists of functions of the form Ce −at . Therefore, kerD + 2D + 4 consists of functions of the form y = C 1 e −2t + C 2 e −4t where C 1 , C 2 are arbitrary constants. In other words, a basis for kerD + 2D + 4 is e −2t , e −4t . 248 Solution: Let A be the linear transformation defined on the vector space of smooth functions (Those which have all derivatives) given by Af = D 2 + 3D − 4f. Find kerA. Note that you could first find kerD − 1 where D is the differentiation operator and then consider kerD − 1D + 4 = kerA and consider Sylvester’s theorem. In this case, the two operators D − 1 and D + 4 commute and are each one to one on the kernel of the other. Also, it is obvious that kera + D consists of functions of the form Ce −at . Therefore, kerD − 1D + 4 consists of functions of the form y = C 1 e t + C 2 e −4t where C 1 , C 2 are arbitrary constants. In other words, a basis for kerD − 1D + 4 is e t , e −4t . 249 Solution: Let A be the linear transformation defined on the vector space of smooth functions (Those which have all derivatives) given by Af = D 2 − 7D + 12f. Find kerA. Note that you could first find kerD − 4 where D is the differentiation operator and then consider kerD − 4D − 3 = kerA and consider Sylvester’s theorem. In this case, the two operators D − 4 and D − 3 commute and are each one to one on the kernel of the other. Also, it is obvious that kera + D consists of functions of the form Ce −at . Therefore, kerD − 4D − 3 consists of functions of the form y = C 1 e 4t + C 2 e 3t where C 1 , C 2 are arbitrary constants. In other words, a basis for kerD − 4D − 3 is e 4t , e 3t . 250 Solution: Let A be the linear transformation defined on the vector space of smooth functions (Those which have all derivatives) given by Af = D 2 + 4D + 4f. Find kerA. First, kerD + a = C 1 e −2t . Therefore, D + ay = C 1 e −2t and so, solving this linear first order equation, y = C 1 te −2t + C 2 e −2t 251 Solution: Suppose Ax = b has a solution where A is a linear transformation. Explain why the solution is unique precisely when Ax = 0 has only the trivial (zero) solution. Suppose first there is a unique solution z. Then there can only be one solution to Ax = 0 because if this last equation had more than one solution, say y i , i = 1, 2, then y i + z, i = 1, 2 would each be solutions to Ay = b because A is linear. Recall also that the general solution to Ax = b consists of the general solution to Ax = 0 added to any solution to Ax = b. Therefore, if there is only one solution to Ax = 0, then there is only one solution to Ax = b. 252 Solution: Verify the linear transformation determined by the matrix A= 1 0 2 0 1 4 maps R 3 onto R 2 but the linear transformation determined by this matrix is not one to one. This is an old problem. The rank is obviously 2 so the matrix maps onto R 2 . It is not one to one because the system Ax = 0 must have a free variable. 253 Solution: Let L be the linear transformation taking polynomials of degree at most three to polynomials of degree at most three given by D 2 + 4D + 3 where D is the differentiation operator. Find the matrix of this linear transformation relative to the basis 1, x, x 2 , x 3 . Also find the matrices with of D + 3 and D + 1 and multiply these. You should get the same thing. Why? 3 3x + 4 3x 2 + 8x + 2 3x 3 + 12x 2 + 6x 3 4 2 0 = 0 3 8 6 1 x x2 x3 0 0 3 12 0 0 0 3 Done the other way, the matrix of D + 3, and D + 1 are respectively 3 1 0 0 0 3 2 0 0 0 3 3 0 0 0 3 1 1 0 0 , 0 1 2 0 0 0 1 3 0 0 0 1 Then since composition of transformation corresponds to multiplication of the matrices, the matrix of the full operator is 3 1 0 0 1 1 0 0 0 3 2 0 0 1 2 0 0 0 3 3 0 0 1 3 0 0 0 3 0 0 0 1 = 3 4 2 0 0 3 8 6 0 0 3 12 0 0 0 3 254 Solution: Let L be the linear transformation taking polynomials of degree at most three to polynomials of degree at most three given by D 2 − 16 where D is the differentiation operator. Find the matrix of this linear transformation relative to the basis 1, x, x 2 , x 3 . Also find the matrices with of D + a and D + b and multiply these. You should get the same thing. Why? −16 −16x 2 − 16x 2 6x − 16x 3 = 1 x x2 x3 −16 0 2 0 0 −16 0 6 0 0 −16 0 0 0 0 −16 Done the other way, the matrix of D − 4, and D + 4 are respectively −4 1 0 0 0 −4 2 0 0 0 −4 3 0 0 0 −4 4 1 0 0 , 0 4 2 0 0 0 4 3 0 0 0 4 Then since composition of transformation corresponds to multiplication of the matrices, the matrix of the full operator is −4 1 0 0 4 1 0 0 0 −4 2 0 0 4 2 0 0 0 −4 3 0 0 4 3 0 0 0 −4 0 0 0 4 = −16 0 2 0 0 −16 0 6 0 0 −16 0 0 0 0 −16 255 Solution: Let L be the linear transformation taking polynomials of degree at most three to polynomials of degree at most three given by D 2 + 2D + 1 where D is the differentiation operator. Find the matrix of this linear transformation relative to the basis 1, x, x 2 , x 3 . Find the matrix directly and then find the matrix with respect to the differential operator D + 1 and multiply this matrix by itself. You should get the same thing. Why? You should get the same thing because the multiplication of matrices corresponds to composition of linear transformations. Lets find the matrix for D + 1 first. 1 x + 1 2x + x 2 3x 2 + x 3 1 1 0 0 = 1 x x2 x3 0 1 2 0 0 0 1 3 0 0 0 1 so the matrix is 1 1 0 0 0 1 2 0 0 0 1 3 0 0 0 1 Then the matrix of the desired transformation is just this one squared. 1 2 2 0 0 1 4 6 0 0 1 6 0 0 0 1 256 Solution: Show that if L ∈ LV, W (linear transformation) where V and W are vector spaces, then if Ly p = f for some y p ∈ V, then the general solution of Ly = f is of the form kerL + y p . This is really old stuff. However, here it applies to an arbitrary linear transformation. Suppose Lz = f. Then L z − y p = Lz − L y p = f − f = 0. Therefore, letting y = z − y p , it follows that y is a solution to Ly = 0 and z = y p + y. Thus every solution to Ly = f is of the form y p + y for some y which solves Ly = 0. 257 Solution: Let L ∈ LV, W where V, W are vector spaces, finite or infinite dimensional, and define x ∼ y if x − y ∈ kerL. Show that ∼ is an equivalence relation. Next define addition and scalar multiplication on the space of equivalence classes as follows. x + y ≡ x + y αx = αx Show that these are well defined definitions and that the set of equivalence classes is a vector space with respect to these operations. The zero is ker L. Denote the resulting vector space by V/ kerL. Now suppose L is onto W. Define a mapping A : V/ kerK → W as follows. Ax ≡ Lx Show that A is well defined, linear, one to one and onto. Now suppose L is onto W. Define a mapping A : V/ kerK → W as follows. Ax ≡ Lx Show that A is well defined, one to one and onto. It is obvious that x ∼ x. If x ∼ y, then y ∼ x is also clear. If x ∼ y and y ∼ z, then z−x = z−y+y−x and by assumption, both z − y and y − x ∈ kerL which is a subspace. Therefore, z − x ∈ kerL also and so ∼ is an equivalence relation. Are the operations well defined? If x = x ′ , y = y ′ , is it true that x + y = y ′ + x ′ ? Of course. x ′ + y ′ − x + y = x ′ − x + y ′ − y ∈ kerL because kerL is a subspace. Similar reasoning applies to the case of scalar multiplication. Now why is A well defined? If x = x ′ , is Lx = Lx ′ ? Of course this is so. x − x ′ ∈ kerL by assumption. Therefore, Lx = Lx ′ . It is clear also that A is linear. Aax + by = Aax + by = aLx + bLy = aAx + bAy If Ax = 0, then Lx = 0 and so x ∈ kerL and so x = 0. Therefore, A is one to one. It is obviously onto LV = W. 258 Solution: If V is a finite dimensional vector space and L ∈ LV, V, show that the minimal polynomial for L equals the minimal polynomial of A where A is the n × n matrix of L with respect to some basis. This is really easy because of the definition of what you mean by the matrix of a linear transformation. L = qAq −1 where q, q −1 are one to one and onto linear transformations from V to F n or from F n to V. Thus if pλ is a polynomial, pL = qpAq −1 Thus the polynomials which send L to 0 are the same as those which send A to 0. 259 Solution: Let A be an n × n matrix. Describe a fairly simple method based on row operations for computing the minimal polynomial of A. Recall, that this is a monic polynomial pλ such that pA = 0 and it has smallest degree of all such monic polynomials. Hint: Consider I, A 2 , ⋯. 2 Regard each as a vector in F n and consider taking the row reduced echelon form or something like this. You might also use the Cayley Hamilton theorem to note that you can stop the above sequence at A n . 2 An easy way to do this is to “unravel" the powers of the matrix making vectors in F n and then making these the columns of a n 2 × n matrix. Look for linear relationships between the columns by obtaining the row reduced echelon form. As an example, consider the following matrix. 1 1 0 −1 0 −1 2 1 3 Lets find its minimal polynomial. We have the following powers 1 0 0 0 1 0 0 0 1 1 , 1 0 0 −1 0 −1 2 1 , 1 −1 −3 −2 −3 3 7 5 8 −3 −1 −4 , −7 −6 −7 18 15 19 By the Cayley Hamilton theorem, I won’t need to consider any higher powers than this. Now I will unravel each and make them the columns of a matrix. 1 1 0 −3 0 1 1 −1 0 0 −1 −4 0 −1 −3 −7 1 0 −2 −6 0 −1 −3 −7 0 2 7 18 0 1 5 15 1 3 8 19 Next you can do row operations and obtain the row reduced echelon form for this matrix and then look for linear relationships. 1 0 0 2 0 1 0 −5 0 0 1 4 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 From this you see that for A denoting the matrix, A 3 = 4A 2 − 5A + 2I and so the minimal polynomial is λ 3 − 4λ 2 + 5λ − 2 No smaller degree polynomial can work either. Since it is of degree 3, this is also the characteristic polynomial. Note how we got this without expanding any determinants or solving any polynomial equations. If you factor this polynomial, you get λ 3 − 4λ 2 + 5λ − 2 = λ − 2λ − 1 2 so this is an easy problem, but you see that this procedure for finding the minimal polynomial will work even when you can’t factor the characteristic polynomial. If you want to work with smaller matrices, you could also look at A k e i for e 1 , e 2 , etc., use similar techniques on each of these and then find the least common multiple of the resulting polynomials. 260 Solution: Let A be an n × n matrix which is non defective. That is, there exists a basis of eigenvectors. Show that if pλ is the minimal polynomial, then pλ has no repeated roots. Hint: First show that the minimal polynomial of A is the same as the minimal polynomial of the diagonal matrix Dλ 1 D= ⋱ Dλ r Where Dλ is a diagonal matrix having λ down the main diagonal and in the above, the λ i are r distinct. Show that the minimal polynomial is ∏ i=1 λ − λ i . If two matrices are similar, then they must have the same minimal polynomial. This is obvious from the fact that for pλ any polynomial and A = S −1 BS, pA = S −1 pBS r So what is the minimal polynomial of the above D? It is obviously ∏ i=1 λ − λ i . Thus there are no repeated roots. 261 Solution: Show that if A is an n × n matrix and the minimal polynomial has no repeated roots, then A is non defective and there exists a basis of eigenvectors. Thus, from the above problem, a matrix may be diagonalized if and only if its minimal polynomial has no repeated roots. It turns out this condition is something which is relatively easy to determine. If A has a minimal polynomial which has no repeated roots, say m pλ = ∏λ − λ i , j=1 then from the material on decomposing into direct sums of generalized eigenspaces, you have F n = kerA − λ 1 I ⊕ kerA − λ 2 I ⊕ ⋯ ⊕ kerA − λ m I and by definition, the basis vectors for kerA − λ 2 I are all eigenvectors. Thus F n has a basis of eigenvectors and is therefore diagonalizable or non defective. Conversely, if F n has a basis of eigenvectors, then A is similar to a diagonal matrix. Say the distinct entries on the diagonal are λ 1 , ⋯, λ m , m ≤ n. Then it is easy to check that the minimal m polynomial is ∏ j=1 λ − λ i because this polynomial will send the diagonal matrix to zero. 262 Solution: Recall the linearization of the Lotka Volterra equations used to model the interaction between predators and prey. It was shown earlier that if x, y are the deviations from the equilibrium point, then x ′ = −bxy − b c y d a ′ y = dxy + dx b If one is interested only in small values of x, y, that is, in behavior near the equilibrium point, these are approximately equal to the system x ′ = −b c y d a ′ y = dx b Written more simply, for α, β > 0, x ′ = −αy y ′ = βx Find the solution to the initial value problem x′ = y′ 0 −α x β y 0 x0 , y0 x0 = y0 where x 0 , y 0 are both real. Also have a computer graph the vector fields −xy − y, xy + x and −y, x which result from setting all the parameters equal to 1. From the discussion in the book, the eigenvalues are ± αβ i. The eigenpairs are −i α β i α ↔ −i α β , β ↔i α β Then the fundamental matrix is −i α i α β β exp −t αβ i 0 0 exp t αβ i 1 2 − 12 i α i α 1 2 β 1 2 β Then the solution desired is xt yt = exp −t αβ i −i α i α β β 1 2 0 − exp t αβ i 0 1 2 i α i α 1 2 β x0 1 2 β y0 That is, xt yt e −it αβ β e −it αβ 1 2 = 1 2 i α + 1 2 − 1 2 αβ e it i α 1 2 β e it α i β αβ 1 2 e it e −it αβ − αβ + 1 2 i 1 2 α β e it e −it αβ αβ x0 y0 Since the solution must be real, this is of the form xt yt αβ t cos = β α sin αβ t α β sin cos αβ t = α β y0 αβ t Thus both x and y oscillate about the origin. For example, letting ω = xt = x 0 cosωt + y 0 x0 αβ , sinωt x 20 + y 20 α cosωt − φ β where φ is a phase shift. A similar formula holds for y. First is the field for the full Lotka Volterra equations and then is the one for the linearized Lotka Volterra equations. These suggest that the behavior of the linearized system and the nonlinear system is similar only for small x, y.