INSTRUCTOR’S SOLUTIONS MANUAL D IFFERENTIAL E QUATIONS AND L INEAR ALGEBRA FOURTH EDITION Stephen W. Goode and Scott A. Annin California State University, Fullerton Boston Columbus Indianapolis New York San Francisco Amsterdam Cape Town Dubai London Madrid Milan Munich Paris Montreal Toronto Delhi Mexico City São Paulo Sydney Hong Kong Seoul Singapore Taipei Tokyo The author and publisher of this book have used their best efforts in preparing this book. These efforts include the development, research, and testing of the theories and programs to determine their effectiveness. The author and publisher make no warranty of any kind, expressed or implied, with regard to these programs or the documentation contained in this book. The author and publisher shall not be liable in any event for incidental or consequential damages in connection with, or arising out of, the furnishing, performance, or use of these programs. Reproduced by Pearson from electronic files supplied by the author. Copyright © 2017, 2011, 2005 Pearson Education, Inc. Publishing as Pearson, 501 Boylston Street, Boston, MA 02116. All rights reserved. No part of this publication may be reproduced, stored in a retrieval system, or transmitted, in any form or by any means, electronic, mechanical, photocopying, recording, or otherwise, without the prior written permission of the publisher. Printed in the United States of America. ISBN-13: 978-0-321-98551-4 ISBN-10: 0-321-98551-6 www.pearsonhighered.com Contents Chapter 1 1 Chapter 2 123 Chapter 3 207 Chapter 4 256 Chapter 5 348 Chapter 6 390 Chapter 7 442 Chapter 8 534 Chapter 9 631 Chapter 10 762 Chapter 11 861 Appendix A 979 Appendix B 982 Appendix C 987 1 Chapter 1 Solutions Solutions to Section 1.1 True-False Review: (a): FALSE. A derivative must involve some derivative of the function y = f (x), not necessarily the first derivative. (b): FALSE. The order of a differential equation is the order of the highest, not the lowest, derivative appearing in the differential equation. (c): FALSE. This differential equation has order two, since the highest order derivative that appears in the equation is the second order expression y . (d): FALSE. The carrying capacity refers to the maximum population size that the environment can support in the long run; it is not related to the initial population in any way. (e): TRUE. The value y(0) is called an initial condition to the differential equation for y(t). (f ): TRUE. According to Newton’s Law of Cooling, the rate of cooling is proportional to the difference between the object’s temperature and the medium’s temperature. Since that difference is greater for the object at 100◦ F than the object at 90◦ F , the object whose temperature is 100◦ F has a greater rate of cooling. (g): FALSE. The temperature of the object is given by T (t) = Tm + ce−kt , where Tm is the temperature of the medium, and c and k are constants. Since e−kt = 0, we see that T (t) = Tm for all times t. The temperature of the object approaches the temperature of the surrounding medium, but never equals it. (h): TRUE. Since the temperature of the coffee is falling, the temperature difference between the coffee and the room is higher initially, during the first hour, than it is later, when the temperature of the coffee has already decreased. (i): FALSE. The slopes of the two curves are negative reciprocals of each other. (j): TRUE. If the original family of parallel lines have slopes k for k = 0, then the family of orthogonal trajectories are parallel lines with slope − k1 . If the original family of parallel lines are vertical (resp. horizontal), then the family of orthogonal trajectories are horizontal (resp. vertical) parallel lines. (k): FALSE. The family of orthogonal trajectories for a family of circles centered at the origin is the family of lines passing through the origin. (l): TRUE. If v(t) denotes the velocity of the object at time t and a(t) denotes the velocity of the object at time t, then we have a(t) = v (t), which is a differential equation for the unknown function v(t). (m): FALSE. The restoring force is directed in the direction opposite to the displacement from the equilibrium position. (n): TRUE. The allometric relationship B = B0 m3/4 , where B0 is a constant, relates the metabolic rate and total body mass for any species. Problems: 1. The order is 2. 2. The order is 1. (c)2017 Pearson Education. Inc. 2 3. The order is 3. 4. The order is 2. 5. We compute the first three derivatives of y(t) = ln t: 1 d2 y = − 2, dt2 t dy 1 = , dt t Therefore, 2 dy dt 3 = d3 y 2 = 3. dt3 t 2 d3 y = 3, 3 t dt as required. 6. We compute the first two derivatives of y(x) = x/(x + 1): dy 1 = dx (x + 1)2 d2 y 2 =− . 2 dx (x + 1)3 and Then y+ d2 y x x3 + 2x2 + x − 2 (x + 1) + (x3 + 2x2 − 3) 1 x3 + 2x2 − 3 2 = = = = + , − 2 3 3 3 2 dx x + 1 (x + 1) (x + 1) (x + 1) (x + 1) (1 + x)3 as required. 7. We compute the first two derivatives of y(x) = ex sin x: dy = ex (sin x + cos x) dx Then 2y cot x − and d2 y = 2ex cos x. dx2 d2 y = 2(ex sin x) cot x − 2ex cos x = 0, dx2 as required. dT d = −k =⇒ (ln |T − Tm |) = −k. The preceding equation can be integrated directly to dt dt yield ln |T − Tm | = −kt + c1 . Exponentiating both sides of this equation gives |T − Tm | = e−kt+c1 , which can be written as T − Tm = ce−kt , 8. (T − Tm )−1 where c = ±ec1 . Rearranging yields T (t) = Tm + ce−kt . 9. After 4 p.m. In the first two hours after noon, the water temperature increased from 50◦ F to 55◦ F, an increase of five degrees. Because the temperature of the water has grown closer to the ambient air temperature, the temperature difference |T − Tm | is smaller, and thus, the rate of change of the temperature of the water grows smaller, according to Newton’s Law of Cooling. Thus, it will take longer for the water temperature to increase another five degrees. Therefore, the water temperature will reach 60◦ F more than two hours later than 2 p.m., or after 4 p.m. 10. The object temperature cools a total of 40◦ F during the 40 minutes, but according to Newton’s Law of Cooling, it cools faster in the beginning (since |T − Tm | is greater at first). Thus, the object cooled half-way (c)2017 Pearson Education. Inc. 3 from 70◦ F to 30◦ F in less than half the total cooling time. Therefore, it took less than 20 minutes for the object to reach 50◦ F. 11. The given family of curves satisfies: x2 + 9y 2 = c =⇒ 2x + 18y dy dy x = 0 =⇒ =− . dx dx 9y Orthogonal trajectories satisfy: dy 9y 1 dy 9 d 9 = =⇒ = =⇒ (ln |y|) = =⇒ ln |y| = 9 ln |x| + c1 =⇒ y = kx9 , where k = ±ec1 dx x y dx x dx x . y(x) 0.8 0.4 x -1.5 -1.0 0.5 -0.5 1.0 1.5 -0.4 -0.8 Figure 0.0.1: Figure for Problem 11 y . Hence, x2 y 2y dy = 2cx = c 2 x = . dx x x 12. Given family of curves satisfies: y = cx2 =⇒ c = Orthogonal trajectories satisfy: 1 x dy d 2 dy =− =⇒ 2y = −x =⇒ (y ) = −x =⇒ y 2 = − x2 + c1 =⇒ 2y 2 + x2 = c2 , dx 2y dx dx 2 where c2 = 2c1 . c dy dy y 13. Given a family of curves satisfies: y = =⇒ x + y = 0 =⇒ =− . x dx dx x Orthogonal trajectories satisfy: 1 1 dy x dy d 1 2 = =⇒ y = x =⇒ y = x =⇒ y 2 = x2 + c1 =⇒ y 2 − x2 = c2 , where c2 = 2c1 . dx y dx dx 2 2 2 y . Hence, x5 y 5y dy = 5cx4 = 5 5 x4 = . dx x x 14. The given family of curves satisfies: y = cx5 =⇒ c = (c)2017 Pearson Education. Inc. 4 y(x) 2.0 1.6 1.2 0.8 0.4 x -2 -1 1 -0.4 2 -0.8 -1.2 -1.6 -2.0 Figure 0.0.2: Figure for Problem 12 y(x) 4 2 x -2 -4 4 2 -2 -4 Figure 0.0.3: Figure for Problem 13 Orthogonal trajectories satisfy: dy x dy d =− =⇒ 5y = −x =⇒ dx 5y dx dx 5 2 y 2 = −x =⇒ 5 2 1 y = − x2 + c1 =⇒ 5y 2 + x2 = c2 , 2 2 where c2 = 2c1 . (c)2017 Pearson Education. Inc. 5 y(x) 0.8 0.4 x -1.5 1.0 -1.0 1.5 -0.4 -0.8 Figure 0.0.4: Figure for Problem 14 dy = cex = y. Orthogonal trajectories satisfy: dx dy 1 1 dy d 1 2 = − =⇒ y = −1 =⇒ y = −1 =⇒ y 2 = −x + c1 =⇒ y 2 = −2x + c2 . dx y dx dx 2 2 15. Given family of curves satisfies: y = cex =⇒ y(x) 2 1 x -1 1 -1 -2 Figure 0.0.5: Figure for Problem 15 16. Given family of curves satisfies: y 2 = 2x + c =⇒ dy 1 = .Orthogonal trajectories satisfy: dx y dy d dy = −y =⇒ y −1 = −1 =⇒ (ln |y|) = −1 =⇒ ln |y| = −x + c1 =⇒ y = c2 e−x . dx dx dx (c)2017 Pearson Education. Inc. 6 y(x) 4 3 2 1 x 1 -1 2 3 4 -1 -2 -3 -4 Figure 0.0.6: Figure for Problem 16 dy y dy my = cmxm−1 , but c = m so = . Orthogonal trajectories satisfy: dx x dx x x 1 2 1 x dy x d 1 2 1 dy =− =⇒ y = − =⇒ y = − =⇒ y 2 = − x + c1 =⇒ y 2 = − x2 + c2 . dx my dx m dx 2 m 2 2m m 17. y = cxm =⇒ dy 18. y = mx + c =⇒ = m. dx Orthogonal trajectories satisfy: 1 1 dy = − =⇒ y = − x + c1 . dx m m dy dy m = m =⇒ = . dx dx 2y Orthogonal trajectories satisfy: 19. y 2 = mx + c =⇒ 2y 2x dy 2y 2 d 2 2 dy =− =⇒ y −1 = − =⇒ (ln |y|) = − =⇒ ln |y| = − x + c1 =⇒ y = c2 e− m . dx m dx m dx m m dy dy mx + 2mx = 0 =⇒ =− . dx dx y Orthogonal trajectories satisfy: 20. y 2 + mx2 = c =⇒ 2y dy y 1 d 1 dy = =⇒ y −1 = =⇒ (ln |y|) = =⇒ m ln |y| = ln |x| + c1 =⇒ y m = c2 x. dx mx dx mx dx mx 21. The given family of curves satisfies: x2 + y 2 = 2cx =⇒ c = 2x + 2y x2 + y 2 . Hence, 2x dy x2 + y 2 = 2c = . dx x (c)2017 Pearson Education. Inc. 7 Therefore, 2y x2 + y 2 y 2 − x2 dy = − 2x = , dx x x so that dy y 2 − x2 = . dx 2xy Orthogonal trajectories satisfy: 2xy dy 2xy = 2 . =− 2 dx y − x2 x − y2 22. u = x2 + 2y 2 =⇒ 0 = 2x + 4y dy dy x =⇒ =− . dx dx 2y Orthogonal trajectories satisfy: dy dy 2y 2 d 2 = =⇒ y −1 = =⇒ (ln |y|) = =⇒ ln |y| = 2 ln |x| + c1 =⇒ y = c2 x2 . dx x dx x dx x y(x) 2.0 1.6 1.2 0.8 0.4 x -2 -1 -0.4 1 2 -0.8 -1.2 -1.6 -2.0 Figure 0.0.7: Figure for Problem 22 23. m1 = tan (a1 ) = tan (a2 − a) = 2 tan (a2 ) − tan (a) m2 − tan (a) = . 1 + tan (a2 ) tan (a) 1 + m2 tan (a) 2 gt 24. ddt2y = g =⇒ dy dt = gt + c1 =⇒ y(t) = 2 + c1 t + c2 . Now impose the initial conditions. y(0) = 0 =⇒ 2 c2 = 0. dy is: y(t) = gt2 . The object hits the dt (0) =⇒ c1 = 0. Hence, the solution to the initial-value problem 2 gt ground at time, t0 , when y(t0 ) = 100. Hence 100 = 20 , so that t0 = 200 g ≈ 4.52 s, where we have taken g = 9.8 ms−2 . (c)2017 Pearson Education. Inc. 8 d2 y = g, we integrate twice to obtain the general equations for the velocity and the position of dt2 dy 1 the ball, respectively: = gt + c and y(t) = gt2 + ct + d, where c, d are constants of integration. Setting dt 2 y = 0 to be at the top of the boy’s head (and positive direction downward), we know that y(0) = 0. Since the object hits the ground 8 seconds later, we have that y(8) = 5 (since the ground lies at the position y = 5). 5 − 32g From the values of y(0) and y(8), we find that d = 0 and 5 = 32g + 8c. Therefore, c = . 8 (a). The ball reaches its maximum height at the moment when y (t) = 0. That is, gt + c = 0. Therefore, 25. From t=− c 32g − 5 = ≈ 3.98 s. g 8g (b). To find the maximum height of the tennis ball, we compute y(3.98) ≈ −253.51 feet. So the ball is 253.51 feet above the top of the boy’s head, which is 258.51 feet above the ground. d2 y = g, we integrate twice to obtain the general equations for the velocity and the position of dt2 dy 1 the rocket, respectively: = gt + c and y(t) = gt2 + ct + d, where c, d are constants of integration. Setting dt 2 y = 0 to be at ground level, we know that y(0) = 0. Thus, d = 0. 26. From (a). The rocket reaches maximum height at the moment when y (t) = 0. That is, gt + c = 0. Therefore, the c time that the rocket achieves its maximum height is t = − . At this time, y(t) = −90 (the negative sign g accounts for the fact that the positive direction is chosen to be downward). Hence, 2 1 c2 c c c c2 c2 = g − = −90 = y − +c − − =− . g 2 g g 2g g 2g √ Solving this for c, we find that c = ± 180g. However, since c represents the initial velocity of the rocket, and the √ initial velocity is negative (relative to the fact that the positive direction is downward), we choose c = − 180g ≈ −42.02 ms−1 , and thus the initial speed at which the rocket must be launched for optimal viewing is approximately 42.02 ms−1 . c −42.02 (b). The time that the rocket reaches its maximum height is t = − ≈ − = 4.28 s. g 9.81 d2 y = g, we integrate twice to obtain the general equations for the velocity and the position of dt2 dy 1 the rocket, respectively: = gt + c and y(t) = gt2 + ct + d, where c, d are constants of integration. Setting dt 2 y = 0 to be at the level of the platform (with positive direction downward), we know that y(0) = 0. Thus, d = 0. 27. From (a). The rocket reaches maximum height at the moment when y (t) = 0. That is, gt + c = 0. Therefore, the c time that the rocket achieves its maximum height is t = − . At this time, y(t) = −85 (this is 85 m above g the platform, or 90 m above the ground). Hence, 2 1 c2 c c c c2 c2 = g − = −85 = y − +c − − =− . g 2 g g 2g g 2g (c)2017 Pearson Education. Inc. 9 √ Solving this for c, we find that c = ± 170g. However, since c represents the initial velocity of the rocket, and the √ initial velocity is negative (relative to the fact that the positive direction is downward), we choose c = − 170g ≈ −40.84 ms−1 , and thus the initial speed at which the rocket must be launched for optimal viewing is approximately 40.84 ms−1 . c −40.84 (b). The time that the rocket reaches its maximum height is t = − ≈ − = 4.16 s. g 9.81 28. If y(t) denotes the displacement of the object from its initial position at time t, the motion of the object can be described by the initial-value problem dy d2 y = g, y(0) = 0, (0) = −2. 2 dt dt dy gt2 d2 y = g =⇒ = gt + c1 =⇒ y(t) = + c1 t + c2 . Now 2 dt dt 2 dy impose the initial conditions. y(0) = 0 =⇒ c2 = 0. (0) = −2 =⇒ c1 = −2. Hence the solution to the dt 2 gt g(10)2 initial-value problem is y(t) = − 2t. We are given that y(10) = h. Consequently, h = − 2 · 10 =⇒ 2 2 −2 h = 10(5g − 2) ≈ 470 m where we have taken g = 9.8 ms . We first integrate this differential equation: 29. If y(t) denotes the displacement of the object from its initial position at time t, the motion of the object can be described by the initial-value problem d2 y dy = g, y(0) = 0, (0) = v0 . 2 dt dt d2 y dy gt2 = g =⇒ =⇒ y(t) = = gt + c + c1 t + c2 . Now impose 1 dt2 dt 2 dy the initial conditions. y(0) = 0 =⇒ c2 = 0. (0) = v0 =⇒ c1 = v0 . Hence the solution to the initial-value dt 2 gt problem is y(t) = + v0 t. We are given that y(t0 ) = h. Consequently, h = gt20 + v0 t0 . Solving for v0 yields 2 2h − gt20 v0 = . 2t0 We first integrate the differential equation: 30. From y(t) = A cos (ωt − φ), we obtain dy = −Aω sin (ωt − φ) dt Hence, and d2 y = −Aω 2 cos (ωt − φ). dt2 d2 y + ω 2 y = −Aω 2 cos (ωt − φ) + Aω 2 cos (ωt − φ) = 0. dt2 Substituting y(0) = a, we obtain a = A cos(−φ) = A cos(φ). Also, from dy dt (0) = 0, we obtain 0 = −Aω sin(−φ) = Aω sin(φ). Since A = 0 and ω = 0 and |φ| < π, we have φ = 0. It follows that a = A. dy d2 y = −c1 ω 2 cos (ωt) − = −c1 ω sin (ωt) + c2 ω cos (ωt) =⇒ dt dt2 d2 y c2 ω 2 sin (ωt) = −ω 2 [c1 cos (ωt) + c2 cos (ωt)] = −ω 2 y. Consequently, + ω 2 y = 0. To determine the dt2 31. y(t) = c1 cos (ωt) + c2 sin (ωt) =⇒ (c)2017 Pearson Education. Inc. 10 amplitude of the motion we write the solution to the differential equation in the equivalent form: c1 c2 y(t) = c21 + c22 2 cos (ωt) + 2 sin (ωt) . 2 c1 + c2 c1 + c22 We can now define an angle φ by c1 cos φ = 2 c1 + c22 and c2 sin φ = 2 . c1 + c22 Then the expression for the solution to the differential equation is y(t) = c21 + c22 [cos (ωt) cos φ + sin (ωt) sin φ] = c21 + c22 cos (ωt + φ). Consequently the motion corresponds to an oscillation with amplitude A = c21 + c22 . 32. In this problem we have m0 = 3g, M = 2700g, a = 1.5. Substituting these values into Equation (1.1.26) yields 4 1/4 1 −1.5t/(4(2700)1/4 ) m(t) = 2700 1 − 1 − e . 900 Therefore the mass of the heron after 30 days is m(30) = 2700 1 − 1 − 1 900 1/4 4 1/4 e−45/(4(2700) ) ≈ 1271.18 g. 33. In this problem we have m0 = 8g, M = 280g, a = 0.25. Substituting these values into Equation (1.1.26) yields 4 1/4 1 −t/(16(280)1/4 ) m(t) = 280 1 − 1 − e . 35 We need to find the time, t when the mass of the rat reaches 75% of its fully grown size. Therefore we need to find t such that 4 1/4 75 1 −t/(16(280)1/4 ) e . · 280 = 280 1 − 1 − 100 35 Solving algebraically for t yields t = 16 · (280) 1/4 · ln 1 − (1/35) 1/4 1 − (75/100)) 1/4 Solutions to Section 1.2 True-False Review: (a): TRUE. This is condition 1 in Definition 1.2.8. (c)2017 Pearson Education. Inc. ≈ 140 days. 11 (b): TRUE. This is the content of Theorem 1.2.12. (c): FALSE. There are solutions to y + y = 0 that do not have the form c1 cos x + 5c2 cos x, such as y(x) = sin x. Therefore, c1 cos x + 5c2 cos x does not meet the second requirement set forth in Definition 1.2.8 for the general solution. (d): FALSE. There are solutions to y + y = 0 that do not have the form c1 cos x + 5c1 sin x, such as y(x) = cos x + sin x. Therefore, c1 cos x + 5c1 sin x does not meet the second requirement set form in Definition 1.2.8 for the general solution. (e): TRUE. Since the right-hand side of the differential equation is a function of x only, we can integrate both sides n times to obtain the formula for the solution y(x). Problems: 1. Linear. 2. Non-linear, because of the y 2 expression on the right side of the equation. 3. Non-linear, because of the term yy on tthe left side of the equation. 4. Non-linear, because of the expression tan y appearing on the left side of the equation. 5. Linear. 6. Non-linear, because of the expression 1 on the left side of the equation. y 7. y(x) = c1 e−5x + c2 e5x =⇒ y = −5c1 e−5x + 5c2 e5x =⇒ y = 25c1 e−5x + 25c2 e5x =⇒ y − 25y = (25c1 e−5x + 25c2 e5x ) − 25(c1 e−5x + c2 e5x ) = 0. Thus y(x) = c1 e−5x + c2 e5x is a solution of the given differential equation for all x ∈ R. 8. y(x) = c1 cos 2x + c2 sin 2x =⇒ y = −2c1 sin 2x + 2c2 cos 2x =⇒ y = −4c1 cos 2x − 4c2 sin 2x =⇒ y + 4y = (−4c1 cos 2x − 4c2 sin 2x) + 4(c1 cos 2x + c2 sin 2x) = 0. Thus y(x) = c1 cos 2x + c2 sin 2x is a solution of the given differential equation for all x ∈ R. 9. y(x) = c1 ex + c2 e−2x =⇒ y = c1 ex − 2c2 e−2x =⇒ y = c1 ex + 4c2 e−2x =⇒ y + y − 2y = (c1 ex + 4c2 e−2x ) + (c1 ex − 2c2 e−2x ) − 2(c1 ex + c2 e−2x ) = 0. Thus y(x) = c1 ex + c2 e−2x is a solution of the given differential equation for all x ∈ R. 1 1 1 = −y 2 . Thus y(x) = =⇒ y = − is a solution of the given differential x+4 (x + 4)2 x+4 equation for x ∈ (−∞, −4) or x ∈ (−4, ∞). 10. y(x) = √ √ y c1 11. y(x) = c1 x =⇒ y = √ = . Thus y(x) = c1 x is a solution of the given differential equation for 2x 2 x all x ∈ {x : x > 0}. 12. y(x) = c1 e−x sin (2x) =⇒ y = 2c1 e−x cos (2x)−c1 e−x sin (2x) =⇒ y = −3c1 e−x sin (2x)−4c1 e−x cos (2x) =⇒ y + 2y + 5y = −3c1 e−x sin (2x) − 4c1 e−x cos (2x) + 2[2c1 e−x cos (2x) − c1 e−x sin (2x)] + 5[c1 e−x sin (2x)] = 0. Thus y(x) = c1 e−x sin (2x) is a solution to the given differential equation for all x ∈ R. 13. y(x) = c1 cosh (3x) + c2 sinh (3x) =⇒ y = 3c1 sinh (3x) + 3c2 cosh (3x) =⇒ y = 9c1 cosh (3x) + (c)2017 Pearson Education. Inc. 12 9c2 sinh (3x) =⇒ y − 9y = [9c1 cosh (3x) + 9c2 sinh (3x)] − 9[c1 cosh (3x) + c2 sinh (3x)] = 0. Thus y(x) = c1 cosh (3x) + c2 sinh (3x) is a solution to the given differential equation for all x ∈ R. c1 12c1 c2 3c1 c2 12c1 2c2 2c2 + + =⇒ y = − 4 − 2 =⇒ y = 5 + 3 =⇒ x2 y + 5xy + 3y = x2 14. y(x) = 3 + x x x x x x x5 x3 3c1 c1 c2 c2 c2 c1 5x − 4 − 2 + 3 3 + = 0. Thus y(x) = 3 + is a solution to the given differential equation x x x x x x for all x ∈ (−∞, 0) or x ∈ (0, ∞). 15. y(x) = c1 x2 ln x =⇒ y = c1 (2x ln x + x) =⇒ y = c1 (2 ln x + 3) =⇒ x2 y − 3xy + 4y = x2 · c1 (2 ln x + 3) − 3x · c1 (2x ln x + x) + 4c1 x2 ln x = c1 x2 [(2 ln x + 3) − 3(2x ln x + 1) + 4 ln x] = 0. Thus y(x) = c1 x2 ln x is a solution of the given differential equation for all x > 0. 16. y(x) = c1 x2 cos(3 ln x) =⇒ y = c1 [2x cos(3 ln x)−3x sin(3 ln x)] =⇒ y = c1 [−7 cos(3 ln x)−6 sin(3 ln x)] =⇒ x2 y −3xy +13y = x2 ·c1 [−7 cos(3 ln x)−9 sin(3 ln x)]−3x·c1 [2x cos(3 ln x)−3x sin(3 ln x)]+13c1 x2 cos(3 ln x) = c1 x2 {[−7 cos(3 ln x) − 9 sin(3 ln x)] − 3[2 cos(3 ln x) − 3 sin(3 ln x)] + 13 cos(3 ln x)} = 0. Thus y(x) = c1 x2 cos(3 ln x) is a solution of the given differential equation for all x > 0. √ c1 c1 c1 2 2 2 17. y(x) = c1 x + 3x =⇒ y = √ + 6x =⇒ y = − √ + 6 =⇒ 2x y − xy + y = 2x − √ + 6 − 2 x 4 x3 4 x3 √ √ c1 √ + 6x +(c1 x+3x2 ) = 9x2 . Thus y(x) = c1 x+3x2 is a solution to the given differential equation x 2 x for all x ∈ {x : x > 0}. 18. y(x) = c1 x2 + c2 x3 − x2 sin x =⇒ y = 2c1 x + 3c2 x2 − x2 cos x − 2x sin x =⇒ y = 2c1 + 6c2 x + x2 sin x − 2x cos x − 2x cos −2 sin x. Substituting these results into the given differential equation yields x2 y − 4xy + 6y = x2 (2c1 + 6c2 x + x2 sin x − 4x cos x − 2 sin x) − 4x(2c1 x + 3c2 x2 − x2 cos x − 2x sin x) + 6(c1 x2 + c2 x3 − x2 sin x) = 2c1 x2 + 6c2 x3 + x4 sin x − 4x3 cos x − 2x2 sin x − 8c1 x2 − 12c2 x3 + 4x3 cos x + 8x2 sin x + 6c1 x2 + 6c2 x3 − 6x2 sin x = x4 sin x. Hence, y(x) = c1 x2 + c2 x3 − x2 sin x is a solution to the differential equation for all x ∈ R. 19. y(x) = c1 eax + c2 ebx =⇒ y = ac1 eax + bc2 ebx =⇒ y = a2 c1 eax + b2 c2 ebx . Substituting these results into the differential equation yields y − (a + b)y + aby = a2 c1 eax + b2 c2 ebx − (a + b)(ac1 eax + bc2 ebx ) + ab(c1 eax + c2 ebx ) = (a2 c1 − a2 c1 − abc1 + abc1 )eax + (b2 c2 − abc2 − b2 c2 + abc2 )ebx = 0. Hence, y(x) = c1 eax + c2 ebx is a solution to the given differential equation for all x ∈ R. 20. y(x) = eax (c1 + c2 x) =⇒ y = eax (c2 ) + aeax (c1 + c2 x) = eax (c2 + ac1 + ac2 x) =⇒ y = eaax (ac2 ) + aeax (c2 + ac1 + ac2 x) = aeax (2c2 + ac1 + ac2 x). Substituting these into the differential equation yields y − 2ay + a2 y = aeax (2c2 + ac1 + ac2 x) − 2aeax (c2 + ac1 + ac2 x) + a2 eax (c1 + c2 x) = aeax (2c2 + ac1 + ac2 x − 2c2 − 2ac1 − 2ac2 x + ac1 + ac2 x) = 0. Thus, y(x) = eax (c1 + c2 x) is a solution to the given differential eqaution for all x ∈ R. (c)2017 Pearson Education. Inc. 13 21. y(x) = eax (c1 cos bx + c2 sin bx) so, y = eax (−bc1 sin bx + bc2 cos bx) + aeax (c1 cos bx + c2 sin bx) = eax [(bc2 + ac1 ) cos bx + (ac2 − bc1 ) sin bx] so, y = eax [−b(bc2 + ac1 ) sin bx + b(ac2 + bc1 ) cos bx] + aeax [(bc2 + ac1 ) cos bx + (ac2 + bc1 ) sin bx] = eax [(a2 c1 − b2 c1 + 2abc2 ) cos bx + (a2 c2 − b2 c2 − abc1 ) sin bx]. Substituting these results into the differential equation yields y − 2ay + (a2 + b2 )y = (eax [(a2 c1 − b2 c1 + 2abc2 ) cos bx + (a2 c2 − b2 c2 − abc1 ) sin bx]) − 2a(eax [(bc2 + ac1 ) cos bx + (ac2 − bc1 ) sin bx]) + (a2 + b2 )(eax (c1 cos bx + c2 sin bx)) = eax [(a2 c1 − b2 c1 + 2abc2 − 2abc2 − 2a2 c1 + a2 c1 + b2 c1 ) cos bx 2 2 2 2 2 + (a c2 − b c2 − 2abc1 + 2abc1 − 2a c2 + a c2 + b c2 ) sin bx] =0 Thus, y(x) = eax (c1 cos bx + c2 sin bx) is a solution to the given differential equation for all x ∈ R. 22. y(x) = erx =⇒ y = rerx =⇒ y = r2 erx . Substituting these results into the given differential equation yields erx (r2 − r − 6) = 0, so that r must satisfy r2 − r − 6 = 0, or (r − 3)(r + 2) = 0. Consequently r = 3 and r = −2 are the only values of r for which y(x) = erx is a solution to the given differential equation. The corresponding solutions are y(x) = e3x and y(x) = e−2x . 23. y(x) = erx =⇒ y = rerx =⇒ y = r2 erx . Substituting these results into the given differential equation yields erx (r2 + 6r + 9) = 0, so that r must satisfy r2 + 6r + 9 = 0, or (r + 3)2 = 0. Consequently r = −3 is the only value of r for which y(x) = erx is a solution to the given differential equation. The corresponding solution are y(x) = e−3x . 24. y(x) = xr =⇒ y = rxr−1 =⇒ y = r(r − 1)xr−2 . Substitution into the given differential equation yields xr [r(r − 1) + r − 1] = 0, so that r must satisfy r2 − 1 = 0. Consequently r = −1 and r = 1 are the only values of r for which y(x) = xr is a solution to the given differential equation. The corresponding solutions are y(x) = x−1 and y(x) = x. 25. y(x) = xr =⇒ y = rxr−1 =⇒ y = r(r − 1)xr−2 . Substitution into the given differential equation yields xr [r(r − 1) + 5r + 4] = 0, so that r must satisfy r2 + 4r + 4 = 0, or equivalently (r + 2)2 = 0. Consequently r = −2 is the only value of r for which y(x) = xr is a solution to the given differential equation. The corresponding solution is y(x) = x−2 . 26. y(x) = 21 x(5x2 − 3) = 12 (5x3 − 3x) =⇒ y = 12 (15x2 − 3) =⇒ y = 15x. Substitution into the Legendre equation with N = 3 yields (1 − x2 )y − 2xy + 12y = (1 − x2 )(15x) + x(15x2 − 3) + 6x(5x2 − 3) = 0. Consequently the given function is a solution to the Legendre equation with N = 3. 27. y(x) = a0 +a1 x+a2 x2 =⇒ y = a1 +2a2 x =⇒ y = 4a2 . Substitution into the given differential equation yields (1−x2 )(2a2 )−x(a1 +2a2 x)+4(a0 +a1 x+a2 x2 ) = 0 =⇒ 3a1 x+2a2 +4a0 = 0. For this equation to hold for all x we require 3a1 = 0, and 2a2 + 4a0 = 0. Consequently a1 = 0, and a2 = −2a0 . The corresponding solution to the differential equation is y(x) = a0 (1 − 2x2 ). Imposing the normalization condition y(1) = 1 requires that a0 = −1. Hence, the required solution to the differential equation is y(x) = 2x2 − 1. 28. x sin y − ex = c =⇒ x cos y dy ex − sin y dy + sin y − ex = 0 =⇒ = . dx dx x cos y (c)2017 Pearson Education. Inc. . 14 29. xy 2 + 2y − x = c =⇒ 2xy dy dy dy 1 − y2 + y2 + 2 − 1 = 0 =⇒ = . dx dx dx 2(xy + 1) dy dy 1 − yexy . Given y(1) = 0 =⇒ + y] − 1 = 0 =⇒ xexy + yexy = 1 =⇒ dx dx xexy ln x e0(1) − 1 = c =⇒ c = 0. Therefore, exy − x = 0, so that y = . x 30. exy + x = c =⇒ exy [x 31. ey/x + xy 2 − x = c =⇒ ey/x x dy −y dy dy x2 (1 − y 2 ) + yey/x dx 2 + 2xy − 1 = 0 =⇒ . + y = x2 dx dx x(ey/x + 2x2 y) dy dy cos x − 2xy 2 1 32. x2 y 2 − sin x = c =⇒ 2x2 y + 2xy 2 − cos x = 0 =⇒ = . Since y(π) = , then dx dx 2x2 y π 2 1 1 + sin x 1 − sin π = c =⇒ c = 1. Hence, x2 y 2 − sin x = 1 so that y 2 = . Since y(π) = , take the π2 2 π x π √ 1 + sin x branch of y where x < 0 so y(x) = . x 33. dy = sin x =⇒ y(x) = − cos x + c for all x ∈ R. dx 34. dy = x−2/3 =⇒ y(x) = 3x1/3 + c for all x = 0. dx 35. dy d2 y = xex =⇒ = xex − ex + c1 =⇒ y(x) = xex − 2ex + c1 x + c2 for all x ∈ R. dx2 dx d2 y = xn , where n is an integer. dx2 dy If n = −1 then = ln |x| + c1 =⇒ y(x) = x ln |x| + c1 x + c2 for all x ∈ (−∞, 0) or x ∈ (0, ∞). dx dy If n = −2 then = −x−1 + c1 =⇒ y(x) = c1 x + c2 − ln |x| for all x ∈ (−∞, 0) or x ∈ (0, ∞). dx dy xn+2 xn+1 If n = −1 and n = −2 then = + c1 =⇒ y = + c1 x + c2 for all x ∈ R. dx n+1 (n + 1)(n + 2) 36. 1 1 1 1 dy = x2 ln x =⇒ y(x) = x3 ln x − x3 + c1 = x3 (3 ln x − 1) + c1 . y(1) = 2 =⇒ 2 = (0 − 1) + c1 =⇒ dx 3 9 9 9 19 19 1 3 1 3 c1 = . Therefore, y(x) = x (3 ln x − 1) + = x (3 ln x − 1) + 19 . 9 9 9 9 37. dy d2 y = cos x =⇒ = sin x + c1 =⇒ y(x) = − cos x + c1 x + c2 . 2 dx dx Thus, y (0) = 1 =⇒ c1 = 1, and y(0) = 2 =⇒ c2 = 3. Thus, y(x) = 3 + x − cos x. 38. d2 y dy d3 y = 6x =⇒ 2 = 3x2 + c1 =⇒ = x3 + c1 x + c2 =⇒ y = 14 x4 + 12 c1 x2 + c2 x + c3 . 3 dx dx dx Thus, y (0) = 4 =⇒ c1 = 4, and y (0) = −1 =⇒ c2 = −1, and y(0) = 1 =⇒ c3 = 1. Thus, y(x) = 1 4 2 4 x + 2x − x + 1. 39. 40. y = xex =⇒ y = xex − ex + c1 =⇒ y = xex − 2ex + c1 x + c2 . (c)2017 Pearson Education. Inc. 15 Thus, y (0) = 4 =⇒ c1 = 5, and y(0) = 3 =⇒ c2 = 5. Thus, y(x) = xex − 2ex + 5x + 5. 41. Starting with y(x) = c1 ex + c2 e−x , we find that y (x) = c1 ex − c2 e−x and y (x) = c1 ex + c2 e−x . Thus, y − y = 0, so y(x) = c1 ex + c2 e−x is a solution to the differential equation on (−∞, ∞). Next we establish that every solution to the differential equation has the form c1 ex + c2 e−x . Suppose that y = f (x) is any solution to the differential equation. Then according to Theorem 1.2.12, y = f (x) is the unique solution to the initial-value problem y − y = 0, y(0) = f (0), y (0) = f (0). However, consider the function y(x) = f (0) + f (0) x f (0) − f (0) −x e + e . 2 2 (0) (0) This is of the form y(x) = c1 ex + c2 e−x , where c1 = f (0)+f and c2 = f (0)−f , and therefore solves the 2 2 differential equation y − y = 0. Furthermore, evaluation this function at x = 0 yields y(0) = f (0) and y (0) = f (0). Consequently, this function solves the initial-value problem above. However, by assumption, y(x) = f (x) solves the same initial-value problem. Owing to the uniqueness of the solution to this initial-value problem, it follows that these two solutions are the same: f (x) = c1 ex + c2 e−x . Consequently, every solution to the differential equation has the form y(x) = c1 ex + c2 e−x , and therefore this is the general solution on any interval I. d2 y dy = e−x =⇒ = −e−x + c1 =⇒ y(x) = e−x + c1 x + c2 . Thus, y(0) = 1 =⇒ c2 = 0, and 2 dx dx y(1) = 0 =⇒ c1 = − 1e . Hence, y(x) = e−x − 1e x. 42. dy d2 y = −6 − 4 ln x =⇒ = −2x − 4x ln x + c1 =⇒ y(x) = −2x2 ln x + c1 x + c2 . Since, y(1) = 0 =⇒ dx2 dx 2e2 2e2 c1 + c2 = 0, and since, y(e) = 0 =⇒ ec1 + c2 = 2e2 . Solving this system yields c1 = , c2 = − . e−1 e−1 2 2e Thus, y(x) = (x − 1) − 2x2 ln x. e−1 43. 44. y(x) = c1 cos x + c2 sin x (a). y(0) = 0 =⇒ 0 = c1 (1) + c2 (0) =⇒ c1 = 0. y(π) = 1 =⇒ 1 = c2 (0), which is impossible. No solutions. (b). y(0) = 0 =⇒ 0 = c1 (1) + c2 (0) =⇒ c1 = 0. y(π) = 0 =⇒ 0 = c2 (0), so c2 can be anything. Infinitely many solutions. 45-50. Use some kind of technology to define each of the given functions. Then use the technology to simplify the expression given on the left-hand side of each differential equation and verify that the result corresponds to the expression on the right-hand side. 51. (a). Use some form of technology to substitute y(x) = a + bx + cx2 + dx3 + ex4 + f x5 where a, b, c, d, e, f are constants, into the given Legendre equation and set the coefficients of each power of x in the resulting equation to zero. The result is: e = 0, 20f + 18d = 0, e + 2c = 0, 3d + 14b = 0, c + 15a = 0. (c)2017 Pearson Education. Inc. 16 9 21 Now solve for the constants to find: a = c = e = 0, d = − 14 3 b, f = − 10 d = 5 b. Consequently the corresponding solution to the Legendre equation is: 14 21 y(x) = bx 1 − x2 + x4 . 3 5 21 15 Imposing the normalization condition y(1) = 1 requires 1 = b(1 − 14 3 + 5 ) =⇒ b = 8 . Consequently the required solution is y(x) = 18 x(15 − 70x2 + 63x4 ). 52. (a). J0 (x) = ∞ (−1)k x 2k 1 4 = 1 − 14 x2 + 64 x + ... 2 2 k=0 (k!) (b). A Maple plot of J(0, x, 4) is given in the accompanying figure. J(0, x, 4) 1 0.8 0.6 Approximation to the first positive zero of J0(x) 0.4 0.2 0 1 2 3 4 x –0.2 Figure 0.0.8: Figure for Problem 52(b) (c). From this graph, an approximation to the first positive zero of J0 (x) is 2.4. Using the Maple internal function BesselJZeros gives the approximation 2.404825558. (c) A Maple plot of the functions J0 (x) and J(0, x, 4) on the interval [0,2] is given in the accompanying figure. We see that to the printer resolution, these graphs are indistinguishable. On a larger interval, for example, [0,3], the two graphs would begin to differ dramatically from one another. (d). By trial and error, we find the smallest value of m to be m = 11. A plot of the functions J(0, x) and J(0, x, 11) is given in the accompanying figure. Solutions to Section 1.3 True-False Review: (a): TRUE. This is precisely the remark after Theorem 1.3.2. (b): FALSE. For instance, the differential equation in Example 1.3.7 has no equilibrium solutions. (c): FALSE. This differential equation has equilibrium solutions y(x) = 2 and y(x) = −2. (d): TRUE. For this differential equation, we have f (x, y) = x2 + y 2 . Therefore, any equation of the form x2 + y 2 = k is an isocline, by definition. (c)2017 Pearson Education. Inc. 17 J0(x), J(0, x, 4) 1 0.8 0.6 0.4 0 0.2 0.4 0.6 0.8 1 1.2 1.4 1.6 1.8 x 2 Figure 0.0.9: Figure for Problem 52(c) J(0, x), J(0, x, 11) 1 0.8 0.6 0.4 0.2 0 2 4 6 8 10 x J(0, x) –0.2 –0.4 J(0, x, 11) Figure 0.0.10: Figure for Problem 52(d) (e): TRUE. Equilibrium solutions are always horizontal lines. These are always parallel to each other. 2 2 +y (f ): TRUE. The isoclines have the form x 2y = k, or x2 +y 2 = 2ky, or x2 +(y −k)2 = k 2 , so the statement is valid. (g): TRUE. An equilibrium solution is a solution, and two solution curves to the differential equation dy dx = f (x, y) do not intersect. Problems: 1. y = ce2x =⇒ c = ye−2x . Hence, 2. y = ecx =⇒ ln y = cx =⇒ c = dy = 2ce2x = 2y. dx ln y y dy , x = 0. Hence, = cecx = ln y, x = 0. x dx x (c)2017 Pearson Education. Inc. 18 3. y = cx2 =⇒ c = y dy 2y y . Hence, = 2cx = 2 2 x = . 2 x dx x x 4. y = cx−1 =⇒ c = xy. Hence, 5. y 2 = cx =⇒ c = dy y = −cx−2 = −(xy)x−2 = − . dx x y2 dy dy c y . Hence, 2y = c, so that, = = . x dx dx 2y 2x x2 + y 2 dy x2 + y 2 dy x2 + y 2 = c. Hence, 2x + 2y = 2c = , so that, y = − x. 2x dx x dx 2x 2 2 dy y −x Consequently, = . dx 2xy 6. x2 + y 2 = 2cx =⇒ x2 + y 2 . Differentiating the given 7. (x − c)2 + (y − c)2 = 2c2 =⇒ x2 − 2cx + y 2 − 2cy = 0 =⇒ c = 2(x + y) 2 2 2 x + y 2 dy dy x +y +2 y− equation yields 2(x − c) + 2(y − c) = 0, so that 2 x − = 0, that is dx 2(x + y) 2(x + y) dx dy x2 + 2xy − y 2 . =− 2 dx y + 2xy − x2 −2y ± 4y 2 + 4x2 dy 2 2 2 2 8. 2cy = x − c =⇒ c + 2cy − x = 0 =⇒ c = = 2x, = −y ± x2 + y 2 . Hence, 2c dx 2 dy x x √ 2 2. so that dx = c = −y± x +y 9. x2 + y 2 = c =⇒ 2x + 2y dy dy x = 0 =⇒ =− . dx dx y y(x) 2.0 1.6 1.2 0.8 0.4 x -2 -1 -0.4 1 2 -0.8 -1.2 -1.6 -2.0 Figure 0.0.11: Figure for Problem 9 dy y 3y = 3cx2 = 3 3 x2 = . The initial condition y(2) = 8 =⇒ 8 = c(2)3 =⇒ c = 1. Thus dx x x the unique solution to the initial value problem is y = x3 . 10. y = cx3 =⇒ (c)2017 Pearson Education. Inc. 19 y(x) 8 (2, 8) 4 x 2 -2 -4 -8 Figure 0.0.12: Figure for Problem 10 dy dy y2 dy = c =⇒ 2y = =⇒ = y2x =⇒ 2x · dy − y · dx = 0. The initial condition dx dx x dx y(1) = 2 =⇒ c = 4, so that the unique solution to the initial value problem is y 2 = 4x. 11. y 2 = cx =⇒ 2y y(x) 3 (1, 2) 1 x -3 -2 1 -1 2 3 -1 -3 Figure 0.0.13: Figure for Problem 11 (c)2017 Pearson Education. Inc. 20 12. (x − c)2 + y 2 = c2 =⇒ x2 − 2cx + c2 + y 2 = c2 , so that x2 − 2cx + y 2 = 0. (0.0.1) dy = 0. dx (0.0.2) Differentiating with respect to x yields 2x − 2c + 2y But from (0.0.1), c = x2 + y 2 which, when substituted into (0.0.2), yields 2x − 2x 2 dy x + y2 + 2y = 0, x dx dy y 2 − x2 = . Imposing the initial condition y(2) = 2 =⇒ from (0.0.1) c = 2, so that the unique dx 2xy solution to the initial value problem is y = + x(4 − x). that is, y(x) 3 (2, 2) 2 1 x 1 2 3 4 5 6 -1 -2 -3 Figure 0.0.14: Figure for Problem 12 13. Let f (x, y) = x sin (x + y), which is continuous for all x, y ∈ R. ∂f = x cos (x + y), which is continuous for all x, y ∈ R. ∂y dy = x sin (x + y), y(x0 ) = y0 has a unique solution for some interval I ∈ R. By Theorem 1.3.2, dx dy x = 2 (y 2 − 9), y(0) = 3. dx x +1 x f (x, y) = 2 (y 2 − 9), which is continuous for all x, y ∈ R. x +1 ∂f 2xy = 2 , which is continuous for all x, y ∈ R. ∂y x +1 So the initial value problem stated above has a unique solution on any interval containing (0, 3). By inspection we see that y(x) = 3 is the unique solution. 14. (c)2017 Pearson Education. Inc. 21 15. The initial-value problem does not necessarily have a unique solution since the hypothesis of the existence ∂f = 12 xy −1/2 and uniqueness theorem are not satisfied at (0,0). This follows since f (x, y) = xy 1/2 , so that ∂y which is not continuous at (0, 0). ∂f = −4xy. Both of these functions are continuous for all (x, y), and ∂y therefore the hypothesis of the uniqueness and existence theorem are satisfied for any (x0 , y0 ). 16. (a). f (x, y) = −2xy 2 =⇒ (b). y(x) = 1 2x = −2xy 2 . =⇒ y = − 2 x2 + c (x + c)2 (c). y(x) = 1 . x2 + c (i). y(0) = 1 =⇒ 1 = 1 1 =⇒ c = 1. Hence, y(x) = 2 . The solution is valid on the interval (−∞, ∞). c x +1 y(x) 1.2 0.8 0.4 x -2 2 Figure 0.0.15: Figure for Problem 16c(i) 1 1 =⇒ c = 0. Hence, y(x) = 2 . This solution is valid on the interval (0, ∞). 1+c x 1 1 (iii). y(0) = −1 =⇒ −1 = =⇒ c = −1. Hence, y(x) = 2 . This solution is valid on the interval c x −1 (−1, 1). (ii). y(1) = 1 =⇒ 1 = (d). Since, by inspection, y(x) = 0 satisfies the given initial-value problem, it must be the unique solution to the initial-value problem. ∂f = 2y − 1 are continuous at all points (x, y). Consequently, the ∂y hypothesis of the existence and uniqueness theorem are satisfied by the given initial-value problem for any x0 , y0 . 17. (a). Both f (x, y) = y(y − 1) and (b). Equilibrium solutions: y(x) = 0, y(x) = 1. dy d2 y = (2y − 1) = (2y − 1)y(y − 1). Hence the dx2 dx 1 solution curves are concave up for 0 < y < 2 , and y > 1, and concave down for y < 0, and 12 < y < 1. (c). Differentiating the given differential equation yields (d). The solutions will be bounded provided 0 ≤ y0 ≤ 1. 18. (a). Equilibrium solutions: y(x) = −2, y(x) = 1. (c)2017 Pearson Education. Inc. 22 y(x) 8 6 4 2 x 1 2 3 4 5 6 Figure 0.0.16: Figure for Problem 16c(ii) y(x) -1.0 -0.5 0.5 1.0 x -1 -2 -3 Figure 0.0.17: Figure for Problem 16c(iii) (b). dy = (y + 2)(y − 1) =⇒ the solutions are increasing when y < −2 and y > 1, and the solutions are dx (c)2017 Pearson Education. Inc. 23 y(x) 2 1 –2 –1 0 1 2 x –1 –2 Figure 0.0.18: Figure for Problem 17(d) decreasing when −2 < y < 1. dy d2 y = (2y + 1) = (2y + 1)(y + 2)(y − 1). Hence 2 dx dx 1 the solution curves are concave up for −2 < y < − 2 , and y > 1, and concave down for y < −2, and − 12 < y < 1. (c). Differentiating the given differential equation yields 19. (a). Equilibrium solution: y(x) = 2. dy = (y − 2)2 =⇒ the solutions are increasing when y < 2 and y > 2. dx d2 y dy (c). Differentiating the given differential equation yields = 2(y − 2) = 2(y − 2)3 . Hence the solution 2 dx dx curves are concave up for y > 2, and concave down for y < 2. (b). 20. (a). Equilibrium solutions: y(x) = 0, y(x) = 1. dy (b). = y 2 (y − 1) =⇒ the solutions are increasing when y < 1, and the solutions are decreasing when dx y < 1. d2 y dy = (3y 2 − 2y) = y 3 (3y − 2)(y − 1). Hence the 2 dx dx solution curves are concave up for 0 < y < − 23 , and y > 1, and concave down for y < 0, and 2/3 < y < 1. (c). Differentiating the given differential equation yields 21. (a). Equilibrium solutions: y(x) = 0, y(x) = 1, y(x) = −1. dy (b). = (y + 2)(y − 1) =⇒ the solutions are increasing when −1 < y < 0 and y > 1, and the solutions are dx decreasing when y < −1, and 0 < y < 1. dy d2 y = (3y 2 − 1) = (3y 2 − 1)y(y − 1)(y + 1). dx2 dx Hence the solution curves are concave up for −1 < y < − √13 , and 0 < y < √13 , and y > 1, and concave down for y < −1, and − √13 < y < 0, and √13 < y < 1. (c). Differentiating the given differential equation yields 22. y = 4x. There are no equilibrium solutions. The slope of the solution curves is positive for x > 0 and is negative for x < 0. The isoclines are the lines x = k4 . (c)2017 Pearson Education. Inc. 24 Slope of Solution Curve -4 -2 0 2 4 Equation of Isocline x = −1 x = −1/2 x=0 x = 1/2 x=1 y(x) 1.5 1 0.5 –1.5 –1 –0.5 0 0.5 1 1.5 x –0.5 –1 –1.5 Figure 0.0.19: Figure for Problem 22 23. y = x1 . There are no equilibrium solutions. The slope of the solution curves is positive for x > 0 and increases without bound as x → 0+ . The slope of the curve is negative for x < 0 and decreases without bound as x → 0− . The isoclines are the lines x1 = k. Slope of Solution Curve ±4 ±2 ±1/2 ±1/4 ±1/10 Equation of Isocline x = ±1/4 x = ±1/2 x = ±2 x = ±4 x = ±10 24. y = x + y. There are no equilibrium solutions. The slope of the solution curves is positive for y > −x, and negative for y < −x. The isoclines are the lines y + x = k. Slope of Solution Curve −2 −1 0 1 2 Equation of Isocline y = −x − 2 y = −x − 1 y = −x y = −x + 1 y = −x + 2 Since the slope of the solution curve along the isocline y = −x − 1 coincides with the slope of the isocline, it follows that y = −x − 1 is a solution to the differential equation. Differentiating the given differential (c)2017 Pearson Education. Inc. 25 y(x) 2 1 –2 0 –1 1 2 x –1 –2 Figure 0.0.20: Figure for Problem 23 equation yields: y = 1 + y = 1 + x + y. Hence the solution curves are concave up for y > −x − 1, and concave down for y < −x − 1. Putting this information together leads to the slope field in the accompanying figure. y(x) 3 2 1 -3 -2 1 -1 2 3 x -1 -2 -3 Figure 0.0.21: Figure for Problem 24 25. y = xy . There are no equilibrium solutions. The slope of the solution curves is zero when x = 0. The solution has a vertical tangent line at all points along the x-axis (except the origin). Differentiating the 1 1 x2 1 x differential equation yields: y = − 2 y = − 3 = 3 (y 2 − x2 ). Hence the solution curves are concave y y y y y up for y > 0 and y 2 > x2 ; y < 0 and y 2 < x2 and concave down for y > 0 and y 2 < x2 ; y < 0 and y 2 > x2 . The isoclines are the lines xy = k. (c)2017 Pearson Education. Inc. 26 Slope of Solution Curve ±2 ±1 ±1/2 ±1/4 ±1/10 Equation of Isocline y = ±x/2 y = ±x y = ±2x y = ±4x y = ±10x Note that y = ±x are solutions to the differential equation. y(x) 2 1 –2 –1 0 1 2 x –1 –2 Figure 0.0.22: Figure for Problem 25 26. y = − 4x y . Slope is zero when x = 0 (y = 0). The solutions have a vertical tangent line at all points along the x-axis(except the origin). The isoclines are the lines − 4x y = k. Some values are given in the table below. Slope of Solution Curve ±1 ±2 ±3 Equation of Isocline y = ±4x y = ±2x y = ±4x/3 4 4 4(y 2 + 4x2 ) 4xy 16x2 + 2 = − − 3 = − . y y y y y Consequently the solution curves are concave up for y < 0, and concave down for y > 0. Putting this information together leads to the slope field in the accompanying figure. Differentiating the given differential equation yields: y = − 27. y = x2 y. Equilibrium solution: y(x) = 0 =⇒ no solution curve can cross the x-axis. Slope: zero when x = 0 or y = 0. Positive when y > 0 (x = 0), negative when y < 0 (x = 0). Differentiating the given d2 y dy differential equation yields: = 2xy +x2 = 2xy +x4 y = xy(2+x3 ). So, when y > 0, the solution curves dx2 dx are concave up for x ∈ (−∞, (−2)1/3 ), and for x > 0, and are concave down for x ∈ ((−2)1/3 , 0). When y < 0, the solution curves are concave up for x ∈ ((−2)1/3 , 0), and concave down for x ∈ (−∞, (−2)1/3 ) and for x > 0. The isoclines are the hyperbolas x2 y = k. (c)2017 Pearson Education. Inc. 27 y(x) 4 3 2 1 x -2 -1 1 2 Figure 0.0.23: Figure for Problem 26 Slope of Solution Curve ±2 ±1 ±1/2 ±1/4 ±1/10 0 Equation of Isocline y = ±2/x2 y = ±1/x2 y = ±1/(2x)2 y = ±1/(4x)2 y = ±1/(10x)2 y=0 28. y = x2 cos y. The slope is zero when x = 0. There are equilibrium solutions when y = (2k + 1) π2 . The slope field is best sketched using technology. The accompanying figure gives the slope field for − π2 < y < 3π 2 . 29. y = x2 + y 2 . The slope of the solution curves is zero at the origin, and positive at all the other points. There are no equilibrium solutions. The isoclines are the circles x2 + y 2 = k. Slope of Solution Curve 1 2 3 4 5 30. dT dt Equation of Isocline x = ±1/4 x = ±1/2 x = ±2 x = ±4 x = ±10 1 = − 80 (T − 70). Equilibrium solution: T (t) = 70. The slope of the solution curves is positive for (c)2017 Pearson Education. Inc. 28 y(x) 2 1 –2 0 –1 1 x 2 –1 –2 Figure 0.0.24: Figure for Problem 27 y(x) 5 4 3 2 1 –3 –2 –1 0 1 2 3 x –1 Figure 0.0.25: Figure for Problem 28 d2 T 1 dT 1 =− = (T − 70). Hence the solution curves are concave dt2 80 dt 6400 1 (T − 70) = k. up for T > 70, and concave down for T < 70. The isoclines are the horizontal lines − 80 T > 70, and negative for T < 70. Slope of Solution Curve −1/4 1/5 0 1/5 1/4 Equation of Isocline T = 90 T = 86 T = 70 T = 54 T = 50 31. y = −2xy. 32. y = x sin x . 1 + y2 (c)2017 Pearson Education. Inc. 29 y(x) 2 1 –2 0 –1 1 2 x –1 –2 Figure 0.0.26: Figure for Problem 29 T(t) 100 80 60 40 20 0 20 40 60 80 100 120 140 t Figure 0.0.27: Figure for Problem 30 33. y = 3x − y. 34. y = 2x2 sin y. 35. y = 2 + y2 . 3 + 0.5x2 36. y = 1 − y2 . 2 + 0.5x2 37. (a). Slope field for the differential equation y = x−1 (3 sin x − y). (b). Slope field with solution curves included. The figure suggests that the solution to the differential equation are unbounded as x → 0+ . (c). Slope field with solution curve corresponding to the initial condition y( π2 ) = π6 . (c)2017 Pearson Education. Inc. 30 y(x) 2 1 –2 –1 0 1 2 x –1 –2 Figure 0.0.28: Figure for Problem 31 y(x) 2 1 –2 –1 0 1 2 x –1 –2 Figure 0.0.29: Figure for Problem 32 y(x) 2 1 –2 –1 0 1 2 x –1 –2 Figure 0.0.30: Figure for Problem 33 This solution curve is bounded as x → 0+ . (c)2017 Pearson Education. Inc. 31 y(x) 3 2 1 –2 –1 0 1 2 x –1 –2 –3 Figure 0.0.31: Figure for Problem 34 y(x) 2 1 –2 –1 0 1 2 x –1 –2 Figure 0.0.32: Figure for Problem 35 y(x) 2 1 –2 –1 0 1 2 x –1 –2 Figure 0.0.33: Figure for Problem 36 (d). In the accompanying figure we have sketched several solution curves on the interval (0,15]. (c)2017 Pearson Education. Inc. 32 y(x) 4 2 0 2 4 6 8 10 x –2 –4 Figure 0.0.34: Figure for Problem 37(a) y(x) 4 2 0 2 4 6 8 10 x –2 –4 Figure 0.0.35: Figure for Problem 37(b) y(x) 4 2 0 2 4 6 8 10 x –2 –4 Figure 0.0.36: Figure for Problem 37(c) The figure suggests that the solution curves approach the x-axis as x → ∞. (c)2017 Pearson Education. Inc. 33 y(x) 4 2 0 2 4 6 8 10 12 14 x –2 –4 Figure 0.0.37: Figure for Problem 37(d) dy y 38. (a). Differentiating the given equation gives = 2kx = 2 . Hence the differential equation of the dx x x dy =− . orthogonal trajectories is dx 2y y(x) 4 2 –4 –2 0 2 4 x –2 –4 Figure 0.0.38: Figure for Problem 38(a) (b). The orthogonal trajectories appear to be ellipses. This can be verified by integrating the differential equation derived in (a). 39. If a > 0, then as illustrated in the following slope field (a = 0.5, b = 1), it appears that limt→∞ i(t) = ab . If a < 0, then as illustrated in the following slope field (a = −0.5, b = 1) it appears that i(t) diverges as t → ∞. If a = 0 and b = 0, then once more i(t) diverges as t → ∞. The accompanying figure shows a representative case when b > 0. Here we see that limt→∞ i(t) = +∞. If b < 0, then limt→∞ i(t) = −∞. If a = b = 0, then the general solution to the differential equation is i(t) = i0 where i0 is a constant. (c)2017 Pearson Education. Inc. 34 y(x) 4 2 –3 –2 –1 0 1 2 3 x –2 –4 Figure 0.0.39: Figure for Problem 39 when a > 0 i(t) 4 2 –3 –2 –1 0 1 2 3 t –2 –4 Figure 0.0.40: Figure for Problem 39 when a < 0 y(x) 4 2 –3 –2 –1 0 1 2 3 x –2 –4 Figure 0.0.41: Figure for Problem 39 when a = 0 Solutions to Section 1.4 (c)2017 Pearson Education. Inc. 35 True-False Review: dy 1 dy (a): TRUE. The differential equation dx = f (x)g(y) can be written g(y) dx = f (x), which is the proper form, according to Definition 1.4.1, for a separable differential equation. (b): TRUE. A separable differential equation is a first-order differential equation, so the general solution contains one constant. The value of that constant can be determined from an initial condition, as usual. (c): TRUE. Newton’s Law of Cooling is usually expressed as dT dt = −k(T − Tm ), and this can be rewritten as 1 dT = −k, T − Tm dt and this form shows that the equation is separable. (d): FALSE. The expression x2 + y 2 cannot be separated in the form f (x)g(y), so the equation is not separable. (e): FALSE. The expression x sin(xy) cannot be separated in the form f (x)g(y), so the equation is not separable. dy (f ): TRUE. We can write the given equation as e−y dx = ex , which is the proper form for a separable equation. dy (g): TRUE. We can write the given equation as (1 + y 2 ) dx = x12 , which is the proper form for a separable equation. (h): FALSE. The expression x+4y 4x+y cannot be separated in the form f (x)g(y), so the equation is not separable. 3 2 2 y dy (i): TRUE. We can write x xy+x = xy, so we can write the given differential equation as y1 dx = x, which 2 +xy is the proper form for a separable equation. Problems: 1. Separating the variables and integrating yields 2 dy = 2 xdx =⇒ ln |y| = x2 + c1 =⇒ y(x) = cex . y 2. Separating the variables and integrating yields dx 1 −2 y dy = =⇒ y(x) = − . x2 + 1 tan−1 x + c 3. Separating the variables and integrating yields ey dy = e−x dx = 0 =⇒ ey + e−x = c =⇒ y(x) = ln (c − e−x ). 4. Separating the variables and integrating yields dy (ln x)−1 = dx =⇒ y(x) = c ln x. y x (c)2017 Pearson Education. Inc. 36 5. Separating the variables and integrating yields dx dy = =⇒ ln |x − 2| − ln |y| = c1 =⇒ y(x) = c(x − 2). x−2 y 6. Separating the variables and integrating yields dy 2x = dx =⇒ ln |y − 1| = ln |x2 + 3| + c1 =⇒ y(x) = c(x2 + 3) + 1. 2 y−1 x +3 7. y − x dy dy dy = 3 − 2x2 =⇒ x(2x − 1) = (3 − y). Separating the variables and integrating yields dx dx dx dx dx 2 dy = =⇒ − ln |y − 3| = − + dx − y−3 x(2x − 1) x 2x − 1 =⇒ − ln |y − 3| = − ln |x| + ln |2x − 1| + c1 x cx − 3 =⇒ = c2 =⇒ y(x) = . (y − 3)(2x − 1) 2x − 1 sin y cos x dy cos (x − y) dy cos x cos y 8. = −1 =⇒ = =⇒ dy = dx =⇒ − ln | cos y| = ln | sin x|+c1 =⇒ dx sin x sin y dx sin x sin y cos y cos y cos y = c csc x. dy x(y 2 − 1) dy 1 xdx = =⇒ = , y = ±1. Thus, dx 2(x − 2)(x − 1) (y + 1)(y − 1) 2 (x − 2)(x − 1) dy dy dx 1 1 dx 1 2 =⇒ − ln |y + 1|+ln |y − 1| = 2 ln |x − 2|−ln |x − 1|+c1 + = − − 2 y+1 2 y−1 2 x−2 x−1 9. y−1 (x − 2)2 (x − 1) + c(x − 2)2 . By inspection we see that y(x) = 1, and y(x) = −1 =c =⇒ y(x) = y+1 x−1 (x − 1) − c(x − 2)2 are solutions of the given differential equation. The former is included in the above solution when c = 0. dy dy 16 x2 y − 32 x2 dx =⇒ ln |y − 2| = 10. + 2 =⇒ dx =⇒ ln |y − 2| = − 1+ 2 = = dx 16 − x2 y−2 16 − x2 x − 16 dx dx dx =⇒ ln |y − 2| = −x + 2 ln |x + 4| − −x − 16 =⇒ ln |y − 2| = −x − 16 − 18 + 18 2 x − 16 x+4 x−4 2 x+4 2 ln |x − 4| + c1 =⇒ y(x) = 2 + c e−x . x−4 dy dy dy dx 1 1 1 dx =⇒ 11. (x−a)(x−b) −(y−c) = 0 =⇒ = =⇒ = − y − c (x − a)(x − b) y−c a−b x−a x−b dx 1/(a−b) 1/(a−b) 1/(a−b) x − a x−b x−a (y − c) = c =⇒ ln |y − c| = ln c1 =⇒ y − c = c =⇒ 1 2 x − b x−a x−b 1/(a−b) x−a . y(x) = c + c2 x−b =⇒ (c)2017 Pearson Education. Inc. 37 dy dx dy π =− =⇒ tan−1 y = tan−1 x + c, but y(0) = 1 so c = . + y 2 = −1 =⇒ 2 2 dx 1+y 1+x 4 π 1−x −1 −1 Thus, tan y = tan x + or y(x) = . 4 1+x 12. (x2 + 1) dy dy 2x dx =⇒ − ln |a − y| = − 12 ln |1 − x2 | + c1 =⇒ y(x) = 13. (1 − x2 ) + xy = ax =⇒ = − 12 − dx a − y 1 − x2 √ √ a + c 1 − x2 , but y(0) = 2a so c = a and therefore, y(x) = a(1 + 1 − x2 ). sin y sin x dy sin (x + y) dy 14. = 1− =⇒ = − tan x cot y =⇒ − dy = dx =⇒ − ln | cos x cos y| = c, but dx sin x sin y dx cos y cos x π π y( ) = so c = ln (2). Hence, − ln | cos x cos y| = ln (2) =⇒ y(x) = cos−1 12 sec x . 4 4 dy dy 1 = sin xdx for y = 0. Thus − 2 = − cos x+c. However, we cannot impose the = y 3 sin x =⇒ dx y3 2y initial condition y(0) = 0 on the last equation since it is not defined at y = 0. But, by inspection, y(x) = 0 is a solution to the given differential equation and further, y(0) = 0; thus, the unique solution to the initial value problem is y(x) = 0. 15. dy dy = 23 dx if y = 1 =⇒ 2(y − 1)1/2 = 23 x + c but y(1) = 1 so = 23 (y − 1)1/2 =⇒ 1/2 dx (y − 1) √ √ c = − 23 =⇒ 2 y − 1 = 23 x − 23 =⇒ y − 1 = 13 (x − 1). This does not contradict the Existence-Uniqueness theorem because the hypothesis of the theorem is not satisfied when x = 1. 16. dv m = mg − kv 2 =⇒ dv = dt. If we let a = mg k then the preceding equation can 2 dt k [(mg/k) − v ] m 1 be written as dv = dt which can be integrated directly to obtain k a2 − v 2 m a+v = t + c, ln 2ak a−v 17. (a). m that is, upon exponentiating both sides, 2ak a+v = c1 e m t . a−v Imposing the initial condition v(0) = 0, yields c = 0 so that 2ak a+v = e m t. a−v Therefore, v(t) = a 2akt e m −1 2akt e m +1 which can be written in the equivalent form v(t) = a tanh gt . a (b). No. As t → ∞, v → a and as t → 0+ , v → 0. (c)2017 Pearson Education. Inc. 38 (c). v(t) = a tanh gt y(0) = 0 then y(t) = a =⇒ dy a2 =⇒ a tanh gt dt =⇒ y(t) = = a tanh gt ln(cosh ( gt a a a )) + c1 and if dt g a2 ln cosh ( gt a) . g dy x = − , y(0) = 12 . Separating dx 4y the variables in the differential equation yields 4y −1 dy = −1dx, which can be integrated directly to obtain x2 2y 2 = − + c. Imposing the initial condition we obtain c = 12 , so that the solution curve has the equation 2 2y 2 = −x2 + 12 , or equivalently, 4y 2 + 2x2 = 1. 18. The required curve is the solution curve to the initial-value problem dy 19. The required curve is the solution curve to the initial-value problem = ex−y , y(3) = 1. Separating dx the variables in the differential equation yields ey dy = ex dx, which can be integrated directly to obtain ey = ex + c. Imposing the initial condition we obtain c = e − e3 , so that the solution curve has the equation ey = ex + e − e3 , or equivalently, y = ln(ex + e − e3 ). dy = x2 y 2 , y(−1) = 1. Separating dx the variables in the differential equation yields y12 dy = x2 dx, which can be integrated directly to obtain − y1 = 13 x3 + c. Imposing the initial condition we obtain c = − 23 , so that the solution curve has the equation 3 y = − 1 x31− 2 , or equivalently, y = 2−x 3. 20. The required curve is the solution curve to the initial-value problem 3 3 1 dv = −dt. Integrating we 1 + v2 −1 −1 obtain tan (v) = −t + c. The initial condition v(0) = v0 implies that c = tan (v0 ), so that tan−1 (v) = −t + tan−1 (v0 ). The object will come to rest if there is time t, at which the velocity is zero. To determine tr , we set v = 0 in the previous equation which yields tan−1 (0) = tr +tan−1 (v0 ). Consequently, tr = tan−1 (v0 ). dv The object does not remain at rest since we see from the given differential equation that < 0 at t = tr , dt and so v is decreasing with time. Consequently v passes through zero and becomes negative for t < tr . 21. (a). Separating the variables in the given differential equation yields dv dx dv dv (b). From the chain rule we have = . Then = v . Substituting this result into the differential dt dt dx dx v dv dv = −dx. Integrating we = −(1 + v 2 ). We now separate the variables: equation (1.4.22) yields v dx 1 + v2 2 obtain ln (1 + v ) = −2x + c. Imposing the initial condition v(0) = v0 , x(0) = 0 implies that c = ln (1 + v02 ), so that ln (1 + v 2 ) = −2x + ln (1 + v02 ). When the object comes to rest the distance travelled by the object is x = 12 ln (1 + v02 ). dv = −kv n =⇒ v −n dv = −kdt. dt 1 1 n = 1 :=⇒ v 1−n = −kt + c. Imposing the initial condition v(0) + v0 yields c = v 1−n , so that 1−n 1−n 0 v = [v01−n + (n − 1)kt]1/(1−n) . The object comes to rest in a finite time if there is a positive value of t for which v = 0. n = 1 :=⇒ Integratingv −n dv = −kdt and imposing the initial conditions yields v = v0 e−kt , and the object does not come to rest in a finite amount of time. dx (b). If n = 1, 2, then = [v01−n + (n − 1)kt]1/(1−n) , where x(t) denotes the distanced travelled by the dt 22. (a). (c)2017 Pearson Education. Inc. 39 1 [v 1−n + (n − 1)kt](2−n)/(1−n) + c. Imposing the initial condition k(2 − n) 0 1 1 1 x(0) = 0 yields c = v 2−n , so that x(t) = − [v 1−n + n(n − 1)kt](2−n)/(1−n) + v 2−n . k(2 − n) 0 k(2 − n) o k(2 − n) 0 2−n 1 For 1 < n < 2, we have < 0, so that limt→∞ x(t) = . Hence the maximum distance that the 1−n k(2 − n) 1 object can travel in a finite time is less than . k(2 − n) v0 If n = 1, then we can integrate to obtain x(t) = (1 − e−kt ), where we have imposed the initial condition k v0 x(0) = 0. Consequently, limt→∞ x(t) = . Thus in this case the maximum distance that the object can k v0 travel in a finite time is less than . k object. Consequently, x(t) = − (c). If n > 2, then x(t) = − 1 1 [v 1−n + n(n − 1)kt](2−n)/(1−n) + v 2−n is still valid. However, k(2 − n) o k(2 − n) 0 2−n > 0, and so limt→∞ x(t) = +∞. Consequently, there is no limit to the distance that the 1−n object can travel. dx = (v0−1 + kt)−1 , which can If n = 2, then we return to v = [v01−n + (n − 1)kt]1/(1−n) . In this case dt 1 be integrated directly to obtain x(t) = ln (1 + v0 kt), where we have imposed the initial condition that k x(0) = 0. Once more we see that limt→∞ x(t) = +∞, so that there is no limit to the distance that the object can travel. in this case ρ 1/γ p ) . Consequently the given differential equation can be written as dp = −gρ0 ( )1/γ dy, ρ0 p0 gρ0 γp(γ−1)/γ gρ0 y or equivalently, p−1/γ dp = − 1/γ dy. This can be integrated directly to obtain = − 1/γ + c. At γ−1 p0 p0 the center of the Earth we have p = p0 . Imposing this initial condition on the preceding solution gives (γ−1)/γ γp0 . Substituting this value of c into the general solution to the differential equation we find, c = γ−1 (γ−1)/γ (γ − 1)ρ0 gy (γ − 1)ρ0 gy (γ−1)/γ (γ−1)/γ 1− , so that p = p0 1 − = p0 . after some simplification, p γp0 γp0 23. Solving p = p0 ( dT dT dT = −k(T − Tm ) =⇒ = −k(T − 75) =⇒ = −kdt =⇒ ln |T − 75| = −kt + c1 =⇒ T (t) = dt dt T − 75 −kt −kt 75 + ce . T (0) = 135 =⇒ c = 60 so T = 75 + 60e . T (1) = 95 =⇒ 95 = 75 + 60e−k =⇒ k = ln 3 =⇒ T (t) = 75 + 60e−t ln 3 . Now if T (t) = 615 then 615 = 75 + 60−t ln 3 =⇒ t = −2h. Thus the object was placed in the room at 2p.m. 24. dT = −k(T − 450) =⇒ T (t) = 450 + Ce−kt .T (0) = 50 =⇒ C = −400 so T (t) = 450 − 400e−kt and dt 1 T (20) = 150 =⇒ k = ln 4 ; hence, T (t) = 450 − 400( 34 )t/20 . 20 3 (i) T (40) = 450 − 400( 34 )2 = 225◦ F. 20 ln 4 (ii) T (t) = 350 = 450 − 400( 34 )t/20 =⇒ ( 34 )t/20 = 14 =⇒ t = ≈ 96.4 minutes. ln(4/3) 25. (c)2017 Pearson Education. Inc. 40 dT dT = −k(T − 34) =⇒ = −kdt =⇒ T (t) = 34 + ce−kt . T (0) = 38 =⇒ c = 4 so that dt T − 34 T (t) = 34 + 4e−kt .T (1) = 36 =⇒ k = ln 2; hence, T (t) = 34 + 4e−t ln 2 . Now T (t) = 98 =⇒ T (t) = 34 + 4e−kt = 98 =⇒ 2−t = 16 =⇒ t = −4h. Thus T (−4) = 98 and Holmes was right, the time of death was 10 a.m. 26. 27. T (t) = 75 + ce−kt . T (10) = 415 =⇒ 75 + ce−10k = 415 =⇒ 340 = ce−10k and T (20) = 347 =⇒ 1 75 + ce−20k = 347 =⇒ 272 = ce−20k . Solving these two equations yields k = 10 ln 54 and c = 425; hence, 4 t/10 T = 75 + 425( 5 ) (a) Furnace temperature: T (0) = 500◦ F. 10 ln 17 ≈ 126.96 minutes. Thus the temperature of (b) If T (t) = 100 then 100 = 75 + 425( 45 )t/10 =⇒ t = ln 54 the coal was 100◦ F at 6:07 p.m. dT dT dT = −k(T − 72) =⇒ = −kdt =⇒ T (t) = 72 + ce−kt . Since = −20, −k(T − 72) = −20 or dt T − 72 dt 10 −10/39 10/39 =⇒ c = 78e ; consequently, T (t) = 72 + 78e10(1−t)/39 . k = 39 . Since T (1) = 150 =⇒ 150 = 72 + ce 28. (i). Initial temperature of the object: t = 0 =⇒ T (t) = 72 + 78e10/30 ≈ 173◦ F (ii). Rate of change of the temperature after 10 minutes: T (10) = 72 + 78e−30/13 so after 10 minutes, dT dT 10 260 −30/13 ≈ 2◦ F per minute. = − (72 + 78e−30/13 − 72) =⇒ =− e dt 39 dt 13 29. Substituting a = 0.5, M = 2000 g, and m0 = 4 g into the initial-value problem (1.4.17) yields m 1/4 dm 3/4 1− , m(0) = 4. = 0.5m dt 2000 Separating the variables in the preceding differential equation gives m3/4 1 − so that m3/4 1 − 1 m 1/4 dm = 0.5 dt 2000 1 m 1/4 dm = 0.5t + c. 2000 To evaluate the integral on the left-hand-side of the preceding equation, we make the change of variable m 1/4 1 1 m −3/4 w= , dw = · dm 2000 4 2000 2000 and simplify to obtain 4 · (2000) 1/4 1 dw = 0.5t + c 1−w which can be integrated directly to obtain −4 · (2000) 1/4 ln(1 − w) = 0.5t + c. Exponentiating both sides of the preceding equation, and solving for w yields 1/4 w = 1 − c1 e−0.125t/(2000) (c)2017 Pearson Education. Inc. 41 or equivalently, m 1/4 2000 Consequently, 1/4 = 1 − c1 e−0.125t/(2000) . 1/4 4 m(t) = 2000 1 − c1 e−0.125t/(2000) . (0.0.3) Imposing the initial condition m(0) = 4 yields 4 = 2000 (1 − c1 ) so that c1 = 1 − 1 500 4 1/4 ≈ 0.7885. Inserting this expression for c1 into Equation (0.0.3) gives 1/4 4 m(t) = 2000 1 − 0.7885e−0.125t/(2000) . Consequently, 1/4 4 m(100) = 2000 1 − 0.7885e−12.5/(2000) ≈ 1190.5 g. 30. Substituting a = 0.10, M = 0.15 g, and m0 = 0.008 g into the initial-value problem (1.4.17) yields m 1/4 dm 3/4 1− , m(0) = 0.008. = 0.1m dt 0.15 Separating the variables in the preceding differential equation gives m3/4 1 − so that m3/4 1 − 1 m 1/4 dm = 0.1 dt 0.15 1 m 1/4 dm = 0.1t + c. 0.15 To evaluate the integral on the left-hand-side of the preceding equation, we make the change of variable w= m 1/4 0.15 , dw = and simplify to obtain 4 · (0.15) 1/4 1 1 m −3/4 dm · 4 0.15 0.15 1 dw = 0.1t + c 1−w which can be integrated directly to obtain −4 · (0.15) 1/4 ln(1 − w) = 0.1t + c. (c)2017 Pearson Education. Inc. 42 Exponentiating both sides of the preceding equation, and solving for w yields w = 1 − c1 e−0.025t/(0.15) or equivalently, m 1/4 0.15 Consequently, 1/4 = 1 − c1 e−0.025t/(0.15) 1/4 . 1/4 4 m(t) = 0.15 1 − c1 e−0.025t/(0.15) . (0.0.4) Imposing the initial condition m(0) = 0.008 yields 0.008 = 0.15 (1 − c1 ) so that c1 = 1 − 4 75 4 1/4 ≈ 0.5194. Inserting this expression for c1 into Equation (0.0.4) gives 1/4 4 . m(t) = 0.15 1 − 0.5194e−0.025t/(0.15) Consequently, 1/4 4 m(30) = 0.15 1 − 0.5194e−0.75/(0.15) ≈ 0.076 g. The guppy will have reached 90% of its fully grown mass at time t where 1/4 4 . 0.9 · 0.15 = 0.15 1 − 0.5194e−0.025t/(0.15) Solving algebraically for t yields t=− (0.15)1/4 1 − (0.9)1/4 ≈ 74.5 days. ln 0.025 0.5194 31. Since the chemicals A and B combine in the ratio 2:1, the amounts of A and B that are unconverted at time t are (20 − 23 Q) grams and (20 − 13 Q) grams, respectively. Thus, according to the law of mass action, the differential equation governing the behavior of Q(t) is dQ 2 1 dQ 1 = k1 (20 − Q)(20 − Q) =⇒ = k(30 − Q)(60 − Q) =⇒ dQ = k dt dt 3 3 dt (30 − Q)(60 − Q) 1 dQ 60 − Q 1 60 − Q 1 = kt + c =⇒ − = kt + c =⇒ ln = c1 e30kt . =⇒ 30 (30 − Q) 60 − Q) dt 30 − Q 30 − Q Imposing the initial condition Q(0) = 0 yields c1 = 2. Further, Q(10) = 15 =⇒ 45/15 = 2e300k , so that 1 k= ln(3/2). Therefore, 300 t/10 60 (3/2)t/10 − 1 t 3 60 − Q =⇒ Q(t) = . = 2e 10 ln(3/2) = 2 30 − Q 2 2(3/2)t/10 − 1 (c)2017 Pearson Education. Inc. 43 Therefore, Q(20) = 60 (3/2)2 − 1 150 150 = . Hence, grams of C are produced in 20 minutes. 2(3/2)2 − 1 7 7 32. Since the chemicals A and B combine in the ratio 2:3, the amounts of A and B that are unconverted at time t are (10 − 25 Q) grams and (15 − 35 Q) grams, respectively. Thus, according to the law of mass action, the differential equation governing the behavior of Q(t) is dQ 2 1 1 3 dQ dQ = k dt =⇒ = k1 (10 − Q)(15 − Q) =⇒ = k(25 − Q)2 =⇒ = kt + c. 2 dt 5 5 dt (25 − Q) 25 − Q Hence, Q(t) = 25 − 1 . Imposing the initial condition Q(0) = 0 yields c = 1/25, so that kt + c 625kt 1 = Q(t) = 25 1 − . 25kt + 1 25kt + 1 50t . Therefore, Q(30) = 1500/75 = 20 grams. Hence, 20 2t + 15 grams of C are produced in 30 minutes. The reaction will be 50% complete when Q(t) = 12.5 this will occur 50t =⇒ t ≈ 7.5 minutes. after t minutes where 12.5 = 2t + 15 Q(5) = 10 =⇒ k = 2/375, so that Q(t) = 33. Since A and B combine in the ratio 3:5 to produce C. Therefore, producing 30 g of C will require 150 5 · 30 = g of A. 8 8 34. (a).Since the chemicals A and B combine in the ratio a : b to produce chemical C, when Q grams of C b a Q grams of A and grams of B. Consequently, the amounts of A and are produced , the consist of a+b a+b a b Q grams and B0 − Q grams, respectively. Therefore, B that are unconverted at time t are A0 − a+b a+b according to the law of mass action, the chemical reaction is governed by the differential equation a b dQ = k A0 − Q B0 − Q . dt a+b a+b a+b a b a+b dQ abk − Q − Q , which can be = k A0 − Q B0 − Q = A B 0 0 dt a+b a+b (a + b)2 a b a+b a+b dQ abk ,α = written as = r(α − Q)(β − Q), where r = A0 , β = B0 . dt (a + b)2 a b 1 dQ 1 1 1 dQ = 35. = r(α − Q)(β − Q) =⇒ dQ = r dt + c =⇒ − dt (α − Q)(β − Q) α−β β−Q α−Q α−Q 1 α − βcer(α−β)t α−Q = rt + c =⇒ rt + c. Hence, . Imposing the ln = cer(α−β)t =⇒ Q(t) = α−β β−Q β−Q 1 − cer(α−β)t r(α−β)t αβ 1 − e initial condition Q(0) = 0 yields c = α/β, so that Q(t) = . When α > β, limt→∞ Q(t) = β − αer(α−β)t αβ = β. α (b). dQ dQ = r(Q − α)2 =⇒ (α − Q)−2 = r =⇒ −(α − Q)−1 = 36. When α = β, Equation (1.4.24) reduces to dt dt rt + c. (c)2017 Pearson Education. Inc. 44 1 α2 rt . Imposing the initial condition Q(0) = 0 yields c = 1/α so that Q(t) = . rt + c αrt + 1 Therefore, limt→∞ Q(t) = α. Hence, Q(t) = α − 37. Separating the variables in the given differential equation yields 1 dQ = kdt (α − Q)(β − Q)(γ − Q) so that 1 1 1 dQ = kt + c, + + (β − α)(γ − α)(α − Q) (α − β)(γ − β)(β − Q) (α − γ)(β − γ)(γ − Q) so that 1 1 1 ln(α − Q) + ln(β − Q) + ln(γ − Q) = −kt − c, (β − α)(γ − α) (α − β)(γ − β) (α − γ)(β − γ) which can be written as (β − γ) ln(α − Q) + (γ − α) ln(α − Q) + (α − β) ln(γ − Q) = (α − β)(β − γ)(γ − α)kt + c1 or equivalently, ln(α − Q)β−γ + ln(α − Q)γ−α + ln(γ − Q)α−β = (α − β)(β − γ)(γ − α)kt + c1 . Exponentiating both sides and simplifying yields (α − Q)β−γ (β − Q)γ−α (γ − Q)α−β = c2 e(α−β)(β−γ)(γ−α)kt . Q(0) = 0 =⇒ c2 = αβ−γ β γ−α γ α−β , so that (1 − Q/α)β−γ (1 − Q/β)γ−α (1 − Q/γ)α−β = e(α−β)(β−γ)(γ−α)kt . Solutions to Section 1.5 True-False Review: (a): TRUE. The differential equation for such a population growth is dP dt = kP , where P (t) is the population as a function of time, and this is the Malthusian growth model described at the beginning of this section. (b): FALSE. The initial population could be greater than the carrying capacity, although in this case the population will asymptotically decrease towards the value of the carrying capacity. (c): TRUE. The differential equation governing the logistic model is (1.5.2), which is certainly separable as D dP = r. P (C − P ) dt Likewise, the differential equation governing the Malthusian growth model is dP dt = kP , and this is separable as P1 dP = k. dt (c)2017 Pearson Education. Inc. 45 (d): TRUE. As (1.5.3) shows, as t → ∞, the population does indeed tend to the carrying capacity C independently of the initial population P0 . As it does so, its rate of change dP dt slows to zero (this is best seen from (1.5.2) with P ≈ C). (e): TRUE. Every five minutes, the population doubles (increase 2-fold). Over 30 minutes, this population will double a total of 6 times, for an overall 26 = 64-fold increase. (f ): TRUE. An 8-fold increase would take 30 years, and a 16-fold increase would take 40 years. Therefore, a 10-fold increase would take between 30 and 40 years. dP (g): FALSE. The growth rate is dP dt = kP , and so as P changes, dt changes. Therefore, it is not always constant. (h): TRUE. From (1.5.2), the equilibrium solutions are P (t) = 0 and P (t) = C, where C is the carrying capacity of the population. (i): FALSE. If the initial population is in the interval ( C2 , C), then although it is less than the carrying capacity, its concavity does not change. To get a true statement, it should be stated instead that the initial population is less than half of the carrying capacity. (j): TRUE. Since P (t) = kP , then P (t) = kP (t) = k 2 P > 0 for all t. Therefore, the concavity is always positive, and does not change, regardless of the initial population. Problems: dP 1. = kP =⇒ P (t) = P0 ekt . Since P (0) = 10, then P = 10ekt . Since P (3) = 20, then 2 = e3k =⇒ k = dt ln 2 . Thus P (t) = 10e(t/3) ln 3 . Therefore, P (24) = 10e(24/3) ln 3 = 10 · 28 = 2560 bacteria. 3 2. Using P (t) = P0 ekt we obtain P (10) = 5000 =⇒ 5000 = p0 e10k and P (12) = 6000 =⇒ 6000 = P0 e12k which implies that e2k = 65 =⇒ k = 12 ln 65 . Hence, P (0) = 5000( 56 )5 = 2009.4. Also, P = 2P0 when 2 ln 2 t = 12 ln 2 = ≈ 7.6h. ln 65 3. From P (t) = P0 ekt and P (0) = 2000 it follows that P (t) = 2000ekt . Since td = 4, k = 14 ln 2 so P = 2000et ln 2/4 . Therefore, P (t) = 106 =⇒ 106 = 2000et ln 2/4 =⇒ t ≈ 35.86 hours. dP = kP =⇒ P (t) = P0 ekt . Since, P (0) = 10000 then P (t) = 10000ekt . Since P (5) = 20000 then dt 20000 = 10000ekt =⇒ k = 15 ln 2. Hence P (t) = 10000e(t ln 2)/5 . 4. (a). P (20) = 10000e4 ln 2 = 160000. (b). 1000000 = 10000e(t ln 2)/5 =⇒ 100 = e(t ln 2)/5 =⇒ t = 5 ln 100 ≈ 33.22 years. ln 2 50C . In formulas (1.5.5) and (1.5.6) we have P0 = 500, P1 = 800, P2 = 1000, t1 = 5, 50 + (C − 50)e−rt 1 (1000)(300) 1 800[(800)(1500) − 2(500)(1000)] and t2 = 10. Hence, r = ln = ln 3, C = ≈ 1142.86, so 5 (500)(200) 5 8002 − (500)(1000) 571430 1142.86)(500) ≈ . Inserting t = 15 into the preceding formula that P (t) = 500 + 642.86e−0.2t ln 3 500 + 642.86e−0.2t ln 3 yields P (15) = 1091. 5. P (t) = (c)2017 Pearson Education. Inc. 46 50C In formulas (1.5.5) and (1.5.6) we have P0 = 50, P1 = 62, P2 = 76, t1 = 2, 50 + (C − 50)e−rt 62[(62)(126) − 2(50)(76)] (76)(12) 1 ≈ 0.132, C = and t2 = 2t1 = 4. Hence, r = ln ≈ 298.727, so that 2 (50)(14) 622 − (50)(76) 14936.35 . Inserting t = 20 into the preceding formula yields P (20) ≈ 221. P (t) = 50 + 248.727e−0.132t 6. P (t) = P2 (P1 − P0 ) > 1. Rearranging the terms in this inequality and P0 (P2 − P1 ) 2P0 P2 P1 (P0 + P2 ) − 2P0 P2 . Further, C > 0 requires that > 0. using the fact that P2 > P1 yields P1 > P0 + P2 P12 − P0 P2 2P0 P2 From P1 > we see that the numerator in the preceding inequality is positive, and therefore the P0 + P2 2P0 P2 denominator must also be positive. Hence in addition to P1 > , we must also have P12 > P0 P2 . P0 + P2 7. From equation (1.5.5) r > 0 requires dy = dt ky(1500 − y), y(0) = 5, y(1) = 10, where k is a positive constant. Separating the differential equation 1 dy = k dt. Using a partial fraction decomposition on the left-hand and integrating yields y(1500 − y) 1 1 y 1 dy = kt + c, so that = kt + c, which upon + ln side gives 1500y 1500(1500 − y) 1500 1500 − y y 1 exponentiation yields = c1 e1500kt . Imposing the initial condition y(0) = 5, we find that c1 = . 1500 − y 299 10 y 1 1500kt 1 1500k . The further condition y(1) = 10 requires . solving Hence, = e = e 1500 − y 299 1490 299 1 299 y 1 t ln (299/149) for k gives k = . Solving algebraically for y we find ln . Therefore, = e 1500 149 1500 − y 299 1500et ln (299/149) 1500 1500 y(t) = = . Hence, y(14) = = 1474. 299 + et ln (299/149) 1 + 299e−t ln (299/149) 1 + 299e−14 ln (299/149) 8. Let y(t) denote the number of passengers who have the flu at time t. Then we must solve 9.(a). Equilibrium solutions: P (t) = 0, P (t) = T . dP dP > 0, 0 < P < T =⇒ < 0. Slope: P > T =⇒ dt dt k rT 2 + 4k 1 2 T± . We see that slope of the Isoclines: r(P − T ) = k =⇒ P − T P − = 0 =⇒ P = r 2 r −rT 2 solution curves satisfies k ≥ . 4 2 dP d P = r(2P − T ) Concavity: = r2 (2P − T )(P − T )P . Hence, the solution curves are concave up for dt2 dt t T P > , and are concave down for 0 < P < . 2 2 (b). See accompanying figure. (c). For 0 < P0 < T , the population dies out with time. For P0 > T , there is a population growth. The term threshold level is appropriate since T gives the minimum value of P0 above which there is a population growth. (c)2017 Pearson Education. Inc. 47 P(t) T t 0 Figure 0.0.42: Figure for Problem 9(b) 1 dP 10. (a). Separating the variables in differential equation (1.5.7) gives = r, which can be P (P − T ) dt 1 1 dP 1 P −T = rt+c, so that written in the equivalent form − = r. Integrating yields ln T (P − T ) T P dt T P P −T P0 − T P −T P0 − T T rt erT t . = c1 , so that = c1 e . The initial condition P (0) = P0 requires = P P0 P P0 T P0 Solving algebraically for P yields P (t) = . P0 − (P0 − T )erT t T P0 is positive, and increases without bound as P0 − (P0 − T )erT t t → ∞. Consequently limt→∞ P (t) = 0. In this case the population dies out as t increases. (b). If P0 < T , then the denominator in T P0 (c). If P0 > T , then the denominator of vanishes when (P0 − T )erT t = P0 , that is when P0 − (P0 − T )erT t 1 P0 t= ln . This means that within a finite time the population grows without bound. We can rT P0 − T interpret this as a mathematical model of a population explosion. dP = r(C − P )(P − T )P, P (0) = P0 , r > 0, 0 < T < C. dt Equilibrium solutions: P (t) = 0, P (t) = T, P (t) = C. The slope of the solution curves is negative for 0 < P < T , and for P > C. It is positive for T < P < C. d2 P Concavity: = r2 [(C − P )(P − T ) − (P − T )P + (C − P )P ](C − P )(P − T )P , which simplifies to dt2 d2 P = r2 (−3P 2 + 2P T + 2CP − CT )(C − P )(P − T ). Hence changes in concavity occur when P = 13 (C + dt2 √ T ± C 2 − CT + T 2 ). A representative slope field with some solution curves is shown in the accompanying figure. We see that for 0 < P0 < T the population dies out, whereas for T < P0 < C the population grows and asymptotes to the equilibrium solution P (t) = C. If P0 > C, then the solution decays towards the equilibrium solution P (t) = C. 11. dP = rP (ln C − ln P ), P (0) = P0 , and r, C, and P0 are positive constants. dt Equilibrium solutions: P (t) = C. The slope of the solution curves is positive for 0 < P < C, and negative 12. (c)2017 Pearson Education. Inc. 48 P(t) C T t 0 Figure 0.0.43: Figure for Problem 11 dP C d2 P C C 2 − 1 − 1 P ln . Hence, the solution curves are concave ln = r ln = r 2 dt P dt P P C C up for 0 < P < and P > C. They are concave down for < P < C. A representative slope field with e e some solution curves are shown in the accompanying figure. for P > C. Concavity: 25 20 15 P(t) 10 5 0 1 2 3 t 4 5 Figure 0.0.44: Figure for Problem 12 1 dP = r which can be integrated directly to P (ln C − ln P ) dt −rt . The initial condition P (0) = P0 requires that obtain − ln (ln C − ln P ) = rt + c so that ln ( C P ) = c1 e −rt C C C −rt ln ( P0 ) = c1 . Hence, ln ( P ) = e ln ( P0 ) so that P (t) = Celn (P0 /k)e . Since limt→∞ e−rt = 0, it follows that limt→∞ P (t) = C. 13. Separating the variables in (1.5.8) yields dP = kP , which is easily integrated to obtain P (t) = P0 ekt . dt kt The initial condition P (0) = 400 requires that P0 = 400, so that P (t) = 400e . We also know that 1 17 . Consequently, ln P (30) = 340. This requires that 340 = 400e30k so that k = 30 20 14. Using the exponential decay model we have 17 P (t) = 400e 30 ln( 20 ) . t (c)2017 Pearson Education. Inc. (0.0.5) 49 17 (a). From (0.0.5), P (60)400e2 ln( 20 ) = 289. 10 17 (b). From (0.0.5), P (100) = 400e 3 ln( 20 ) ≈ 233 (c). From (0.0.5), the half-life, tH , is determine from tH 17 200 = 400e 30 ln( 20 ) =⇒ tH = 30 ln 2 ≈ 128 days. ln 20 17 15. (a). More. dP = kP , which is easily integrated to obtain P (t) = P0 ekt . dt 100,000ekt . We also know The initial condition P (0) = 100, 000 requires that P0 = 100, 000, so that P (t) = 1 4 . Consequently, that P (10) = 80, 000. This requires that 100, 000 = 80, 000e10k so that k = ln 10 5 (b). Using the exponential decay model we have 4 P (t) = 100, 000e 10 ln( 5 ) . t (0.0.6) Using (0.0.6), the half-life is determined from tH 4 50, 000 = 100, 000e 10 ln( 5 ) =⇒ tH = 10 ln 2 ≈ 31.06 min. ln 54 (c). Using (0.0.6) there will be 15,000 fans left in the stadium at time t0 , where 15, 000 = 100, 000e t0 10 ln 3 ( 45 ) =⇒ t = 10 ln 20 ≈ 85.02 min. 0 4 ln 15 dP 16. Using the exponential decay model we have = kP , which is easily integrated to obtain P (t) = P0 ekt . dt Since the half-life is 5.2 years, we have ln 2 1 P0 = P0 e5.2k =⇒ k = − . 2 5.2 Therefore, ln 2 P (t) = P0 e−t 5.2 . Consequently, only 4% of the original amount will remain at time t0 where ln 2 ln 25 4 P0 = P0 e−t0 5.2 =⇒ t0 = 5.2 ≈ 24.15 years. 100 ln 2 17. Maple, or even a TI 92 plus, has no problem in solving these equations. 18. (a). Malthusian model is P (t) = 151.3ekt . Since P (1) = 179.4, then 179.4 = 151.3e10k =⇒ k = t ln (179.4/151.1) 1 179.4 10 . 10 ln 151.3 . Hence, P (t) = 151.3e (c)2017 Pearson Education. Inc. 50 P(t) 320 300 280 260 240 220 200 180 160 0 10 20 30 40 t Figure 0.0.45: Figure for Problem 18(c) 151.3C . Imposing the initial conditions P (10) = 179.4 and P (20) = 203.3 151.3 + (C − 151.3)e−rt 10 ln (179.4/151.1) 20 ln (179.4/151.1) 10 10 and 203.3 = 151.3e whose solution is gives the pair of equations 179.4 = 151.3e 39935.6 . C ≈ 263.95, r ≈ 0.046. Using these values for C and r gives P (t) = 151.3 + 112.65e−0.046t (b). P (t) = (c). Malthusian model: P (30) ≈ 253 million; P (40) ≈ 300 million. Logistics model: P (30) ≈ 222 million; P (40) ≈ 236 million. The logistics model fits the data better than the Malthusian model, but still gives a significant underestimate of the 1990 population. 50C . Imposing the conditions P (5) = 100, P (15) = 250 gives the pair of equations 50 + (C − 50)e−rt 50C 50C and 250 = whose positive solutions are C ≈ 370.32, r ≈ 0.17. 100 = 50 + (C − 50)e−5r 50 + (C − 50)e−15r 18500 Using these values for C and r gives P (t) = . From the figure we see that it will take 50 + 18450e−0.17t approximately 52 years to reach 95% of the carrying capacity. 19. P (t) = Solutions to Section 1.6 True-False Review: (a): FALSE. Any solution to the differential equation (1.6.7) serves as an integrating factor for the differ ential equation. There are infinitely many solutions to (1.6.7), taking the form I(x) = c1 e p(x)dx , where c1 is an arbitrary constant. (b): TRUE. Any solution to the differential equation (1.6.7) serves as an integrating factor for the differ ential equation. There are infinitely many solutions to (1.6.7), taking the form I(x) = c1 e p(x)dx , where c1 is an arbitrary constant. The most natural choice is c1 = 1, giving the integrating factor I(x) = e p(x)dx . (c): TRUE. Multiplying y + p(x)y = q(x) by I(x) yields y I + pIy = qI. Assuming that I = pI, the requirement on the integrating factor, we have y I + I y = qI, or by the product rule, (I · y) = qI, as requested. (c)2017 Pearson Education. Inc. 51 P(t) Carrying Capacity 350 300 250 200 150 100 50 0 10 20 30 40 50 60 70 t Figure 0.0.46: Figure for Problem 19 (d): FALSE. Rewriting the differential equation as dy − x2 y = sin x, dx 2 3 we have p(x) = −x2 , and so an integrating factor must have the form I(x) = e p(x)dx = e (−x )dx = e−x /3 , 3 2 or any constant multiple of e−x /3 . Since ex is not of this form, then it is not an integrating factor. (e): FALSE. Rewriting the differential equation as dy 1 + y = x, dx x 1 we have p(x) = , and so an integrating factor must have the form I(x) = e p(x)dx = e (1/x)dx = x, or any x constant multiple of x. Since x + 5 is not of this form, then it is not an integrating factor. Problems: In this section the function I(x) = e p(x)dx will represent the integrating factor for a differential equation of the form y + p(x)y = q(x). 1. y + y = 4ex . I(x) = e dx = ex =⇒ 2. y + d(ex y) = 4e2x =⇒ ex y = 2e2x + c =⇒ y(x) = e−x (2e2x + c). dx 2 d(x2 y) y = 5x2 . I(x) = e (2/x) dx = x2 =⇒ = 5x4 =⇒ x2 y = x5 + c =⇒ y(x) = x−2 (x5 + c). x dx 3. x2 y − 4xy = x7 sin x, x > 0 =⇒ y − x4 y = x5 sin x. I(x) = x−4 =⇒ sin x − x cos x + c =⇒ y(x) = x4 (sin x − x cos x + c). 4. y + 2xy = 2x3 . I(x) = e2 2 xdx 2 = ex =⇒ 2 ex (x2 − 1) + c =⇒ y(x) = x2 − 1 + ce−x . d(x−4 y) = x sin x =⇒ x−4 y = dx 2 d x2 2 2 2 (e y) = 2ex x3 =⇒ ex y = 2 ex x3 dx =⇒ ex y = dx (c)2017 Pearson Education. Inc. 52 2x 1 d y = 4x, −1 < x < 1. I(x) = =⇒ 2 2 1−x 1−x dx c =⇒ y(x) = (1 − x2 )[− ln (1 − x2 )2 + c]. 5. y + y 1 − x2 = 4x y =⇒ = − ln(1 − x2 )2 + 2 1−x 1 − x2 2x 2x 4 d 4 dx 1+x2 y = . I(x) = e = 1 + x2 =⇒ =⇒ (1 + x2 )y = [(1 + x2 )y] = 2 2 2 1+x (1 + x ) dx (1 + x2 )2 dx 1 4 =⇒ (1 + x2 )y = 4 tan−1 x + c =⇒ y(x) = (4 tan−1 x + c). 2 1+x 1 + x2 6. y + dy sin 2x 1 d + y sin 2x = 4 cos4 x, 0 ≤ x ≤ π2 =⇒ y + y = 2 cos2 x. I(x) = =⇒ (y sec x) = dx 2 cos2 x cos x dx cos x =⇒ y(x) = cos x(2 sin x + c) =⇒ y(x) = sin 2x + c cos x. 7. 2 cos2 x dx 1 d y = 9x2 . I(x) = e x ln x = ln x =⇒ (y ln x) = 9 x2 ln xdx =⇒ y ln x = 3x3 ln x − x3 + c =⇒ x ln x dx 3x3 ln x − x3 + c y(x) = . ln x 8. y + d 9. y − y tan x = 8 sin3 x. I(x) = cos x =⇒ (y cos x) = 8 cos x sin3 x =⇒ y cos x = 8 cos x sin3 xdx + c =⇒ dx 1 y cos x = 2 sin4 x + c =⇒ y(x) = (2 sin4 x + c). cos x dt 2 d dx 4et + 2x = 4et =⇒ x + x = . I(x) = e2 t = t2 =⇒ (t2 x) = 4tet =⇒ t2 x = 4 tet dt + c =⇒ dt t t dt 4et (t − 1) + c 2 t t x = 4e (t − 1) + c =⇒ x(t) = . t2 10. t d y = (sin x sec x)y − 2 sin x =⇒ y − (sin x sec x)y = −2 sin x. I(x) = cos x =⇒ (y cos x) = dx 1 1 1 cos 2x + c . −2 sin x cos x =⇒ y cos x = −2 sin x cos xdx + c = cos 2x + c =⇒ y(x) = 2 cos x 2 11. d (y sec x) = 12. (1−y sin x)dx−cos xdy = 0 =⇒ y +(sin x sec x)y = sec x. I(x) = e sin x sec xdx = sec x =⇒ dx 2 2 sec x =⇒ y sec x = sec xdx+c =⇒ y sec x = tan x+c =⇒ y(x) = cos x(tan x+c) =⇒ y(x) = sin x+c cos x. 1 d −1 13. y − x−1 y = 2x2 ln x. I(x) = e− x dx = x−1 =⇒ (x y) = 2x ln x =⇒ x−1 y = 2 x ln xdx + c =⇒ dx 1 3 1 2 −1 x y = x (2 ln x − 1) + c. Hence, y(x) = x (2 ln x − 1) + cx. 2 2 d αx (e y) = e(α+β)x =⇒ eαx y = e(α+β)x dx + c. If α + β = 0, dx e(α+β)x eβx αx −αx (x + c). If α + β = 0, then eαx y = then e y = x + c =⇒ y(x) = e + c =⇒ y(x) = + ce−αx . α+β α+β 14. y + αy = eβx . I(x) = eα dx = eαx =⇒ m d m y = ln x. I(x) = xm =⇒ (x y) = xm ln x =⇒ xm y = xm ln xdx + c. If m = −1, then x dx (ln x)2 xm+1 (ln x)2 xm+1 xm y = + c =⇒ y(x) = + c =⇒ y(x) = x + c . If m = −1, then xm y = ln x − 2 2 m+1 (m + 1)2 15. y + (c)2017 Pearson Education. Inc. 53 c x x + m. ln x − m+1 (m + 1)2 x dx 2 d 2 16. y + y = 4x. I(x) = e2 x = e2 ln x = x2 =⇒ (x y) = 4x3 =⇒ x2 y = 4 x3 dx + c =⇒ x2 y = x4 + c, x dx x4 + 1 but y(1) = 2 so c = 1; thus, y(x) = . x2 d (y csc x) = 2 csc x cos x =⇒ dx π y csc x = 2 ln (sin x) + c, but y( 2 ) = 2 so c = 2; thus, y(x) = 2 sin x[ln (sin x) + 1]. 17. y sin x − y cos x = sin 2x =⇒ y − y cot x = 2 cos x. I(x) = csc x =⇒ dt 2 d 18. x + x = 5. I(t) = e2 4−t = e−2 ln(4−t) = (4−t)t−2 =⇒ ((4−t)−2 x) = 5(4−t)−2 =⇒ (4−t)−2 x = 4−t dt 5 (4 − t)−2 dt + c =⇒ (4 − t)−2 x = 5(4 − t)−1 + c, but x(0) = 4 so c = −1; thus, x(t) = (4 − t)2 [5(4 − t)−1 − 1] or x(t) = (4 − t)(1 + t). d x e2x (e y) = e2x =⇒ ex y = + c, but y(0) = 1 so 19. (y − e−x )dx + dy = 0 =⇒ y + y = ex . I(x) = ex =⇒ dx 2 1 1 x c = ; thus, y(x) = (e + e−x ) = cosh x. 2 2 20. y + y = f (x), y(0) = 3, f (x) = 1, 0, if x ≤ 1, if x > 1. x x d x I(x) = e dx = ex =⇒ (e y) = ex f (x) =⇒ [ex y]x0 = 0 ex f (x)dx =⇒ ex y − y(0) = 0 ex f (x)dx =⇒ dx x x ex y − 3 = 0 ex dx =⇒ y(x) = e−x 3 + 0 ex f (x)dx . x x x x If x ≤ 1, 0 e f (x)dx = 0 e dx = ex − 1 =⇒ y(x) = e−x (2 + ex ) x x If x > 1, 0 ex f (x)dx = 0 ex dx = e − 1 =⇒ y(x) = e−x (2 + e). 21. y − 2y = f (x), y(0) = 1, f (x) = 1 − x, 0, if x < 1, if x ≥ 1. x d −2x y) = e−2x f (x) =⇒ [e−2x y]x0 = 0 e−2x f (x)dx =⇒ e−2x y − y(0) = I(x) = e− 2dx = e−2x =⇒ (e dx x x x −2x e f (x)dx =⇒ e−2x y − 1 = 0 e−2x f (x) =⇒ y(x) = e2x 1 + 0 e−2x f (x)dx . 0 x −2x x −2x 1 −2x 1 −2x 2x 2x 2x 1+ e f (x)dx = 0 e (1 − x)dx = e (2x−1+e ) =⇒ y(x) = e (2x − 1 + e ) = If x < 1, 0 e 4 4 1 2x (5e + 2x − 1). 4 x x 1 1 1 If x ≥ 1, 0 e−2x f (x)dx = 0 e−2x (1 − x)dx = (1 + e−2 ) =⇒ y(x) = e2x 1 + (1 + e−2 ) = e2x (5 + e−2 ). 4 4 4 22. On (−∞, 1), y − y = 1 =⇒ I(x) = e−x =⇒ y(x) = c1 ex − 1. Imposing the initial condition y(0) = 0 requires c1 = 1, so that y(x) = ex − 1, for x < 1. d −x On [1, ∞), y − y = 2 − x =⇒ I(x) = e−x =⇒ (e y) = (2 − x)e−x =⇒ y(x) = x − 1 + c2 e−x . dx Continuity at x = 1 requires that limx→1 y(x) = y(1). Consequently we must choose c2 to satisfy c2 e = e − 1, so that c2 = 1 − e−1 . Hence, for x ≥ 1, y(x) = x − 1 + (1 − e−1 )ex . (c)2017 Pearson Education. Inc. 54 d2 y 1 dy du 1 dy du d2 y + = 9x, x > 0. Let u = so = 2 . The first equation becomes + u = 9x which is 2 dx x dx dx dx dx dx x d (xu) = 9x2 =⇒ xu = x2 dx + c =⇒ xu = first-order linear. An integrating factor for this is I(x) = x so dx dy dy 3x3 + c1 =⇒ u = 3x2 + c1 x−1 , but u = so = 3x2 + c1 x−1 =⇒ y = (3x2 + c1 x−1 )dx + c2 =⇒ y(x) = dx dx x3 + c1 ln x + c2 . 23. 24. The differential equation for Newton’s Law of Cooling is dT dt = −k(T − Tm ). We can re-write this equation in the form of a first-order linear differential equation: dT dt + kT = kTm . An integrating factor d for this differential equation is I = e k dt = ekt . Thus, (T ekt ) = kTm ekt . Integrating both sides, we get dt T ekt = Tm ekt + c, and hence, T = Tm + ce−kt , which is the solution to Newton’s Law of Cooling. dT dT dTm = α =⇒ Tm = αt + c1 so = −k(T − αt − c1 ) =⇒ + kT = k(αt + c1 ). An integrating dt dt dt d kt factor for this differential equation is I = ek dt = ekt . Thus, (e T ) = kekt (αt + c1 ) =⇒ ekt T = dt α α 1 ekt (αt − + c1 ) + c2 =⇒ T = αt − + c1 + c2 e−kt =⇒ T (t) = α(t − ) + β + T0 e−kt where β = c1 and k k k T0 = c 2 . 25. dTm dT = 10 =⇒ Tm = 10t + c1 but Tm = 65 when t = 0 so c1 = 65 and Tm = 10t + 65. = dt dt dT dT 1 = −k(T − 10t − 65), but (1) = 5, so k = . The last differential equation can be −k(T − Tm ) =⇒ dt dt 8 2 13 d kt 2 dT kt + c =⇒ + kT = k(10t + 65) =⇒ (e T ) = 5ke (2t + 13) =⇒ ekt T = 5kekt t− 2 + written dt dt k k k 2 1 t 1 T = 5(2t − + 13) + ce−kt , but k = so T (t) = 5(2t − 3) + ce− 8 . Since T (1) = 35, c = 40e 8 . Thus, k 8 1 T (t) = 10t − 15 + 40e 8 (1−t) . 26. dT 1 27. (a). In this case, Newton’s law of cooling is = − (t − 80e−t/20 ). This linear differential equation dt 40 dT 1 has standard form + T = 2e−t/20 , with integrating factor I(t) = et/40 . Consequently the differential dt 40 d t/40 T ) = 2e−t/40 , so that T (t) = −80e−t/20 + ce−t/40 . equation can be written in the integrable form (e dt Then T (0) = 0 =⇒ c = 80, so that T (t) = 80(e−t/40 − e−t/20 ). (b). We see that limt→∞ = 0. This is a reasonable result since the temperature of the surrounding medium also approaches zero as t → ∞. We would expect the temperature of the object to approach to the temperature of the surrounding medium at late times. dT 1 −t/20 1 −t/40 −t/40 −t/20 (c). T (t) = 80(e . So T (t) has only one critical point −e ) =⇒ + e = 80 − e 40 20 dt 1 1 when 80 − e−t/40 + e−t/20 = 0 =⇒ t = 40 ln 2. Since T (0) = 0, and limt→∞ T (t) = 0 the function 40 20 assumes a maximum value at tmax = 40 ln 2. T (tmax ) = 80(e− ln 2 − e−2 ln 2 ) = 20◦ F, Tm (tmax ) = 80e−2 ln 2 = 20◦ F. (d). The behavior of T (t) and Tm (t) is given in the accompanying figure. (c)2017 Pearson Education. Inc. 55 T(t), Tm(t) 80 Tm(t) 60 40 T(t) 20 t 0 0 20 40 60 80 100 120 Figure 0.0.47: Figure for Problem 27(d) 28. (a). The temperature varies from a minimum of A − B at t = 0 to a maximum of A + B when t = 12. T A+B A-B t 5 10 15 20 Figure 0.0.48: Figure for Problem 28(a) dT + k1 T = k1 (A − B cos ωt) + T0 . Multiplying dt k1 t by the integrating factor I = e reduces this differential equation to the integrable form (b). First write the differential equation in the linear form d k1 t (e T ) = k1 ek1 t (A − B cos ωt) + T0 ek1 t . dt Consequently, ek1 t T (t) = Aek1 t − Bk1 ek1 t cos ωtdt + (c)2017 Pearson Education. Inc. T0 k 1 t e +c k1 56 so that T (t) = A + T0 Bk1 − 2 (k1 cos ωt + ω sin ωt) + ce−k1 t . k1 k1 + ω 2 This can be written in the equivalent form T (t) = A + T0 Bk1 − 2 cos (ωt − α) + ce−k1 t k1 k1 + ω 2 for an approximate phase constant α. dy dy dy + p(x)y = 0 =⇒ = −p(x)dx =⇒ = − p(x)dx =⇒ ln |y| = − p(x)dx + c =⇒ yH = dx y y c1 e− p(x)dx . dy dv du =u + v . Substituting this (b). Replace c1 in part (a) by u(x) and let v = e− p(x)dx . y = uv =⇒ dx dx dx dv du dy + p(x) = q(x), we obtain u + v + p(x)y = q(x), but last result into the original differential equation, dx dx dx du dv = −vp, the last equation reduces to v = q(x) =⇒ du = v −1 (x)q(x)dx =⇒ u = v −1 (x)q(x)dx + since dx dx e p(x)dx q(x)dx + c . c. Substituting the values for u and v into y = uv, we obtain y = e− p(x)dx 29. (a). dy +x−1 y = 0, with solution yH = cx−1 . According to Problem dx 29, we determine the function u(x) such that y(x) = x−1 u(x) is a solution to the given differential equation. du du 1 dy dy We have = x−1 −x−2 u. Substituting into +x−1 y = cos x yields x−1 − 2 u+x−1 (x−1 u) = cos x, dx dx dx dx x du so that = x cos x. Integrating we obtain u = x sin x + cos x + c, so that y(x) = x−1 (x sin x + cos x + c). dx 30. The associated homogeneous equation is dy + y = 0, with solution yH = ce−x . According to Problem dx 29, we determine the function u(x) such that y(x) = e−x u(x) is a solution to the given differential equation. dy du −x dy du −x We have = e − e−x u. Substituting into + y = e−2x yields e − e−x u + e−x u(x) = e−2x , so dx dx dx dx du that = e−x . Integrating we obtain u = −e−x + c, so that y(x) = e−x (−e−x + c). dx 31. The associated homogeneous equation is dy + cot x · y = 0, with solution yH = c · csc x. According dx to Problem 29, we determine the function u(x) such that y(x) = csc x · u(x) is a solution to the given du dy dy = csc x · − csc x · cot x · u. Substituting into + cot x · y = 2 cos x differential equation. We have dx dx dx du du − csc x · cot x · u + csc x · cot x · u = cos x, so that = 2 cos x sin x. Integrating we obtain yields csc x · dx dx 2 2 u = sin x + c, so that y(x) = csc x(sin x + c). 32. The associated homogeneous equation is dy 1 − y = 0, with solution yH = cx. We determine the dx x dy du function u(x) such that y(x) = xu(x) is a solution of the given differential equation. We have = x + u. dx dx du dy 1 − y = x ln x and simplifying yields = ln x, so that u = x ln x − x + c. Consequently, Substituting into dx x dx 33. The associated homogeneous equation is (c)2017 Pearson Education. Inc. 57 y(x) = x(x ln x − x + c). Problems 34 - 39 are easily solved using a differential equation solver such as the dsolve package in Maple. Solutions to Section 1.7 True-False Review: (a): TRUE. Concentration of chemical is defined as the ratio of mass to volume; that is, c(t) = VA(t) (t) . Therefore, A(t) = c(t)V (t). (b): FALSE. The rate of change of volume is “rate in” − “rate out”, which is r1 − r2 , not r2 − r1 . (c): TRUE. This is reflected in the fact that c1 is always assumed to be a constant. (d): FALSE. The concentration of chemical leaving the tank is c2 (t) = VA(t) (t) , and since both A(t) and V (t) can be nonconstant, c2 (t) can also be nonconstant. (e): FALSE. Kirchhoff’s second law states that the sum of the voltage drops around a closed circuit is zero, not that it is independent of time. (f ): TRUE. This is essentially Ohm’s law, (1.7.10). (g): TRUE. Due to the negative exponential in the formula for the transient current, iT (t), it decays to zero as t → ∞. Meanwhile, the steady-state current, iS (t), oscillates with the same frequency ω as the alternating current, albeit with a phase shift. 0 (h): TRUE. The amplitude is given in (1.7.19) as A = √R2E , and so as ω gets larger, the amplitude +ω 2 L2 A gets smaller. Problems: A(60) . ΔV = r1 Δt − V (60) dV dA r2 Δt =⇒ = 3 =⇒ V (t) = 3(t + 200) since V (0) = 600. ΔA ≈ c1 r1 Δt − c2 r2 Δt =⇒ = 30 − 3c2 = dt dt A A 30 − 3 = 30 − =⇒ (t + 200)A = 15(t + 200)2 + c. Since A(0) = 1500, c = −300000 and therefore V t + 200 15 A(60) 596 A(t) = [(t + 200)2 − 20000]. Thus = g/L. t + 200 V (60) 169 1. Given V (0) = 600, A(0) = 1500, c1 = 5, r1 = 6, and r2 = 3. We need to find dV 2. Given V (0) = 10, A(0) = 20, c1 = 4, r1 = 2, and r2 = 1. Then ΔV = r1 Δt−r2 Δt =⇒ = 1 =⇒ V (t) = dt dA A A dA 1 = 8 − c2 = 8 − = 8 − =⇒ + A= t + 10 since V (0) = 10. ΔA ≈ c1 r1 Δt − c2 r2 Δt =⇒ dt V t + 10 dt t + 10 4 8 =⇒ (t + 10)A = 4(t + 10)2 + c1 . Since A(0) = 20 =⇒ c1 = −200 so A(t) = [(t + 10)2 − 50]. Therefore, t + 10 A(40) = 196 g. dV 3. Given V (0) = 100, A(0) = 100, c1 = 0.5, r1 = 6, and r2 = 4. Then ΔV = r1 Δt − r2 Δt =⇒ = 2 =⇒ dt 4A dA 2A d dA + = 3 =⇒ + = 3 =⇒ [(t + 50)2 A] = V (t) = 2(t + 50) since V (0) = 100. Then dt 2(t + 50) dt t + 50 dt (c)2017 Pearson Education. Inc. 58 125000 . (t + 50)2 The tank is full when V (t) = 200, that is when 2(t+50) = 200 so that t = 50 min. Therefore the concentration A(50) 9 just before the tank overflows is: = g/L. V (50) 16 3(t+50)2 =⇒ (t+50)2 A = (t+50)3 +c but A(0) = 100 so c = 125000 and therefore A(t) = t+50+ dV = 2 =⇒ V = dt 2(t + 10) since V (0) = 20. Thus V (t) = 40 for t = 10, so we must find A(10).ΔA ≈ c1 r1 Δt − c2 r2 Δt =⇒ 2A dA A dA 1 d = 40 − 2c2 = 40 − = 40 − =⇒ + A = 40 =⇒ [(t + 10)A] = 40(t + 10)dt =⇒ dt V t + 10 dt t + 10 dt 20 (t + 10)A = 20(t + 10)2 + c. Since A(0) = 0 =⇒ c = −2000 so A(t) = [(t + 10)2 − 100] and A(10) = 300 t + 10 g. 4. Given V (0) = 20, A(0) = 0, c1 = 10, r1 = 4, and r2 = 2. Then ΔV = r1 Δt − r2 Δt =⇒ dV = 1 =⇒ dt (20 + t)3 + c dA 2 d V = t+20 since V (0) = 20. Then + A = 3 =⇒ [(t+20)2 A] = 3(t+20)2 =⇒ A(t) = dt t + 20 dt (t + 20)2 3 3 (t + 20) − 20 and since A(0) = 0, c = −203 which means that A(t) = . (t + 20)2 5. (a). We are given that V (0) = 20, c1 = 1, r1 = 3, and r2 = 2. Then ΔV = r1 Δt − r2 Δt =⇒ A(t) A(t) or c2 = so from part (a), V (t) t + 20 √ 1 (t + 20)3 − 203 1 (t + 20)3 − 203 . Therefore c2 = g/l when = =⇒ t = 20( 3 2 − 1)minutes. c2 = 3 3 (t + 20) 2 2 (t + 20) (b). The concentration of chemical in the tank, c2 , is given by c2 = 6. We are given that V (0) = 10, A(0) = 0, c1 = 0.5, r1 = 3, r2 =, and A(5) = 0.2. V (5) dV = 1 =⇒ V (t) = t + 10 since V (0) = 10. Then ΔA ≈ c1 r1 Δt − c2 r2 Δt =⇒ (a). ΔV = r1 Δt − r2 Δt =⇒ dt A dA 2A dA 2dt = −2c2 = −2 = − =⇒ =− =⇒ ln |A| = −2 ln |t + 10| + c =⇒ A = k(t+ 10)−2 . Then dt V t + 10 A t + 10 A(5) 675 A(5) = 3 since V (5) = 15 and . In particular, A(0) = 6.75 = 0.2. Thus, k = 675 and A(t) = V (5) (t + 10)2 g. A(t) A(t) 675 675 and V (t) = t + 10 =⇒ . = 0.1. From part (a) A(t) = = 2 V (t) (t + 10) V (t) (t + 10)3 √ √ A(t) Since = 0.1 =⇒ (t + 10)3 = 6750 =⇒ t + 10 = 15 3 2 so V (t) = t + 10 = 15 3 2 L. V (t) (b). Find V (t) when 7. (a). We are given that V (0) = w, c1 = k, r1 = r, r2 = r, and A(0) = A0 . Then ΔV = r1 Δr − r2 Δt =⇒ dA dV A A = 0 =⇒ V (t) = V (0) = w for all t. Then ΔA = c1 r1 Δt − c2 r2 Δt =⇒ = kr − r = kr − r = dt dt V V dA r d r + A = kr =⇒ (e−rt/w A) = kre−rt/w =⇒ A(t) = kw + ce−rt/w . Since A(0) = A0 so kr − A =⇒ w dt w dt c = A0 − kw =⇒ A(t) = e−rt/w [kw(ert/w − 1) + A0 ]. A(t) e−rt/w A0 (b). limt→∞ = limt→∞ [kw(ert/w − 1) + A0 ] = limt→∞ [k + − k e−rt/w ] = k. This is V (t) w w reasonable since the volume remains constant, and the solution in the tank is gradually mixed with and (c)2017 Pearson Education. Inc. 59 replaced by the solution of concentration k flowing in. A1 (t) dA1 dA1 dA1 = c1 r1 − c2 r2 =⇒ = c 1 r1 − r 2 =⇒ = c 1 r1 − dt dt V1 (t) dt r2 dA1 r2 A1 (t) =⇒ A1 = c1 r1 . + (r1 − r2 )t + V1 dt (r1 − r2 )t + v1 A1 A2 (t) dA2 dA2 dA2 For the bottom tank we have: − r3 = c2 r2 − c3 r3 =⇒ = r2 =⇒ = dt dt (r1 − r2 )t + V1 V2 (t) dt A1 A2 (t) dA2 r2 A1 r3 − r3 =⇒ A2 = . r2 + (r1 − r2 )t + V1 (r2 − r3 )t + V2 dt (r1 − r2 )t + V2 (r1 − r2 )t + V1 8. (a). For the top tank we have: dA1 dA1 dA1 r2 4 2 A1 = c1 r1 =⇒ + + A1 = 3 =⇒ + A1 = dt (r1 − r2 )t + v1 dt 2t + 40 dt t + 20 d c 3 =⇒ [(t + 20)2 A] = 3(t + 20)2 =⇒ A1 = t + 20 + but A1 (0) = 4 so c = −6400. Consequently dt (t + 20)2 6400 dA2 6400 dA2 2 3 3 t + 20 − =⇒ . Then A1 (t) = t + 20 − + A2 = + A2 = 2 2 (t + 20) dt t + 20 t + 20 (t + 20) dt t + 20 2[(t + 20)3 − 6400] d t + 20 2[(t + 20)3 − 6400] 12800t =⇒ A2 (t) = =⇒ [(t + 20)3 A2 ] = (t + 20)3 + − 3 (t + 20) dt (t + 20)3 2 (t + 20)3 t + 20 80000 k 12800t but A2 (0) = 20 so k = 80000. Thus A2 (t) = + and in particular − (t + 20)3 2 (t + 20)3 (t + 20)3 119 A2 (10) = ≈ 13.2 g. 9 (b). From part (a) di R 1 di d 40t + i = E(t) =⇒ + 40i = 200 =⇒ (e i) = dt L L dt dt 40t −40t −40t 200e =⇒ i(t) = 5 + ce . But i(0) = 0 =⇒ c = −5. Consequently i(t) = 5(1 − e ). 1 . Then 9. Let E(t) = 20, R = 4 and L = 10 dq 1 E dq d + q= =⇒ +10q = 20 =⇒ (qe10t ) = 20e10t =⇒ dt RC R dt dt q(t) = 2 + ce−10t . But q(0) = 0 =⇒ c = −2 so q(t) = 2(1 − e−40t ). 1 and E(t) = 100. Then 10. Let R = 5, C = 50 di R 1 di + i = E(t) =⇒ + 3i = 15 sin 4t =⇒ dt L L dt 3t 3e d 3t 3 4 (e i) = 15e3t sin 4t =⇒ e3t i = (3 sin 4t − 4 cos 4t) + c =⇒ i = 3 sin 4t − cos 4t + ce−3t , but dt 5 5 5 12 3 −3t i(0) = 0 =⇒ c = so i(t) = (3 sin 4t − 4 cos 4t + 4e ). 5 5 11. Let R = 2, L = 23 and E(t) = 10 sin 4t. Then 12. Let R = 2, C = 18 and E(t) = 10 cos 3t. Then dq 1 E dq d + q= =⇒ + 4q = 5 cos 3t =⇒ (e4t q) = dt RC R dt dt e4t 5e4t cos 3t =⇒ e4t q = (4 cos 3t + 3 sin 3t) + c =⇒ q(t) = 15 (4 cos 3t + 3 sin 3t) + ce4t , but q(0) = 1 =⇒ c = 5 dq 1 1 1 (4 cos 3t + 3 sin 3t) + e−4t and i(t) = = (9 cos 3t − 12 sin 3t − 4e−4t ). 5 5 dt 5 dq 1 E 13. In an RC circuit for t > 0 the differential equation is given by + q = . If E(t) = 0 then dt RC R 1 d t/RC dq t/RC −t/RC q) = 0 =⇒ q = ce and if q(0) = 5 then q(t) = 5e . Then limt→∞ q(t) = + q = 0 =⇒ (e dt RC dt 0. Yes, this is reasonable. As the time increases and E(t) = 0, the charge will dissipate to zero. (c)2017 Pearson Education. Inc. 60 1 E0 14. In an RC circuit the differential equation is given by i + q= . Differentiating this equation with RC R 1 dq dq di 1 d di + = 0, but = i so + i = 0 =⇒ (et/RC i) = 0 =⇒ i(t) = ce−t/RC . respect to t we obtain dt RC dt dt dt RC dt E0 E0 −t/RC Since q(0) = 0, i(0) = =⇒ d = E0 k so q(t) = E0 k(1 − e−t/RC ). Then and so i(t) = e R R limt→∞ q(t) = E0 k, and lim t → ∞i(t) = 0. q(t) E0k t Figure 0.0.49: Figure for Problem 14 di di R E(t) R E0 15. In an RL circuit, + i = and since E(t) = E0 sin ωt, then + i = sin ωt =⇒ dt L L dt L L E0 Rt/L E0 d Rt/L i) = sin ωt =⇒ i(t) = 2 [R sin ωt − ωL cos ωt] + Ae−Rt/L . We can write this (e e dt L R + L2 ω 2 R ωL E0 √ sin ωt − √ cos ωt + Ae−Rt/L . Defining the phase φ by as i(t) = √ R 2 + L2 ω 2 R 2 + L2 ω 2 R 2 + L2 ω 2 R ωL E0 cos φ = √ , sin φ = √ , we have i(t) = √ [cos φ sin ωt−sin φ cos ωt]+Ae−Rt/L . 2 2 2 2 2 2 2 R +L ω R +L ω R + L2 ω 2 E0 That is, i(t) = √ sin (ωt − φ) + Ae−Rt/L . 2 R + L2 ω 2 Transient part of the solution: iT (t) = Ae−Rt/L . E0 Steady state part of the solution: iS (t) = √ sin (ωt − φ). 2 R + L2 ω 2 di E0 R + ai = , i(0) = 0, where a = , and E0 denotes the dt L L constant EMF. An integrating factor for the differential equation is I = eat , so that the differential equation E0 at E0 d at (e i) = e . Integrating yields i(t) = + c1 e−at . The given initial can be written in the form dt L aL E0 E0 E0 E0 condition requires c1 + = 0, so that c1 = − . Hence i(t) = (1 − e−at ) = (1 − e−at ). aL aL aL R 16. We must solve the initial value problem 17. d E0 (1/RC−a)t 1 E(t) dq 1 E0 −at dq =⇒ (et/RC q) = =⇒ + q= =⇒ + q= e e dt RC R dt RC R dt R (c)2017 Pearson Education. Inc. 61 E0 C E0 C e(1/RC−a)t + k =⇒ q(t) = e−at + ke−t/RC . Imposing the initial condition 1 − aRC 1 − aRC E0 C E0 C q(0) = 0 (capacitor initially uncharged) requires k = − , so that q(t) = (e−at − e−t/RC ). 1 − aRC 1 − aRC 1 −t/RC dq E0 C − ae−at . Thus i(t) = = e dt 1 − aRC RC q(t) = e−t/RC d2 q 1 1 2 di 1 dq 1 + q = 0 =⇒ i + q = 0, since i = . Then idi = − qdq =⇒ i2 = − q + k but 2 dt LC dq LC dt LC LC q2 − q2 q2 − q2 q2 1 2 q2 dq =⇒ =⇒ q(0) = q0 and i(0) = 0 so k = 0 =⇒ i2 = − q + 0 =⇒ i = ± √0 = ± √0 LC LC LC LC LC dt dt t t dq = ±√ =⇒ sin−1 ( qq0 ) = ± √ + k1 =⇒ q = q0 sin ± √ + k1 but q(0) = q0 so q0 = 2 2 LC LC LC qo − q 1 t π =⇒ q(t) = q0 cos √ q0 sin k1 =⇒ k1 = π2 + 2nπ where n is an integer =⇒ q = q0 sin ± √ + 2 LC LC t q0 dq sin √ . = −√ and i(t) = dt LC LC 18. 1 di dq d2 q E0 dq d2 di + = q = . Since i = then = i . Hence the original equation can be written 2 2 dt LC L dt dt dq dt dq di 1 E0 1 E0 i2 q2 E0 q as i + q = =⇒ idi + qdq = dq or + = + A. Since i(0) = 0 and q(0) = q0 dq LC L LC L 2 2LC L 1/2 q02 E0 q 0 i2 q2 E0 q 2E0 q q2 then A = =⇒ i = − . From + = + A we get that i = 2A + − 2LC L 2 2LC L L LC 1/2 2 1/2 (E0 C)2 (2E0 C)2 (q − E0 C)2 dq q − E0 C 2 √ and we let D = 2A + =⇒ 2A + − then i =D 1− LC LC LC dt D LC √ √ q − E0 C q − E0 C √ √ LC sin−1 = t + B. Then since q(0) = 0 so B = LC sin−1 and therefore D LC D LC √ q − E0 C dq t+B t+B t+B √ = sin √ =⇒ q(t) = D LC sin √ + E0 c =⇒ i = . Since = D cos √ dt D LC LC LC LC 2 2 2A + (E0 C) q E0 q 0 |q0 − E0 C| . D2 = and A = 0 − we can substitute to eliminate A and obtain D = ± √ LC LC 2LC L t+B + E0 c. Thus q(t) = ±|q0 − E0 C| sin √ LC 19. Solutions to Section 1.8 True-False Review: (a): TRUE. We have f (tx, ty) = 3(yt)2 − 5(xt)(yt) 3y 2 t2 − 5xyt2 3y 2 − 5xy = = = f (x, y), 2(xt)(yt) + (yt)2 2xyt2 + y 2 t2 2xy + y 2 so f is homogeneous of degree zero. (c)2017 Pearson Education. Inc. 62 (b): FALSE. We have f (tx, ty) = y 2 t2 + xt y2 t + x (yt)2 + xt y2 + x = 2 2 = 2 , = 2 2 2 2 2 2 (xt) + 2(yt) x t + 2y t x t + 2y t x + 2y 2 so f is not homogeneous of degree zero. 3 2 (c): FALSE. Setting f (x, y) = xy+xy 3 +1 , we have f (tx, ty) = (xt)3 + (xt)(yt)2 x3 t3 + xy 2 t3 = = f (x, y), (yt)3 + 1 y 3 t3 + 1 so f is not homogeneous of degree zero. Therefore, the differential equation is not homogeneous. 4 −2 y (d): TRUE. Setting f (x, y) = xx2 +y 2 , we have f (tx, ty) = x4 y −2 t2 x4 y −2 (xt)4 (yt)−2 = = = f (x, y). (xt)2 + (yt)2 x 2 t 2 + y 2 t2 x2 + y 2 Therefore, f is homogeneous of degree zero, and therefore, the differential equation is homogeneous. (e): TRUE. This is verified in the calculation leading to Theorem 1.8.5. (f ): TRUE. This is verified in the calculation leading to (1.8.12). (g): TRUE. We can rewrite the equation as y − √ xy = √ xy 1/2 , √ √ which is the proper form for a Bernoulli equation, with p(x) = − x, q(x) = x, and n = 1/2. (h): FALSE. The presence of an exponential exy involving y prohibits this equation from having the proper form for a Bernoulli equation. (i): TRUE. After dividing the differential equation through by y, it becomes Bernoulli equation with p(x) = x, q(x) = x2 , and n = 2/3. Unless otherwise indicated in this section v = dy + xy = x2 y 2/3 , which is a dx y dy dv , =v+x and t > 0. x dx dx Problems: 5(xt) + 2(yt) t(5x + 2y) 5x + 2y = = = f (x, y). Thus, f is homogeneous of degree zero. 9(xt) − 4(yt) t(9x − 4y) 9x − 4y y 5 + 2x 5 + 2v = f (x, y) = = F (v). 9 − 4 xy 9 − 4v 1. f (tx, ty) = 2. f (tx, ty) = 2(xt) − 5(yt) = t(2x − 5y) = f (x, y). Thus, f is not homogeneous of degree zero. 3. f (tx, ty) = zero. f (x, y) = ty (tx) sin ( tx ty ) − (ty) cos ( tx ) y x sin xy − xy cos xy y x = = x sin xy − y cos xy y = f (x, y). Thus f is homogeneous of degree sin v1 − v cos v = F (v). v (c)2017 Pearson Education. Inc. 63 3(tx)2 + 5(ty)2 3x2 + 5y 2 4. f (tx, ty) = = = f (x, y). Thus f is homogeneous of degree zero. f (x, y) = 2(tx) + 5(ty) √ 2x + 5y y 3 + 5( x )2 3x2 + 5y 2 3 + 5V 2 = = = F (v). y 2x + 5y 2 + 5( x ) 2 + 5V 5. f (tx, ty) = tx + 7 x+7 = . Thus f is not homogeneous of degree zero. 2ty 2y tx − 2 5(ty) + 3 (3tx − 6) + (10ty + 6) t(3x + 10y) 3x + 10y x−2 5y + 3 + = = = = + . 2(ty) 3(ty) 6ty 6ty 6y 2y 3y 3x + 10y x 5 1 5 Thus, f is homogeneous of degree zero. We have f (x, y) = = + = + = F (v). 6y 2y 3 2v 3 6. f (tx, ty) = (tx)2 + (ty)2 x2 + y 2 7. f (tx, ty) = = = f (x, y). Thus f is homogeneous of degree zero. f (x, y) = tx x y 2 2 2 2 √ |x| 1 + ( x ) x +y y = − 1 + v 2 = F (v). = =− 1+ x x x (tx)2 + 4(ty)2 − (tx) + (ty) x2 + 4y 2 − x + y 8. f (tx, ty) = = = f (x, y). Thus f is homogeneous of x + 3y (tx) + 3(ty) √ 1 + 4( xy )2 − 1 + xy x2 + 4y 2 − x + y 1 + 4v 2 − 1 + v = = = F (v). degree zero. f (x, y) = y x + 3y 1 + 3x 1 + 3v 9. By inspection the differential equation is first-order homogeneous. We therefore let y = xV in which case y = xV + V . Substituting these results into the given differential equation yields xV + V = V 2 + V + 1, or equivalently, xV = V 2 + 1. Separating the variables and integrating yields 1 dV = V2+1 y 1 dx =⇒ arctan = ln |x| + c1 =⇒ y(x) = tan(x ln cx). x x dy y dy y dv dv 3v = 3y =⇒ (3 − 2 ) = 3 =⇒ (3 − 2v) v + x = 3v =⇒ x = − v =⇒ dx x dx x dx dx 3 − 2v dx 3 − 2v 3x 3x 3 y dv = =⇒ − − ln |v| = ln |x| + c1 =⇒ − − ln | | = ln |x| + c1 =⇒ ln y = − + c2 =⇒ 2 2v x 2v 2y x 2y y 2 = ce−3x/y . 10. (3x − 2y) dv dx dy dy dv (x + y)2 1 y 2 1 =⇒ =⇒ v + x = = 1+ = (1 + v)2 =⇒ = =⇒ tan−1 v = 2 2 dx 2x dx 2 x dx 2 v + 1 x y 1 1 ln |x| + c =⇒ tan−1 = ln |x| + c. 2 x 2 y y dy y dy x y dv = cos 12. sin x − y = x cos =⇒ sin − =⇒ sin v v + x −v = x dx x y dx x x dx y sin v dx dv = cos v =⇒ cos v =⇒ sin v x dv = =⇒ − ln | cos v| = ln |x| + c1 =⇒ x cos = c2 =⇒ dx cos v x x c y(x) = x cos−1 . x 11. (c)2017 Pearson Education. Inc. 64 √ dy 16x2 − y 2 + y dv dv dy 13. = = =⇒ = 16 − ( xy )2 + xy =⇒ v + x = 16 − v 2 + v =⇒ √ dx x dx dx 16 − v 2 dx y ) = ln |x| + c. =⇒ sin−1 ( v4 ) = ln |x| + c =⇒ sin−1 ( 4x x (9x2 + y 2 ) + y 14. We first rewrite the given differential equation in the equivalent form y = . Factoring x y 2 |x| 9 + ( x ) + y out an x2 from the square root yields y = . Since we are told to solve the differential x equation on the interval x > 0 we have |x| = x, so that y = 9 + ( xy )2 + xy , which we recognize as being homogeneous. We therefore√let y = xV , so that y = xV√ + V . Substitution into the preceding differential equation yields xV + V = 9 + V 2 + V , that is xV = 9 + V 2 . Separating the variables in this equation √ 1 1 dV = dx. Integrating we obtain ln (V + 9 + V 2 ) = ln c1 x. Exponentiating both sides we obtain √ 2 x 9+V √ y yields V + 9 + V 2 = c1 x. Substituting = V and multiplying through by x yields the general solution x y + 9x2 + y 2 = c1 x2 . 15. The given differential equation can be written in the equivalent form dy y(x2 − y 2 ) = , dx x(x2 + y 2 ) which we recognize as being first order homogeneous. The substitution y = xv yields v+x so that dv v(1 − v 2 ) 2v 3 dv =⇒ x , = =− 2 dx 1+v dx 1 + v2 1 + v2 dv = −2 v3 dx v −2 =⇒ − + ln |v| = −2 ln |x| + c1 . x 2 Consequently, − x2 + ln |xy| = c1 . 2y 2 dx dy dy y y dv dv + y ln x = y ln y =⇒ = ln =⇒ v + x = v ln v =⇒ =⇒ ln | ln v − 1| = dx dx x x dx v(ln v − 1) x ln xy − 1 ln |x| + c1 =⇒ = c =⇒ y(x) = xe1+cx . x 16. x dv dv dy y 2 + 2xy − 2x2 v 2 + 2v − 2 −v 3 + 2v 2 + v − 2 v2 − v + 1 =⇒ v+x =⇒ x = 2 = = =⇒ dv = 3 2 dx x − xy + y 2 dx 1 − v + v2 v2 − v + 1 v − 2v − v + 2 dx 2 dx dx dx v −v+1 1 1 1 dv = − − =⇒ dv = − =⇒ − + =⇒ x (v − 1)(v + 2)(v + 1) x v − 2 2(v − 1) 2(v + 1) x 2 (v − 2)2 (v + 1) 1 1 = −2 ln x+c2 =⇒ (y − 2x) (y + x) = ln |v − 2|− ln |v − 1|+ ln |v + 1| = − ln |x|+c1 =⇒ ln 2 2 v−1 y−x c. y 2 dv y dy 2 2 2 2 2 − (e−v + = 0 =⇒ 2v v + x − e−y /x + 2 18. 2xydy − (x2 e−y /x + 2y 2 )dx = 0 =⇒ 2 x dx x dx 17. (c)2017 Pearson Education. Inc. 65 2v 2 ) = 0 =⇒ 2vx x2 ln (ln (cx)). 2 dx dv 2 2 2 2 = e−v =⇒ ev (2vdv) = =⇒ ev = ln |x| + c1 =⇒ ey /x = ln (cx) =⇒ y 2 = dx x dy dy dv dv = y 2 + 3xy + x2 =⇒ = ( xy )2 + 3 xy + 1 =⇒ v + x = v 2 + 3v + 1 =⇒ x = (v + 1)2 =⇒ 19. x2 dx dx dx dx dx 1 y dv 1 1 = =⇒ − = ln |x| + c1 =⇒ − y = ln |x| + c1 =⇒ = − =⇒ y(x) = (v + 1)2 x v + 1 + 1 x ln (cx) x 1 . −x 1 + ln (cx) √ √ 1 + ( xy )2 − 1 dy x2 + y 2 − x 1 + v2 − 1 + v2 − x dv dy dv 20. =⇒ v + x = =⇒ = = =⇒ x = =⇒ y dx y dx dx v dx v x dx c v √ dv = =⇒ ln |1 − u| = ln |x| + c1 =⇒ |x(1 − u)| = c2 =⇒ 1 − u = =⇒ u2 = x x 1 + v2 − 1 − v2 c2 c c2 c 2 − 2 = − 2 =⇒ y 2 = c2 − 2cx. + 1 =⇒ v x2 x x2 x y dy dv dv dy = v(4 − v) =⇒ 2x 21. 2x(y + 2x) = y(4x − y) =⇒ 2 +2 = xy (4 − xy ) =⇒ 2(v + 2) v + x = dx x dx dx dx dx v+2 3v 2 dv = −3 =⇒ 2 =⇒ 2 ln |v| − v4 = −3 ln |x| + c1 =⇒ y 2 = cxe4x/y . − v+2 v2 x dx dv dy dv = x tan ( xy ) + y =⇒ v + x = tan v + v =⇒ x = tan v =⇒ cot vdv = =⇒ ln | sin v| = dx dx dx x −1 −1 ln |x| + c1 =⇒ sin v = cx =⇒ v = sin (cx) =⇒ y(x) = x sin (cx). 22. x 2 x x2 + y 2 + y 2 y dv dy dv +1 + ( v1 )2 + 1 + v =⇒ 23. =⇒ = =⇒ v + x = = xy dx y x dx dx √ √ dv dx ( v1 )2 + 1 =⇒ = =⇒ ln |v + 1 + v 2 | = ln |x|+c =⇒ v + 1 + v 2 = cx =⇒ xy + 1 + ( xy )2 = x ( 1 )2 dy x = dx v [(cx)2 − 1] y xc =⇒ 2( x ) + 1 = (cx)2 =⇒ y 2 = x2 . 2 24. The given differential equation can be written as (x−4y)dy = (4x+y)dx. Converting to polar coordinates we have x = r cos θ =⇒ dx = cos θdr − r sin θdθ, and y = r sin θdr + r cos θdθ. Substituting these results into the preceding differential equation and simplifying yields the separable equation 4r−1 dr = dθ which can be integrated directly to yield 4 ln r = θ + c, so that r = c1 eθ/4 . x dy 2(2y − x) dy y +1 = . Since x = 0, divide the numerator and denominator by y yields = . dx x+y dx 2(2 − xy ) 2(2 − v) dy dx dv dv v+1 = v+y =⇒ v + y = =⇒ dv = =⇒ Now let v = xy so that 2 dy dy dy 2(2 − v) 2v − 3v + 1 y dv dv (v − 1)2 −6 = ln (c2 |y|) =⇒ (x − y)2 = c(y − 2x)3 . Since y(0) = 2 +2 = ln |y| + c1 =⇒ ln 2v − 1 v−1 |2v − 1|3 1 1 then c = . Thus, (x − y)2 = (y − 2x)3 . 2 2 25. (c)2017 Pearson Education. Inc. 66 2 − xy dy dv 2x − y dy 2−v dv 2 − 2v − 4v 2 1 1 + 4v = =⇒ = = =⇒ x = =⇒ dv = y =⇒ v+x 2 dx x + 4y dx 1 + 4x dx 1 + 4v dx 1 + 4v 2 2v + v − 1 dx 1 1 1 =⇒ ln |2v 2 + v − 1| = − ln |x| + c =⇒ ln |x2 (2v 2 + v − 1)| = c =⇒ ln |2y 2 + yx − x2 | = c, but − x 2 2 2 1 1 1 2 2 y(1) = 1 so c = ln 2. Thus ln |2y + yx − x | = ln 2 and since y(1) = 1 it must be the case that 2 2 2 2y 2 + yx − x2 = 2. 26. √ dx y dv dv dy y − x2 + y 2 dy y =− 27. = =⇒ = − 1 + ( )2 =⇒ x = − 1 + v 2 =⇒ √ =⇒ 2 dx x dx x x dx x 1+v √ |x| 2 |x| ln (v + 1 + v 2 ) = − ln |x| + c1 =⇒ y + x + y 2 = c2 . Since y(3) = 4 then c2 = 9. Then take =1 x x since we must have y(3) = 4; thus y + x2 + y 2 = 9. y 2 √ dx dy v dv dv y 28. =⇒ v + x = − = 4− = v + 4 − v 2 =⇒ √ =⇒ sin−1 = ln |x| + c =⇒ dx x x dx x 2 4 − v2 −1 y sin = ln x + c since x > 0. 2x x + ay dv 1 + v2 dy = . Substituting y = xv and simplifying yields x = . Separating the variables dx ax − y dx a−v 1 and integrating we obtain a tan−1 v − ln (1 + v 2 ) = ln x + ln c or equivalently, a tan−1 xy − 12 ln (x2 + y 2 ) = 2 ln c. Substituting for x = r cos θ, y = r sin θ yields aθ − ln r = ln c. Exponentiating then gives r = keaθ . √ (b). The initial condition y(1) = 1 corresponds to r( π4 ) = 2. Imposing this condition on the polar form √ −π/8 of the solution obtained in (a) yields k = 2e . Hence, the solution to the initial value problem is √ (θ−π/4)/2 1 dy 2x + y . When a = , the differential equation is r = 2e = . Consequently every solution 2 dx x − 2y x curve has a vertical tangent line at points of intersection with the line y = . The maximum interval of 2 x existence for the solution of the initial value problem can be obtained by determining where y = intersects 2 √ (θ−π/4)/2 x 1 . The line y = has a polar equation tan θ = . The corresponding values of θ the curve r = 2e 2 2 1 are θ = θ1 = tan−1 ≈ 0.464, θ = θ2 = θ1 + π ≈ 3.61. Consequently, the x-coordinates of the intersection 2 √ √ points are x1 = r cos θ1 = 2e(θ1 −π/4)/2 cos θ1 ≈ 1.08, x2 = r cos θ2 = 2e(θ2 −π/4)/2 cos θ2 ≈ −5.18. Hence the maximum interval of existence for the solution is approximately (−5.18, 1.08). 29. (a). (c). See the accompanying figure. x2 + y 2 dy dy dy . Hence 2x + 2y = 2c =⇒ = 2y dx dx dx x dy 2xy y 2 − x2 dy dv . Orthogonal trajectories satisfies: = 2 = . Let y = vx so that = v+x . 2 c−y x −y dx 2xy dx dx dv v2 + 1 1 Substituting these results into the last equation yields x =− =⇒ ln |v + 1| = − ln |x| + c1 =⇒ dx 2v 2 y c2 +1= =⇒ x2 + y 2 = 2kx. 2 x x 30. Given family of curves satisfies: x2 + y 2 = 2cy =⇒ c (c)2017 Pearson Education. Inc. 67 y(x) 2 1 -6 -5 -4 -3 -2 -1 1 0 x -1 -2 -3 -4 Figure 0.0.50: Figure for Problem 29(c) y(x) x Figure 0.0.51: Figure for Problem 30 x2 + y 2 dy . Hence 2(x−c)+2(y −c) = 2(x + y) dx c−x dy y 2 − 2xy − x2 y 2 + 2xy − x2 0 =⇒ . Orthogonal trajectories satisfies: . Let y = vx so = 2 = y−c y + 2xy − x2 dx x2 + 2xy − y 2 dy dv dv v 2 + 2v − 1 that =⇒ = v + x . Substituting these results into the last equation yields v + x = dx dx dx 1 + 2v − v 2 31. Given family of curves satisfies: (x−c)2 +(y −c)2 = 2c2 =⇒ c = (c)2017 Pearson Education. Inc. 68 1 + 2v − v 2 1 dv = dx =⇒ 3 2 v −v +v−1 x 2k 2 . 1 2v − 2 v−1 v +1 dv = 1 dx =⇒ x2 +y 2 = 2k(x−y) =⇒ (x−k)2 +(y +k)2 = x y(x) x Figure 0.0.52: Figure for Problem 31 32. (a). Let r represent the radius √ of one of the circles with center at (a.ma) and passing through (0, 0). r =√ (a − 0)2 + (ma − 0)2 = |a| 1 + m2 . Thus, the circle’s equation can be written as (x−a)2 +(y−ma)2 = (|a| 1 + m2 )2 or (x − a)2 + (y − ma)2 = a2 (1 + m2 ). x2 + y 2 . Differentiating the first equation with respect 2(x + my) dy a−x dy y 2 − x2 − 2mxy x and solving we obtain = . Substituting for a and simplifying yields = . dx y − ma dx my 2 − mx2 + 2xy y y 2 m − m( ) − 2 x mx2 − my 2 − 2xy dy dy . Let y = vx so that = 2 =⇒ = y 2 x Orthogonal trajectories satisfies: 2 dx y − x − 2mxy dx ( x ) − 1 − 2m xy dy dv dv m − mv 2 − 2v xdv = v + x . Substituting these results into the last equation yields v + x = 2 =⇒ = dx dx dx v − 1 − 2mv dx 2 2 dx v − 2mv − 1 dx dv 2v (m − v)(1 + v ) dv = =⇒ dv = =⇒ − =⇒ ln |v − m| − v 2 − 2mv − 1 (m − v)(1 + v 2 ) x v−m 1 + v2 x 2 2 2 2 2 2 ln (1 + v ) = ln |x| + c1 =⇒ v − m = c2 x(1 + v ) =⇒ y − mx = c2 x + c2 y =⇒ x + y + cmx − cy = 0. Completing the square we obtain (x + cm/2)2 + (y − c/2)2 = c2 /4(m2 + 1). Now letting b = c/2, the last equation becomes (x + bm)2 + (y − b)2 = b2 (m2 + 1) which is a family or circles lying on the line y = −my and passing through the origin. (b). (x − a)2 + (y − ma)2 = a2 (1 + m2 ) =⇒ a = (c). See the accompanying figure. 33. x2 + y 2 = c =⇒ m2 − tan ( π4 ) dy x −x/y − 1 x+y = − = m2 . m1 = = . Let y = vx so that π = dx y 1 + m2 tan ( 4 ) 1 − x/y x−y (c)2017 Pearson Education. Inc. 69 y(x) x Figure 0.0.53: Figure for Problem 32(c) dy dv dv 1+v 1−v dv = = v + x . Substituting these results into the last equation yields v + x = =⇒ dx dx 1−v 1 + v2 dx dx 1 dx v 1 dv = + 2 =⇒ − =⇒ − ln (1 + v 2 )+tan−1 v = ln |x|+c1 =⇒ Oblique trajectories: 2 x 1+v v +1 x 2 ln (x2 + y 2 ) − 2 tan−1 (y/x) = c2 . m2 − tan ( π4 ) dy 6y/x − 1 6y − x = 6y/x = m2 . m1 = = = . Let y = vx so that dx 1 + m2 tan ( π4 ) 1 + 6y/x 6y + x dv dv 6v − 1 dv dy = v + x . Substitute these results into the last equation yields v + x = =⇒ x = dx dx dx 6v + 1 dx dx 9 8 (3v − 1)(1 − 2v) dv = =⇒ − =⇒ 3 ln |3v − 1| − 4 ln |2v − 1| = ln |x| + c1 =⇒ 6v + 1 3v − 1 2v − 1 x Oblique trajectories (3y − x)3 = k(2y − x)4 . 34. y = cx6 =⇒ 35. x2 + y 2 = 2cx =⇒ c = y 2 − x2 − 2xy . y 2 − x2 + 2xy 2 x +y 2x 2 2 2 m2 − tan ( π4 ) dy y −x and = = m2 . m1 = = dx 2xy 1 + m2 tan ( π4 ) y 2 − x2 −1 2xy = y 2 − x2 1+ 2xy dy dv = v + x . Substituting these results into the last equadx dx dv v 2 − 2v − 1 dv −v 3 − v 2 − v − 1 −v 2 − 2v + 1 dx tion yields v + x = 2 =⇒ x = =⇒ dv = =⇒ 2 3 2 dx v + 2v − 1 dx v + 2v − 1 v +v +v+1 x dx 2v 1 dv = − 2 =⇒ ln |v + 1| − ln (v 2 + 1) = ln |x| + c1 =⇒ ln |y + x| = ln |y 2 + x2 | + c1 =⇒ v+1 v +1 x Oblique trajectories: x2 + y 2 = 2k(x + y) or, equivalently, (x − k)2 + (y − k)2 = 2k 2 . Let y = vx so that (c)2017 Pearson Education. Inc. 70 dy y m2 − tan α0 −y/x − tan α0 = . Let y = vx so = −cx−2 = − . m1 = dx x 1 + m2 tan α0 1 − y/x tan α0 dy dv dv tan α0 + v that = v + x . Substituting these results into the last equation yields v + x = =⇒ dx dx dx v tan α0 − 1 2v tan α0 − 2 2dx dv = − =⇒ ln |v 2 tan α0 − 2v − tan α0 | = −2 ln |x| + c1 =⇒ (y 2 − x2 ) tan α0 − v 2 tan α0 − 2v − tan α0 x 2xy = k. 36. (a). y = cx−1 =⇒ (b). See the accompanying figure. y(x) x Figure 0.0.54: Figure for Problem 36(b) dy m2 − tan α0 −x/y − m x x + my dy = = − . m1 = = . Let y = vx so that = dx y 1 + m2 tan α0 1 − (x/y)m mx − y dx dv dv 1 + mv dv 1 + v2 v + x . Substituting these results into the last equation yields v + x = =⇒ x = =⇒ dx dx m−v dx m−v dx dx m v 1 v−m dv = − dv = − − =⇒ =⇒ ln (1 + v 2 ) − m tan v = − ln |x| + c1 . In polar 1 + v2 x 1 + v2 1 + v2 x 2 coordinates, r = x2 + y 2 and θ = tan−1 y/x, so this result becomes ln r − mθ = c1 =⇒ r = emθ where k is an arbitrary constant. 37. (a). x2 +y 2 = c =⇒ (b). See the accompanying figure. dy 1 dy 1 2 − y = 4x2 y −1 cos x. This is a Bernoulli equation. Multiplying both sides y results in y − y = dx x dx x du dy dy 1 du dy 1 4x2 cos x. Let u = y 2 so = 2y or y = . Substituting these results into y − y 2 = 4x2 cos x dx dx dx 2 dx dx x d −2 du 2 yields − u = 8x2 cos x which has an integrating factor I(x) = x−2 =⇒ (x u) = 8 cos x =⇒ x−2 u = dx x dx 8 cos xdx + c =⇒ x−2 u = 8 sin x + c =⇒ u = x2 (8 sin x + c) =⇒ y 2 = x2 (8 sin x + c). 38. 39. y −3 dy dy du 1 + y −2 tan x = 2 sin x. This is a Bernoulli equation. Let u = y −2 so = −2y −3 or dx 2 dx dx (c)2017 Pearson Education. Inc. 71 y(x) x Figure 0.0.55: Figure for Problem 37(b) dy 1 du du = . Substituting these results into the last equation yields −u tan x = −4 sin x. An integrating dx 2 dx dx d (u cos x) = −4 cos x sin x =⇒ u cos x = 4 cos x sin xdx =⇒ factor for this equation is I(x) = cos x. Thus, dx c 1 (cos2 x + c) =⇒ y −2 = 2 cos x + . u(x) = cos x cos x y −3 dy 1 dy 3 2/3 dy dy 3 2 = 6x2 ln x. Let u = y 2/3 =⇒ − y = 6y 1/3 x2 ln x or 1/3 − y = y −1/3 . Substituting dx 2x dx 2x dx 3 dx y du 3 2/3 1 1 dy = 6x2 ln x yields − y − u = 4x2 ln x. An integrating factor for this these results into 1/3 dx 2x dx x y 1 d −1 −1 equation is I(x) = so (x u) = 4x ln x =⇒ x u = 4 x ln xdx + c =⇒ x−1 u = 2x2 ln x − x2 + c =⇒ x dx u(x) = x(2x2 ln x − x2 + c) =⇒ y 2/3 = x(2x2 ln x − x2 + c). 40. √ √ dy dy dy du 2 2 + y = 6 1 + x2 y 1/2 or y −1/2 + y 1/2 = 6 1 + x2 . Let u = y 1/2 =⇒ 2 = y −1/2 . dx x dx x dx dx √ √ dy du 2 1 + y 1/2 = 6 1 + x2 yields + u = 3 1 + x2 . An integrating Substituting these results into y −1/2 dx x dx x √ √ d 2 factor for this equation is I(x) = x so (xu) = 3x 1 + x =⇒ xu = x 1 + x2 dx + c =⇒ xu = dx 1 c 1 c (1 + x2 )3/2 + c =⇒ u = (1 + x2 )3/2 + =⇒ y 1/2 = (1 + x2 )3/2 + . x x x x 41. dy dy du dy 2 2 + y = 6y 2 x4 or y −2 + y −1 = 6x4 . Let u = y −1 =⇒ − = y −2 . Substituting these results dx x dx x dx dx du 2 2 dy + y −1 = 6x4 yields − u = −6x4 . An integrating factor for this equation is I(x) = x−2 so into y −2 dx x dx x 1 d −2 (x u) = −6x2 =⇒ x−2 u = −2x3 + c =⇒ u = −2x5 + cx2 =⇒ y −1 = −2x5 + cx2 =⇒ y(x) = 2 . dx x (c − 2x3 ) 42. (c)2017 Pearson Education. Inc. 72 dy + y 3 x2 dx dy dy 1 du 1 −2 + y = −x−2 . Let u = y −2 =⇒ − = y −3 . Substituting dx 2x 2 dx dx du 1 1 −2 −3 dy −2 2 + y = −x yields − u = 2x . An integrating factor for this equation is these results into y dx 2x dx x 1 d −1 I(x) = so (x u) = 2x =⇒ x−1 u = x2 + c =⇒ u = x3 + cx =⇒ y −2 = x3 + cx. x dx dy dy du 2(b − a) 44. (x − a)(x − b) − y 1/2 = 2(b − a)y or y −1/2 − y 1/2 = 1. Let u = y 1/2 =⇒ 2 = dx dx (x − a)(x − b) dx dy dy du 2(b − a) (b − a) 1 y 1/2 = 1 yields − u= . y −1/2 . Substituting these results into y −1/2 − dx dx (x − a)(x− b) dx (x − 1)(x − b) 2 d x−a x−a x−a 1 x−a so u = =⇒ u = [x+(b− An integrating factor for this equation is I(x) = x−b dx x − b 2(x − b) x−b 2 2 x−b 1 x−b [x+(b−a) ln |x − b|+c]2 . a) ln |x − b|+c] =⇒ y 1/2 = [x+(b−a) ln |x − b|+c] =⇒ y(x) = 2(x − a) 4 x−a 43. 2x + y = 0 or y −3 dy cos x dy dy cos x du 6 6 + y = 3y 2/3 or y −2/3 + y 1/3 = 3 . Let u = y 1/3 =⇒ 3 = y −2/3 . Substituting dx x x dx x x dx dx dy 6 1/3 cos x du 2 cos x =3 + y yields + u= . An integrating factor for this equation these results into y −2/3 dx x x dx x x d 2 cos x + x sin x + c . (x u) = x cos x =⇒ x2 u = cos x + x sin x + c =⇒ y 1/3 = is I(x) = x2 so dx x2 45. dy dy dy du + 4xy = 4x3 y 1/2 or y −1/2 + 4xy 1/2 = 4x3 . Let u = y 1/2 =⇒ 2 = y −1/2 . Substituting these dx dx dx dx dy du + 4xy 1/2 = 4x3 yields + 2xu = 2x3 . An integrating factor for this equation is I(x) = results into y −1/2 dx dx d x2 2 2 2 2 2 2 ex so (e u) = 2ex x3 =⇒ ex u = ex (x2 −1)+c =⇒ y 1/2 = x2 −1+ce−x =⇒ y(x) = [(x2 −1)+ce−x ]2 . dx 46. dy dy dy 1 du 1 1 − = 2xy 3 or y −3 − y −2 = 2x. Let u = y −2 =⇒ = y −3 . Substituting these dx 2x ln x dx 2x ln x 2 dx dx dy du 1 1 − y −2 = 2x yields + u = −4x. An integrating factor for this equation is results into y −3 dx 2x ln x dx x ln x ln x d . (u ln x) = −4x ln x =⇒ u ln x = x2 − 2x2 ln x + c =⇒ y 2 = 2 I(x) = ln x so dx x (1 − 2 ln x) + c 47. dy dy 3x 1 du 1 3 1 dy − y= xy π or y −π − y 1−π = . Let u = y 1−π =⇒ = y −π . dx (π − 1)x (1 − π) dx (π − 1)x 1−π 1 − π dx dx 3x 1 du 1 −π dy 1−π = Substituting these results into y − y yields + u = 3x. An integrating factor for dx (π − 1)x 1−π dx x 1/(1−π) 3 x3 + c x +c d 2 3 1−π = . (xu) = 3x =⇒ xu = x + c =⇒ y =⇒ y(x) = this equation is I(x) = x so dx x x 48. dy dy du dy +y cot x = 8y −1 cos3 x or 2y +y 2 cot x = 8 cos2 x. Let u = y 2 =⇒ = 2y . Substituting these dx dx dx dx du dy + y 2 cot x = 8 cos2 x yields + u sec x = sec x. An integrating factor for this equation is results into 2y dx dx −2 cos4 x + c d 3 I(x) = sin x so (u sin x) = 8 cos x sin x =⇒ u sin x = −2 cos4 x + c =⇒ y 2 = . dx sin x 49. 2 (c)2017 Pearson Education. Inc. 73 √ √ dy √ √ √ √ dy du 50. (1− 3) +y sec x = y 3 sec x or (1− 3)y − 3 +y 1− 3 sec x = sec x. Let u = y 1− 3 =⇒ = (1− dx dx dx √ √ −√3 dy √ −√3 dy du 3)y . Substituting these results into (1− 3)y +y 1− 3 sec x = sec x yields +u sec x = sec x. dx dx dx d [(sec x+tan x)u] = sec x(sec x+tan x) =⇒ An integrating factor for this equation is I(x) = sec x+tan x so dx 1/(1−√3) √ 1 c 1− 3 =1+ . =⇒ y(x) = 1 + (sec x + tan x)u = tan x + sec x + c =⇒ y sec x + tan x sec x + tan x dy dy 1 dy du 2x 2x 1 y = xy 2 or 2 + + = x. Let u = y −1 so = −y −2 . Substituting these dx 1 + x2 y dx 1 + x2 y dx dx 2x 1 du 2x 1 dy u = −x. An integrating factor for this equation is + = x yields − results into 2 y dx 1 + x2 y dx 1 + x2 xdx u x d u u 1 1 =− so =⇒ =− +c =⇒ = − ln (1 + x2 )+c =⇒ I(x) = 2 2 2 2 1 + x2 dx 1 + x2 1 + x 1 + x 1 + x 1 + x 2 1 1 u = (1 + x2 ) − ln (1 + x2 ) + c =⇒ y −1 = (1 + x2 ) − ln (1 + x2 ) + c . Since y(0) = 1 =⇒ c = 1 so 2 2 1 1 = (1 + x2 ) − ln (1 + x2 ) + 1 . y 2 51. dy dy 1 du dy + y cot x = y 3 sin3 x or y −3 + y −2 cot x = sin3 x. Let u = y −2 =⇒ − = y −3 . Substituting dx dx 2 dx dx dy du + y −2 cot x = sin3 x yields − 2u cot x = −2 sin3 x. An integrating factor for this these results into y −3 dx dx d equation is I(x) = csc2 x so (u csc2 x) = − sin x =⇒ u csc2 x = 2 cos x + c. Since y(π/2) = 1 =⇒ c = 1. dx 1 Thus y 2 = . sin2 x(2 cos x + 1) 52. dy dv dy dy dv dy 53. = F (ax + by + c). Let v = ax + by + c so that = a+b =⇒ b = − a =⇒ = dx dx dx dx dx dx dv dv dv 1 dv − a = F (v) =⇒ − a = bF (v) =⇒ = bF (v) + a =⇒ = dx. b dx dx dx bf (v) + a 54. dv dy dy dv dv = dx =⇒ = (9x − y)2 . Let v = 9x − y so that = 9− =⇒ = 9 − v 2 =⇒ 2 dx dx dx dx 9−v 1 tanh−1 (v/3) = x + c1 but y(0) = 0 so c = 0. Thus, tanh−1 (3x − y/3) = 3x or y(x) = 3(3x − tanh 3x). 3 55. dv dy dv dy dv = (4x + y + 2)2 . Let v = 4x + y + 2 so that = 4+ =⇒ 2 = dx =⇒ = dx =⇒ 2 dx dx dx v +4 v +4 1 tan−1 v/2 = x + c1 =⇒ tan−1 (2x + y/2 + 1) = 2x + c =⇒ y(x) = 2[tan (2x + c) − 2x − 1]. 2 dy dy 1 dv 1 dv 56. = sin2 (3x − 3y + 1). Let v = 3x − 3y + 1 so that 1− =⇒ 1 − = sin2 v =⇒ dx dx 3 dx 3 dx dv = 3 cos2 v =⇒ sec2 vdv = 3 dx =⇒ tan v = 3x + c =⇒ tan (3x − 3y + 1) = 3x + c =⇒ y(x) = dx 1 [3x − tan−1 (3x + c) + 1]. 3 (c)2017 Pearson Education. Inc. 74 57. V = xy =⇒ V = xy + y =⇒ y = (V − y)/x. Substitution into the differential equation yields 1 dV 1 (V − y)/x = yF (V )/x =⇒ V = y[F (V ) + 1] =⇒ V = V [F (V ) + 1]/x, so that = . V [F (V ) + 1] dx x 1 dV 1 1 1 = for F (V ) = ln V − 1 yields dV = dx =⇒ ln ln V = V [F (V ) + 1] dx x V ln V x 1 ln cx =⇒ V = ecx =⇒ y(x) = ecx . x 58. Substituting into 59. (a). y(x) = w(x)−x =⇒ y = w −1. Substituting these results into (1.8.18) yields w −1 = 2xw2 −1 =⇒ w = 2xw2 . (b). Separating the variables in the preceding differential equation and integrating yields 1 1 dw = 2 x dx + c =⇒ −w−1 = x2 + c =⇒ w(x) = , w2 c 1 − x2 where c1 = −c. Hence, the general solution to (1.8.18) is y(x) = y(0) = 1 requires that 1 = 1 − x. Imposing the initial condition c 1 − x2 1 1 =⇒ c1 = 1. Therefore, y(x) = − x. c1 1 − x2 dy dv 60. (a). x = u − 1, y = v + 1 =⇒ = . Substitution into the given differential equation yields dx du u + 2v dv = . du 2u − v (b). The differential equation obtained in (a) is first order homogeneous. We therefore let W = v/u, and 1 + 2W 1 + W2 substitute into the differential equation to obtain W u + W = =⇒ W u = . Separating the 2−W 2−W 1 W 2 dW = du. This can be integrated directly to obtain 2 tan−1 W − − variables yields 1 + W2 1 + W2 u 1 −1 −1 ln (1 + W 2 ) = ln u + ln c. Simplifying we obtain cu2 (1 + W 2 ) = e4 tan W =⇒ c(u2 + v 2 ) = etan (v/u) . 2 −1 Substituting back in for x and y yields c[(x + 1)2 + (y − 1)2 ] = etan [(y−1)/(x+1)] . 61. (a). y = Y (x) + v −1 (x) =⇒ y = Y (x) − v −2 (x)v (x). Now substitute into the given differential equation and simplify algebraically to obtain Y (x) + p(x)Y (x) + q(x)Y 2 (x) − v −2 (x)v (x) + v −1 (x)p(x) + q(x)[2Y (x)v −1 (x) + v −2 (x)] = r(x). We are told that Y (x) is a particular solution to the given differential equation, and therefore Y (x) + p(x)Y (x) + q(x)Y 2 (x) = r(x). Consequently the transformed differential equation reduces to −v −2 (x)v (x) + v −1 p(x) + q(x)[2Y (x)v −1 (x) + v −2 (x)] = 0, or equivalently v − [p(x) + 2Y (x)q(x)]v = q(x). (b). The given differential equation can be written as y − x−1 y − y 2 = x−2 , which is a Riccati differential equation with p(x) = −x−1 , q(x) = −1, and r(x) = x−2 . Since y(x) = −x−1 is a solution to the given differential equation, we make a substitution y(x) = −x−1 + v −1 (x). According to the result from part (a), the given differential equation then reduces to v −(−x−1 +2x−1 )v = −1, or equivalently v −x−1 v = −1. This d −1 linear differential equation has an integrating factor I(x) = x−1 , so that v must satisfy (x v) = −x−1 =⇒ dx 1 1 1 1 v(x) = x(c−ln x). Hence the solution to the original equation is y(x) = − + = −1 . x x(c − ln x) x c − ln x 62. (a). If y = axr , then y = arxr−1 . Substituting these expressions into the given differential equation (c)2017 Pearson Education. Inc. 75 yields arxr−1 + 2axr−1 − a2 x2r = −2x−2 . For this to hold for all x > 0, the powers of x must match up on either side of the equation. Hence, r = −1. Then a is determined from the quadratic −a + 2a − a2 = −1 ⇐⇒ a2 − a − 2 = 0 ⇐⇒ (a − 2)(a + 1) = 0. Consequently, a = 2, −1 in order for us to have a solution to the given differential equation. Therefore, two solutions to the differential equation are y1 (x) = 2x−1 , y2 (x) = −x−1 . (b). Taking Y (x) = 2x−1 and using the result from Problem 61(a), we now substitute y(x) = 2x−1 + v −1 into the given Riccati equation. The result is (−2x−2 −v −2 v )+2x−1 (2x−1 +v −1 )−(4x−2 +4x−1 v −1 +v −2 ) = −2x−2 . Simplifying this equation yields the linear equation v + 2x−1 v = −1. Multiplying by the integrating d 2 −1 (x v) = −x2 . Integrating this factor I(x) = e 2x dx = x2 results in the integrable differential equation dx 1 1 differential equation we obtain v(x) = x2 − x3 + c = x2 (c1 − x3 ). Consequently, the general solution 3 3 2 3 to the Riccati equation is y(x) = + 2 . x x (c1 − x3 ) 63. (a). y = x−1 + w(x) =⇒ y = −x−2 + w . Substituting into the given differential equation yields (−x−2 + w ) + 7x−1 (x−1 + w) − 3(x−2 + 2x−1 w + w2 ) = 3x−2 which simplifies to w + x−1 w − 3w2 = 0. (b). The preceding equation can be written in the equivalent form w−2 w + x−1 w−1 = 3. We let u = w−1 , so that u = −w−2 w . Substitution into the differential equation gives, after simplification, u − x−1 u = −3. An integrating factor for this linear differential equation is I(x) = x−1 , so that the differential equation d −1 can be written in the integrable form (x u) = −3x−1 . Integrating we obtain u(x) = x(−3 ln x + dx 1 c), so that w(x) = . Consequently the solution to the original Riccati equation is y(x) = x(c − 3 ln x) 1 1 1+ . x c − 3 ln x dy du 1 dy du + p(x) ln y = q(x). If we let u = ln y, then = and the given equation becomes + dx dx y dx dx e p(x)dx q(x)dx + c . p(x)u = q(x) which is a first order linear and has a solution of the form u = e− p(x)dx p(x)dx −1 q(x)dx + c into u = ln y we obtain y(x) = eI [ I(t)q(t)dt+c] where e Substituting ln y = e− p(x)dx 64. y −1 I(x) = e p(t)dt and c is an arbitrary constant. dy 2 1 − 2 ln x du 2 − ln y = . Let u = ln y so using the technique of the preceding problem: − u= dx x x⎡ dx x ⎤ dx dx 1 − 2 ln x −2 2 1 − 2 ln x 1 − 2 ln x 2 x ⎣ e x dx + c1 ⎦ = x2 dx + c and u = e 1 = ln x + cx , x x x3 65. y −1 2 and since u = ln y, ln y = ln x + cx2 . Now y(1) = e so c = 1 =⇒ y(x) = xex . du dy dy du = f (y) and the given equation f (y) + p(x)f (y) = q(x) becomes + dx dx dx dx e p(x)dx q(x)dx + c . Substituting p(x)u = q(x) which has a solution of the form u(x) = e− p(x)dx f (y) = e− p(x)dx e p(x)dx q(x)dx + c into u = f (y) and using the fact that f is invertible, we obtain y(x) = f −1 I −1 I(t)q(t)dt + c where I(x) = e p(t)dt and c is and arbitrary constant. 66. If u = f (y), then (c)2017 Pearson Education. Inc. 76 dy dy 1 du 1 tan y = √ . Let u = tan y so that + √ = sec2 y and the given equation dx dx dx 2 1+x 2 1+x 1 1 du u= √ which is first order linear. An integrating factor for this equation is + √ becomes dx 2 1+x 2 1+x√ √ √ √ √ √ e 1+x d √1+x e 1+x 1+x 1+x √ √ I(x) = e =⇒ u) = u = =⇒ e =⇒ e 1+x u = e 1+x + c =⇒ u = (e dx 2 1+x √ 2 1+x √ √ − 1+x − 1+x 1 + ce . But u = tan y so tan y = 1 + ce or y(x) = tan−1 (1 + ce− 1+x ). 67. sec2 y Solutions to Section 1.9 True-False Review: (a): FALSE. The requirement, as stated in Theorem 1.9.4, is that My = Nx , not Mx = Ny , as stated. (b): FALSE. A potential function φ(x, y) is not an equation. The general solution to an exact differential equation takes the form φ(x, y, ) = c, where φ(x, y) is a potential function. (c): FALSE. According to Definition 1.9.2, M (x)dx + N (y)dy = 0 is only exact if there exists a function φ(x, y) such that φx = M and φy = N for all (x, y) in a region R of the xy-plane. (d): TRUE. This is the content of part 1 of Theorem 1.9.11. (e): FALSE. If φ(x, y) is a potential function for M (x, y)dx + N (x, y)dy = 0, then so is φ(x, y) + c for any constant c. (f ): TRUE. We have My = 2e2x − cos y and Nx = 2e2x − cos y, and so since My = Nx , this equation is exact. (g): FALSE. We have My = and Nx = (x2 + y)2 (−2x) + 4xy(x2 + y) (x2 + y)4 (x2 + y)2 (2x) − 2x2 (x2 + y)(2x) . (x2 + y)4 Thus, My = Nx , and so this equation is not exact. (h): FALSE. We have My = 2y and Nx = 2y 2 , and since My = Nx , we conclude that this equation is not exact. (i): FALSE. We have My = ex sin y cos y + xex sin y cos y and Nx = cos y sin yex sin y , and since My = Nx , we conclude that this equation is not exact. Problems: 1. yexy dx + (2y − xe−xy )dy = 0. M = yexy and N = 2y − xe−xy =⇒ My = yxexy + exy and Nx = xye−xy − e−xy =⇒ My = Nx =⇒ the differential equation is not exact. (c)2017 Pearson Education. Inc. 77 2. [cos (xy) − xy sin (xy)]dx − x2 sin (xy)dy = 0 =⇒ M = cos (xy) − xy sin (xy) and N = −x2 sin (xy) =⇒ My = −2x sin (xy) − x2 y cos (xy) and Nx = −2x sin (xy) − x2 y cos (xy) =⇒ My = Nx =⇒ the differential equation is exact. 3. (y + 3x2 )dx + xdy = 0. M = y + 3x2 and N = x =⇒ My = 1 and Nx = 1 =⇒ My = Nx =⇒ the differential equation is exact. 4. 2xey dx + (3y 2 + x2 ey )dy = 0. M = 2xey and N = 3y 2 + x2 ey =⇒ My = 2xey and Nx = 2xey =⇒ My = Nx =⇒ the differential equation is exact. 5. 2xydx + (x2 + 1)dy = 0. M = 2xy and N = x2 + 1 =⇒ My = 2x and Nx = 2x =⇒ My = Nx =⇒ ∂φ ∂φ the differential equation is exact so there exists a potential function φ such that (a) = 2xy and (b) = ∂x ∂x dh(x) dh(x) dh(x) so from (a), 2xy = 2xy + =⇒ = 0 =⇒ h(x) is a constant. Since we need just one 2xy + dx dx dx 2 potential function, let h(x) = 0. Thus, φ(x, y) = (x + 1)y; hence, (x2 + 1)y = c. 6. Given (y 2 − 2x)dx + 2xydy = 0 then My = Nx = 2xy so the differential equation is exact and there exists ∂φ ∂φ a potential function φ such that (a) = y 2 − 2x and (b) = 2xy. From (b) φ(x, y) = xy 2 + h(x) =⇒ ∂x ∂y ∂φ dh(x) dh(x) dh(x) = y2 + so from (a) y 2 + = y 2 − 2x =⇒ = −2x =⇒ h(x) = −2x where the constant of ∂x dx dx dx integration has been set to zero since we just need one potential function. φ(x, y) = xy 2 −x2 =⇒ xy 2 −x2 = c. 7. Given (4e2x + 2xy − y 2 )dx + (x − y)2 dy = 0 then My = Nx = 2y so the differential equation is exact ∂φ ∂φ and there exists a potential function φ such that (a) = 4e2x + 2xy − y 2 and (b) = (x − y)2 . ∂x ∂y y3 dh(x) dh(x) ∂φ From (b) φ(x, y) = x2 y − xy 2 + + h(x) =⇒ = 2xy − y 2 + so from (a) 2xy − y 2 + = 3 ∂x dx dx dh(x) = 4e2x =⇒ h(x) = 2e2x where the constant of integration has been set to zero 4e2x + 2xy − y 2 =⇒ dx y3 y3 since we need just one potential function. φ(x, y) = x2 y − xy 2 + + 2e2x =⇒ x2 y − xy 2 + + 2e2x = 3 3 c1 =⇒ 6e2x + 3x2 y − 3xy 2 + y 3 = c. x y 2 − x2 dy = 0 then M = N = so the differential equation y x x2 + y 2 (x2 + y 2 )2 ∂φ ∂φ 1 y x and (b) . is exact and there exists a potential function φ such that (a) = − 2 = 2 2 ∂x x x +y ∂y x + y2 ∂φ dh(x) dh(x) y y From (b) φ(x, y) = tan−1 (y/x) + h(x) =⇒ + + = − 2 so from (a) − 2 = 2 2 ∂x x +y dx x +y dx dh 1 y =⇒ − = x−1 =⇒ h(x) = ln |x| where the constant of integration is set to zero since we only x x2 + y 2 dx need one potential function. φ(x, y) = tan−1 (y/x) + ln |x| =⇒ tan−1 (y/x) + ln |x| = c. 8. Given 1 y − 2 x x + y2 dx + 9. Given [y cos (xy) − sin x]dx + x cos (xy)dy = 0 then My = Nx = −xy sin (xy) + cos (xy) so the differential ∂φ equation is exact so there exists a potential function φ such that (a) = y cos (xy) − sin x and (b) ∂x ∂φ dh(x) ∂φ = x cos (xy). From (b) φ(x, y) = sin (xy) + h(x) =⇒ = y cos (xy) + so from (a) y cos (xy) + ∂x ∂x dx (c)2017 Pearson Education. Inc. 78 dh(x) dh = y cos (xy) − sin x =⇒ = − sin x =⇒ h(x) = cos x where the constant of integration is set to zero dx dx since we only need one potential function. φ(x, y) = sin (xy) + cos x =⇒ sin (xy) + cos x = c. 10. (2y 2 e2x + 3x2 )dx + 2ye2x dy = 0. M = 2y 2 e2x + 3x2 and N = 2ye2x =⇒ My = 4ye2x and Nx = 4ye2x =⇒ My = Nx =⇒ the differential equation is exact so there exists a potential function φ such that ∂φ ∂φ ∂φ dh(y) (a) = 2y 2 e2x + 3x2 and (b) = 2ye2x . From (a) φ(x, y) = y 2 e2x + x3 + h(y) =⇒ = 2ye2x + ∂x ∂y ∂y dy dh(y) dh 2x 2x so from (b) 2ye + = 2ye =⇒ = 0 =⇒ h(y) = c1 . Since we only need one potential function we dy dy 2 2x 3 can set c1 = 0. Then φ(x, y) = y e + x =⇒ y 2 e2x + x3 = c. 11. (y 2 + cos x)dx + (2xy + sin y)dy = 0. M = y 2 + cos x and N = 2xy + sin y =⇒ My = 2y and Nx = 2y =⇒ My = Nx =⇒ the differential equation is exact so there exists a potential function φ such that ∂φ ∂φ ∂φ dh(y) (a) = y 2 + cos x and (b) = 2xy + sin y. From (a) φ(x, y) = xy 2 + sin x + h(y) =⇒ = 2xy + so ∂x ∂y ∂y dy dh(y) dh from (b) 2xy + = 2xy + sin y =⇒ = sin y =⇒ h(y) = − cos y where the constant of integration has dy dy been set to zero since we just need one potential function. φ(x, y) = xy 2 +sin x−cos y =⇒ xy 2 +sin x−cos y = c. 12. (sin y + y cos x)dx + (x cos y + sin x)dy = 0. M = sin y + y cos x and N = x cos y + sin x =⇒ My = cos y + cos x and Nx = cos y + cos x =⇒ My = Nx =⇒ the differential equation is exact so there exists ∂φ ∂φ a potential function φ such that (a) = sin y + y cos x and (b) = x cos y + sin x. From (a) φ(x, y) = ∂x ∂y ∂φ dh(y) dh(y) x sin y + y sin x + h(y) =⇒ = x cos y + sin x + so from (b) x cos y + sin x + = x cos y + ∂y dy dy dh sin x =⇒ = 0 =⇒ h(y) = c1 . Since we only need one potential function we can set c1 = 0. φ(x, y) = dy x sin y + y sin x =⇒ x sin y + y sin x = c. x 13. Given [1+ln (xy)]dx+ dy = 0 then My = Nx = y −1 so the differential equation is exact and there exists y ∂φ ∂φ dh(x) a potential function φ such that (a) = 1+ln (xy) and (b) φ(x, y) = x ln y +h(x) =⇒ = ln y + so ∂x ∂x dx dh dh(x) = 1 ln (xy) =⇒ = 1+ln x =⇒ h(x) = c ln x where the constant of integration is set to from (a) ln y+ dx dx zero since we only need one potential function. φ(x, y) = x ln y+x ln x =⇒ x ln y+x ln x = c =⇒ x ln (xy) = c. xy + 1 xy − 1 dx + dy = 0 then My = Nx = 1 =⇒ then the differential equation is exact so x y ∂φ xy − 1 ∂φ xy + 1 there exists a potential function φ such that (a) = and (b) = . From (a) φ(x, y) = ∂x x ∂y y ∂φ dh(y) dh(y) xy + 1 dh(y) xy − ln |x| + h(y) =⇒ =x+ so from (b), x + = =⇒ = y −1 =⇒ h(y) = ln |y| ∂y dy dy y dy where the constant of integration has been set to zero since we need just one potential function. φ(x, y) = xy + ln |y/x| =⇒ xy + ln |x/y| = c. 14. Given 15. Given (2xy + cos y)dx + (x2 − x sin y − 2y)dy = 0 then My = Nx = 2x − sin y so the differential equation ∂φ ∂φ is exact so there is a potential function φ such that (a) = 2xy + cos y and (b) = x2 − x sin y − 2y. ∂x ∂y (c)2017 Pearson Education. Inc. 79 From (a) φ(x, y) = x2 y + x cos y + h(y) =⇒ ∂φ dh(y) dh(y) = x2 − x sin y + so from (b) x2 − x sin y + = ∂y dy dy dh = −2y =⇒ h(y) = −y 2 where the constant of integration has been set to zero since dy we only need one potential function. φ(x, y) = x2 y + x cos y − y 2 =⇒ x2 y + x cos y − y 2 = c. x2 − x sin y − 2y =⇒ dx + 4xy = 3 sin x =⇒ (4xy − 3 sin x)dx + 2x2 dy = 0 then My = Nx = 4x so the differential dy ∂φ ∂φ equation is exact so there exists a potential function φ such that (a) = 4xy − 3 sin x and (b) = 2x2 . ∂x ∂y ∂φ dh(x) dh(x) dh(x) From (b) φ(x, y) = 2x2 y + h(x) =⇒ = 4xy + so from (a) 4xy + = 4xy − 3 sin x =⇒ = ∂x dx dx dx −3 sin x =⇒ h(x) = 3 cos x where the constant of integration has been set to zero since we only need one potential function. φ(x, y) = 2x2 y + 3 cos x =⇒ 2x2 y + 3 cos x = c. Now since y(2π) = 0, c = 3; thus, 3 − 3 cos x . 2x2 y + 3 cos x = 3 or y(x) = 2x2 16. Given 2x2 17. Given (3x2 ln x + x2 − y)dx − xdy = 0 then My = Nx = −1 so the differential equation is exact so ∂φ ∂φ there exists a potential function φ such that (a) = 3x2 ln x + x2 − y and (b) = −x. From (b) ∂x ∂y dh(x) ∂φ dh(x) dh(x) φ(x, y) = −xy + h(x) =⇒ = −y + so from (a) −y + = 3x2 ln x + x2 − y =⇒ = ∂x dx dx dx 2 2 3 3x ln x + x =⇒ h(x) = x ln x where the constant of integration has been set to zero since we only need one potential function. φ(x, y) = −xy + x3 ln x =⇒ −xy + x3 ln x = c. Now since y(1) = 5, c = −5; thus, x3 ln x + 5 x3 ln x − xy = −5 or y(x) = . x 18. Given (yexy + cos x)dx + xexy dy = 0 then My = Nx = xyexy + exy so the differential equation is ∂φ ∂φ exact so there exists a potential function φ such that (a) = yexy + cos x and (b) = xexy . From (b) ∂x ∂y ∂φ dh(x) dh(x) φ(x, y) = exy +h(x) =⇒ = yexy + so from (a) yexy +cos x =⇒ = cos x =⇒ h(x) = sin x where ∂x dx dx the constant of integration is set to zero since we only need one potential function. φ(x, y) = exy + sin x =⇒ ln (2 − sin x) exy + sin x = c. Now since y(π/2) = 0, c = 2; thus, exy + sin x = 2 or y(x) = . x 19. If φ(x, y) is a potential function for M dx + N dy = 0 =⇒ d(φ(x, y)) = 0 so d(φ(x, y) + c) = d(φ(x, y)) + d(c) = 0 + 0 = 0 =⇒ φ(x, y) + c is also a potential function. 20. M = cos (xy)[tan (xy) + xy] and N = x2 cos (xy) =⇒ My = 2x cos (xy) − x2 y sin (xy) = Nx =⇒ My = Nx =⇒ M dx = N dy = 0 is exact so I(x, y) = cos (xy) is an integrating factor for [tan (xy)+xy]dx+x2 dy = 0. 21. M = e−x/y (x2 y −1 − 2x) and N = −e−x/y x3 y −2 =⇒ My = e−x/y (x3 y −3 − 3x2 y −2 ) = Nx =⇒ M dx + N dy = 0 is exact so I(x, y) = y −2 e−x/y is an integrating factor for y[x2 − 2xy]dx − x3 dy = 0. 22. M = sec x[2x − (x2 + y 2 ) tan x] and N = 2y sec x =⇒ My = −2y sec x tan x and Nx = 2y sec x tan x =⇒ My = Nx =⇒ M dx + N dy = 0 is not exact so I(x) = sec x is not an integrating factor for [2x − (x2 + y 2 ) tan x]dx + 2ydy = 0. 23. Given (y − x2 )dx + 2xdy = 0 then M = y − x2 and N = 2x. Thus My = 1 and Nx = 2 so (c)2017 Pearson Education. Inc. My − N x = N 80 1 1 = f (x) is a function of x alone so I(x) = e f (x)dx = √ is an integrating factor for the given equation. 2x x Multiplying the given equation by I(x) results in the exact equation (x−1/2 y−x3/2 )dx+2x1/2 dy = 0. We find 2x5/2 2x5/2 that φ(x, y) = 2x1/2 y − and hence the general solution of our differential equation is 2x1/2 y − =c 5 5 c + 2x5/2 √ . or y(x) = 10 x − 24. Given (3xy − 2y −1 )dx + x(x + y −2 )dy = 0 then M = 3xy − 2y −1 and N = x(x + y −2 ). Thus My − N x 1 My = 3x + 2y −2 and Nx = 2x + y −2 so = = f (x) is a function of x alone so I(x) = e f (x)dx = x N x is an integrating factor for the given equation. Multiplying the given equation by I(x) results in the exact equation (3x2 y − 2xy −1 )dx + x2 (x + y −2 )dy = 0. We find that φ(x, y) = x3 y − x2 y −1 and hence the general solution of our differential equation is x3 y − x2 y −1 = c. 25. Given x2 ydx + y(x3 + e−3y sin y)dy = 0 then M = x2 y and N = y(x3 + e−3y sin y). Thus My = x2 My − N x and Nx = 3x2 y so = y −1 − 3 = g(y) is a function of y alone so I(y) = e g(y)dy = e3y /y is an M integrating factor for the given equation. Multiplying the equation by I(y) results in the exact equation x3 e3y x2 e3y dx + e3y (x3 + e−3y sin y)dy = 0. We find that φ(x, y) = − cos y and hence the general solution 3 3 3y x e of our differential equation is − cos y = c. 3 26. Given (xy − 1)dx + x2 dy = 0 then M = xy − 1 and N = x2 . Thus My = x and Nx = 2x so My − Nx = −x−1 = f (x) is a function of x alone so I(x) = e f (x)dx = x−1 is an integrating factor for the N given equation. Multiplying the given equation by I(x) results in the exact equation (y − x−1 )dx + xdy = 0. We find that φ(x, y) = xy −ln |x| and hence, the general solution of our differential equation is xy −ln |x| = c. dy 1 2xy = =⇒ (2xy + 2x3 y − 1)dx + (1 + x2 )2 dy = 0 then M = 2xy + 2x3 y − 1 and + 2 dx 1 + x (1 + x2 )2 My − N x 2x = f (x) is a function of N = (1 + x2 )2 . Thus My = 2x + 2x3 and Nx = 4x(1 + x2 ) so =− N 1 + x2 1 x alone so I(x) = e f (x)dx = is an integrating factor for the given equation. Multiplying the given 1 + x2 1 dx + (1 + x2 )dy = 0. We find that φ(x, y) = equation by I(x) yields the exact equation 2xy − 1 + x2 (1 + x2 )y − tan−1 x and hence the general solution of our differential equation is (1 + x2 )y − tan−1 x = c or tan−1 x + c y(x) = . 1 + x2 27. Given 28. Given xy[2 ln (xy) + 1]dx + x2 dy = 0 then M = xy[2 ln (xy) + 1] and N + x2 . Thus My = 3x + 2x ln (xy) MY − N x 1 = y −1 = g(y) is a function of y only so I(y) = e g(y)dy = is an integrating factor and Nx = 2x so M y for the given equation. Multiplying the given equation by I(y) results in the exact equation x[2 ln (xy) + 1]dx + x2 y −1 dy = 0. We find that φ(x, y) = x2 ln y + x2 ln x and hence the general solution of our differential 2 equation is x2 ln y + x2 ln x = c or y(x) = xec/x . 29. Given ydx − (2x + y 4 )dy = 0 then M = y and N = −(2x + y 4 ). Thus My = 1 and Nx = −2 so (c)2017 Pearson Education. Inc. 81 My − N x = 3y −1 = g(y) is a function of y alone so I(y) = e− g(y)dy = 1/y 3 is an integrating factor M for the given differential equation. Multiplying the given equation by I(y) results in the exact equation y −2 dx − (2xy −3 + y)dy = 0. We find that φ(x, y) = xy −2 − y 2 /2 and hence, the general solution of our differential equation is xy −2 − y 2 /2 = c1 =⇒ 2x − y 4 = cy 2 . 30. Given (y −1 − x−1 )dx + (xy −2 − 2y −1 )dy = 0 =⇒ xr y s (y −1 − x−1 )dx + xr y s (xy −2 − 2y −1 )dy = 0 =⇒ (xr y s−1 −xr−1 y s )dx+(xr+1 y s−2 −2xr y s−1 )dy = 0. Then M = xr y s−1 −xr−1 y s and N = xr+1 y s−2 −2xr y s−1 so My = xr (s − 1)y s−2 − xr−1 sy s−1 and Nx = (r + 1)xr y s−2 − 2rxr−1 y s−1 . The equation is exact if and s−1 s 2r r+1 only if My = Nx =⇒ xr y s−1 − xr−1 y s = (r + 1)xr y s−2 − 2rxr−1 y s−1 =⇒ − − = =⇒ 2 2 y xy y xy s − 2r s−r−2 = . From the last equation we require that s − r − 2 = 0 and s − 2r = 0. Solving this system y2 xy yields r = 2 and s = 4. 31. Given 2y(y + 2x2 )dx + x(4y + 3x2 )dy = 0 =⇒ xr y s 2y(y + 2x2 )dx + xr y s x(4y + 3x2 )dy = 0. Then M = 2xr y s+2 + 4xr+2 y s+1 and N = 4xr+1 y s+1 + 3xr+3 y s so My = 2xr (s + 2)y s+1 + 4xr+2 (s + 1)y s and Nx = 4(r + 1)xr y s+1 + 3(r + 3)xr+2 y s . The equation is exact if and only if My = Nx =⇒ 2xr (s + 2)y s+1 + 4xr+2 (s + 1)y s = 4(r + 1)xr y s+1 + 3(r + 3)xr+2 y s =⇒ 2(s + 2)y + 4x2 (s + 1) = 4(r + 1)y + 3(r + 3)x2 . From this last equation we require that 2(s + 2) = 4(r + 1) and 4(s + 1) = 3(r + 3). Solving this system yields r = 1 and s = 2. 32. Given y(5xy 2 + 4)dx + x(xy 2 − 1)dy = 0 =⇒ xr y s y(5xy 2 + 4)dx + xr y s x(xy 2 − 1)dy = 0. Then M = xr y s+1 (5xy 2 + 4) and N = xr+1 y s (xy 2 − 1) so My = 5(s + 3)xr+1 y s+2 + 4(s + 1)xr y s and Nx = (r + 2)xr+1 y s−2 − (r + 1)xr y s . The equation is exact if and only if My = Nx =⇒ 5(s + 3)xr+1 y s+2 + 4(s + 1)xr y s = (r + 2)xr+1 y s+2 − (r + 1)xr y s =⇒ 5(s + 3)xy 2 + 4(s + 1) = (r + 2)xy 2 − (r + 1). From the last equation we require that 5(s + 3) = r + 2 and 4(s + 1) = −(r + 1). Solving this system yields r = 3 and s = −2. My − N x = g(y) is a function of y only. Then dividing the equation (1.9.21) by M , M it follows that I is an integrating factor for M (x, y)dx + N (x, y)dy = 0 if and only if it is a solution of ∂I N ∂I − = Ig(y) (30.1). We must show that this differential equation has a solution I = I(y). However, M ∂x ∂y dI = −Ig(y), which is a separable equation with solution I(y) = e− g(t)dt . if I = I(y), then (30.1) reduces to dy 33. Suppose that dy + py = q can be written in the differential form as (py − q)dx + dy = 0 (34.1). This dx My − N x has M = py − q and N = 1 so that = p(x). Consequently, an integrating factor for (34.1) is N x I(x) = e p(t)dt . 34. (a). Note x x x yields the exact equation e p(t)dt (py − q)dx + e p(t)dt dy = 0. x x ∂φ ∂φ Hence, there exists a potential function φ such that (i) = e p(t)dt (py − q) and (ii) = e p(t)dt . From ∂x ∂y x x x x dh(x) dh(x) p(t)dt p(t)dt p(t)dt (i), p(x)ye + (py − q) =⇒ =⇒ h(x) = − q(x)e p(t)dt dx, =e = −q(x)e dx dx where the constant of integration just need one potential function. Consequently, x x has been set to zero since x we x φ(x, y) = ye p(t)dt − q(x)e p(t)dt dx =⇒ y(x) = I −1 Iq(t)dt + c , where I(x) = e p(t)dt . (b). Multiplying (34.1) by I(x) = e p(t)dt (c)2017 Pearson Education. Inc. 82 Solutions to Section 1.10 True-False Review: (a): TRUE. This is well-illustrated by the calculations shown in Example 1.10.1. (b): TRUE. The equation y1 = y0 + f (x0 , y0 )(x1 − x0 ) dy = f (x, y) at the point (x0 , y0 ). Once the point (x1 , y1 ) is determined, the is the tangent line to the curve dx procedure can be iterated over and over at the new points obtained to carry out Euler’s method. (c): FALSE. It is possible, depending on the circumstances, for the errors associated with Euler’s method to decrease from one step to the next. (d): TRUE. This is illustrated in Figure 1.10.3. Problems: 1. Applying Euler’s method with y = 4y −1, x0 = 0, y0 = 1, and h = 0.05 we have yn+1 = yn +0.05(4yn −1). This generates the sequence of approximants given in the table below. n 1 2 3 4 5 6 7 8 9 10 xn 0.05 0.10 0.15 0.20 0.25 0.30 0.35 0.40 0.45 0.50 yn 1.15 1.33 1.546 1.805 2.116 2.489 2.937 3.475 4.120 4.894 Consequently the Euler approximation to y(0.5) is y10 = 4.894. (Actual value: y(.05) = 5.792 rounded to 3 decimal places). 2xy x n yn , x0 = 0, y0 = 1, and h = 0.1 we have yn+1 = yn − 0.2 . 1 + x2 1 + x2n This generates the sequence of approximants given in the table below. 2. Applying Euler’s method with y = − n 1 2 3 4 5 6 7 8 9 10 xn 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0 yn 1 0.980 0.942 0.891 0.829 0.763 0.696 0.610 0.569 0.512 (c)2017 Pearson Education. Inc. 83 Consequently the Euler approximation to y(1) is y10 = 0.512. (Actual value: y(1) = 0.5). 3. Applying Euler’s method with y = x−y 2 , x0 = 0, y0 = 2, and h = 0.05 we have yn+1 = yn +0.05(xn −yn2 ). This generates the sequence of approximants given in the table below. n 1 2 3 4 5 6 7 8 9 10 xn 0.05 0.10 0.15 0.20 0.25 0.30 0.35 0.40 0.45 0.50 yn 1.80 1.641 1.511 1.404 1.316 1.242 1.180 1.127 1.084 1.048 Consequently the Euler approximation to y(0.5) is y10 = 1.048. (Actual value: y(.05) = 1.0477 rounded to four decimal places). 4. Applying Euler’s method with y = −x2 y, x0 = 0, y0 = 1, and h = 0.2 we have yn+1 = yn − 0.2x2n yn . This generates the sequence of approximants given in the table below. n 1 2 3 4 5 xn 0.2 0.4 0.6 0.8 1.0 yn 1 0.992 0.960 0.891 0.777 Consequently the Euler approximation to y(1) is y5 = 0.777. (Actual value: y(1) = 0.717 rounded to 3 decimal places). 5. Applying Euler’s method with y = 2xy 2 , x0 = 0, y0 = 1, and h = 0.1 we have yn+1 = yn + 0.1xn yn2 . This generates the sequence of approximants given in the table below. n 1 2 3 4 5 6 7 8 9 10 xn 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0 yn 0.5 0.505 0.515 0.531 0.554 0.584 0.625 0.680 0.754 0.858 Consequently the Euler approximation to y(1) is y10 = 0.856. (Actual value: y(1) = 1). (c)2017 Pearson Education. Inc. 84 ∗ = 6. Applying the modified Euler method with y = 4y − 1, x0 = 0, y0 = 1, and h = 0.05 we have yn+1 yn + 0.05(4yn − 1) ∗ yn+1 = yn + 0.025(4yn − 1 + 4yn+1 − 1). This generates the sequence of approximants given in the table below. n 1 2 3 4 5 6 7 8 9 10 xn 0.05 0.10 0.15 0.20 0.25 0.30 0.35 0.40 0.45 0.50 yn 1.165 1.3663 1.6119 1.9115 2.2770 2.7230 3.2670 3.9308 4.7406 5.7285 Consequently the modified Euler approximation to y(0.5) is y10 = 5.7285. (Actual value: y(.05) = 5.7918 rounded to 4 decimal places). 2xy ∗ 7. Applying the modified Euler method with y = − , x0 = 0, y0 = 1, and h = 0.1 we have yn+1 = 1 + x2 x n yn yn − 0.2 1 + x2n ∗ xn+1 yn+1 x n yn . This generates the sequence of approximants given in the table yn+1 = yn + 0.05 − − 2 1 + x2n 1 + x2n+1 below. n 1 2 3 4 5 6 7 8 9 10 xn 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0 yn 0.9900 0.9616 0.9177 0.8625 0.8007 0.7163 0.6721 0.6108 0.5536 0.5012 Consequently the modified Euler approximation to y(1) is y10 = 0.5012. (Actual value: y(1) = 0.5). ∗ 8. Applying the modified Euler method with y = x − y 2 , x0 = 0, y0 = 2, and h = 0.05 we have yn+1 = yn − 0.05(xn − yn2 ) ∗ yn+1 = yn + 0.025(xn − yn2 + xn+1 − (yn+1 )2 ). This generates the sequence of approximants given in the table below. (c)2017 Pearson Education. Inc. 85 n 1 2 3 4 5 6 7 8 9 10 xn 0.05 0.10 0.15 0.20 0.25 0.30 0.35 0.40 0.45 0.50 yn 1.8203 1.6725 1.5497 1.4468 1.3600 1.2866 1.2243 1.1715 1.1269 1.0895 Consequently the modified Euler approximation to y(0.5) is y10 = 1.0895. (Actual value: y(.05) = 1.0878 rounded to 4 decimal places). ∗ 9. Applying the modified Euler method with y = −x2 y, x0 = 0, y0 = 1, and h = 0.2 we have yn+1 = 2 yn − 0.2xn yn ∗ yn+1 = yn − 0.1[x2n yn + x2n+1 yn+1 ]. This generates the sequence of approximants given in the table below. n 1 2 3 4 5 xn 0.2 0.4 0.6 0.8 1.0 yn 0.9960 0.9762 0.9266 0.8382 0.7114 Consequently the modified Euler approximation to y(1) is y5 = 0.7114. (Actual value: y(1) = 0.7165 rounded to 4 decimal places). ∗ 10. Applying the modified Euler method with y = 2xy 2 , x0 = 0, y0 = 1, and h = 0.1 we have yn+1 = 2 yn + 0.1xn yn ∗ yn+1 = yn +0.05[xn yn2 +xn+1 (yn+1 )2 ]. This generates the sequence of approximants given in the table below. n 1 2 3 4 5 6 7 8 9 10 xn 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0 yn 0.5025 0.5102 0.5235 0.5434 0.5713 0.6095 0.6617 0.7342 0.8379 0.9941 Consequently the modified Euler approximation to y(1) is y10 = 0.9941. (Actual value: y(1) = 1). 11. We have y = 4y −1, x0 = 0, y0 = 1, and h = 0.05. So, k1 = 0.05(4yn −1), k2 = 0.05[4(yn + 12 k1 )−1], k3 = 0.05[4(yn + 12 k2 ) − 1], k4 = 0.05[4(yn + 12 k3 ) − 1], (c)2017 Pearson Education. Inc. 86 yn+1 = yn + 16 (k1 + k2 + k3 + k4 ). This generates the sequence of approximants given in the table below (computations rounded to five decimal places). n 1 2 3 4 5 6 7 8 9 10 xn 0.05 0.10 0.15 0.20 0.25 0.30 0.35 0.40 0.45 0.50 yn 1.16605 1.36886 1.61658 1.91914 2.28868 2.74005 3.29135 3.96471 4.78714 5.79167 Consequently the Runge-Kutta approximation to y(0.5) is y10 = 5.79167. (Actual value: y(.05) = 5.79179 rounded to 5 decimal places). (xn + 0.05)(yn + k21 ) xy x n yn , x = 0, y = 1, and h = 0.1. So, k = −0.2 , k = −0.2 ), k3 = 0 0 1 2 1 + x2 1 + x2n [1 + (xn + 0.05)2 ] (xn + 0.05)(yn + k22 ) xn+1 (yn + k3 ) −0.2 , k4 = −0.2 , [1 + (xn + 0.05)2 ] [1 + (xn+1 )2 ] 1 yn+1 = yn + 6 (k1 + k2 + k3 + k4 ). This generates the sequence of approximants given in the table below (computations rounded to seven decimal places). 12. We have y = −2 n 1 2 3 4 5 6 7 8 9 10 xn 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0 yn 0.9900990 0.9615383 0.9174309 0.8620686 0.7999996 0.7352937 0.6711406 0.6097558 0.5524860 0.4999999 Consequently the Runge-Kutta approximation to y(1) is y10 = 0.4999999. (Actual value: y(.05) = 0.5). 13. We have y = x − y 2 , x0 = 0, y0 = 2, and h = 0.05. So, k1 = 0.05(xn − yn2 ), k2 = 0.05[xn + 0.025 − (yn + k1 2 k2 2 2 2 ) ], k3 = 0.05[xn + 0.025 − (yn + 2 ) ], k4 = 0.05[xn+1 − (yn + k3 ) ]], 1 yn+1 = yn + 6 (k1 + k2 + k3 + k4 ). This generates the sequence of approximants given in the table below (computations rounded to six decimal places). (c)2017 Pearson Education. Inc. 87 n 1 2 3 4 5 6 7 8 9 10 xn 0.05 0.10 0.15 0.20 0.25 0.30 0.35 0.40 0.45 0.50 yn 1.1.81936 1.671135 1.548079 1.445025 1.358189 1.284738 1.222501 1.169789 1.125263 1.087845 Consequently the Runge-Kutta approximation to y(0.5) is y10 = 1.087845. (Actual value: y(0.5) = 1.087845 rounded to 6 decimal places). 14. We have y = −x2 y, x0 = 0, y0 = 1, and h = 0.2. So, k1 = −0.2x2n yn , k2 = −0.2(xn + 0.1)2 (yn + k21 ), k3 = −0.2(xn + 0.1)2 (yn + k22 ), k4 = −0.2(xn+1 )2 (yn + k3 ), yn+1 = yn + 16 (k1 + k2 + k3 + k4 ). This generates the sequence of approximants given in the table below (computations rounded to six decimal places). n 1 2 3 4 5 xn 0.2 0.4 0.6 0.8 1.0 yn 0.997337 0.978892 0.930530 0.843102 0.716530 Consequently the Runge-Kutta approximation to y(1) is y10 = 0.716530. (Actual value: y(1) = 0.716531 rounded to 6 decimal places). 15. We have y = 2xy 2 , x0 = 0, y0 = 1, and h = 0.1. So, k1 = 0.2xn − yn2 , k2 = 0.2(xn + 0.05)(yn + k21 )2 , k3 = 0.2(xn + 0.05)(yn + k22 )2 , k4 = 0.2xn+1 (yn + k3 )2 ]], yn+1 = yn + 16 (k1 + k2 + k3 + k4 ). This generates the sequence of approximants given in the table below (computations rounded to six decimal places). n 1 2 3 4 5 6 7 8 9 10 xn 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0 yn 0.502513 0.510204 0.523560 0.543478 0.571429 0.609756 0.662252 0.735295 0.840336 0.999996 Consequently the Runge-Kutta approximation to y(1) is y10 = 0.999996. (Actual value: y(1) = 1). (c)2017 Pearson Education. Inc. 88 1 1 16. We have y + y = e−x/10 cos x, x0 = 0, y0 = 0, and h = 0.5 Hence, k1 = 0.5 − yn + exn /10 cos xn , k2 = 10 10 1 1 1 1 (−xn +0.25)/10 yn + k 1 + e yn + k2 + e(−xn +0.25)/10 cos (xn + 0.25) , cos (xn + 0.25) , k3 = 0.5 − 0.5 − 10 2 10 2 1 k4 = 0.5 − (yn + k3 ) + e−xn+1 /10 cos xn+1 , yn+1 = yn + 16 (k1 + 2k2 + 2k3 + k4 ). This generates the se10 quence of approximants plotted in the accompanying figure. We see that the solution appears to be oscillating with a diminishing amplitude. Indeed, the exact solution to the initial value problem is y(x) = e−x/10 sin x. The corresponding solution curve is also given in the figure. y(x) 0.75 0.5 0.25 5 10 x 15 20 25 -0.25 -0.5 Figure 0.0.56: Figure for Problem 16 Solutions to Section 1.11 Problems: dy dy d2 y du d2 y 3x − 2 . Let u = . Substituting these results into the first equation = 6e so that = dx2 dx dx dx dx2 du yields − 2u = 6e3x . An appropriate integrating factor for this equation is I(x) = e−2 dx = e−2x =⇒ dx dy d −2x u) = 6ex =⇒ e−2x u = 6ex + c1 =⇒ u = 6e3x + c1 e2x =⇒ (e = 6e3x + c1 e2x =⇒ y(x) = dx dx 2e3x + c1 e2x + c2 , where we have redefined the constant c1 in the last step. 1. d2 y 2 dy dy du d2 y 2 = . Let u = . Substituting these results into the first equation + 4x so that = dx2 x dx dx dx dx2 du du 2 2 yields = u + 4x2 =⇒ − u = 4x2 . An appropriate integrating factor for this equation is I(x) = dx x dx x 2. (c)2017 Pearson Education. Inc. 89 dx dy x = x−2 =⇒ d(x−2 u) = 4 =⇒ x−2 u = 4 dx =⇒ x−2 u = 4x + c1 =⇒ u = 4x3 + c1 x2 =⇒ = dx 3 2 3 4 4x + c1 x =⇒ y(x) = c1 x + x + c2 . e −2 dy d2 y 1 dy du d2 y = . Substituting these results into − 1 . Let u = so that = 2 dx (x − 1)(x − 2) dx dx dx dx2 1 du 1 1 du = (u − 1) =⇒ − u = − . An the first equation yields dx (x − 1)(x − 2) dx (x − 1)(x − 2) (x − 1)(x − 2) 1 − dx x−1 d x−1 (x − 1)(x − 2) = =⇒ u = appropriate integrating factor for this equation is I(x) = e x−2 dx x − 2 x−1 1 dy 1 1 =⇒ u = (x − 2)−2 dx =⇒ u = − + c1 =⇒ =− + c1 =⇒ y(x) = − ln |x − 1| + 2 (x − 2) x−2 x−1 dx x−1 c1 x + c2 . 3. d2 y 2 4. + dx2 y 2 dy dy du du d2 y . Substituting these results into the first . Let u = so that =u = dx dx dx dy dx2 du 2 2 du + u2 = u =⇒ u = 0 or − u = 1. An appropriate integrating factor for the last equation yields u dy y dy y 2 dy d 2 y3 dy y c1 (y u) = y 2 =⇒ y 2 u = y 2 dy =⇒ y 2 u = +c1 =⇒ = + 2 =⇒ equation is I(y) = e y = y 2 =⇒ 3 dx 3 y √ dy ln |y 3 + c2 | = x + c3 =⇒ y(x) = 3 c4 ex + c5 . = 2 dy du du d2 y . Substituting these results into the first so that =u = dx dx dy dx2 du du equation yields u = u2 tan y. If u = 0 then = 0 =⇒ y equals a constant and this is a solution to the dy dx du dy du = u tan y =⇒ = tan ydy =⇒ u = c1 sec y =⇒ = equation. Now suppose that u = 0. Then dy u dx c1 sec y =⇒ y(x) = sin−1 (c1 x + c2 ). 5. d2 y = dx2 dy dx dy dx tan y. Let u = 2 d2 y dy dy dy du d2 y 6. + tan x . Let u = . Substituting these results into the first = so that = dx2 dx dx dx dx dx2 1 du du dz equation yields + tan xu = u2 which is a Bernoulli equation. Letting z = u−1 gives 2 = =− . dx u dx dx dz − tan xz = −1. Then an integrating factor for Substituting these results into the last equation yields dx d − tan xdx = cos x =⇒ (z cos x) = − cos x =⇒ z cos x = − cos xdx =⇒ z = this equation is I(x) = e dx cos x dy cos x − sin x + c1 =⇒ u = =⇒ = =⇒ y(x) = c2 − ln |c1 − sin x|. cos x c1 − sin x dx c1 − sin x 2 dx dx d2 x dx du d2 x = + 2 . Substituting these results into the first equation 7. . Let u = so that = dt2 dt dt dt dt dt2 du du = u2 + 2u =⇒ − 2u = u2 which is a Bernoulli equation. If u = 0 then x is a constant which yields dt dt dz 1 du satisfies the equation. Now suppose that u = 0. Let z = u−1 so that =− 2 . Substituting these dt u dt (c)2017 Pearson Education. Inc. 90 dz + 2z = −1. An integrating factor for this equation is I(x) = e2t =⇒ dt 2e2t d 2t 1 2e2t =⇒ x = dt =⇒ x(t) = c2 − ln |c1 − e2t |. (e z) = −e2t =⇒ z = ce−2t − =⇒ u = dt 2 2c − e2t 2c − e2t results into the last equation yields d2 y dy dy du d2 y 4 − 2 . Let u = . Substituting these results into the first equation = 6x so that = dx2 dx dx dx dx2 dx −2 du 2 x = x−2 =⇒ yields − u = 6x4 . An appropriate integrating factor for this equation is I(x) = e dx x dy 1 d −2 (x u) = 6x2 =⇒ x−2 u = 6 x2 dx =⇒ u = 2x5 + cx2 =⇒ = 2x5 + cx2 =⇒ y(x) = x6 + c1 x3 + c2 . dx dx 3 8. dx d2 x dx du d2 x . Let u = 9. t 2 = 2 t + so that = 2 . Substituting these results into the first equation dt dt dt dt dt d 2 du − u = 2. An integrating factor for this equation is I(x) = t−2 =⇒ (t−2 u) = 2t−2 =⇒ u = yields dt t dt dx −2t + ct2 =⇒ = −2t + ct2 =⇒ x(t) = c1 t3 − t2 + c2 . dt d2 y 10. −α dx2 dy dx 2 −β dy dy du d2 y . Substituting these results into the = 0. Let u = so that = dx dx dx dx2 du − βu = αu2 which is a Bernoulli equation. If u = 0 then y is a constant and dx du dz satisfies the equation. Now suppose that u = 0. Let z = u−1 so that = −u−2 . Substituting dx dx dz + βz = −α. The an integrating factor for this equation is these results into the last equation yields dx βx α βeβx dy βeβx β dx βx βx = e =⇒ e z = −α e dx =⇒ − + ce−βx =⇒ u = =⇒ =⇒ = I(x) = e βx β cβ − αe dx cβ − αeβx 1 βeβx dx =⇒ y(x) = − ln |c1 + c2 eβx |. y= cβ − αeβx α first equation yields d2 y 2 dy dy du d2 y 4 − . Let u = . Substituting these results into the first equation = 18x so that = dx2 x dx dx dx dx2 d −2 du 2 yields − u = 18x4 which has I(x) = x−2 as an integrating factor so (x u) = 18x2 =⇒ u = dx x dx dy 6x5 + cx2 =⇒ = 6x5 + cx2 =⇒ y(x) = x6 + c1 x3 + c2 . dx 11. d2 y 2x dy dy du = − . Let u = so that = df racd2 ydx2 . If u = 0 then y is a constant and dx2 1 + x2 dx dx dx satisfies the equation. Now suppose that u = 0. Substituting these results into the first equation yields c1 dy 2x c1 du u =⇒ ln |u| = − ln (1 + x2 ) + c =⇒ u = =⇒ =⇒ y(x) = c1 tan−1 x + c2 . =− = dx 1 + x2 1 + x2 dx 1 + x2 12. 3 dy dy du du d2 y . Let u = so that =u = 2 . Substituting these results into dx dx dx dy dx du 1 2 the first equation yields u + u = ye−y u3 . If u = 0 then y is a constant and satisfies the equation. Now dx y 13. d2 y 1 + dx2 y dy dx 2 = ye−3 (c)2017 Pearson Education. Inc. 91 suppose that u = 0. Substituting these results into the first equation yields du u + = ye−y u2 which is a dx y du dv = −u−2 . Substituting these results into the last equation dy dy d −1 dv v yields − = −ye−y . Then I(y) = y −1 is an integrating factor for the equation thus (y v) = −e−1 =⇒ dy y dy ey dy ey =⇒ =⇒ (ye−y + cy)dy = dx =⇒ e−y (y + 1) + c1 y 2 − x. v = y(e−1 + c) =⇒ u = = y y + cye dx y + cyey Bernoulli equation. Let v = u−1 so that dy d2 y dy du d2 y − tan x . Substituting these results into the first equation = 1. Let u = so that u = 2 dx dx dx dy dx2 du yields − u tan x = 1. An appropriate integrating factor for this equation is I(x) = e− tan xdx = cos x =⇒ dx d dy (u cos x) = cos x =⇒ u cos x = sin x + c =⇒ u(x) = tan x + c sec x =⇒ = tan x + c sec x =⇒ y(x) = dx dx ln sec x + c1 ln (sec x + tan x) + c2 . 14. 2 dy dy du du d2 y + y 2 . Let u = . Substituting these results into the first so that =u = dx dx dx dy dx2 du du 2 1 dz equation yields u − u2 = y, a Bernoulli equation. Let z = u2 so that u = . Substituting these dy y dy 2 dy dz 4 − z = 2y which has I(y) = y −4 as an integrating factor. Therefore, results into the last equation yields dy y dy d −4 (y z) = 2y −3 =⇒ z = c1 y 4 − y 2 =⇒ u2 = c1 y 4 − y 2 =⇒ u = ± c1 y 4 − y 2 =⇒ = ± c1 y 4 − y 2 =⇒ dy dx 1 −1 = ±x + c2 . Using the facts that f (0) = 1 and y (0) = 0 we find that c1 = 1 and c2 = 0; thus cos √ y c1 y(x) = sec x. 15. y d2 y =2 dx2 d2 y dy du du d2 y 2 = ω y where ω > 0. Let u = . Substituting these results into the first so that = u = dx2 dx dx dy dx2 du equation yields u = ω 2 y =⇒ u2 = ω 2 y 2 + c2 . Using the given that y(0) = a and y (0) = 0 we find that dy dy 1 c2 = a2 ω 2 . Then = ±ω y 2 − a2 =⇒ cosh−1 (y/a) = ±x + c =⇒ y(x) = a cosh [ω(c ± x)] =⇒ y = dx ω ±aω sinh [ω(c ± x)] and since y (0) = 0, c = 0; hence, y(x) = a cosh (ωx). 16. dy du d2 y 17. Let u = . Substituting these results into the differential equation yields so that u = dx dx dx2 √ 1 du 1√ 1 + u2 . Separating the variables and integrating we obtain 1 + u2 = y +c. Imposing the initial u = dy a a √ 1 1 dy 2 conditions y(0) = a, (0) = 0 gives c = 0. Hence, 1 + u = y so that 1 + u2 = 2 y 2 or equivalently, dx a a dy 1 1 = ± dx which u = ± y 2 /a2 − 1. Substituting u = and separating the variables gives dx |a| y 2 − a2 can be integrated to obtain cosh−1 (y/a) = ±x/a + c1 so that y = a cosh (±x/a + c1 ). Imposing the initial conditions y(0) = a gives c1 = 0 so that y(x) = a cosh (x/a). 18. d2 y dy dy du d2 y + p(x) . Substituting these results into the first equation = q(x). Let u = so that = dx2 dx dx dx dx2 (c)2017 Pearson Education. Inc. 92 du e− p(x)dx q(x)dx + c1 gives us the equivalent system: +p(x)u = q(x) which has a solution u = e− p(x)dx dx # " − p(x)dx − p(x)dx dy e− p(x)dx q(x)dx + c1 . Thus y = e e q(x)dx + c1 dx + c2 is so = e− p(x)dx dx a solution to the original equation. du1 du2 du3 d3 y d2 y dy d2 y d3 y since =⇒ ; thus = F x, = =⇒ u3 = = = dx dx dx dx2 dx dx3 dx3 dx2 du3 the latter equation is equivalent to = F (x, u3 ). dx d3 y 1 d2 y du1 du2 (b). = − 1 . Replace this equation by the equivalent first order system: = u2 , = u3 , dx3 x dx2 dx dx du3 dx du2 K 1 du3 = (u3 − 1) =⇒ = =⇒ u3 = Kx + 1 =⇒ = Kx + 1 =⇒ u2 = x2 + x + c2 =⇒ and dx x u3 − 1 x dx 2 du1 K 3 1 2 1 2 K 2 3 = x + x + c2 =⇒ u1 = x + x + c2 x + c3 =⇒ y(x) = u1 = c1 x + x + c2 x + c3 . dx 2 6 2 2 19. (a). u1 = y =⇒ u2 = 19. Given d2 θ g dθ + sin θ = 0, θ(0) = θ0 , and (0) = 0. 2 dt L dt g du dθ d2 θ dθ du d2 θ du + = θ = 0. Let u = so that = = u . Substituting these results into 2 2 dt L dt dt dt dθ dt dθ g dθ g du + θ = 0 =⇒ u2 = − θ2 + c21 , but (0) = 0 and θ(0) = θ0 so c21 = the last equation yields u dθ L L dt g 2 θ g 2 g g = ± θ0 − θ2 =⇒ sin−1 θ0 =⇒ u2 = (θ02 − θ2 ) =⇒ u = ± t + c2 , but θ(0) = θ0 so L L L θ0 L θ π g g g π π c2 = =⇒ sin−1 = ± t =⇒ θ = θ0 sin ± t =⇒ θ = θ0 cos t . Yes, the predicted 2 θ0 2 L 2 L L motion is reasonable. (a). du dθ dθ du d2 θ du d2 θ g + sin θ = 0. Let u = so that = 2 = = u . Substituting these results into the last 2 dt L dt dt dt dθ dt dθ 2g dθ 2g du g 2 cos θ + c. Since θ(0) = θ0 and (0) = 0, then c = − cos θ0 equation yields u + sin θ = 0 =⇒ u = dθ L L dt L 2g 2g 2g dθ dθ 2g 2g and so u2 = cos θ − cos θ0 =⇒ =± cos θ − cos θ0 =⇒ =± [cos θ − cos θ0 ]1/2 . L L dt L L dt L L dθ = ±dt. When the pendulum goes from θ = θ0 to θ = 0 (c). From part (b), 2g [cos θ − cos θ0 ]1/2 dθ (which corresponds to one quarter of a period) is negative; hence, choose the negative sign. Thus, dt 0 L dθ L θ0 dθ =⇒ T = . T =− 1/2 2g θ0 [cos θ − cos θ0 ] 2g 0 [cos θ − cos θ0 ]1/2 θ0 L dθ =⇒ (d). T = 2g 0 [cos θ − cos θ0 ]1/2 (b). T = L 2g θ0 0 2 sin 2 θ0 2 dθ − 2 sin 2 1 1/2 = 2 θ L 2g θ0 0 2 (c)2017 Pearson Education. Inc. sin 2 θ0 2 dθ 1/2 . θ − sin 2 2 93 Let k = sin θ0 2 so that 1 T = 2 L 2g θ0 0 dθ 1/2 . θ 2 k 2 − sin 2 (0.0.7) 2k cos (u)du Now let sin θ/2 = k sin u. When θ = 0, u = 0 and when θ = θ0 , u = π/2; moreover, dθ = =⇒ cos (θ/2) 2 k 2 − sin2 (θ/2)du 2k 1 − sin2 (u)du 2 k 2 − (k sin (u))2 du =⇒ dθ = =⇒ dθ = . Making this dθ = 1 − sin2 (θ/2) 1 − k 2 sin2 (u) 1 − k 2 sin2 (u) change of variables in equation (0.0.7) yields L π/2 du T = where k = sin θ0 /2. g 0 1 − k 2 sin2 (u) Solutions to Section 1.12 1. The acceleration of gravity is a = 9.8 meters/sec2 . Integrating, we find that the vertical component of the velocity of the rocket is v(t) = 4.9t + c1 . We are given that v(0) = −10, so that c1 = −10. Thus, v(t) = 4.9t − 10. Integrating again, we find the position s(t) = 2.495t2 − 10t + c2 . Setting s = 0 at two meters above the ground, we have s(0) = 0 so that s(t) = 2.495t2 − 10t. 10 ≈ 2.04 seconds. Thus, (a). The highest point above the ground is obtained when v(t) = 0. That is, t = 4.9 2 the highest point is approximately s(2.04) = 2.495 · (2.04) − 10(2.04) ≈ −10.02, which is 12.02 meters above the ground. (b). The rocket hits the ground when s(t) = 2. That is 2.495t2 − 10t − 2 = 0. Solving for t with the quadratic formula, we find that t = −0.19 or t = 4.27. Since we must report a positive answer, we conclude that the rocket hits the ground 4.27 seconds after launch. 2. The acceleration of gravity is a = 32 ft/sec2 . Integrating, we find that the vertical component of the velocity of the ball is v(t) = 16t + c1 . Since the ball is initially hit horizontally, we have v(0) = 0, so that c1 = 0. Hence, v(t) = 16t. Integrating again, we find the position s(t) = 8t2 + c2 . Setting s = 0 at two feet above the ground, we have s(0) = 0 so that c2 = 0. Thus, s(t) = 8t2 . The ball hits the ground when s(t) = 2, so that t2 = 14 . Therefore, t = 12 . Since 80 miles per hour equates to over 117 ft/sec. In one-half second, the horizontal change in position of the ball is therefore more than 117 2 = 58.5 feet, more than enough to span the necessary 40 feet for the ball to reach the front wall. Therefore, the ball does reach the front wall before hitting the ground. 3. We first determine the slope of the given family at the point (x, y). Differentiating y = cx3 (0.0.8) with respect to x yields dy = 3cx2 . dx From (0.0.8) we have c = xy3 which, when substituted into Equation (0.0.9) yields 3y dy = . dx x (c)2017 Pearson Education. Inc. (0.0.9) 94 Consequently, the differential equation for the orthogonal trajectories is dy x =− . dx 3y Separating the variables and integrating gives 1 3 2 y = − x2 + C, 2 2 which can be written in the equivalent form x2 + 3y 2 = k. 4. We first determine the slope of the given family at the point (x, y). Differentiating y = ln(cx) (0.0.10) with respect to x yields 1 dy = . dx x Consequently, the differential equation for the orthogonal trajectories is dy = −x. dx which can be integrated directly to obtain 1 y = − x2 + k. 2 5. We first determine the slope of the given family at the point (x, y). Differentiating y 2 = cx3 (0.0.11) with respect to x yields 2y so that dy = 3cx2 dx dy 3cx2 = . dx 2y 2 From (0.0.11) we have c = xy 3 which, when substituted into Equation (0.0.12) yields dy 3y = . dx 2x Consequently, the differential equation for the orthogonal trajectories is 2x dy =− . dx 3y (c)2017 Pearson Education. Inc. (0.0.12) 95 Separating the variables and integrating gives 3 2 y = −x2 + C, 2 which can be written in the equivalent form 2x2 + 3y 2 = k. 6. We first determine the slope of the given family at the point (x, y). Differentiating x4 + y 4 = c (0.0.13) with respect to x yields 4x3 + 4y 3 so that dy =0 dx dy x3 = − 3. dx y (0.0.14) Consequently, the differential equation for the orthogonal trajectories is dy y3 = 3. dx x Separating the variables and integrating gives 1 1 − y −2 = − x−2 + C, 2 2 which can be written in the equivalent form y 2 − x2 = kx2 y 2 . 7. (a). We first determine the slope of the given family at the point (x, y). Differentiating x2 + 3y 2 = 2cy (0.0.15) with respect to x yields 2x + 6y so that dy dy = 2c dx dx dy x = . dx c − 3y 2 (0.0.16) 2 From (0.0.15) we have c = x +3y which, when substituted into Equation (0.0.16) yields 2y dy 2xy x = 2 , = x2 +3y2 dx x − 3y 2 − 3y 2y as required. (c)2017 Pearson Education. Inc. (0.0.17) 96 (b). It follows from Equation(0.0.17) that the differential equation for the orthogonal trajectories is dy 3y 2 − x2 =− . dx 2xy This differential equation is first-order homogeneous. Substituting y = xV into the preceding differential equation gives 3V 2 − 1 dV +V = x dx 2V which simplifies to dV V2−1 = . dx 2V Separating the variables and integrating we obtain ln(V 2 − 1) = ln x + C, or, upon exponentiation, V 2 − 1 = kx. Inserting V = y/x into the preceding equation yields y2 − 1 = kx, x2 that is, y 2 − x2 = kx3 . 8. Slope field. 9. Slope field. 10. Slope field. 11. See accompanying figure. 12. Slope field. 13. (a). If v(t) = 25, then 1 dv = 0 = (25 − v). dt 2 (b). The accompanying figure suggests that lim v(t) = 25. t→∞ 14. (a). The equilibrium solutions are any constant values of m that satisfy m 1/4 = 0. am3/4 1 − M (c)2017 Pearson Education. Inc. 97 y(x) x Figure 0.0.57: Figure for Problem 11 v(t) 25 20 15 10 5 t 5 10 Figure 0.0.58: Figure for Problem 13 Hence, there are two equilibrium solutions, namely, m = 0 and m = M . (b).This follows since a > 0, and 0 < m(t) < M . (c)2017 Pearson Education. Inc. 98 (c). The given differential equation can be written in the equivalent form dm 1 3/4 = a m − 1/4 m , dt M dm 3 −1/4 1 − 1/4 m 4 dt M m 1/4 1 2 3/4 3/4 =a m m − 1/4 m 1 − M M 1/4 m m 1/4 1 2 1/2 3−4 1− . = a m 4 M M so that d2 m =a dt2 Since 0 < m < M , the expression on the right-hand side of the preceding equation is positive when 3−4 m 1/4 M > 0, m 1/4 3 81 < , or equivalently, m < M . Consequently, the solution curves are concave up for M 4 256 81 81 0<m< M , and concave down for M < m < M. 256 256 81 (d). From the results of (c), there is a change in concavity when m = M . Substituting this value of 256 m into the right-hand side of the given differential equation yields the following value for the slope of the solutions curves at the point of inflection: 3/4 1/4 81 27 81 1− = a. a 256 256 256 that is, (e). Slope field. 15. (a). Separating the variables in Equation (1.12.6) yields mv dv =1 mg − kv 2 dy which can be integrated to obtain m ln(mg − kv 2 ) = y + c. 2k Multiplying both sides of this equation by −1 and exponentiating gives − 2k mg − kv 2 = c1 e− m y . The initial condition v(0) = 0 requires that c1 = mg, which, when inserted into the preceding equation yields 2k mg − kv 2 = mge− m y , or equivalently, v2 = 2k mg 1 − e− m y , k (c)2017 Pearson Education. Inc. 99 v2(y) mg/k y Figure 0.0.59: Figure for Problem 15 as required. (b). See accompanying figure. 16. By inspection the differential equation is separable, but not first-order homogeneous. Further, the differential equation can be re-written as dy + x2 y = x2 y 2 dx which reveals that it is also a Bernoulli equation, but is not linear. Finally, a different rearrangement of terms yields 1 x2 dx − dy = 0, y(y − 1) which is an exact differential equation. Separating the variables in the given differential equation yields 1 1 1 2 dy = x dx =⇒ dy = x3 + c y(y − 1) y(y − 1) 3 Using a partial fraction decomposition of the integrand on the left-hand side of the preceding equation we obtain 3 1 y−1 y−1 1 1 1 − dy = x3 + c =⇒ ln = x3 + c =⇒ = c1 ex /3 y−1 y 3 y 3 y so that 1 . y(x) = 1 − c1 ex3 /3 41. Writing the differential equation in the form e−y dy = xex · dx 1+y (c)2017 Pearson Education. Inc. 100 we see that it is separable, but not homogeneous. It is also neither a linear differential equation nor a Bernoulli differential equation. Rearranging the differential equation we have xex dx − ey (1 + y)dy = 0 which is exact. Separating the variables in the given differential equation and integrating yields ey (1 + y)dy = xex dx + c =⇒ yey = ex (x − 1) + c. 17. The given differential equation is separable. Separating the variables gives y dy ln x =2 , dx x which can be integrated directly to obtain 1 2 y = (ln x)2 + c, 2 or, equivalently, y 2 = 2(ln x)2 + c1 . 18. The given differential equation is first-order linear. We first divide by x to put the differential equation in standard form: dy 2 − y = 2x ln x. (0.0.18) dx x An integrating factor for this equation is I = e (−2/x)dx = x−2 . Multiplying Equation (0.0.18) by x−2 reduces it to d −2 (x y) = 2x−1 ln x, dx which can be integrated to obtain x−2 y = (ln x)2 + c so that y(x) = x2 [(ln x)2 + c]. 19. We first re-write the given differential equation in the differential form 2xy dx + (x2 + 2y)dy = 0. (0.0.19) Then My = 2x = Nx so that the differential equation is exact. Consequently, there exists a potential function φ satisfying ∂φ = 2xy, ∂x ∂φ = x2 + 2y. ∂y Integrating these two equations in the usual manner yields φ(x, y) = x2 y + y 2 . (c)2017 Pearson Education. Inc. 101 Therefore Equation (0.0.19) can be written in the equivalent form d(x2 y + y 2 ) = 0 with general solution x2 y + y 2 = c. 20. We first rewrite the given differential equation as dy y 2 + 3xy + x2 , = dx x2 which is first order homogeneous. Substituting y = xV into the preceding equation yields x so that x dV + V = V 2 + 3V + 1 dx dV = V 2 + 2V + 1 = (V + 1)2 , dx or, in separable form, 1 dV 1 = . 2 (V + 1) dx x This equation can be integrated to obtain −(V + 1)−1 = ln x + c so that V +1= 1 . c1 − ln x Inserting V = y/x into the preceding equation yields 1 y +1= , x c1 − ln x so that y(x) = x − x. c1 − ln x 21. We first rewrite the given differential equation in the equivalent form dy + y · tan x = −y 2 sin x, dx which is a Bernoulli equation. Dividing this equation by y 2 yields y −2 dy + y −1 tan x = − sin x. dx Now make the change of variables u = y −1 in which case Equation (0.0.20) gives the linear differential equation − (0.0.20) dy du = −y −2 . Substituting these results into dx dx du + u · tan x = − sin x dx (c)2017 Pearson Education. Inc. 102 or, in standard form, du − u · tan x = sin x. dx (0.0.21) An integrating factor for this differential equation is I = e− tan x dx = cos x. Multiplying Equation (0.0.21) by cos x reduces it to d (u · cos x) = sin x cos x dx which can be integrated directly to obtain 1 u · cos x = − cos2 x + c1 , 2 so that u= − cos2 x + c2 . cos x Inserting u = y −1 into the preceding equation and rearranging yields y(x) = −2 cos x 2 cos x = . − cos2 x + c2 cos2 x + c 22. The given differential equation is linear with integrating factor I=e 2e2x dx 1+e2x 2x = eln(1+e ) = 1 + e2x . Multiplying the given differential equation by 1 + e2x yields e2x + 1 d 2e2x (1 + e2x )y = 2x = −1 + 2x dx e −1 e −1 which can be integrated directly to obtain (1 + e2x )y = −x + ln |e2x − 1| + c, so that y(x) = −x + ln |e2x − 1| + c . 1 + e2x 23. We first rewrite the given differential equation in the equivalent form dy y + x2 − y 2 = , dx x which we recognize as being first order homogeneous. Inserting y = xV into the preceding equation yields x |x| dV 1 − V 2, +V =V + dx x that is, √ dV 1 1 =± . x 1 − V 2 dx (c)2017 Pearson Education. Inc. 103 Integrating we obtain sin−1 V = ± ln |x| + c, so that V = sin(c ± ln |x|). Inserting V = y/x into the preceding equation yields y(x) = x sin(c ± ln |x|). 24. We first rewrite the given differential equation in the equivalent form (sin y + y cos x + 1)dx − (1 − x cos y − sin x)dy = 0. Then My = cos y + cos x = Nx so that the differential equation is exact. Consequently, there is a potential function satisfying ∂φ = sin y + y cos x + 1, ∂x ∂φ = −(1 − x cos y − sin x). ∂y Integrating these two equations in the usual manner yields φ(x, y) = x − y + x sin y + y sin x, so that the differential equation can be written as d(x − y + x sin y + y sin x) = 0, and therefore has general solution x − y + x sin y + y sin x = c. 25. Writing the given differential equation as dy 1 25 −1 2 + y= y x ln x, dx x 2 we see that it is a Bernoulli equation with n = −1. We therefore divide the equation by y −1 to obtain y dy 25 2 1 + y2 = x ln x. dx x 2 dy We now make the change of variables u = y 2 , in which case, du dx = 2y dx . Inserting these results into the preceding differential equation yields 1 du 1 25 2 + u= x ln x, 2 dx x 2 or, in standard form, du 2 + u = 25x2 ln x. dx x An integrating factor for this linear differential equation is I = e (2/x)dx = x2 . Multiplying the previous differential equation by x2 reduces it to d 2 (x u) = 25x4 ln x dx (c)2017 Pearson Education. Inc. 104 which can be integrated directly to obtain 2 x u = 25 so that 1 5 1 x ln x − x5 5 25 +c u = x3 (5 ln x − 1) + cx−2 . Making the replacement u = y 2 in this equation gives y 2 = x3 (5 ln x − 1) + cx−2 . 26. The given differential equation can be written in the equivalent form dy ex−y = 2x+y = e−x e−2y , dx e which we recognize as being separable. Separating the variables gives e2y dy = e−x dx which can be integrated to obtain 1 2y e = −e−x + c 2 so that y(x) = 1 ln(c1 − 2e−x ). 2 27. The given differential equation is linear with integrating factor I = e cot x dx = sin x. Multiplying the given differential equation by sin x reduces it to d sin x (y sin x) = dx cos x which can be integrated directly to obtain y sin x = − ln(cos x) + c, so that y(x) = c − ln(cos x) . sin x 28. Writing the given differential equation as 1 2ex dy y = 2y 2 e−x , + x dx 1 + e 1 we see that it is a Bernoulli equation with n = 1/2. We therefore divide the equation by y 2 to obtain 1 y− 2 dy 2ex 1 y 2 = 2e−x . + dx 1 + ex (c)2017 Pearson Education. Inc. 105 1 1 1 − 2 dy We now make the change of variables u = y 2 , in which case, du dx = 2 y dx . Inserting these results into the preceding differential equation yields du 2ex u = 2e−x , 2 + dx 1 + ex or, in standard form, ex du u = e−x . + dx 1 + ex An integrating factor for this linear differential equation is I=e ex 1+ex dx = eln(1+e ) = 1 + ex . x Multiplying the previous differential equation by 1 + ex reduces it to d [(1 + ex )u] = e−x (1 + ex ) = e−x + 1 dx which can be integrated directly to obtain (1 + ex )u = −e−x + x + c so that u= x − e−x + c . 1 + ex 1 Making the replacement u = y 2 in this equation gives 1 y2 = x − e−x + c . 1 + ex 29. We first rewrite the given differential equation in the equivalent form dy y y = ln +1 . dx x x The function appearing on the right of this equation is homogeneous of degree zero, and therefore the differential equation itself is first order homogeneous. We therefore insert y = xV into the differential equation to obtain dV x + V = V (ln V + 1), x so that dV = V ln V. x dx Separating the variables yields 1 dV 1 = V ln V dx x which can be integrated to obtain ln(ln V ) = ln x + c. Exponentiating both side of this equation gives ln V = c1 x, (c)2017 Pearson Education. Inc. 106 or equivalently, V = e c1 x . Inserting V = y/x in the preceding equation yields y = xec1 x . 30. For the given differential equation we have M (x, y) = 1 + 2xey , N (x, y) = −(ey + x), so that My − N x 1 + 2xey = 1. = M 1 + 2xey Consequently, an integrating factor for the given differential equation is I = e− dy = e−y . Multiplying the given differential equation by e−y yields the exact differential equation (2x + e−y )dx − (1 + xe−y )dy = 0. (0.0.22) Therefore, there exists a potential function φ satisfying ∂φ = 2x + e−y , ∂x ∂φ = −(1 + xe−y ). ∂y Integrating these two equations in the usual manner yields φ(x, y) = x2 − y + xe−y . Therefore Equation (0.0.22) can be written in the equivalent form d(x2 − y + xe−y ) = 0 with general solution x2 − y + xe−y = c. 31. The given differential equation is first-order linear. However, it can also e written in the equivalent form dy = (1 − y) sin x dx which is separable. Separating the variables and integrating yields − ln |1 − y| = − cos x + c, so that Hence, 1 − y = c1 ecos x . y(x) = 1 − c1 ecos x . (c)2017 Pearson Education. Inc. 107 32. For the given differential equation we have M (x, y) = 3y 2 + x2 , N (x, y) = −2xy, so that My − N x 4 =− . N x Consequently, an integrating factor for the given differential equation is I = e− 4 x dx = x−4 . Multiplying the given differential equation by x−4 yields the exact differential equation (3y 2 x−4 + x−2 )dx − 2yx−3 dy = 0. (0.0.23) Therefore, there exists a potential function φ satisfying ∂φ = 3y 2 x−4 + x−2 , ∂x ∂φ = −2yx−3 . ∂y Integrating these two equations in the usual manner yields φ(x, y) = −y 2 x−3 − x−1 . Therefore Equation (0.0.23) can be written in the equivalent form d(−y 2 x−3 − x−1 ) = 0 with general solution −y 2 x−3 − x−1 = c, or equivalently, x2 + y 2 = c 1 x3 . Notice that the given differential equation can be written in the equivalent form dy 3y 2 + x2 = , dx 2xy which is first-order homogeneous. Another equivalent way of writing the given differential equation is dy 3 1 − y = xy −1 , dx 2x 2 which is a Bernoulli equation. 33. The given differential equation can be written in the equivalent form dy 1 9 − y = − x2 y 3 , dx 2x ln x 2 which is a Bernoulli equation with n = 3. We therefore divide the equation by y 3 to obtain y −3 dy 9 1 − y −2 = − x2 . dx 2x ln x 2 (c)2017 Pearson Education. Inc. 108 −3 dy We now make the change of variables u = y −2 , in which case, du dx = −2y dx . Inserting these results into the preceding differential equation yields 1 du 1 9 − u = − x2 , 2 dx 2x ln x 2 − or, in standard form, du 1 + u = 9x2 . dx x ln x An integrating factor for this linear differential equation is 1 x ln x dx I=e = eln(ln x) = ln x. Multiplying the previous differential equation by ln x reduces it to d (ln x · u) = 9x2 ln x dx which can be integrated to obtain ln x · u = x3 (3 ln x − 1) + c so that x3 (3 ln x − 1) + c . ln x Making the replacement u = y 3 in this equation gives u= y3 = x3 (3 ln x − 1) + c . ln x 34. Separating the variables in the given differential equation yields 1 dy 2+x 1 = =1+ , y dx 1+x 1+x which can be integrated to obtain ln |y| = x + ln |1 + x| + c. Exponentiating both sides of this equation gives y(x) = c1 (1 + x)ex . 35. The given differential equation can be written in the equivalent form 2 dy + y=1 dx x2 − 1 (0.0.24) which is first-order linear. An integrating factor is I=e 2 dx x2 −1 1 1 = e ( x−1 − x+1 )dx = e[ln(x−1)−ln(x+1)] = Multiplying (0.0.24) by (x − 1)/(x + 1) reduces it to the integrable form x−1 2 d x−1 ·y = =1− . dx x + 1 x+1 x+1 (c)2017 Pearson Education. Inc. x−1 . x+1 109 Integrating both sides of this differential equation yields x−1 · y = x − 2 ln(x + 1) + c x+1 so that y(x) = x+1 x−1 [x − 2 ln(x + 1) + c]. 36. The given differential equation can be written in the equivalent form [y sec2 (xy) + 2x]dx + x sec2 (xy)dy = 0 Then My = sec2 (xy) + 2xy sec2 (x) tan(xy) = Nx so that the differential equation is exact. Consequently, there is a potential function satisfying ∂φ = y sec2 (xy) + 2x, ∂x ∂φ = x sec2 (xy). ∂y Integrating these two equations in the usual manner yields φ(x, y) = x2 + tan(xy), so that the differential equation can be written as d(x2 + tan(xy)) = 0, and therefore has general solution x2 + tan(xy) = c, or equivalently, y(x) = tan−1 (c − x2 ) . x 37. The given differential equation is first-order homogeneous. Inserting y = xV into the given equation yields dV 1 x + V, +V = dx 1+V2 that is, (1 + V 2 ) dV 1 = . dx x Integrating we obtain 1 V + V 3 = ln |x| + c. 3 Inserting V = y/x into the preceding equation yields y3 y + 3 = ln |x| + c, x 3x (c)2017 Pearson Education. Inc. 110 or equivalently, 3x2 y + y 3 = 3x3 (ln |x| + c). dy 38. The differential equation is first-order homogeneous. We therefore let y = xv, in which case = dx dv + v. Substituting these results into the given differential equation yields x dx x so that x 1 + v2 dv , +v = dx 1 − 3v 2 dv (v + 1)(3v 2 − 2v + 1) 1 + v 2 − v + 3v 3 = . = dx 1 − 3v 2 1 − 3v 2 Separating the variables gives 1 − 3v 2 1 dv = dx. (v + 1)(3v 2 − 2v + 1) x Decomposing the left-hand side into partial fractions yields 1 1 2(3v − 2) dv = dx, − − 2 3(v + 1) 3(3v − 3v + 1) x or equivalently, − 1 1 2(3v − 2) − 2 2 dv = dx. 1 3(v + 1) x + v− 3 9 This can be integrated to yield 1 3v − 1 3 2 1 3 √ − ln(v + 1) = ln x + c. − ln (3v − 1)2 + 2 √ arctan 2 9 9 3 2 2 Therefore, 2 3 √ arctan 2 3y − x √ 2x − 3 2 (3y − x)2 1 y+x + ln − ln = ln x + c. 2 9x2 9 3 x 39. The given differential equation is a Bernoulli equation with n = −1. We therefore divide the equation by y −1 to obtain 25 ln x 1 dy . + y2 = y dx x 2x3 dy We now make the change of variables u = y 2 , in which case, du dx = 2y dx . Inserting these results into the preceding differential equation yields 1 du 1 25 ln x , + u= 2 dx x 2x3 or, in standard form, du 2 + u = 25x−3 ln x. dx x An integrating factor for this linear differential equation is I=e 2 x dx = x2 . (c)2017 Pearson Education. Inc. 111 Multiplying the previous differential equation by x2 reduces it to d 2 (x u) = 25x−1 ln x, dx which can be integrated directly to obtain x2 u = so that u= 25 (ln x)2 + c 2 25(ln x)2 + c . 2x2 Making the replacement u = y 2 in this equation gives y2 = 25(ln x)2 + c . 2x2 dy 40. The differential equation is first-order homogeneous. We therefore let y = xv, in which case = dx dv + v. Substituting these results into the given differential equation yields x dx x so that x dv 1 + v2 , +v = dx 1 − 3v 2 dv (v + 1)(3v 2 − 2v + 1) 1 + v 2 − v + 3v 3 = . = dx 1 − 3v 2 1 − 3v 2 Separating the variables gives 1 − 3v 2 1 dv = dx. 2 (v + 1)(3v − 2v + 1) x Decomposing the left-hand side into partial fractions yields 2(3v − 2) 1 1 − dv = dx, − 2 3(v + 1) 3(3v − 3v + 1) x or equivalently, − 1 1 2(3v − 2) dv = dx. − 1 2 2 3(v + 1) x + v− 3 9 This can be integrated to yield 1 3 2 1 3v − 1 3 2 √ − ln(v + 1) = ln x + c. − ln (3v − 1) + 2 √ arctan 2 9 9 3 2 2 Therefore, 2 3 √ arctan 2 3y − x √ 2x − 1 3 2 (3y − x)2 y+x − = ln x + c. + ln ln 2 9x2 9 3 x (c)2017 Pearson Education. Inc. 112 41. The given differential equation can be written in the equivalent form ey (1 + y) dy = xex dx which is separable. Integrating both sides of this equation gives yey = ex (x − 1) + c. 42. The given differential equation can be written in the equivalent form dy cos x − y = − cos x dx sin x which is first order linear with integrating factor I = e− cos x sin x dx = e− ln(sin x) = 1 . sin x Multiplying the preceding differential equation by sin1 x reduces it to 1 d cos x ·y =− dx sin x sin x which can be integrated directly to obtain 1 · y = − ln(sin x) + c sin x so that y(x) = sin x[c − ln(sin x)]. 43. The given differential equation is linear, and therefore can be solved using an appropriate integrating factor. However, if we rearrange the terms in the given differential equation then it can be written in the equivalent form 1 dy = x2 1 + y dx which is separable. Integrating both sides of the preceding differential equation yields 1 3 x +c 3 ln(1 + y) = so that 1 3 y(x) = c1 e 3 x − 1. Imposing the initial condition y(0) = 5 we find c1 = 6. Therefore the solution to the initial-value problem is 1 3 y(x) = 6e 3 x − 1. 44. The given differential equation can be written in the equivalent form e−6y dy = −e−4x dx (c)2017 Pearson Education. Inc. 113 which is separable. Integrating both sides of the preceding equation yields 1 1 − e−6y = e−4x + c 6 4 so that 3 −4x 1 . y(x) = − ln c1 − e 6 2 Imposing the initial condition y(0) = 0 requires that 3 . 0 = ln c1 − 2 Hence, c1 = 52 , and so 1 y(x) = − ln 6 5 − 3e−4x 2 . 45. For the given differential equation we have My = 4xy = Nx so that the differential equation is exact. Consequently, there is a potential function satisfying ∂φ = 3x2 + 2xy 2 , ∂x ∂φ = 2x2 y. ∂y Integrating these two equations in the usual manner yields φ(x, y) = x2 y 2 + x3 , so that the differential equation can be written as d(x2 y 2 + x3 ) = 0, and therefore has general solution x2 y 2 + x3 = c. Imposing the initial condition y(1) = 3 yields c = 10. Therefore, x2 y 2 + x3 = 10 so that 10 − x3 . x2 Note that the given differential equation can be written in the equivalent form y2 = 1 3 dy + y = − y −1 , dx x 2 which is a Bernoulli equation with n = −1. Consequently, the Bernoulli technique could also have been used to solve the differential equation. (c)2017 Pearson Education. Inc. 114 46. The given differential equation is linear with integrating factor I = e− sin x dx = ecos x . Multiplying the given differential equation by ecos x reduces it to the integrable form d cos x · y) = 1, (e dx which can be integrated directly to obtain ecos x · y = x + c.. Hence, y(x) = e− cos x (x + c). Imposing the given initial condition y(0) = 1e requires that c = 1. Consequently, y(x) = e− cos x (x + 1). 47. (a). For the given differential equation we have My = my m−1 , Nx = −nxn−1 y 3 . We see that the only values for m and n for which My = Nx are m = n0. Consequently, these are the only values of m and n for which the differential equation is exact. (b). We rewrite the given differential equation in the equivalent form dy x5 + y m , = dx xn y 3 (0.0.25) from which we see that the differential equation is separable provided m = 0. In this case there are no restrictions on n. (c). From Equation (0.0.25) we see that the only values of m and n for which the differential equation is first-order homogeneous are m = 5 and n = 2. (d). We now rewrite the given differential equation in the equivalent form dy − x−n y m−3 = x5−n y −3 . dx (0.0.26) Due to the y −3 term on the right-hand side of the preceding differential equation, it follows that there are no values of m and n for which the equation is linear. (e). From Equation (0.0.26) we see that the differential equation is a Bernoulli equation whenever m = 4. There are no constraints on n in this case. 48. In Newton’s Law of Cooling we have Tm = 180◦ F, T (0) = 80◦ F, T (3) = 100◦ F. We need to determine the time, t0 when T (t0 ) = 140◦ F. The temperature of the sandals at time t is governed by the differential equation dT = −k(T − 180). dt (c)2017 Pearson Education. Inc. 115 This separable differential equation is easily integrated to obtain T (t) = 180 + ce−kt . Since T (0) = 80 we have 80 = 180 + c =⇒ c = −100. Hence, T (t) = 180 − 100e−kt . Imposing the condition T (3) = 100 requires 100 = 180 − 100e−3k . Solving for k we find k = 13 ln 5 4 . Inserting this value for k into the preceding expression for T (t) yields t 5 t0 5 T (t) = 180 − 100e− 3 ln( 4 ) . We need to find t0 such that 140 = 180 − 100e− 3 ln( 4 ) . Solving for t0 we find ln 52 ≈ 12.32 min. t0 = 3 ln 54 49. In Newton’s Law of Cooling we have Tm = 70◦ F, T (0) = 150◦ F, T (10) = 125◦ F. We need to determine the time, t0 when T (t0 ) = 100◦ F. The temperature of the plate at time t is governed by the differential equation dT = −k(T − 70). dt This separable differential equation is easily integrated to obtain T (t) = 70 + ce−kt . Since T (0) = 150 we have 150 = 70 + c =⇒ c = 80. Hence, T (t) = 70 + 80e−kt . Imposing the condition T (10) = 125 requires 125 = 70 + 80e−10k . 1 ln Solving for k we find k = 10 16 11 . Inserting this value for k into the preceding expression for T (t) yields t 16 t0 16 T (t) = 70 + 80e− 10 ln( 11 ) . We need to find t0 such that 100 = 70 + 80e− 10 ln( 11 ) . (c)2017 Pearson Education. Inc. 116 Solving for t0 we find ln 83 ≈ 26.18 min. t0 = 10 ln 16 11 50. Let T (t) denote the temperature of the object at time t, and let Tm denote the temperature of the surrounding medium. Then we must solve the initial-value problem dT = k(T − Tm )2 , dt T (0) = T0 , where k is a constant. The differential equation can be written in separated form as 1 dT = k. (T − Tm )2 dt Integrating both sides of this differential equation yields − 1 = kt + c T − Tm so that T (t) = Tm − 1 . kt + c Imposing the initial condition T (0) = T0 we find that c= 1 Tm − T0 which, when substituted back into the preceding expression for T (t) yields T (t) = Tm − 1 kt + Tm1−T0 = Tm − Tm − T0 . k(Tm − T0 )t + 1 As t → ∞, T (t) approaches Tm . dv (0) = 2, the velocity is increasing at the rate of 2 m/s2 at t = 0. dt (b). Evaluating the given differential equation at t = 0, using the given initial conditions yields 51.(a). Since 2 + 20k = 80k =⇒ k = 1 . 3 (c). An integrating factor for the given differential equation is I = e k dt = ekt . Multiplying the given d differential equation by this integrating factor reduces is to (ekt ·v) = 80k =⇒ v(t) = e−kt (80t+c). Imposing dt 4 the initial condition v(0) = 20 yields c = 20, so that v(t) = 20e−kt (4kt + 1) =⇒ v(t) = e−t/30 (2t + 15). 3 4 −t/30 (d). v(t) = e (2t + 15) =⇒ there is no finite t > 0 when v(t) = 0. Hence the object does not come to 3 rest in a finite time. (e). limt→∞ v(t) = 0. (c)2017 Pearson Education. Inc. 117 52. We are given the differential equation dT = −k(T − 5 cos 2t) dt (0.0.27) dT (0) = 5. dt (0.0.28) together with the initial conditions T (0) = 0; (a). Setting t = 0 in (0.0.27) and using (0.0.28) yields 5 = −k(0 − 5) so that k = 1. (b). Substituting k = 1 into the differential equation (0.0.27) and rearranging terms yields dT + T = 5 cos t. dt An integrating factor for this linear differential equation is I = e dt = et . Multiplying the preceding differential equation by et reduces it to d t (e · T ) = 5et cos 2t dt which upon integration yields et · T = et (cos 2t + 2 sin 2t) + c, so that T (t) = ce−t + cos 2t + 2 sin 2t. Imposing the initial condition T (0) = 0 we find that c = −1. Hence, T (t) = cos 2t + 2 sin 2t − e−t . (c). For large values of t we have T (t) ≈ cos 2t + 2 sin 2t, which can be written in phase-amplitude form as T (t) ≈ √ 5 cos(2t − φ), where tan φ√= 2. consequently, for large t, the temperature is approximately oscillatory with period π and amplitude 5. 53. If we let C(t) denote the number of sandhill cranes in the Platte River valley t days after April 1, then C(t) is governed by the differential equation dC = −kC dt (0.0.29) together with the auxiliary conditions C(0) = 500, 000; C(15) = 100, 000. (c)2017 Pearson Education. Inc. (0.0.30) 118 Separating the variables in the differential equation (0.0.29) yields 1 dC = −k, C dt which can be integrated directly to obtain ln C = −kt + c. Exponentiation yields C(t) = c0 e−kt . The initial condition C(0) = 500, 000 requires c0 = 500, 000, so that C(t) = 500, 000e−kt . (0.0.31) Imposing the auxiliary condition C(15) = 100, 000 yields 100, 000 = 500, 000e−15k . 1 Taking the natural logarithm of both sides of the preceding equation and simplifying we find that k = 15 ln 5. Substituting this value for k into (0.0.31) gives C(t) = 500, 000e− 15 ln 5 . t (0.0.32) 1 = 20, 000 sandhile cranes. (a). C(3) = 500, 000e−2 ln 5 = 500, 000 · 25 35 (b). C(35) = 500, 000e− 15 ln 5 ≈ 11696 sandhile cranes. (c). We need to determine t0 such that t0 1000 = 500, 000e− 15 ln 5 that is, 1 . 500 Taking the natural logarithm of both sides of this equation and simplifying yields t0 e− 15 ln 5 = t0 = 15 · ln 500 ≈ 57.9 days after April 1. ln 5 54. Substituting P0 = 200, 000 into Equation (1.5.3) in the text yields P (t) = 200, 000C . 200, 000 + (C − 200, 000)e−rt (0.0.33) We are given P (3) = P (t1 ) = 230, 000, P (6) = P (t2 ) = 250, 000. Since t2 = 2t1 we can use the formulas (1.5.5) and (1.5.6) of the text to obtain r and C directly as follows: 1 1 25(23 − 20) 15 = ln ≈ 0.21. r = ln 3 20(25 − 23) 3 8 (c)2017 Pearson Education. Inc. 119 230, 000[(23)(45) − (40)(25)] = 277586. (23)2 − (20)(25) C= Substituting these values for r and C into (0.0.33) yields P (t) = 55517200000 . 200, 000 + (77586)e−0.21t Therefore, P (10) = 55517200000 ≈ 264, 997, 200, 000 + (77586)e−2.1 P (20) = 55517200000 ≈ 275981. 200, 000 + (77586)e−4.2 and 55. The differential equation for determining q(t) is dq 5 3 + q = cos 2t, dt 4 2 which has integrating factor I = e it to the integrable form 5 4 dt 5 5 = e 4 t . Multiplying the preceding differential equation by e 4 t reduces d 5t 3 5t e 4 · q = e 4 cos 2t. dt 2 Integrating and simplifying we find q(t) = 5 6 (5 cos 2t + 8 sin 2t) + ce− 4 t . 89 (0.0.34) The initial condition q(0) = 3 requires 3= 30 + c, 89 so that c = 237 89 . Making this replacement in (0.0.34) yields q(t) = 6 237 − 5 t (5 cos 2t + 8 sin 2t) + e 4 . 89 89 The current in the circuit is i(t) = dq 12 1185 − 5 t = (8 cos 2t − 5 sin 2t) − e 4 . dt 89 356 56. The current in the circuit is governed by the differential equation di 100 + 10i = , dt 3 which has integrating factor I = e 10 dt = e10t . Multiplying the preceding differential equation by e10t reduces it to the integrable form d 10t 100 10t e ·i = e . dt 3 (c)2017 Pearson Education. Inc. 120 Integrating and simplifying we find i(t) = 10 + ce−10t . 3 (0.0.35) The initial condition i(0) = 3 requires 10 + c, 3 so that c = − 13 . Making this replacement in (0.0.35) yields 3= i(t) = 1 (10 − e−10t ). 3 57. We are given: r1 = 6 L/min, c1 = 3 g/L, r2 = 4 L/min, V (0) = 30 L, A(0) = 0 g, and we need to determine the amount of salt in the tank when V (t) = 60L. Consider a small time interval Δt. Using the preceding information we have: ΔV = 6Δt − 4Δt = 2Δt, and A Δt. V Dividing both of these equations by Δt and letting Δt → 0 yields ΔA ≈ 18Δt − 4 dV = 2. dt (0.0.36) A dA + 4 = 18. dt V Integrating (0.0.36) and imposing the initial condition V (0) = 30 yields V (t) = 2(t + 15). (0.0.37) (0.0.38) We now insert this expression for V (t) into (0.0.37) to obtain dA 2 + A = 18. dt t + 15 2 An integrating factor for this differential equation is I = e t+15 dt = (t + 15)2 . Multiplying the preceding differential equation by (t + 15)2 reduces it to the integrable form d (t + 15)2 A = 18(t + 15)2 . dt Integrating and simplifying we find A(t) = 6(t + 15)3 + c . (t + 15)2 Imposing the initial condition A(0) = 0 requires 0= 6(15)3 + c , (15)2 (c)2017 Pearson Education. Inc. 121 so that c = −20250. Consequently, A(t) = 6(t + 15)3 − 20250 . (t + 15)2 We need to determine the time when the solution overflows. Since the tank can hold 60 L of solution, from (0.0.38) overflow will occur when 60 = 2(t + 15) =⇒ t = 15. The amount of chemical in the tank at this time is A(15) = 6(30)3 − 20250 ≈ 157.5 g. (30)2 58. Applying Euler’s method with y = x2 + 2y 2 , x0 = 0, y0 = −3, and h = 0.1 we have yn+1 = yn + 0.1(x2n + 2yn2 ). This generates the sequence of approximants given in the table below. n 1 2 3 4 5 6 7 8 9 10 xn 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0 yn −1.2 −0.911 −0.74102 −0.62219 −0.52877 −0.44785 −0.371736 −0.29510 −0.21368 −0.12355 Consequently the Euler approximation to y(1) is y10 = −0.12355. 3x + 2, x0 = 1, y0 = 2, and h = 0.05 we have y 3xn +2 . yn+1 = yn + 0.05 yn 59. Applying Euler’s method with y = This generates the sequence of approximants given in the table below. n 1 2 3 4 5 6 7 8 9 10 xn 1.05 1.10 1.15 1.20 1.25 1.30 1.35 1.40 1.45 1.50 yn 2.1750 2.34741 2.51770 2.68622 2.85323 3.01894 3.18353 3.34714 3.50988 3.67185 (c)2017 Pearson Education. Inc. 122 Consequently, the Euler approximation to y(1.5) is y10 = 3.67185. 60. Applying the modified Euler method with y = x2 + 2y 2 , x0 = 0, y0 = −3, and h = 0.1 generates the sequence of approximants given in the table below. n 1 2 3 4 5 6 7 8 9 10 xn 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0 yn −1.9555 −1.42906 −1.11499 −0.90466 −0.74976 −0.62555 −0.51778 −0.41723 −0.31719 −0.21196 Consequently, the modified Euler approximation to y(1) is y10 = −0.21196. Comparing this to the corresponding Euler approximation from Problem 58 we have |yME − yE | = |0.21196 − 0.12355| = 0.8841. 61. Applying the modified Euler method with y = 3x + 2, x0 = 1, y0 = 2, and h = 0.05 generates the y sequence of approximants given in the table below. n 1 2 3 4 5 6 7 8 9 10 xn 1.05 1.10 1.15 1.20 1.25 1.30 1.35 1.40 1.45 1.50 yn 2.17371 2.34510 2.51457 2.68241 2.84886 3.01411 3.17831 3.34159 3.50404 3.66576 Consequently, the modified Euler approximation to y(1.5) is y10 = 3.66576. Comparing this to the corresponding Euler approximation from Problem 59 we have |yME − yE | = |3.66576 − 3.67185| = 0.00609. 62. Applying the Runge-Kutta method with y = x2 + 2y 2 , x0 = 0, y0 = −3, and h = 0.1 generates the sequence of approximants given in the table below. (c)2017 Pearson Education. Inc. 123 n 1 2 3 4 5 6 7 8 9 10 xn 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0 yn −1.87392 −1.36127 −1.06476 −0.86734 −0.72143 −0.60353 −0.50028 −0.40303 −0.30541 −0.20195 Consequently the Runge-Kutta approximation to y(1) is y10 = −0.20195. Comparing this to the corresponding Euler approximation from Problem 58 we have |yRK − yE | = |0.20195 − 0.12355| = 0.07840. 63. Applying the Runge-Kutta method with y = 3x + 2, x0 = 1, y0 = 2, and h = 0.05 generates the y sequence of approximants given in the table below. n 1 2 3 4 5 6 7 8 9 10 xn 1.05 1.10 1.15 1.20 1.25 1.30 1.35 1.40 1.45 1.50 yn 2.17369 2.34506 2.51452 2.68235 2.84880 3.01404 3.17823 3.34151 3.50396 3.66568 Consequently the Runge-Kutta approximation to y(1.5) is y10 = 3.66568. Comparing this to the corresponding Euler approximation from Problem 59 we have |yRK − yE | = |3.66568 − 3.67185| = 0.00617. Chapter 2 Solutions Solutions to Section 2.1 True-False Review: (a): TRUE. A diagonal matrix has no entries below the main diagonal, so it is upper triangular. Likewise, it has no entries above the main diagonal, so it is also lower triangular. (b): FALSE. An m × n matrix has m row vectors and n column vectors. (c)2017 Pearson Education. Inc. 124 (c): TRUE. This is a square matrix, and all entries off the main diagonal are zero, so it is a diagonal matrix (the entries on the diagonal also happen to be zero, but this is not required). (d): FALSE. The main diagonal entries of a skew-symmetric matrix must be zero. In this case, a11 = 4 = 0, so this matrix is not skew-symmetric. (e): FALSE. The form presented uses the same number along the entire main diagonal, but a symmetric matrix need not have identical entries on the main diagonal. (f ): TRUE. Since A is symmetric, A = AT . Thus, (AT )T = A = AT , so AT is symmetric. (g): FALSE. The trace of a matrix is the sum of the entries along the main diagonal. (h): TRUE. If A is skew-symmetric, then AT = −A. But A and AT contain the same entries along the main diagonal, so for AT = −A, both A and −A must have the same main diagonal. This is only possible if all entries along the main diagonal are 0. (i): TRUE. If A is both symmetric and skew-symmetric, then A = AT = −A, and A = −A is only possible if all entries of A are zero. (j): TRUE. Both matrix functions are defined for values of t such that t > 0. (k): FALSE. The (3, 2)-entry contains a function that is not defined for values of t with t ≤ 3. So for example, this matrix functions is not defined for t = 2. (l): TRUE. Each numerical entry of the matrix function is a constant function, which has domain R. (m): FALSE. For instance, the matrix function A(t) = [t] and B(t) = [t2 ] satisfy A(0) = B(0), but A and B are not the same matrix function. Problems: 1(a). a31 = 0, a24 = −1, a14 = 2, a32 = 2, a21 = 7, a34 = 4. 1(b). (1, 4) and (3, 2). 2(a). b12 = −1, b33 = 4, b41 = 0, b43 = 8, b51 = −1, and b52 = 9. 2(b). (1, 2), (1, 3), (2, 1), (3, 2), and (5, 1). 1 5 3. ; 2 × 2 matrix. −1 3 2 1 −1 4. ; 2 × 3 matrix. 0 4 −2 ⎤ ⎡ −1 ⎢ 1 ⎥ ⎥ 5. ⎢ ⎣ 1 ⎦; 4 × 1 matrix. −5 ⎤ ⎡ 1 −3 −2 ⎢ 3 6 0 ⎥ ⎥; 4 × 3 matrix. 6. ⎢ ⎣ 2 7 4 ⎦ −4 −1 5 ⎡ ⎤ 0 −1 2 0 3 ⎦; 3 × 3 matrix. 7. ⎣ 1 −2 −3 0 (c)2017 Pearson Education. Inc. 125 ⎡ 0 ⎢ 1 ⎢ 8. ⎣ 2 3 ⎡ 2 ⎢ 3 9. ⎢ ⎣ 4 5 ⎤ −1 −2 −3 0 −1 −2 ⎥ ⎥; 4 × 4 matrix. 1 0 −1 ⎦ 2 1 −0 ⎤ 3 4 5 4 5 6 ⎥ ⎥; 4 × 4 matrix. 5 6 7 ⎦ 6 7 8 10. tr(A) = 1 + 3 = 4. 11. tr(A) = 1 + 2 + (−3) = 0. 12. tr(A) = 2 + 2 + (−5) = −1. 1 −1 13. Column vectors: , . 3 5 Row vectors: [1 − 1], [3 5]. ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 1 3 −4 14. Column vectors: ⎣ −1 ⎦ , ⎣ −2 ⎦ , ⎣ 5 ⎦. 2 6 7 Row vectors: [1 3 − 4], [−1 − 2 5], [2 6 7]. 2 10 6 15. Column vectors: , , . Row vectors: [2 10 6], [5 − 1 3]. 5 −1 3 ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 1 2 1 2 16. A = ⎣ 3 4 ⎦. Column vectors: ⎣ 3 ⎦ , ⎣ 4 ⎦. 5 1 5 1 −1 −1 −2 0 4 −1 −1 −2 0 4 , . 17. A = ; column vectors: , , , 0 8 9 −4 −4 0 8 9 −4 −4 ⎤ ⎡ −2 −4 ⎢ −6 −6 ⎥ ⎥ ⎢ 0 ⎥ 18. B = ⎢ ⎥; row vectors: −2 −4 , −6 −6 , 3 0 , −1 0 , −2 1 . ⎢ 3 ⎣ −1 0 ⎦ −2 1 ⎡ ⎤ 2 5 0 1 7 0 2 ⎦. Row vectors: [2 5 0 1], [−1 7 0 2], [4 − 6 0 3]. 19. B = ⎣ −1 4 −6 0 3 20. A = [a1 , a2 , . . . , ap ] has p columns and each column q-vector has q rows, so the resulting matrix has dimensions q × p. ⎡ ⎤ 2 0 0 0 ⎦. 21. One example: ⎣ 0 3 0 0 −1 ⎤ ⎡ 2 3 1 2 ⎢ 0 5 6 2 ⎥ ⎥ 22. One example: ⎢ ⎣ 0 0 3 5 ⎦. 0 0 0 1 (c)2017 Pearson Education. Inc. 126 ⎤ 1 3 −1 2 ⎢ −3 0 4 −3 ⎥ ⎥. 23. One example: ⎢ ⎣ 1 −4 0 1 ⎦ −2 3 −1 0 ⎡ ⎤ 3 0 0 24. One example: ⎣ 0 2 0 ⎦. 0 0 5 ⎡ ⎡ 0 25. The only possibility here is the zero matrix: ⎣ 0 0 ⎡ ⎤ 0 0 0 26. ⎣ 0 0 0 ⎦. 0 0 0 ⎤ ⎡ 2 t −t 0 ⎢ 0 0 ⎥ ⎥. 27. One example: ⎢ ⎣ 0 0 ⎦ 0 0 √ 1 √ t+2 0 3−t 28. One example: . 0 0 0 1 t2 +1 29. One example: . 0 30. One example: t2 + 1 1 1 1 0 0 0 ⎤ 0 0 ⎦. 0 1 . 31. One example: Let A and B be 1 × 1 matrix functions given by A(t) = [t] and B(t) = [t2 ]. 32. Let A be a symmetric upper triangular matrix. Then all elements below the main diagonal are zeros. Consequently, since A is symmetric, all elements above the main diagonal must also be zero. Hence, the only nonzero entries can occur along the main diagonal. That is, A is a diagonal matrix. 33. Since A is skew-symmetric, we know that aij = −aji for all (i, j). But since A is symmetric, we know that aij = aji for all (i, j). Thus, for all (i, j), we must have −aji = aji . That is, aji = 0 for all (i, j). That is, every element of A is zero. Solutions to Section 2.2 True-False Review: (a): FALSE. The correct statement is (AB)C = A(BC), the associative law. A counterexample to the particular statement given in this review item can be found in Problem 5. (b): TRUE. Multiplying from left to right, we note that AB is an m × p matrix, and right multiplying AB by the p × q matrix C, we see that ABC is an m × q matrix. (c): TRUE. We have (A + B)T = AT + B T = A + B, so A + B is symmetric. (c)2017 Pearson Education. Inc. 127 ⎡ ⎤ ⎡ 0 1 0 0 (d): FALSE. For example, let A = ⎣ −1 0 0 ⎦ , B = ⎣ 0 0 0 0 −3 ⎡ ⎤ 0 0 0 symmetric, but AB = ⎣ 0 0 −3 ⎦ is not symmetric. 0 0 0 0 0 0 ⎤ 3 0 ⎦. Then A and B are skew0 2 . The is false since AB (e): FALSE. The correct equation is (A+B)2 = A2+AB +BA+B +BA statement 1 1 1 0 0 1 2 does not necessarily equal 2AB. For instance, if A = and B = , then (A+B) = 0 0 0 0 0 0 1 2 = (A + B)2 . and A2 + 2AB + B 2 = 0 0 0 1 1 0 (f ): FALSE. For example, let A = and B = . Then AB = 0 even though A = 0 and 0 0 0 0 B = 0. 0 0 0 0 (g): FALSE. For example, let A = and let B = . Then A is not upper triangular, 1 0 0 0 despite the fact that AB is the zero matrix, hence automatically upper triangular. 1 0 (h): FALSE. For instance, the matrix A = is neither the zero matrix nor the identity matrix, 0 0 and yet A2 = A. (i): TRUE. The derivative of each entry of the matrix is zero, since in each entry, we take the derivative of a constant, thus obtaining zero for each entry of the derivative of the matrix. (j): FALSE. The correct statement is given in Problem 45. The problem with the statement as given is dA that the second term should be dA dt B, not B dt . t 0 2e satisfies A = dA (k): FALSE. For instance, the matrix function A = dt , but A does not have 0 3et t 0 ce . the form 0 cet (l): TRUE. This follows by exactly the same proof as given in the text for matrices of numbers (see part 3 of Theorem 2.2.23). Problems: −10 30 5 . −5 0 −15 −6 −3 3 1(b). −3B = . 0 −12 12 ⎡ ⎤ −1 + i −1 + 2i 1(c). iC = ⎣ −1 + 3i −1 + 4i ⎦. −1 + 5i −1 + 6i −6 11 3 1(d). 2A − B = . −2 −4 −2 1 + 3i 15 + 3i 16 + 3i . 1(e). A + 3C T = 5 + 3i 12 + 3i 15 + 3i 1(a). 5A = (c)2017 Pearson Education. Inc. 128 ⎡ ⎤ 8 10 7 1(f ). 3D − 2E = ⎣ 1 4 9 ⎦. 1 7 12 ⎡ 12 −3 − 3i 1(g). D + E + F = ⎣ 3 + i 3 − 2i 6 4 + 2i ⎤ −1 + i ⎦. 8 2 1(h). Solving for G and simplifying, we have that 3 G=− A−B = 2 1 −10 −1/2 3/2 −4 17/2 . 1(i). Solving for H and simplifying, we have that H = 4E − D − 2F = ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ 8 −20 −8 4 0 1 12 4 − 6i 2i −8 ⎣ 4 4 12 ⎦ − ⎣ 1 2 5 ⎦ − ⎣ 2 + 2i −4i 0 ⎦ = ⎣ 1 − 2i 16 −8 −12 3 1 2 −2 10 + 4i 6 15 −24 + 6i 2 + 4i −19 − 4i ⎤ −9 − 2i ⎦. 7 −20 ⎤ ⎡ 2+i −13 + i 4 + i ⎦ = ⎣ 21 + 3i 6+i 5 + 5i ⎤ −5 + 2i −9 + 4i ⎦ . −5 + 6i 1(j). We have K T = 2B − 3A, so that K = (2B − 3A)T = 2B T − 3AT . Thus, ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 2 0 −2 −1 10 3 4 ⎦ − 3⎣ 6 0 ⎦ = ⎣ −16 8 ⎦ . K = 2⎣ 1 −1 −4 1 −3 −5 1 ⎡ ⎤ −4 0 −1 2(a). −D = ⎣ −1 −2 −5 ⎦. −3 −1 −2 ⎡ ⎤ ⎡ ⎤ 2 0 8 0 4 ⎦=⎣ 4 16 ⎦. 2(b). 4B T = 4 ⎣ 1 −1 −4 −4 −16 ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ −2 −1 1+i 2+i 5+i 4+i 0 ⎦ + ⎣ 3 + i 4 + i ⎦ = ⎣ −9 + i 4 + i ⎦. 2(c). −2AT + C = −2 ⎣ 6 1 −3 5+i 6+i 3 + i 12 + i ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 10 −25 −10 4 0 1 14 −25 −9 5 15 ⎦ + ⎣ 1 2 5 ⎦ = ⎣ 6 7 20 ⎦. 2(d). 5E + D = ⎣ 5 20 −10 −15 3 1 2 23 −9 −13 2(e). We have ⎡ ⎤ ⎡ ⎤ ⎡ −2 −1 2 0 1+i 0 ⎦ − 2⎣ 1 4 ⎦ + i⎣ 3 + i 4AT − 2B T + iC = 4 ⎣ 6 1 −3 −1 −4 5+i 2(f ). We have ⎡ ⎤ ⎡ 8 −20 −8 12 4 12 ⎦ − ⎣ 0 4E − 3DT = ⎣ 4 16 −8 −12 3 3 6 15 ⎤ ⎡ ⎤ 9 −4 −23 −17 3 ⎦ = ⎣ 4 −2 9 ⎦. 6 13 −23 −18 (c)2017 Pearson Education. Inc. 129 2(g). We have (1 − 6i)F + iD = ⎡ ⎤ ⎡ 6 − 36i −16 − 15i 6+i 4i ⎣ 7 − 5i ⎦+⎣ i −12 − 2i 0 −1 + 6i 17 − 28i 3 − 18i 3i 0 2i i ⎤ ⎡ i 6 − 32i 5i ⎦ = ⎣ 7 − 4i 2i −1 + 9i −16 − 15i −12 17 − 27i ⎤ 6 + 2i ⎦. 5i 3 − 16i 2(h). Solving for G, we have G = A + (1 − i)C = T −2 −1 6 1 0 −3 1+i 2+i + (1 − i) −2 6 1 2 4 − 2i = + −1 0 −3 3 − i 5 − 3i 0 10 − 2i 7 − 4i = . 2 − i 5 − 3i 4 − 5i 3+i 4+i 5+i 6+i 6 − 4i 7 − 5i 2(i). Solve for H, we have 3 3 D − E + 3I3 2 2 ⎡ ⎤ ⎡ ⎤ ⎡ 6 0 3/2 3 −15/2 −3 1 ⎣ ⎦ ⎣ ⎦ ⎣ 3/2 3 15/2 3/2 3/2 9/2 = − + 0 9/2 3/2 3 6 −3 −9/2 0 ⎡ ⎤ 6 15/2 9/2 9/2 3 ⎦. =⎣ 0 −3/2 9/2 21/2 H= 2(j). We have K T = DT + E T − F T = (D + E − F )T , so that ⎡ 0 −7 + 3i K = D + E − F = ⎣ 1 − i 3 + 2i 8 −6 − 2i 3(a). AB = 3(b). 5 10 −3 27 22 3 0 1 0 ⎤ −1 − i ⎦. 8 −4 ⎡ ⎤ 9 BC = ⎣ 8 ⎦ −6 3(c). CA cannot be computed. 3(d). ⎡ 1 AT E = ⎣ −1 2 ⎤ 3 2−i 1 ⎦ −i 4 1+i 2 + 4i ⎡ 2 − 4i = ⎣ −2 4 − 6i (c)2017 Pearson Education. Inc. ⎤ 7 + 13i 1 + 3i ⎦ 10 + 18i ⎤ 0 0 ⎦ 1 130 3(e). ⎤ 2 −2 3 2 −3 ⎦ . CD = ⎣ −2 4 −4 6 3(f ). ⎡ ⎡ 1 −1 C T AT = 3(g). 2 F = 3(h). i 1 − 3i 0 4+i 2 ⎤ 1 3 ⎣ −1 1 ⎦ = 2 4 i 1 − 3i 0 4+i = 6 10 −1 10 − 10i 0 15 + 8i ⎡ ⎤⎡ ⎤ ⎡ ⎤ 2 −1 3 2 15 1 2 ⎦ ⎣ −2 ⎦ = ⎣ 14 ⎦ BDT = ⎣ 5 4 6 −2 3 −10 3(i). ⎡ ⎤ ⎡ 3 10 1 −1 2 1 ⎦ =⎣ 2 3 1 4 4 14 1 AT A = ⎣ −1 2 3(j). FE = i 1 − 3i 0 4+i 4(a). AC = 2−i −i 1 −1 2 3 1 4 1+i 2 + 4i = ⎤ 2 14 2 2 ⎦ 2 20 −2 + i 1 − 4i 13 − i 4 + 18i ⎡ ⎤ 1 6 ⎣ −1 ⎦ = 10 2 4(b). DC = [10] 4(c). DB = [6 14 − 4] 4(d). AD cannot be computed. i 2−i 1+i 4(e). EF = −i 2 + 4i 0 1 − 3i 4+i = 1 + 2i 1 2 − 2i 1 + 17i . 4(f ). Since AT is a 3 × 2 matrix and B is a 3 × 3 matrix, the product AT B cannot be constructed. 4(g). Since C is a 3 × 1 matrix, it is impossible to form the product C · C = C 2 . 2−i 1+i 2−i 1+i 4 − 5i 1 + 7i 4(h). E 2 = = . −i 2 + 4i −i 2 + 4i 3 − 4i −11 + 15i (c)2017 Pearson Education. Inc. 131 4(i). ADT = T 4(j). E A = 5. We have ⎤ 2 ⎣ −2 ⎦ = 10 . 16 3 −i 1 −1 2 2 − 4i = 2 + 4i 3 1 4 7 + 13i 1 −1 2 3 1 4 2−i 1+i ⎡ ⎛ ⎜ −3 ABC = (AB)C = ⎜ ⎝ 6 = = 2 7 −1 0 −3 −5 CAB = C(AB) = = = 6. Ac = 1 3 −5 4 15 −95 −9 65 −185 −460 119 316 and −6 1 1 5 −6 1 1 5 −99 −30 −2 1 + 3i ⎛ ⎜ ⎜ ⎝ −3 6 1 5 15 −95 −9 65 635 . 230 =6 1 −5 . ⎤⎞ −2 8 ⎢ 8 −3 ⎥⎟ −6 ⎥⎟ ⎢ ⎣ −1 −9 ⎦⎠ 1 0 2 ⎡ 2 7 −1 0 −3 −5 6 −2 −6 1 4 − 6i 10 + 18i + (−2) 3 4 1 5 ⎤⎞ −2 8 ⎢ 8 −3 ⎥⎟ ⎥⎟ ⎢ ⎣ −1 −9 ⎦⎠ 0 2 ⎡ = 0 −38 . 7. ⎡ ⎤⎡ ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 3 −1 4 2 3 −1 4 −13 1 5 ⎦ ⎣ 3 ⎦ = 2 ⎣ 2 ⎦ + 3 ⎣ 1 ⎦ + (−4) ⎣ 5 ⎦ = ⎣ −13 ⎦ . Ac = ⎣ 2 7 −6 3 −4 7 −6 3 −16 8. ⎤ ⎡ ⎤ ⎤ ⎡ ⎤ ⎡ −7 −1 2 −1 2 5 7 ⎦ Ac = ⎣ 4 = 5 ⎣ 4 ⎦ + (−1) ⎣ 7 ⎦ = ⎣ 13 ⎦ . −1 −4 29 5 −4 5 ⎡ 9. We have Ac = x a e +y b f +z c g +w d h = xa + yb + zc + wd xe + yf + zg + wh . 10(a). The dimensions of B should be n × r in order that ABC is defined. 10(b). The elements of the ith row of A are ai1 , ai2 , . . . , ain and the elements of the jth column of BC are r , m=1 b1m cmj , r , m=1 b2m cmj , . . . , r , m=1 (c)2017 Pearson Education. Inc. bnm cmj , 132 so the element in the ith row and jth column of ABC = A(BC) is ai1 r , b1m cmj + ai2 m=1 = n , aik r , bkm cmj bnm cmj m=1 m=1 r , = n , r , k=1 m=1 aik bkm cmj . 1 −1 −1 −4 = . 2 3 8 7 −1 −4 1 −1 −9 −11 = . A3 = A2 A = 8 7 2 3 22 13 −9 −11 1 −1 −31 −24 = . A4 = A3 A = 22 13 2 3 48 17 A2 = AA = 11(b). b2m cmj + · · · + ain m=1 k=1 11(a). r , 1 −1 2 3 ⎡ ⎤⎡ ⎤ ⎡ ⎤ 0 1 0 0 1 0 −2 0 1 0 1 ⎦ ⎣ −2 0 1 ⎦ = ⎣ 4 −3 0 ⎦. A2 = AA = ⎣ −2 4 −1 0 4 −1 0 2 4 −1 ⎡ ⎤⎡ ⎤ ⎡ ⎤ −2 0 1 0 1 0 4 −3 0 0 ⎦ ⎣ −2 0 1 ⎦=⎣ 6 4 −3 ⎦ . A3 = A2 A = ⎣ 4 −3 2 4 −1 4 −1 0 −12 3 4 ⎡ ⎤⎡ ⎤ ⎡ ⎤ 4 −3 0 0 1 0 6 4 −3 6 4 −3 ⎦ ⎣ −2 0 1 ⎦ = ⎣ −20 9 4 ⎦. A4 = A3 A = ⎣ −12 3 4 4 −1 0 10 −16 3 12(a). We apply the distributive property of matrix multiplication as follows: (A+2B)2 = (A+2B)(A+2B) = A(A+2B)+(2B)(A+2B) = (A2 +A(2B))+((2B)A+(2B)2 ) = A2 +2AB+2BA+4B 2 , where scalar factors of 2 are moved in front of the terms since they commute with matrix multiplication. 12(b). We apply the distributive property of matrix multiplication as follows: (A + B + C)2 = (A + B + C)(A + B + C) = A(A + B + C) + B(A + B + C) + C(A + B + C) = A2 + AB + AC + BA + B 2 + BC + CA + CB + C 2 = A2 + B 2 + C 2 + AB + BA + AC + CA + BC + CB, as required. 12(c). We can use the formula for (A + B)3 found in Example 2.2.20 and substitute −B for B throughout the expression: (A − B)3 = A3 + A(−B)A + (−B)A2 + (−B)2 A + A2 (−B) + A(−B)2 + (−B)A(−B) + (−B)3 = A3 − ABA − BA2 + B 2 A − A2 B + AB 2 + BAB − B 3 , as needed. (c)2017 Pearson Education. Inc. 133 13. We have 2 A = so that 2 A + 4A + 18I2 = 2 −5 6 −6 −26 −24 20 6 2 −5 6 −6 + ⎡ ⎤⎡ −1 0 4 −1 A2 = ⎣ 1 1 2 ⎦ ⎣ 1 −2 3 0 −2 14. We have = 8 −20 24 −24 0 1 3 −26 −24 + 20 6 18 0 ⎤ ⎡ 4 −7 2 ⎦ = ⎣ −4 0 5 , 0 18 = 0 0 0 0 . ⎤ 12 −4 7 6 ⎦ 3 −2 and ⎡ −1 A3 = ⎣ 1 −2 0 1 3 ⎤⎡ 4 −1 2 ⎦⎣ 1 0 −2 ⎤⎡ ⎤ ⎡ ⎤⎡ 4 −1 0 4 −7 12 −4 −1 2 ⎦ ⎣ 1 1 2 ⎦ = ⎣ −4 7 6 ⎦⎣ 1 0 −2 3 0 5 3 −2 −2 0 1 3 0 1 3 ⎤ ⎡ ⎤ 4 27 0 −4 2 ⎦ = ⎣ −1 25 −2 ⎦ . 0 2 −3 26 Therefore, we have ⎡ ⎤ ⎡ 27 0 −4 −1 A3 + A − 26I3 = ⎣ −1 25 −2 ⎦ + ⎣ 1 2 −3 26 −2 0 1 3 ⎤ ⎡ 4 26 2 ⎦−⎣ 0 0 0 0 26 0 ⎤ ⎡ 0 0 0 ⎦=⎣ 0 26 0 1 1 0 ⎤ 0 1 ⎦. 1 ⎡ 15. ⎡ 1 Substituting A = ⎣ 0 0 ⎤ ⎡ ⎤ ⎡ 1 0 0 0 −1 0 1 0 −1 ⎦ = ⎣ 0 A2 = ⎣ 0 1 0 ⎦ − ⎣ 0 0 0 1 0 0 0 0 ⎤ x z 1 y ⎦ for A, we have 0 1 ⎡ 1 ⎣ 0 0 that is, ⎡ x 1 0 1 ⎣ 0 0 ⎤⎡ z 1 x y ⎦⎣ 0 1 1 0 0 2x 1 0 ⎤ ⎡ z 1 y ⎦=⎣ 0 1 0 ⎤ ⎡ 2z + xy 1 ⎦=⎣ 0 2y 1 0 1 1 0 1 1 0 0 0 0 ⎤ 0 0 ⎦. 0 ⎤ 0 1 ⎦, 1 ⎤ 0 1 ⎦. 1 Since corresponding elements of equal matrices are equal, we obtain the following implications: 2y = 1 =⇒ y = 1/2, 2x = 1 =⇒ x = 1/2, 2z + xy = 0 =⇒ 2z + (1/2)(1/2) = 0 =⇒ z = −1/8. ⎡ 1 Thus, A = ⎣ 0 0 1/2 1 0 ⎤ −1/8 1/2 ⎦. 1 (c)2017 Pearson Education. Inc. 134 x+y x 1 x 1 x 1 x2 − 2 16. In order that A2 = A, we require = = , that is, −2x − 2y −2 + y 2 −2 y −2 y −2 y x 1 x2 − x − 2 x+y−1 , or equivalently, = 02 . Since corresponding elements of equal ma−2 y −2x − 2y + 2 y 2 − y − 2 trices are equal, it follows that x2 − x − 2 = 0 =⇒ x = −1 or x = 2, and y 2 − y − 2 = 0 =⇒ y = −1 or y = 2. Two cases arise from x + y − 1 = 0: (a): If x = −1, then y = 2. (b): If x = 2, then y = −1. Thus, −1 −2 1 2 0 i −i 0 A= 17. σ1 σ 2 = σ2 σ3 = σ3 σ1 = 0 1 1 0 0 −i i 0 1 0 0 −1 1 0 0 1 or A = 2 1 −2 −1 . i 0 1 0 = =i = iσ3 . 0 −i 0 −1 0 0 i 0 1 = =i = iσ1 . −1 i 0 1 0 1 0 1 0 −i = =i = iσ2 . 0 −1 0 i 0 18. [A, B] = AB − BA 1 −1 3 1 3 = − 2 1 4 2 4 −1 −1 5 −2 = − 10 4 8 −2 −6 1 = = 02 . 2 6 1 2 1 −1 2 1 19. [A1 , A2 ] = A1 A2 − A2 A1 1 0 0 1 0 1 1 0 = − 0 1 0 0 0 0 0 1 0 1 0 1 = − = 02 , thus A1 and A2 commute. 0 0 0 0 [A1 , A3 ] = A1 A3 − A3 A1 1 0 0 0 0 0 1 0 = − 0 1 1 0 1 0 0 1 0 0 0 0 = − = 02 , thus A1 and A3 commute. 1 0 1 0 (c)2017 Pearson Education. Inc. 135 [A2 , A3 ] = A2 A3 − A3 A2 0 1 0 0 0 0 0 1 = − 0 0 1 0 1 0 0 0 1 0 0 0 1 0 = − = = 02 . 0 0 0 1 0 −1 −1 0 Then [A3 , A2 ] = −[A2 , A3 ] = = 02 . Thus, A2 and A3 do not commute. 0 1 20. [A1 , A2 ] = A1 A2 − A2 A1 1 0 −1 1 0 i 0 −1 0 i − = 1 0 0 i 0 4 i 0 4 1 1 −i 0 1 i 0 − = 0 −i 0 i 4 4 1 i 1 2i 0 0 = = A3 . = 4 0 −2i 2 0 −i [A2 , A3 ] = A2 A3 − A3 A2 1 i 1 0 −1 i 0 0 0 −1 − = 0 0 −i 1 0 4 1 4 0 −1 1 1 0 i 0 −i − = 0 4 i 0 4 −i 1 0 i 1 0 2i = = A1 . = 4 2i 0 2 i 0 [A3 , A1 ] = A3 A1 − A1 A3 1 0 i 1 i 0 0 i i 0 − = i 0 0 −i 4 0 −i 4 i 0 1 0 1 1 0 −1 − = 0 4 −1 4 1 0 1 0 −1 1 0 −2 = = A2 . = 0 0 4 2 2 1 21. [A, [B, C]] + [B, [C, A]] + [C, [A, B]] = [A, BC − CB] + [B, CA − AC] + [C, AB − BA] = A(BC − CB) − (BC − CB)A + B(CA − AC) − (CA − AC)B + C(AB − BA) − (AB − BA)C = ABC − ACB − BCA + CBA + BCA − BAC − CAB + ACB + CAB − CBA − ABC + BAC = 0. 22. Proof that A(BC) = (AB)C: Let A = [aij ] be of size m × n, B = [bjk ] be of size n × p, and C = [ckl ] be of size p × q. Consider the (i, j)-element of (AB)C: n p p n , , , , [(AB)C]ij = aih bhk ckj = aih bhk ckj = [A(BC)]ij . k=1 h=1 h=1 k=1 (c)2017 Pearson Education. Inc. 136 Proof that A(B + C) = AB + AC: We have [A(B + C)]ij = = = n , k=1 n , aik (bkj + ckj ) (aij bkj + aik ckj ) k=1 n , k=1 aik bkj + n , aik ckj k=1 = [AB + AC]ij . 23. Proof that (AT )T = A: Let A = [aij ]. Then AT = [aji ], so (AT )T = [aji ]T = aij = A, as needed. Proof that (A + C)T = AT + C T : Let A = [aij ] and C = [cij ]. Then [(A + C)T ]ij = [A + C]ji = [A]ji + [C]ji = aji + cji = [AT ]ij + [C T ]ij = [AT + C T ]ij . Hence, (A + C)T = AT + C T . 24. We have (IA)ij = m , δik akj = δii aij = aij , k=1 for 1 ≤ i ≤ m and 1 ≤ j ≤ p. Thus, Im Am×p = Am×p . 25. Let A = [aij ] and B = [bij ] be n × n matrices. Then n n n n n n , , , , , , aki bik = bik aki = bik aki = tr(BA). tr(AB) = k=1 26(a). B T AT = ⎡ ⎢ 26(b). C T B T = ⎢ ⎣ i=1 i=1 i=1 k=1 ⎤ −3 1 ⎣ −1 ⎦ = . −7 6 ⎤ ⎡ ⎤ −4 64 6 1 ⎢ −4 1 −3 ⎥ 0 −7 −1 1 ⎥ ⎥ ⎥ =⎢ ⎣ −20 −16 −18 ⎦. 1 −3 5 ⎦ −4 8 12 8 −2 0 −7 −1 −4 1 −3 −9 0 3 −2 k=1 ⎡ 26(c). Since DT is a 3 × 3 matrix and A is a 1 × 3 matrix, it is not possible to compute the expression DT A. ⎡ ⎤ −2 0 1 27(a). ADT = −3 −1 6 ⎣ 1 0 −2 ⎦ = 35 42 −7 . 5 7 −1 ⎤ ⎡ ⎤ ⎡ 82 1 −22 16 −9 1 ⎢ ⎢ 0 1 1 5 −2 ⎥ 1 ⎥ ⎥. Therefore, ⎥ −9 0 3 −2 = ⎢ 27(b). First note that C T C = ⎢ ⎣ ⎣ 3 ⎦ −22 5 34 −16 ⎦ 1 1 5 −2 5 16 −2 −16 8 −2 −2 ⎤ ⎤ ⎡ ⎤⎡ ⎡ 7465 −59 −2803 1790 82 1 −22 16 82 1 −22 16 ⎢ ⎢ ⎢ −59 31 185 −82 ⎥ 1 1 5 −2 ⎥ 1 1 5 −2 ⎥ ⎥. ⎥=⎢ ⎥⎢ (C T C)2 = ⎢ ⎦ ⎣ ⎦ ⎣ ⎣ −22 −2803 185 1921 −1034 ⎦ −22 5 34 −16 5 34 −16 1790 −82 −1034 580 16 −2 −16 8 16 −2 −16 8 (c)2017 Pearson Education. Inc. 137 ⎡ ⎤⎡ ⎤ ⎡ ⎤ −2 0 1 0 −4 −1 5 1 ⎦=⎣ 2 2 ⎦. 27(c). DT B = ⎣ 1 0 −2 ⎦ ⎣ −7 5 7 −1 −1 −3 −48 −10 28(a). We have ⎡ ⎤ z 2z ⎦ , z −x −y y S = [s1 , s2 , s3 ] = ⎣ 0 x −y ⎡ so 2 AS = ⎣ 2 1 2 5 2 28(b). ⎤⎡ 1 −x −y 2 ⎦⎣ 0 y 2 x −y ⎡ −x S T AS = S T (AS) = ⎣ −y z 0 y 2z ⎤ ⎡ z −x −y 2z ⎦ = ⎣ 0 y z x −y ⎤⎡ x −x −y −y ⎦ ⎣ 0 y z x −y ⎤ 7z 14z ⎦ = [s1 , s2 , 7s3 ]. 7z ⎤ ⎡ 7z 2x2 14z ⎦ = ⎣ 0 0 7z 0 3y 2 0 ⎤ 0 0 ⎦, 42z 2 but S T AS = diag(1, 1, 7), so we have the following √ 2 2 √ 3 3y 2 = 1 =⇒ y = ± 3 √ 6 2 6z = 1 =⇒ z = ± . 6 2 2x = 1 =⇒ x = ± 29(a). We have ⎡ ⎤⎡ 1 −4 0 0 2x 7 0 ⎦⎣ 0 x AS = ⎣ −4 0 0 5 z 0 ⎡ ⎤ 0 −2x 9y −x −18y ⎦ =⎣ 0 5z 0 0 ⎤ y −2y ⎦ 0 = [5s1 , −s2 , 9s3 ]. 29(b). We have ⎡ 0 S T AS = ⎣ 2x y ⎤⎡ 0 z 0 x 0 ⎦⎣ 0 −2y 0 5z −2x −x 0 ⎤ ⎡ 9y 5z 2 −18y ⎦ = ⎣ 0 0 0 0 −5x2 0 ⎤ 0 0 ⎦, 45y 2 so in order for this to be equal to diag(5, −1, 9), we must have 5z 2 = 5, −5x2 = −1, 45y 2 = 9. Thus, we must have z 2 = 1, x2 = 15 , and y 2 = 15 . Therefore, the values of x, y, and z that we are looking for are x = ± 15 , y = ± 15 , and z = ±1. (c)2017 Pearson Education. Inc. 138 ⎡ 2 ⎢ 0 ⎢ 30(a). ⎣ 0 0 ⎡ 7 30(b). ⎣ 0 0 ⎤ 0 0 ⎥ ⎥. 0 ⎦ 2 ⎤ 0 2 0 0 0 0 2 0 0 7 0 0 0 ⎦. 7 31. Suppose A is an n × n scalar matrix with trace k. If A = aIn , then tr(A) = na = k, so we conclude that a = k/n. So A = nk In , a uniquely determined matrix. 32. We have 1 B = (A + AT ) 2 T T and CT = 1 (A − AT ) 2 = T = 1 1 (A + AT )T = (AT + A) = B 2 2 1 T 1 (A − A) = − (A − AT ) = −C. 2 2 Thus, B is symmetric and C is skew-symmetric. 33. We have B+C = 1 1 1 1 1 1 (A + AT ) + (A − AT ) = A + AT + A − AT = A. 2 2 2 2 2 2 34. We have ⎛⎡ ⎤ ⎡ ⎡ ⎤⎞ 4 −1 0 8 8 4 9 2 1 1 1 B = (A + AT ) = ⎝⎣ 9 −2 3 ⎦ + ⎣ −1 −2 5 ⎦⎠ = ⎣ 8 −4 2 2 2 2 5 5 2 8 0 3 5 ⎤ ⎡ ⎤ 2 4 4 1 8 ⎦ = ⎣ 4 −2 4 ⎦ 10 1 4 5 and ⎛⎡ ⎤ ⎡ ⎡ ⎤ ⎡ ⎤⎞ ⎤ 4 −1 0 0 −10 −2 4 9 2 0 −5 −1 1 1 1 0 −2 ⎦ = ⎣ 5 0 −1 ⎦ . C = (A − AT ) = ⎝⎣ 9 −2 3 ⎦ − ⎣ −1 −2 5 ⎦⎠ = ⎣ 10 2 2 2 2 5 5 2 2 0 0 3 5 1 1 0 35. ⎛⎡ 1 1 ⎝⎣ 3 B= 2 7 ⎛⎡ 1 1 C = ⎝⎣ 3 2 7 ⎤ ⎡ −5 3 1 2 4 ⎦ + ⎣ −5 −2 6 3 ⎤ ⎡ −5 3 1 2 4 ⎦ − ⎣ −5 −2 6 3 ⎡ ⎤ ⎡ ⎤⎞ ⎤ 2 −2 10 3 7 1 −1 5 1 4 2 ⎦ = ⎣ −1 2 −2 ⎦⎠ = ⎣ −2 2 1 ⎦. 2 10 2 12 4 6 5 1 6 ⎤ ⎡ ⎤ ⎤⎞ ⎡ 3 7 0 −8 −4 0 −4 −2 1 2 −2 ⎦⎠ = ⎣ 8 0 6 ⎦=⎣ 4 0 3 ⎦. 2 4 −6 0 2 −3 0 4 6 36(a). If A is symmetric, then AT = A, so that B= 1 1 1 (A + AT ) = (A + A) = (2A) = A 2 2 2 C= 1 1 1 (A − AT ) = (A − A) = (0n ) = 0n . 2 2 2 and (c)2017 Pearson Education. Inc. 139 36(b). If A is skew-symmetric, then AT = −A, so that B= 1 1 1 (A + AT ) = (A + (−A)) = (0n ) = 0n 2 2 2 C= 1 1 1 (A − AT ) = (A − (−A)) = (2A) = A. 2 2 2 and 37. If A = [aij ] and D = diag(d1 , d2 , . . . , dn ), then we must show that the (i, j)-entry of DA is di aij . In index notation, we have n , (DA)ij = di δik akj = di δii aij = di aij . k=1 Hence, DA is the matrix obtained by multiplying the ith row vector of A by di , where 1 ≤ i ≤ n. 38. If A = [aij ] and D = diag(d1 , d2 , . . . , dn ), then we must show that the (i, j)-entry of AD is dj aij . In index notation, we have n , (AD)ij = aik dj δkj = aij dj δjj = aij dj . k=1 Hence, AD is the matrix obtained by multiplying the jth column vector of A by dj , where 1 ≤ j ≤ n. 39. Since A and B are symmetric, we have that AT = A and B T = B. Using properties of the transpose operation, we therefore have (AB)T = B T AT = BA = AB, and this shows that AB is symmetric. 40(a). We have (AAT )T = (AT )T AT = AAT , so that AAT is symmetric. 40(b). We have (ABC)T = [(AB)C]T = C T (AB)T = C T (B T AT ) = C T B T AT , as needed. 1 cos t 41. A (t) = . − sin t 4 −2e−2t . 42. A (t) = cos t ⎡ ⎤ cos t − sin t 0 43. A (t) = ⎣ sin t cos t 1 ⎦. 0 3 0 t 2e2t 2t e . 44. A (t) = 2et 8e2t 10t 45. We that the (i, j)-entry of both sides of the equation agree. First, recall that the (i, j)-entry of show n d AB is k=1 aik bkj , and therefore, the (i, j)-entry of dt (AB) is (by the product rule) n , k=1 aik bkj + aik bkj = n , k=1 aik bkj + n , aik bkj . k=1 The former term is precise the (i, j)-entry of the matrix dA dt B, while the latter term is precise the (i, j)-entry d of the matrix A dB . Thus, the (i, j)-entry of (AB) is precisely the sum of the (i, j)-entry of dA dt dt dt B and the dB (i, j)-entry of A dt . Thus, the equation we are proving follows immediately. (c)2017 Pearson Education. Inc. 140 46. We have 1 t e t 2e 0 e−t 5e−t 47. We have π/2 0 cos t sin t 48. We have 1 t e t 2e 0 dt = e2t 4e2t t2 5t2 1 0 = −e−t −5e−t sin t − cos t π/2 0 = dt = et 2et dt = = = sin(π/2) − cos(π/2) 1 2t 2e 2t t3 3 5 3 3t et 2et 2e e 2e e2 /2 1/3 2e2 5/3 49. We have ⎡ ⎤ ⎡ 1 2t 1 sin 2t e2t 2e ⎣ t2 − 5 ⎦ dt = ⎣ t3 − 5t tet 3 0 sec2 t 3t − sin t tan t ⎡ e2 2 −1/e −5/e e 2e = ⎣ −14/3 tan 1 − − sin 0 − cos 0 1 2 1/2 0 2 0 = = 1 0 e−1 2e − 2 1 − 1/e 5 − 5/e 0 −1 − = . 1 1 . 1 0 − ⎤ − 12 cos 2t tet − et ⎦ 10 3 2 2 t + cos t ⎤ ⎡ 1 − cos2 2 2 ⎦−⎣ 0 0 3 0 2 + cos 1 3t 1 e dt = −5t tan−1 (t) −5dt A(t)dt = t2 +1 dt 2 t 2t dt = . 51. t3 3t2 ⎡ ⎤ ⎤ ⎡ sin t cos t 0 − cos t sin t 0 ⎣ − cos t sin t t ⎦ dt = ⎣ − sin t − cos t t2 /2 ⎦. 52. 0 3t2 /2 0 3t 1 t t e−t −e−t et e dt = . 53. t −t t 2e 5e 2e −5e−t ⎤ ⎤ ⎡ ⎡ 1 2t e − 12 cos 2t sin 2t e2t 2 ⎦ dt = ⎣ t3 − 5t ⎣ t2 − 5 tet 54. tet − et ⎦. 3 2 3 2 sec t 3t − sin t tan t 2 t + cos t 50. 1 −1 2 −5 = e−1 2e − 2 1/3 2e − 2 5/3 ⎤ ⎡ e2 −1 − 12 2 −1 ⎦ = ⎣ −14/3 1 tan 1 1 3t 3e e2 −1 2 2 1−cos 2 2 1 . ⎤ ⎦. 1 2 + cos 1 . Solutions to Section 2.3 True-False Review: (a): FALSE. The last column of the augmented matrix corresponds to the constants on the right-hand side of the linear system, so if the augmented matrix has n columns, there are only n − 1 unknowns under consideration in the system. (c)2017 Pearson Education. Inc. 141 (b): FALSE. Three distinct planes can intersect in a line (e.g. Figure 2.3.1, lower right picture). For instance, the xy-plane, the xz-plane, and the plane y = z intersect in the x-axis. (c): FALSE. The right-hand side vector must have m components, not n components. (d): TRUE. If a linear system has two distinct solutions x1 and x2 , then any point on the line containing x1 and x2 is also a solution, giving us infinitely many solutions, not exactly two solutions. (e): TRUE. The augmented matrix for a linear system has one additional column (containing the constants on the right-hand side of the equation) beyond the matrix of coefficients. (f ): FALSE. Because the vector (x1 , x2 , x3 , 0, 0) has five entries, this vector belongs to R5 . Vectors in R3 can only have three slots. (g): FALSE. The two column vectors given have different numbers of components, so they are not the same vectors. Problems: 1. 2 · 1 − 3(−1) + 4 · 2 = 13, 1 + (−1) − 2 = −2, 5 · 1 + 4(−1) + 2 = 3. 2. 2 + (−3) − 2 · 1 = −3, 3 · 2 − (−3) − 7 · 1 = 2, 2 + (−3) + 1 = 0, 2 · 2 + 2(−3) − 4 · 1 = −6. 3. (1 − t) + (2 + 3t) + (3 − 2t) = 6, (1 − t) − (2 + 3t) − 2(3 − 2t) = −7, 5(1 − t) + (2 + 3t) − (3 − 2t) = 4. 4. s + (s − 2t) − (2s + 3t) + 5t = 0, 2(s − 2t) − (2s + 3t) + 7t = 0, 4s + 2(s − 2t) − 3(2s + 3t) + 13t = 0. 5. The two given lines are the same line. Therefore, since this line contains an infinite number of points, there must be an infinite number of solutions to this linear system. 6. These two lines are parallel and distinct, and therefore, there are no common points on these lines. In other words, there are no solutions to this linear system. 7. These two lines have different slopes, and therefore, they will intersect in exactly one point. Thus, this system of equations has exactly one solution. (c)2017 Pearson Education. Inc. 142 8. The first and third equations describe lines that are parallel and distinct, and therefore, there are no common points on these lines. In other words, there are no solutions to this linear system. ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 1 2 −3 1 1 2 −3 1 9. A = ⎣ 2 4 −5 ⎦ , b = ⎣ 2 ⎦ , A# = ⎣ 2 4 −5 2 ⎦. 7 2 −1 7 2 −1 3 3 1 1 1 −1 3 1 1 1 −1 3 . 10. A = ,b = , A# = 2 4 −3 7 2 2 4 −3 7 2 ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 1 2 −1 0 1 2 −1 0 11. A = ⎣ 2 3 −2 ⎦ , b = ⎣ 0 ⎦ , A# = ⎣ 2 3 −2 0 ⎦. 5 6 −5 0 5 6 −5 0 12. It is acceptable to use any variable names. We will use x1 , x2 , x3 , x4 : x1 − x2 x1 + x2 3x1 + x2 +2x3 + 3x4 −2x3 + 6x4 +4x3 + 2x4 = 1, = −1, = 2. 13. It is acceptable to use any variable names. We will use x1 , x2 , x3 : 2x1 + x2 4x1 − x2 7x1 + 6x2 +3x3 = 3, +2x3 = 1, +3x3 = −5. 14. The system of equations here only contains one equation: 4x1 − 2x2 − 2x3 − 3x5 = −9. 15. This system of equations has three equations: −3x2 = −1, 2x1 − 7x2 = 6, 16. Given Ax = 0 and Ay = 0, and an arbitrary constant c, (a). we have Az = A(x + y) = Ax + Ay = 0 + 0 = 0 and Aw = A(cx) = c(Ax) = c0 = 0. (b). No, because A(x + y) = Ax + Ay = b + b = 2b = b, and A(cx) = c(Ax) = cb = b in general. −4 3 x1 4t x1 = + . 17. x2 6 −4 x2 t2 −t x1 t2 x1 = . 18. − sin t 1 x2 x2 0 e2t x1 0 x1 = + . 19. x2 − sin t 0 x2 1 (c)2017 Pearson Education. Inc. 5x1 + 5x2 = 7. 143 ⎤ ⎡ x1 0 20. ⎣ x2 ⎦ = ⎣ −et x3 −t ⎡ − sin t 0 t2 ⎤ ⎡ ⎤⎡ ⎤ x1 1 t t2 ⎦ ⎣ x 2 ⎦ + ⎣ t3 ⎦ . 0 x3 1 21. We have 4e4t −2(4e4t ) x (t) = and Ax + b = 2 −1 −2 3 22. We have x (t) = e4t −2e4t 0 0 + = = 4(−2e−2t ) + 2 cos t 3(−2e−2t ) + sin t 4e4t −8e4t 2e4t + (−1)(−2e4t ) + 0 −2e4t + 3(−2e4t ) + 0 = −8e−2t + 2 cos t −6e−2t + sin t = 4e4t −8e4t . and Ax + b = = 1 −4 −3 2 4e−2t + 2 sin t 3e−2t − cos t + −2(cos t + sin t) 7 sin t + 2 cos t 4e−2t + 2 sin t − 4(3e−2t − cos t) − 2(cos t + sin t) −3(4e−2t + 2 sin t) + 2(3e−2t − cos t) + 7 sin t + 2 cos t 23. We compute x = 3et + 2tet et + 2tet = −8e−2t + 2 cos t −6e−2t + sin t . and Ax + b = 2 −1 −1 2 2tet + et 2tet − et + 0 4et = 2(2tet + et ) − (2tet − et ) + 0 −(2tet + et ) + 2(2tet − et ) + 4et = 2tet + 3et 2tet + et . Therefore, we see from these calculations that x = Ax + b. 24. We compute ⎤ −tet − et ⎦ −9e−t x = ⎣ t te + et − 6e−t ⎡ and ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ ⎤⎡ −et −tet −tet − et −et 1 0 0 −tet ⎦+⎣ 6e−t ⎦ = ⎣ 2(−tet ) − 3(9e−t ) + 2(tet + 6e−t ) ⎦+⎣ 6e−t ⎦ = ⎣ 9e−t −9e−t Ax+b = ⎣ 2 −3 2 ⎦ ⎣ −tet − 2(9e−t ) + 2(tet + 6e−t ) tet + 6e−t et et tet + et − 6e− 1 −2 2 ⎡ Therefore, we see from these calculations that x = Ax + b. Solutions to Section 2.4 True-False Review: (a): TRUE. The precise row-echelon form obtained for a matrix depends on the particular elementary row operations (and their order). However, Theorem 2.4.15 states that there is a unique reduced row-echelon form for a matrix. (c)2017 Pearson Education. Inc. 144 (b): FALSE. Upper triangular matrices could have pivotentries that are not 1. For instance, the following 2 0 matrix is upper triangular, but not in row echelon form: . 0 0 (c): TRUE. The pivots in a row-echelon form of an n × n matrix must move down and to the right as we look from one row to the next beneath it. Thus, the pivots must occur on or to the right of the main diagonal of the matrix, and thus all entries below the main diagonal of the matrix are zero. (d): FALSE. This would not be true, for example, if A was a zero matrix with 5 rows and B was a nonzero matrix with 4 rows. (e): FALSE. If A is a nonzero matrix and B = −A, then A + B = 0, so rank(A + B) = 0, but rank(A), rank(B) ≥ 1 so rank(A)+ rank(B) ≥ 2. 0 1 (f ): FALSE. For example, if A = B = , then AB = 0, so rank(AB) = 0, but rank(A)+ 0 0 rank(B) = 1 + 1 = 2. (g): TRUE. A matrix of rank zero cannot have any pivots, hence no nonzero rows. It must be the zero matrix. (h): TRUE. The matrices A and 2A have the same reduced row-echelon form, since we can move between the two matrices by multiplying the rows of one of them by 2 or 1/2, a matter of carrying out elementary row operations. If the two matrices have the same reduced row-echelon form, then they have the same rank. (i): TRUE. The matrices A and 2A have the same reduced row-echelon form, since we can move between the two matrices by multiplying the rows of one of them by 2 or 1/2, a matter of carrying out elementary row operations. Problems: 1. Neither. 2. Reduced row-echelon form. 3. Neither. 4. Row-echelon form. 5. Row-echelon form. 6. Reduced row-echelon form. 7. Reduced row-echelon form. 8. Reduced row-echelon form. 9. 10. 2 1 2 −4 −4 8 1 −3 1 ∼ 1 ∼ 1 −3 2 1 1 −2 −4 8 2 ∼ 1 −2 0 0 1. M1 ( 12 ) 2. A12 (4) 1. P12 2 ∼ 1 −3 0 7 , Rank (A) = 1. ∼ 1 −3 0 1 2. A12 (−2) 3. M2 ( 17 ) 3 (c)2017 Pearson Education. Inc. , Rank (A) = 2. 145 ⎡ 11. 0 ⎣ 0 0 1 1 3 ⎤ ⎡ 3 0 1 1 4 ⎦∼⎣ 0 0 5 0 0 ⎤ ⎡ 3 0 2 1 ⎦∼⎣ 0 4 0 1. A12 (−1), A13 (−3) 2. A21 (−1) 3. A12 (−2), A13 (−3) 4. P23 5. M2 (−1) 6. A32 (5) ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 2 −1 3 3 1 −2 1 2 −5 1 2 −5 1 2 −5 1 2 3 4 ⎣ 3 1 −2 ⎦ ∼ ⎣ 2 −1 3 ⎦ ∼ ⎣ 2 −1 3 ⎦ ∼ ⎣ 0 −5 13 ⎦ ∼ ⎣ 0 −1 −2 ⎦ 2 −2 1 2 −2 1 0 −1 −2 0 −1 −2 0 −5 13 ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 1 2 −5 1 2 −5 1 2 −5 5 6 7 1 2 ⎦∼⎣ 0 1 2 ⎦∼⎣ 0 1 2 ⎦ , Rank (A) = 3. ∼⎣ 0 0 −5 13 0 0 23 0 0 1 1. P12 2. A21 (−1), A23 (−1) 3. A12 (−2) 4. P23 5. M2 (−1) ⎡ 6. A23 (5) ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ 2 −1 3 2 1 3 1 3 1 3 1 1 2 3 4 5 ⎣ 3 2 ⎦ ∼ ⎣ 2 −1 ⎦ ∼ ⎣ 2 −1 ⎦ ∼ ⎣ 0 −7 ⎦ ∼ ⎣ 0 −1 ⎦ ∼ ⎣ 0 2 5 2 5 0 −1 0 −7 2 5 0 1. P12 15. 2. A23 (−4) ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 2 1 4 3 −2 6 1 1 2 1 1 2 1 1 2 1 2 3 4 ⎣ 2 −3 4 ⎦ ∼ ⎣ 2 −3 4 ⎦ ∼ ⎣ 2 −3 4 ⎦ ∼ ⎣ 0 −5 0 ⎦ ∼ ⎣ 0 −1 0 ⎦ 3 −2 6 2 1 4 2 1 4 0 −1 0 0 −5 0 ⎡ ⎤ ⎡ ⎤ 1 1 2 1 1 2 5 6 1 0 ⎦ ∼ ⎣ 0 1 0 ⎦ , Rank (A) = 2. ∼⎣ 0 0 −5 0 0 0 0 1. P13 14. ⎤ 3 1 ⎦ , Rank (A) = 2. 0 ⎡ 12. 13. 1 0 0 ⎡ 2 ⎢ 3 ⎢ ⎣ 1 2 −2 −2 −1 −1 2. A21 (−1) 3. A12 (−2), A13 (−2) 4. P23 7. M3 (1/23). ⎤ 3 1 ⎦ , Rank (A) = 2. 0 5. M2 (−1), A23 (7). ⎤ ⎡ ⎤ ⎡ ⎤ ⎤ ⎡ 1 −1 1 0 1 −1 1 0 1 −1 1 0 −1 3 3 1 ⎥ 1 0 1 ⎥ 1 0 1 ⎥ 3 1 ⎥ 1 ⎢ 2 ⎢ 3 ⎢ ⎥∼ ⎥∼ ⎥ ⎥∼ ⎢ 3 −2 ⎢ 0 ⎢ 0 ⎦ ⎦ ⎦ ⎣ ⎣ ⎣ 2 −2 −1 3 0 0 −3 3 0 0 −3 3 ⎦ 1 0 2 −1 2 2 0 1 0 2 0 0 0 1 2 2 ⎤ ⎡ 1 −1 1 0 0 1 0 1 ⎥ 4 ⎢ ⎥ , Rank (A) = 4. ⎢ ∼⎣ 0 0 1 −1 ⎦ 0 0 0 1 (c)2017 Pearson Education. Inc. 146 1. P13 2. A12 (−3), A13 (−2), A14 (−2) 3. A24 (−1) 4. M3 (1/3) 16. ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ 1 −2 2 −1 3 4 1 −2 1 3 1 −2 1 3 1 2 3 ⎣ 1 −2 1 3 ⎦ ∼ ⎣ 2 −1 3 4 ⎦ ∼ ⎣ 0 3 1 −2 ⎦ ∼ ⎣ 0 1 1 −5 0 5 1 −5 0 5 0 0 0 0 0 0 1. P12 17. ⎡ 2 ⎣ 1 2 1 0 3 3 2 1 4 1 5 ⎤ ⎡ 2 1 1 3 ⎦∼⎣ 2 7 2 0 1 3 2. A12 (−2), A13 (−1) 2 3 1 1 4 5 ⎡ 1 ∼⎣ 0 0 18. 2. A12 (−2), A13 (−2), 4. M3 (− 13 ) 3. A23 (−3) 3 2 1 −1 1 ∼ 2. A12 (−3), A13 (−2), A14 (−5) −4 −6 2 3 1 ∼ 1 −1 3 2 2 ∼ 1. P12 − 23 ⎦ , Rank (A) = 2. 0 ⎤ ⎡ ⎤ 0 2 1 3 1 0 2 1 3 3 1 −1 2 −4 ⎦ ∼ ⎣ 0 1 −1 2 −4 ⎦ 3 −3 3 1 0 0 0 −3 1 ⎤ ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ 1 2 1 2 1 2 1 2 1 4 7 4 7 ⎥ 2 ⎢ 0 −1 0 −1 ⎥ 3 ⎢ 0 ⎥ 1 ⎢ 3 ⎢ 3 5 3 5 5 3 5 ⎥ ⎢ ⎥ ⎢ ⎥ ⎢ ⎢ ⎣ 2 −2 2 −2 ⎦ ∼ ⎣ 2 −2 2 −2 ⎦ ∼ ⎣ 0 −6 0 −6 ⎦ ∼ ⎣ 0 5 −2 5 −2 0 −12 0 −12 0 5 −2 5 −2 ⎤ ⎡ 1 2 1 2 4 ⎢ 0 1 0 1 ⎥ ⎥ ∼⎢ ⎣ 0 0 0 0 ⎦ , Rank (A) = 2. 0 0 0 0 19. ⎤ 3. M2 (1/3) ⎡ 1. A21 (−1) 20. 1 3 0 3 0 2 1 3 1 −1 2 −4 ⎦ , Rank (A) = 3. 0 0 1 − 13 4 1. P12 ⎤ ⎡ 3 1 2 2 ⎦∼⎣ 0 7 0 1 1 −1/2 −6 3 3. M2 (−1) ∼ 1. M1 (− 14 ) 2. A12 (6) 1 −1 0 5 3 ∼ 2. A12 (−3) 1 −1 0 1 4 ∼ 3. M2 ( 15 ) (c)2017 Pearson Education. Inc. 4. A23 (6), A24 (12) 1 −1/2 0 0 2 ⎤ 2 1 2 1 0 1 ⎥ ⎥ −6 0 −6 ⎦ −12 0 −12 , Rank(A) = 1. 1 0 0 1 = I2 , Rank (A) = 2. 4. A21 (1) 147 ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ 1 2 1 1 2 1 1 2 1 1 3 7 10 1 2 3 4 ⎣ 2 3 −1 ⎦ ∼ ⎣ 0 −1 −3 ⎦ ∼ ⎣ 0 1 3 ⎦∼ ⎣ 0 ⎣ 2 3 −1 ⎦ ∼ 3 7 10 0 1 7 0 1 7 0 1 2 1 ⎡ ⎤ ⎡ ⎤ 1 0 −5 1 0 0 5 6 3 ⎦ ∼ ⎣ 0 1 0 ⎦ = I3 , Rank (A) = 3. ∼⎣ 0 1 0 0 1 0 0 1 ⎡ 21. 1. P13 2. A12 (−2), A13 (−3) ⎡ 22. 3 −3 ⎣ 2 −2 6 −6 3. M2 (−1) 5. M3 ( 41 ) 4. A21 (−2), A23 (−1) ⎤ 0 −5 1 3 ⎦ 0 4 6. A31 (5), A32 (−3) ⎤ ⎡ ⎤ 6 1 −1 2 1 4 ⎦∼⎣ 0 0 0 ⎦ , Rank (A) = 1. 12 0 0 0 1. M1 ( 13 ), A12 (−2), A13 (−6) 23. ⎡ ⎤ ⎡ ⎤ ⎡ 3 5 −12 1 2 −5 1 1 2 ⎣ 2 3 −7 ⎦ ∼ ⎣ 0 −1 3 ⎦∼⎣ 0 −2 −1 1 0 3 −9 0 1. A21 (−1), A12 (−2), A13 (2) ⎡ 24. 1 ⎢ 3 ⎢ ⎣ 2 4 3. A21 (−2), A23 (−3) 2. A21 (1), A23 (−1), A24 (−2) 4. M4 (−1) 1 2 3 2. M2 (−1) ⎤ 0 1 1 −3 ⎦ , Rank (A) = 2. 0 0 ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ 1 −1 −1 2 1 0 2 3 1 0 −1 −1 2 ⎥ ⎥ ⎢ ⎢ ⎢ 0 1 3 1 0 1 3 1 −2 0 7 ⎥ 1 ⎢ 2 ⎢ 3 ⎢ 0 1 ⎥∼ ⎥∼ ⎥∼ 1 4 0 ⎦ ⎣ 0 0 1 −1 ⎦ ⎣ 0 0 −1 2 4 ⎦ ⎣ 0 0 2 7 0 0 0 1 −2 0 0 −2 3 8 ⎤ ⎡ ⎤ ⎡ 1 0 0 5 1 0 0 0 ⎥ 5 ⎢ 0 1 0 0 ⎥ 0 1 0 4 4 ⎢ ⎥ ⎢ ⎥ ∼⎢ ⎣ 0 0 1 −1 ⎦ ∼ ⎣ 0 0 1 0 ⎦ = I4 , Rank (A) = 4. 0 0 0 1 0 0 0 1 1. A12 (−3), A13 (−2), A14 (−4) 25. ⎡ 1 −2 ⎣ 3 −6 4 −8 ⎤ ⎡ 2 −5 1 3 1 −3 ⎦ ∼ ⎣ 0 3 −9 0 ⎤ 0 5 0 4 ⎥ ⎥ 1 −1 ⎦ 0 −1 3. A31 (−2), A32 (−3), A34 (−1) 5. A41 (−5), A42 (−4), A43 (1) ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 3 1 −2 1 3 1 −2 1 3 1 −2 0 1 1 2 3 7 ⎦∼⎣ 0 0 −1 −2 ⎦ ∼ ⎣ 0 0 1 2 ⎦∼⎣ 0 0 1 2 ⎦ , Rank (A) = 2. 10 0 0 −1 −2 0 0 −1 −2 0 0 0 0 1. A12 (−3), A13 (−4) 2. M2 (−1) 3. A21 (−1), A23 (1) (c)2017 Pearson Education. Inc. 148 26. ⎡ 0 ⎣ 0 0 1 3 2 2 1 0 ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ 1 0 1 2 1 0 1 2 1 0 1 1 2 3 2 ⎦ ∼ ⎣ 0 0 −6 −2 ⎦ ∼ ⎣ 0 0 1 1/3 ⎦ ∼ ⎣ 0 0 1 0 0 −4 −1 0 0 −4 −1 0 0 ⎡ ⎤ ⎡ ⎤ 0 1 0 1/3 0 1 0 0 4 5 ∼ ⎣ 0 0 1 1/3 ⎦ ∼ ⎣ 0 0 1 0 ⎦ , Rank (A) = 3. 0 0 0 1 0 0 0 1 1. A12 (−3), A13 (−2) 2. M2 (− 16 ) 3. A21 (−2), A23 (4) 4. M3 (3) 0 1 0 ⎤ 1/3 1/3 ⎦ 1/3 5. A32 (− 13 ), A31 (− 13 ) Solutions to Section 2.5 True-False Review: (a): FALSE. This process is known as Gaussian elimination. Gauss-Jordan elimination is the process by which a matrix is brought to reduced row echelon form via elementary row operations. (b): TRUE. A homogeneous linear system always has the trivial solution x = 0, hence it is consistent. (c): TRUE. The columns of the row-echelon form that contain leading 1s correspond to leading variables, while columns of the row-echelon form that do not contain leading 1s correspond to free variables. (d): TRUE. If the last column of the row-reduced augmented matrix for the system does not contain a pivot, then the system can be solved by back-substitution. On the other hand, if this column does contain a pivot, then that row of the row-reduced matrix containing the pivot in the last column corresponds to the impossible equation 0 = 1. (e): FALSE. The linear system x = 0, y = 0, z = 0 has a solution in (0, 0, 0) even though none of the variables here is free. (f ): FALSE. The columns containing the leading 1s correspond to the leading variables, not the free variables. Problems: For the problems of this section, A will denote the coefficient matrix of the given system, and A# will denote the augmented matrix of the given system. 1. Converting the given system of equations to an augmented matrix and using Gaussian elimination we obtain the following equivalent matrices: 1 −5 3 1 −5 3 2 1 −5 3 1 . ∼ ∼ 3 −9 15 0 6 6 0 1 1 1. A12 (−3) 2. M2 ( 16 ) By back substitution, we find that x2 = 1, and then x1 = 8. Therefore, the solution is (8, 1). 2. Converting the given system of equations to an augmented matrix and using Gaussian elimination we obtain the following equivalent matrices: 2 3 1 − 14 4 −1 8 1 1 − 14 2 2 1 − 14 2 ∼ ∼ ∼ . 3 2 1 1 2 1 1 0 1 −2 −3 0 2 (c)2017 Pearson Education. Inc. 149 1. M1 ( 14 ) 2. A12 (−2) 3. M2 ( 23 ) 3 . Therefore, the solution is ( 32 , −2). 2 3. Converting the given system of equations to an augmented matrix and using Gaussian elimination we obtain the following equivalent matrices: 7 −3 5 7 −3 5 2 1 − 37 57 1 . ∼ ∼ 14 −6 10 0 0 0 0 0 0 By back substitution, we find that x2 = −2, and then x1 = 1. A12 (−2) 2. M2 ( 17 ) Observe that x2 is a free variable, so we set x2 = t. Then by back substitution, we have x1 = 37 t + 57 . Therefore, the solution set to this system is 3 5 t + ,t : t ∈ R . 7 7 4. Converting the given system of equations to an augmented matrix and using Gaussian elimination we obtain the following equivalent matrices: ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 1 1 1 1 2 1 1 1 2 1 1 2 1 1 2 1 1 2 3 ⎣ 3 5 1 3 ⎦ ∼ ⎣ 0 −1 −2 0 ⎦∼⎣ 0 1 2 0 ⎦∼⎣ 0 1 2 0 ⎦. 2 6 7 1 0 2 5 −1 0 2 5 −1 0 0 1 −1 1. A12 (−3), A13 (−2) 2. M2 (−1) 3. A23 (−2) The last augmented matrix results in the system: x1 + 2x2 + x3 = 1, x2 + 2x3 = 0, x3 = −1. By back substitution we obtain the solution (−2, 2, −1). 5. Converting the given system of equations to an augmented matrix and using Gaussian elimination, we obtain the following equivalent matrices: ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 1 3 −1 0 1 −2 −5 −3 1 −2 −5 −3 1 2 ⎣ 2 1 5 1 5 5 15 10 ⎦ 4 ⎦∼⎣ 2 4 ⎦∼⎣ 0 7 −5 −8 −3 7 −5 −8 −3 0 9 27 18 ⎡ ⎤ ⎡ ⎤ 1 −2 −5 −3 1 0 1 1 3 4 1 3 2 ⎦ ∼ ⎣ 0 1 3 2 ⎦. ∼⎣ 0 0 9 27 18 0 0 0 0 1. A21 (−1) 2. A12 (−2), A13 (−7) 3. M2 ( 15 ) (c)2017 Pearson Education. Inc. 4. A21 (2), A23 (−9) 150 The last augmented matrix results in the system: x1 + x3 = 1, x2 + 3x3 = 2. Let the free variable x3 = t, a real number. By back substitution we find that the system has the solution set {(1 − t, 2 − 3t, t) : for all real numbers t}. 6. Converting the given system of equations to an augmented matrix and using Gaussian elimination we obtain the following equivalent matrices: ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 3 3 3 5 −1 14 1 2 1 3 1 2 1 1 2 1 1 2 3 ⎣ 1 2 1 3 ⎦ ∼ ⎣ 3 5 −1 4 ⎦ ∼ ⎣ 0 −1 −4 −5 ⎦ ∼ ⎣ 0 1 4 5 ⎦ 2 5 6 2 2 5 6 2 0 1 4 −4 0 1 4 −4 ⎡ ⎤ ⎡ ⎤ 3 1 2 1 1 2 1 3 4 5 5 ⎦ ∼ ⎣ 0 1 4 5 ⎦. ∼⎣ 0 1 4 0 0 0 −9 0 0 0 1 1. P12 2. A12 (−3), A13 (−2) 3. M2 (−1) 4. A23 (−1) 5. M4 (− 19 ) This system of equations is inconsistent since 2 = rank(A) < rank(A# ) = 3. 7. Converting the given system of equations to an augmented matrix and using Gaussian elimination we obtain the following equivalent matrices: ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 6 −3 3 12 2 1 − 12 − 12 1 − 12 12 2 1 1 ⎣ 2 −1 1 4 ⎦ ∼ ⎣ 2 −1 1 0 0 0 ⎦. 4 ⎦∼⎣ 0 −4 2 −2 −8 −4 2 −2 −8 0 0 0 0 1. M1 ( 16 ) 2. A12 (−2), A13 (4) Since x2 and x3 are free variables, let x2 = s and x3 = t. The single equation obtained from the augmented matrix is given by x1 − 12 x2 + 12 x3 = 2. Thus, the solution set of our system is given by {(2 + s t − , s, t) : s, t any real numbers }. 2 2 8. Converting the given system of equations to an augmented matrix and using Gaussian elimination we obtain the following equivalent matrices: ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 3 1 −2 −1 1 2 −5 −15 1 2 −5 −15 2 −1 3 14 ⎢ 3 1 −2 −1 ⎥ 3 14 ⎥ 3 −14 ⎥ −5 13 44 ⎥ 1 ⎢ 2 ⎢ 3 ⎢ ⎢ 2 −1 ⎢ 2 −1 ⎢ 0 ⎥∼ ⎥∼ ⎥∼ ⎥ ⎢ ⎣ 7 2 −3 2 −3 2 −3 3 ⎦ ⎣ 7 3 ⎦ ⎣ 7 3 ⎦ ⎣ 0 −12 32 108 ⎦ 5 −1 −2 5 −1 −2 5 −1 −2 0 −11 23 5 5 5 80 ⎡ ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 1 2 −5 −15 1 2 −5 −15 1 2 −5 −15 32 108 ⎥ 9 1 −9 −28 ⎥ 28 ⎥ 4 ⎢ 0 −12 5 ⎢ 6 ⎢ ⎢ 0 −1 ⎢ 0 ⎥∼ ⎥∼ ⎥ ∼⎢ ⎣ 0 −5 13 ⎣ ⎣ ⎦ ⎦ 0 −5 13 0 −5 13 44 44 44 ⎦ 0 −11 23 0 −11 23 0 −11 23 80 80 80 (c)2017 Pearson Education. Inc. 151 ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 1 2 −5 −15 1 2 −5 −15 1 2 −5 −15 ⎢ ⎥ 9 ⎢ 0 1 −9 −28 ⎥ −9 −28 ⎥ 7 ⎢ 0 1 ⎢ ⎥ 8 ⎢ 0 1 −9 −28 ⎥ ∼ ⎥. ∼⎢ ⎣ 0 0 −32 −96 ⎦ ∼ ⎣ 0 0 32 1 96 ⎦ ⎣ 0 0 3 ⎦ 0 0 −76 −228 0 0 −76 −228 0 0 0 0 1. P12 5. A42 (−1) 2. A21 (−1) 6. M2 (−1) 3. A12 (−2), A13 (−7), A14 (−5) 7. A23 (5), A24 (11) 4. P23 8. M3 (−1) 1 9. M3 ( 32 ), A34 (76). The last augmented matrix results in the system of equations: x1 − 2x2 − 5x3 = −15, x2 − 9x3 = −28, 3. x3 = Thus, using back substitution, the solution set for our system is given by {(2, −1, 3)}. 9. Converting the given system of equations to an augmented matrix and using Gaussian elimination we obtain the following equivalent matrices: ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 2 −1 −4 5 1 1 −3 −3 1 1 −3 −3 1 1 −3 −3 ⎢ 3 2 −5 2 −5 4 17 ⎥ 1 −4 −17 ⎥ 8 ⎥ 8 ⎥ 1 ⎢ 2 ⎢ 3 ⎢ ⎢ ⎢ 3 ⎢ 0 −1 ⎢ 0 ⎥∼ ⎥∼ ⎥∼ ⎥ ⎣ 5 ⎣ ⎣ ⎣ ⎦ ⎦ ⎦ 6 −6 20 5 6 −6 20 0 1 9 35 0 1 9 35 ⎦ 1 1 −3 −3 2 −1 −4 −5 0 −3 2 11 0 −3 2 11 ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 1 1 −3 −3 1 1 −3 −3 1 1 −3 −3 ⎥ 5 ⎢ 0 1 −4 −17 ⎥ 6 ⎢ 0 1 −4 −17 ⎥ 0 1 −4 −17 4 ⎢ ⎥∼⎢ ⎥∼⎢ ⎥. ∼⎢ ⎣ 0 0 13 1 1 52 ⎦ ⎣ 0 0 4 ⎦ ⎣ 0 0 4 ⎦ 0 0 −10 −40 0 0 −10 −40 0 0 0 0 1. P14 2. A12 (−3), A13 (−5), A14 (−2) 3. M2 (−1) 4. A23 (−1), A24 (3) 1 5. M3 ( 13 ) 6. A34 (10) The last augmented matrix results in the system of equations: x1 + x2 − 3x3 = − 3, x2 − 4x3 = −17, x3 = 4. By back substitution, we obtain the solution set {(10, −1, 4)}. 10. Converting the given system of equations to an augmented matrix and using Gaussian elimination we obtain the following equivalent matrices: ⎡ ⎤ ⎡ ⎤ 1 2 −1 1 1 1 2 −1 1 1 1 ⎣ 2 4 −2 2 2 ⎦ ∼ ⎣ 0 0 0 0 0 ⎦. 5 10 −5 5 5 0 0 0 0 0 1. A12 (−2), A13 (−5) The last augmented matrix results in the equation x1 + 2x3 − x3 + x4 = 1. Now x2 , x3 , and x4 are free variables, so we let x2 = r, x3 = s, and x4 = t. It follows that x1 = 1 − 2r + s − t. Consequently, the solution set of the system is given by {(1 − 2r + s − t, r, s, t) : r, s, t and real numbers }. (c)2017 Pearson Education. Inc. 152 11. Converting the given system of equations to an augmented matrix and using Gaussian elimination we obtain the following equivalent matrices: ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 1 1 2 −1 1 1 2 −1 1 1 1 2 −1 1 1 3 ⎢ 2 −3 1 −1 2 ⎥ 3 −3 0 ⎥ 0 ⎥ 1 − 37 1 ⎢ 2 ⎢ 7 ⎢ ⎢ 0 −7 ⎢ 0 ⎥∼ ⎥∼ ⎥ ⎣ 1 −5 2 −2 1 ⎦ ⎣ 0 −7 3 −3 0 ⎦ ⎣ 0 −7 3 −3 0 ⎦ 4 1 −1 1 3 0 −7 3 −3 −1 0 −7 3 −3 −1 ⎡ 1 2 −1 ⎢ 0 1 − 37 3 ∼⎢ ⎣ 0 0 0 0 0 0 1 3 7 0 0 ⎤ ⎡ 1 1 2 −1 3 ⎢ 0 ⎥ 4 ⎢ 0 1 −7 ⎥∼ 0 0 ⎦ ⎣ 0 0 0 0 0 −1 1. A12 (−2), A13 (−1), A14 (−4) 2. M2 (− 17 ) 1 3 7 0 0 ⎤ ⎡ 1 1 2 −1 3 ⎢ 0 ⎥ 5 ⎢ 0 1 −7 ⎥∼ 0 −1 ⎦ ⎣ 0 0 0 0 0 0 3. A23 (7), A24 (7) 4. P34 1 3 7 0 0 ⎤ 1 0 ⎥ ⎥. 1 ⎦ 0 5. M3 (−1) The given system of equations is inconsistent since 2 = rank(A) < rank(A# ) = 3. 12. Converting the given system of equations to an augmented matrix and using Gauss-Jordan elimination we obtain the following equivalent matrices: ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 1 2 1 1 −2 3 1 2 1 1 −2 3 1 2 1 1 −2 3 1 2 ⎣ 0 0 1 4 −3 2 ⎦ ∼ ⎣ 0 0 1 4 −3 2 ⎦ ∼ ⎣ 0 0 1 4 −3 2 ⎦ . 2 4 −1 −10 5 0 0 0 −3 −12 9 −6 0 0 0 0 0 0 1. A13 (−2) 2. A23 (3) The last augmented matrix indicates that the first two equations of the initial system completely determine its solution. We see that x4 and x5 are free variables, so let x4 = s and x5 = t. Then x3 = 2 − 4x4 + 3x5 = 2−4s+3t. Moreover, x2 is a free variable, say x2 = r, so then x1 = 3−2r−(2−4s+3t)−s+2t = 1−2r+3s−t. Hence, the solution set for the system is {(1 − 2r + 3s − t, r, 2 − 4s + 3t, s, t) : r, s, t any real numbers }. 13. Converting the given system of equations to an augmented matrix and using Gauss-Jordan elimination we obtain the following equivalent matrices: ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ −1 −2 2 4 4 4 1 4 1 1 4 1 1 4 1 1 2 3 3 −2 −1 ⎦ ∼ ⎣ 4 3 −2 −1 ⎦ ∼ ⎣ 0 −13 −6 −17 ⎦ ∼ ⎣ 0 −9 −3 −6 ⎦ 4 1 2 −1 −1 0 −9 −3 −6 0 −13 −6 −17 4 2 2 ⎣ 4 1 ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 4 4 4 1 4 1 1 4 1 1 4 1 1 4 1 4 5 6 7 12 4 4 8 ⎦ ∼ ⎣ 0 12 8 ⎦ ∼ ⎣ 0 −1 −2 −9 ⎦ ∼ ⎣ 0 1 2 9 ⎦ ∼⎣ 0 0 −13 −6 −17 0 −1 −2 −9 0 12 4 0 12 4 8 8 ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 3 1 0 −7 −32 1 0 −7 −32 1 0 0 8 9 10 2 2 9 ⎦∼⎣ 0 1 9 ⎦ ∼ ⎣ 0 1 0 −1 ⎦ . ∼⎣ 0 1 0 0 −20 −100 0 0 1 0 0 1 5 5 ⎡ 4 (c)2017 Pearson Education. Inc. 153 1. P13 6. P23 2. A12 (−4), A13 (−2) 7. M2 (−1) 4. M2 (− 43 ) 3. P23 8. A21 (−4), A23 (−12) 5. A23 (1) 1 9. M3 (− 20 ) 10. A31 (7), A32 (−2) The last augmented matrix results in the solution (3, −1, 5). 14. Converting the given system of equations to an augmented matrix and using Gauss-Jordan elimination we obtain the following equivalent matrices: ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 1 3 1 5 2 1 1 −1 1 1 1 −1 1 2 ⎣ 1 1 −1 1 ⎦ ∼ ⎣ 3 1 5 2 ⎦ ∼ ⎣ 0 −2 8 −1 ⎦ 2 1 2 3 2 1 2 3 0 −1 4 1 ⎡ 1 ∼⎣ 0 0 3 1 −1 1 −4 −1 4 ⎤ 1 1 −1 1 1 ⎦ ∼ ⎣ 0 1 −4 1/2 ⎦ . 2 0 0 0 3/2 1 1 ⎡ ⎤ 4 We can stop here, since we see from this last augmented matrix that the system is inconsistent. In particular, 2 = rank(A) < rank(A# ) = 3. 1. P12 2. A12 (−3), A13 (−2) 3. M2 (− 12 ) 4. A23 (1) 15. Converting the given system of equations to an augmented matrix and using Gauss-Jordan elimination we obtain the following equivalent matrices: ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 1 0 −2 −3 1 0 −2 −3 1 0 −2 −3 1 0 −2 −3 1 2 3 ⎣ 3 −2 4 −9 ⎦ ∼ ⎣ 0 −2 2 1 −1 0 ⎦∼⎣ 0 0 ⎦ ∼ ⎣ 0 1 −1 0 ⎦ 1 −4 2 −3 0 −4 4 0 −4 4 0 0 0 0 0 0 . 1. A12 (−3), A13 (−1) 2. M2 (− 12 ) 3. A23 (4) The last augmented matrix results in the following system of equations: x1 − 2x3 = −3 and x2 − x3 = 0. Since x3 is free, let x3 = t. Thus, from the system we obtain the solutions {(2t − 3, t, t) : t any real number }. 16. Converting the given system of equations to an augmented matrix and using Gauss-Jordan elimination we obtain the following equivalent matrices: ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 3 6 6 2 −1 3 −1 1 −2 3 1 1 −2 3 1 1 2 ⎣ 3 2 1 −5 −6 ⎦ ∼ ⎣ 3 2 1 −5 −6 ⎦ ∼ ⎣ 0 8 −8 −8 −24 ⎦ 1 −2 3 1 2 −1 3 −1 0 3 −3 −3 −9 6 3 ⎡ ⎤ ⎡ ⎤ 6 0 1 −2 3 1 1 0 1 −1 3 4 1 −1 −1 −3 ⎦ ∼ ⎣ 0 1 −1 −1 −3 ⎦ . ∼⎣ 0 0 3 −3 −3 −9 0 0 0 0 0 1. P13 2. A12 (−3), A13 (−2) 3. M2 ( 18 ) (c)2017 Pearson Education. Inc. 4. A21 (2), A23 (−3) 154 The last augmented matrix results in the following system of equations: x1 + x3 − x4 = 0 and x2 − x3 − x4 = −3. Since x3 and x4 are free variables, we can let x3 = s and x4 = t, where s and t are real numbers. It follows that the solution set of the system is given by {(t − s, s + t − 3, s, t) : s, t any real numbers }. 17. Converting the given system of equations to an augmented matrix and using Gauss-Jordan elimination we obtain the following equivalent matrices: ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 1 1 1 −1 1 1 1 −1 1 1 1 −1 4 4 4 ⎢ 1 −1 −1 −1 0 −2 ⎥ 0 1 ⎥ 2 ⎥ 1 ⎢ 2 ⎢ ⎢ ⎢ 0 −2 −2 ⎢ 0 1 1 ⎥∼ ⎥∼ ⎥ ⎣ 1 1 −1 1 −2 ⎦ ⎣ 0 0 −2 2 −6 ⎦ ⎣ 0 0 1 −1 3 ⎦ 1 −1 1 1 −8 0 −2 0 2 −12 0 1 0 −1 6 ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 1 0 0 −1 3 1 0 0 −1 1 0 0 −1 1 0 0 0 −1 3 3 1 0 1 ⎥ 1 −2 ⎥ 1 −2 ⎥ 2 ⎥ 3 ⎢ 0 1 4 ⎢ 5 ⎢ 6 ⎢ ⎢ 0 1 0 ⎢ 0 1 0 ⎢ 0 1 0 0 ⎥∼ ⎥∼ ⎥∼ ⎥. ∼⎢ ⎣ 0 0 1 −1 3 ⎦ ⎣ 0 0 1 −1 3 ⎦ ⎣ 0 0 1 −1 3 ⎦ ⎣ 0 0 1 0 −1 ⎦ 0 0 −1 −1 5 0 0 0 −2 0 0 0 1 −4 0 0 0 1 −4 8 1. A12 (−1), A13 (−1), A14 (−1) 4. A32 (−1), A34 (1) 2. M2 (− 12 ), M3 (− 12 ), M4 (− 12 ) 5. M4 (− 12 ) 3. A24 (−1) 6. A41 (1), A42 (−1), A43 (1) It follows from the last augmented matrix that the solution to the system is given by (−1, 2, −1, −4). 18. Converting the given system of equations to an augmented matrix and using Gauss-Jordan elimination we obtain the following equivalent matrices: ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 2 −1 3 1 −1 11 2 2 1 −3 −2 −1 −2 1 −3 −2 −1 −2 ⎢ 1 −3 −2 −1 −2 ⎢ ⎢ 3 1 −1 11 ⎥ 5 7 3 3 2 ⎥ 7 ⎥ ⎢ ⎥ 1 ⎢ 2 −1 ⎥ 2 ⎢ 0 ⎥ ⎢ 3 ⎥∼⎢ 3 ⎥ ∼ ⎢ 0 10 ⎥ 1 −2 −1 1 1 −2 −1 1 4 2 7 −2 −2 −8 ⎢ ⎥ ⎢ ⎥ ⎢ ⎥ ⎣ 1 2 1 2 3 −3 ⎦ ⎣ 1 2 1 2 3 −3 ⎦ ⎣ 0 5 3 3 5 −5 ⎦ 5 −3 −3 1 2 5 −3 −3 1 2 0 12 7 6 12 −8 2 2 ⎡ ⎤ ⎡ ⎤ ⎤ ⎡ 4 11 31 4 11 31 1 0 − 15 1 −3 −2 −1 −2 1 0 2 − 15 5 5 5 5 5 5 3 3 7 7 ⎥ 7 7 ⎥ 7 7 ⎥ 3 3 3 3 ⎢ 0 ⎢ ⎢ 1 5 5 5 5 ⎥ 5 5 5 5 ⎥ 4 ⎢ 0 1 5 5 5 5 ⎥ 5 ⎢ 0 1 3 ⎢ 2 1 11 ⎢ ⎢ ⎥ 4 2 7 −8 ⎥ 1 −22 ⎥ 1 − 10 ∼⎢ 5 5 ⎥ ⎢ 0 10 ⎥∼⎢ 0 0 ⎥ ∼ ⎢ 0 0 −10 −4 ⎣ 0 5 3 3 5 −5 ⎦ ⎣ 0 0 −4 0 2 −12 ⎦ ⎣ 0 0 −4 0 2 −12 ⎦ 24 24 0 12 7 6 12 −8 − 65 − 124 0 0 − 49 − 65 − 124 0 0 − 49 5 5 5 5 5 5 ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 1 2 34 1 2 34 1 6 1 0 0 − 25 1 0 0 − 25 1 0 0 0 10 50 25 50 25 5 37 1 42 ⎥ 37 1 42 ⎥ 7 ⎢ 0 1 0 ⎢ 0 1 0 ⎢ 0 1 0 0 − − − 85 ⎥ 25 50 25 ⎥ 7 ⎢ 25 50 25 ⎥ 8 ⎢ 10 ⎥ 6 ⎢ 2 1 11 ⎥ 2 1 11 ⎥ 1 ⎢ ⎢ − 10 − 10 3 ⎥ ∼⎢ 5 5 ⎥∼⎢ 0 0 1 5 5 ⎥ ∼ ⎢ 0 0 1 0 −2 ⎢ 0 0 1 ⎥ 8 8 ⎣ 0 0 0 ⎦ ⎣ 0 0 0 1 −2 ⎦ 1 1 −2 ⎦ ⎣ 0 0 0 1 − 16 5 5 5 191 68 11 191 68 − 81 0 0 0 0 0 0 0 11 0 0 0 − 81 25 50 25 10 5 25 50 25 ⎤ ⎡ ⎡ ⎤ 1 6 1 0 0 0 10 1 0 0 0 0 1 5 7 ⎢ 0 1 0 0 ⎢ 0 1 0 0 0 −3 ⎥ − 85 ⎥ 10 ⎥ ⎢ ⎢ ⎥ 9 10 ⎢ 1 4 ⎥ 3 ⎥ ∼⎢ ⎥∼⎢ 0 0 1 0 0 ⎢ 0 0 1 0 −2 ⎥ ⎣ 0 0 0 1 1 −2 ⎦ ⎣ 0 0 0 1 0 −4 ⎦ 0 0 0 0 1 2 0 0 0 0 1 2 (c)2017 Pearson Education. Inc. 155 1. P12 2. A12 (−2), A13 (−3), A14 (−1), A15 (−5) 3. M2 ( 51 ) 4. A21 (3), A23 (−10), A24 (−5), A25 (−12) 1 11 7 5. M3 (− 10 ) 6. A31 (− 5 ), A32 (− 5 ), A34 (4), A35 ( 49 7. M4 ( 58 ) 5 ) 2 1 2 68 10 1 7 8. A41 ( 25 ), A42 (− 25 ), A43 (− 5 ), A45 (− 25 ) 9. M5 ( 11 ) 10. A51 (− 10 ), A52 (− 10 ), A53 ( 12 ), A54 (−1) It follows from the last augmented matrix that the solution to the system is given by (1, −3, 4, −4, 2). 19. The equation Ax = b reads ⎤ ⎡ ⎤⎡ ⎤ 1 −3 1 x1 8 ⎣ 5 −4 1 ⎦ ⎣ x2 ⎦ = ⎣ 15 ⎦ . 2 4 −3 x3 −4 ⎡ Converting the given system of equations to an augmented matrix and using Gauss-Jordan elimination we obtain the following equivalent matrices: ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ 1 −3 1 8 8 8 1 −3 1 1 −3 1 1 2 ⎣ 5 −4 1 15 ⎦ ∼ ⎣ 0 11 −4 −25 ⎦ ∼ ⎣ 0 1 1 −5 ⎦ 2 4 −3 −4 0 10 −5 −20 0 10 −5 −20 ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 1 1 0 4 −7 1 0 4 −7 1 0 0 3 4 5 1 −5 ⎦ ∼ ⎣ 0 1 1 −5 ⎦ ∼ ⎣ 0 1 0 −3 ⎦ . ∼⎣ 0 1 0 0 −15 30 0 0 1 −2 0 0 1 −2 1. A12 (−5), A13 (−2) 2. A32 (−1) 3. A21 (3), A23 (−10) 1 4. M3 (− 15 ) 5. A31 (−4), A32 (−1) Thus, from the last augmented matrix, we see that x1 = 1, x2 = −3, and x3 = −2. 20. The equation Ax = b reads ⎤ ⎡ ⎤⎡ ⎤ 1 0 5 x1 0 ⎣ 3 −2 11 ⎦ ⎣ x2 ⎦ = ⎣ 2 ⎦ . 2 −2 6 x3 2 ⎡ Converting the given system of equations to an augmented matrix and using Gauss-Jordan elimination we obtain the following equivalent matrices: ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ 0 1 0 5 0 1 0 5 1 0 5 0 1 2 ⎣ 0 −2 −4 2 ⎦ ∼ ⎣ 0 ⎣ 3 −2 11 2 ⎦ ∼ 1 2 −1 ⎦ 2 −2 6 2 0 −2 −4 2 0 −2 −4 2 ⎡ ⎤ 0 1 0 5 3 ∼ ⎣ 0 1 2 −1 ⎦ . 0 0 0 0 1. A12 (−3), A13 (−2) 2. M2 (−1/2) 3. A23 (2) Hence, we have x1 + 5x3 = 0 and x2 + 2x3 = −1. Since x3 is a free variable, we can let x3 = t, where t is any real number. It follows that the solution set for the given system is given by {(−5t, −2t − 1, t) : t ∈ R}. 21. The equation Ax = b reads ⎡ 0 ⎣ 0 0 ⎤⎡ ⎤ ⎡ ⎤ x1 1 −1 −2 5 1 ⎦ ⎣ x2 ⎦ = ⎣ 8 ⎦ . 2 1 5 x3 (c)2017 Pearson Education. Inc. 156 Converting the given system of equations to an augmented matrix using Gauss-Jordan elimination we obtain the following equivalent matrices: ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 0 1 −1 −2 0 1 −1 −2 0 1 −1 −2 0 1 0 1 1 2 3 ⎣ 0 5 1 6 18 ⎦ ∼ ⎣ 0 0 1 8 ⎦∼⎣ 0 0 3 ⎦ ∼ ⎣ 0 0 1 3 ⎦. 0 2 1 0 0 3 0 0 3 0 0 0 0 5 9 9 1. A12 (−5), A13 (−2) 2. M2 (1/6) 3. A21 (1), A23 (−3) Consequently, from the last augmented matrix it follows that the solution set for the matrix equation is given by {(t, 1, 3) : t ∈ R}. 22. The equation Ax = b reads ⎤⎡ ⎤ ⎡ ⎤ x1 1 −1 0 −1 2 ⎣ 2 1 3 7 ⎦ ⎣ x2 ⎦ = ⎣ 2 ⎦ . 3 −2 1 0 4 x3 ⎡ Converting the given system of equations to an augmented matrix and using Gauss-Jordan elimination we obtain the following equivalent matrices: ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 2 2 0 1 −1 0 −1 2 1 −1 0 −1 1 −1 0 −1 1 0 1 2 1 2 3 ⎣ 2 1 3 7 2 ⎦∼⎣ 0 3 3 9 −2 ⎦ ∼ ⎣ 0 1 1 3 −2 ⎦ ∼ ⎣ 0 1 1 3 −2 ⎦ . 3 −2 1 0 4 0 1 1 3 −2 0 3 3 9 −2 0 0 0 0 4 1. A12 (−2), A13 (−3) 2. P23 3. A21 (1), A23 (−3) From the last row of the last augmented matrix, it is clear that the given system is inconsistent. 23. The equation Ax = b reads ⎤⎡ ⎤ ⎤ ⎡ x1 1 1 0 −1 2 ⎢ ⎥ ⎢ 3 1 −2 ⎥ ⎢ 3 ⎥ ⎥ ⎢ x2 ⎥ = ⎢ 8 ⎥ . ⎢ ⎣ 2 3 1 1 ⎦ ⎣ x3 ⎦ ⎣ 3 ⎦ −2 3 5 −2 −9 x4 ⎡ Converting the given system of equations to an augmented matrix and using Gauss-Jordan elimination we obtain the following equivalent matrices: ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 1 1 0 1 1 1 0 1 1 0 −1 1 2 2 2 3 1 1 0 1 ⎢ 3 1 −2 3 1 1 0 −1 ⎥ 1 0 −1 ⎥ 8 ⎥ 2 ⎥ 1 ⎢ 2 ⎢ 3 ⎢ ⎢ ⎢ 0 −2 −2 0 ⎢ 0 ⎢ 0 1 ⎥∼ ⎥∼ ⎥∼ ⎥. ⎣ 2 3 ⎣ ⎣ ⎣ ⎦ ⎦ ⎦ ⎦ 1 2 0 1 1 0 −1 0 −2 −2 0 0 0 0 0 3 2 −2 3 5 −2 −9 0 5 5 0 −5 0 5 5 0 −5 0 0 0 0 0 1. A12 (−3), A13 (−2), A14 (2) 2. P23 3. A21 (−1), A23 (2), A24 (−5) From the last augmented matrix, we obtain the system of equations: x1 − x3 + x4 = 3, x2 + x3 = −1. Since both x3 and x4 are free variables, we may let x3 = r and x4 = t, where r and t are real numbers. The solution set for the system is given by {(3 + r − t, −r − 1, r, t) : r, t ∈ R}. 24. Converting the given system of equations to an augmented matrix and using Gauss-Jordan elimination we obtain the following equivalent matrices: ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 3 3 3 1 2 −1 1 2 −1 1 2 −1 1 2 ⎣ 2 5 ⎦∼⎣ 0 1 ⎦. 1 1 3 3 7 ⎦∼⎣ 0 1 1 2 2 2 1 1 −k −k 0 −1 1 − k −3 − k 0 0 4 − k −2 − k (c)2017 Pearson Education. Inc. 157 1. A12 (−2), A13 (−1) 2. A23 (1) (a). If k = 2, then the last row of the last augmented matrix reveals an inconsistency; hence the system has no solutions in this case. (b). If k = −2, then the last row of the last augmented matrix consists entirely of zeros, and hence we have only two pivots (first two columns) and a free variable x3 ; hence the system has infinitely many solutions. (c). If k = ±2, then the last augmented matrix above contains a pivot for each variable x1 , x2 , and x3 , and can be solved for a unique solution by back-substitution. 25. Converting the given system of equations to an augmented matrix and using Gauss-Jordan elimination we obtain the following equivalent matrices: ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ 1 1 1 −1 0 1 1 1 −1 0 2 1 −1 1 0 ⎢ 1 1 1 −1 0 ⎥ 1 −1 1 0 ⎥ 3 0 ⎥ 1 ⎢ 2 ⎢ ⎢ 2 ⎢ 0 −1 −3 ⎥∼ ⎥∼ ⎥ ⎢ ⎣ ⎣ ⎦ ⎦ ⎣ 4 2 −1 1 0 4 2 −1 1 0 0 −2 −5 5 0 ⎦ 3 −1 1 k 0 3 −1 1 k 0 0 −4 −2 k + 3 0 ⎡ ⎤ ⎡ 1 1 0 ⎢ 0 1 0 ⎥ 4 ⎥∼⎢ 0 ⎦ ⎣ 0 0 0 0 0 1 −1 3 −3 1 −1 10 k − 9 1. P12 2. A12 (−2), A13 (−4), A14 (−3) 3. M2 (−1) 1 1 1 −1 0 1 3 −3 3 ⎢ ∼⎢ ⎣ 0 −2 −5 5 0 −4 −2 k + 3 ⎤ ⎡ 1 1 1 −1 0 ⎢ 0 1 3 −3 0 ⎥ 5 ⎥∼⎢ 0 ⎦ ⎣ 0 0 1 −1 0 0 0 k+1 0 4. A23 (2), A24 (4) ⎤ 0 0 ⎥ ⎥. 0 ⎦ 0 5. A34 (−10) (a). Note that the trivial solution (0, 0, 0, 0) exists under all circumstances, so there are no values of k for which there is no solution. (b). From the last row of the last augmented matrix, we see that if k = −1, then the variable x4 corresponds to an unpivoted column, and hence it is a free variable. In this case, therefore, we have infinitely solutions. (c). Provided that k = −1, then each variable in the system corresponds to a pivoted column of the last augmented matrix above. Therefore, we can solve the system by back-substitution. The conclusion from this is that there is a unique solution, (0, 0, 0, 0). 26. Converting the given system of equations to an augmented matrix and using Gauss-Jordan elimination we obtain the following equivalent matrices: ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 1 1 −2 4 4 4 2 1 1 −2 1 1 −2 1 0 −3 1 2 3 ⎣ 3 5 −4 16 ⎦ ∼ ⎣ 0 2 ⎦. 2 1 1 4 ⎦∼⎣ 0 1 2 ⎦∼⎣ 0 1 2 2 3 −a 0 1 4−a b−8 0 1 4−a b−8 0 0 3 − a b − 10 b 1. A12 (−3), A13 (−2) 2. M2 ( 12 ) 3. A21 (−1), A23 (−1) (a). From the last row of the last augmented matrix above, we see that there is no solution if a = 3 and b = 10. (b). From the last row of the augmented matrix above, we see that there are infinitely many solutions if a = 3 and b = 10, because in that case, there is no pivot in the column of the last augmented matrix corresponding to the third variable x3 . (c)2017 Pearson Education. Inc. 158 (c). From the last row of the augmented matrix above, we see that if a = 3, then regardless of the value of b, there is a pivot corresponding to each variable x1 , x2 , and x3 . Therefore, we can uniquely solve the corresponding system by back-substitution. 27. Converting the given system of equations to an augmented matrix and using Gauss-Jordan elimination we obtain the following equivalent matrices: ⎡ ⎤ ⎡ ⎤ 1 −a 3 3 1 −a 1 ⎣ 2 1 6 ⎦ ∼ ⎣ 0 1 + 2a 0 ⎦ . −3 a + b 1 0 b − 2a 10 From the middle row, we see that if a = − 12 , then we must have x2 = 0, but this leads to an inconsistency in solving for x1 (the first equation would require x1 = 3 while the last equationwould require x1 = − 13 . Now 1 −1/2 3 . If b = −1, suppose that a = − 12 . Then the augmented matrix on the right reduces to 0 b + 1 10 then once more we have an inconsistency in the last row. However, if b = −1, then the row-echelon form obtained has full rank, and there is a unique solution. Therefore, we draw the following conclusions: (a). There is no solution to the system if a = − 12 or if a = − 12 and b = −1. (b). Under no circumstances are there an infinite number of solutions to the linear system. (c). There is a unique solution if a = − 12 and b = −1. 28. The corresponding augmented matrix for this linear system can be reduced to row-echelon form via ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ y1 y1 1 1 1 y1 1 1 1 1 1 1 1 2 ⎣ 2 3 1 y2 ⎦ ∼ ⎦. ⎣ 0 1 −1 y2 − 2y1 ⎦ ∼ ⎣ 0 1 −1 y2 − 2y1 3 5 1 y3 0 2 −2 y3 − 3y1 0 0 0 y1 − 2y2 + y3 1. A12 (−2), A13 (−3) 2. A23 (−2) For consistency, we must have rank(A) = rank(A# ), which requires (y1 , y2 , y3 ) to satisfy y1 − 2y2 + y3 = 0. If this holds, then the system has an infinite number of solutions, because the column of the augmented matrix corresponding to y3 will be unpivoted, indicating that y3 is a free variable in the solution set. 29. Converting the given system of equations to an augmented matrix and using Gaussian elimination we obtain the following row-equivalent matrices. Since a11 = 0: a12 b1 b1 1 1 aa12 a11 a12 b1 1 2 a a a11 11 11 11 . ∼ ∼ a11 b2 −a21 b1 21 a12 a21 a22 b2 0 aΔ11 aΔ112 0 a22 a11a−a a11 11 1. M1 (1/a11 ), A12 (−a21 ) 2. Definition of Δ and Δ2 (a). If Δ = 0, then rank(A) = rank(A# ) = 2, so the system a unique solution (of course, we are assuming has Δ a11 = 0 here). Using the last augmented matrix above, a11 x2 = aΔ112 , so that x2 = ΔΔ2 . Using this, we can 1 x2 = ab11 for x1 to obtain x1 = ΔΔ1 , where we have used the fact that Δ1 = a22 b1 − a12 b2 . solve x1 + aa12 11 b1 1 aa12 a 11 11 , so it follows that (b). If Δ = 0 and a11 = 0, then the augmented matrix of the system is 0 0 Δ2 the system has (i) no solution if Δ2 = 0, since rank(A) < rank(A# ) = 2, and (ii) an infinite number of solutions if Δ2 = 0, since rank(A# ) < 2. (c)2017 Pearson Education. Inc. 159 (c). An infinite number of solutions would be represented as one line. No solution would be two parallel lines. A unique solution would be the intersection of two distinct lines at one point. 30. We first use the partial pivoting algorithm to reduce the augmented matrix of the system: ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 3 1 2 1 1 3 5 1 3 3 5 1 1 2 ⎣ 3 5 1 3 ⎦ ∼ ⎣ 1 2 1 1 ⎦ ∼ ⎣ 0 1/3 2/3 0 ⎦ 2 6 7 1 2 6 7 1 0 8/3 19/3 −1 ⎤ ⎡ ⎤ ⎡ 3 3 3 5 1 3 5 1 3 4 ∼ ⎣ 0 8/3 19/3 −1 ⎦ ∼ ⎣ 0 8/3 19/3 −1 ⎦ . 0 1/3 2/3 0 0 −1/8 1/8 0 1. P12 2. A12 (−1/3), A13 (−2/3) 3. P23 4. A23 (−1/8) Using back substitution to solve the equivalent system yields the unique solution (−2, 2, −1). 31. We first use the partial pivoting algorithm to reduce the augmented matrix of the system: ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ 3 3 7 2 −3 7 2 −3 2 −1 3 14 ⎢ 3 1 −2 −1 ⎥ 1 −2 −1 ⎥ 1/7 −5/7 −16/7 ⎥ 1 ⎢ 2 ⎢ ⎢ 3 ⎢ 0 ⎥∼ ⎥∼ ⎥ ⎢ ⎣ 7 2 −3 3 14 ⎦ ⎣ 0 −11/7 27/7 3 ⎦ ⎣ 2 −1 92/7 ⎦ 5 −1 −2 5 −1 −2 0 −17/7 1/7 5 5 20/7 ⎡ ⎤ ⎡ ⎤ 3 3 7 2 −3 7 2 −3 1/7 1/7 20/7 ⎥ 20/7 ⎥ 3 ⎢ 0 −17/7 4 ⎢ ⎢ 0 −17/7 ⎥∼ ⎥ ∼⎢ ⎣ 0 −11/7 27/7 ⎣ ⎦ 0 0 64/17 92/7 192/17 ⎦ 0 1/7 −5/7 −16/7 0 0 −12/17 −36/17 ⎡ ⎤ 3 7 2 −3 ⎥ 0 −17/7 1/7 20/7 5 ⎢ ⎥. ∼⎢ ⎣ 0 0 64/17 192/17 ⎦ 0 0 0 0 1. P13 2. A12 (−3/7), A13 (−2/7), A14 (−5/7) 4. A23 (−11/17), A24 (1/17) 3. P24 5. A34 (3/16) Using back substitution to solve the equivalent system yields the unique solution (2, −1, 3). 32. We first use the partial pivoting algorithm to reduce the augmented matrix of the system: ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ 5 6 −6 −20 5 6 −6 5 20 2 −1 −4 ⎥ ⎢ 3 2 −5 2 −5 8 ⎥ 8 ⎥ 1 ⎢ 2 ⎢ ⎢ 3 ⎢ 0 −8/5 −7/5 −4 ⎥ ⎥∼ ⎥∼ ⎢ ⎣ ⎣ ⎦ ⎦ ⎣ 5 6 −6 20 2 −1 −4 0 −17/5 −8/5 −3 ⎦ 5 1 1 −3 −3 1 1 −3 −3 0 −1/5 −9/5 −7 ⎡ ⎤ ⎡ ⎤ 20 20 5 6 −6 5 6 −6 ⎢ ⎥ −3 3 ⎢ 0 −17/5 −8/5 −3 ⎥ ⎥ 4 ⎢ 0 −17/5 −8/5 ⎥ ∼⎢ ⎣ 0 −8/5 −7/5 −4 ⎦ ∼ ⎣ 0 0 −11/17 −44/17 ⎦ 0 −1/5 −9/5 −7 0 0 −29/17 −116/17 ⎡ ⎤ ⎡ ⎤ 5 6 −6 5 6 −6 20 20 ⎥ 6 ⎢ 0 −17/5 −8/5 ⎥ −8/5 −3 −3 5 ⎢ 0 −17/5 ⎥∼⎢ ⎥. ∼⎢ ⎣ ⎣ 0 ⎦ 0 −29/17 −116/17 0 0 −29/17 −116/17 ⎦ 0 0 −11/17 −44/17 0 0 0 0 (c)2017 Pearson Education. Inc. 160 1. P13 2. A12 (−3/5), A13 (−2/5), A14 (−1/5) 4. A23 (−8/17), A24 (−1/17) 5. P34 3. P23 6. A34 (−11/29) Using back substitution to solve the equivalent system yields the unique solution (10, −1, 4). 33. We first use the partial pivoting algorithm to reduce the augmented matrix of the system: ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 2 2 −1 −1 4 3 −2 −1 4 3 −2 −1 1 2 ⎣ 4 3 −2 −1 ⎦ ∼ ⎣ 2 −1 −1 0 2 ⎦ ∼ ⎣ 0 −5/2 5/2 ⎦ 1 4 1 1 4 1 0 13/4 3/2 17/4 4 4 ⎡ 4 3 ∼ ⎣ 0 13/4 0 −5/2 3 1. P12 ⎤ ⎡ −2 −1 4 4 3/2 17/4 ⎦ ∼ ⎣ 0 0 0 5/2 ⎤ −1 3 −2 13/4 3/2 17/4 ⎦ . 0 15/13 75/13 2. A12 (−1/2), A13 (−1/4) 3. P23 4. A23 (10/13) Using back substitution to solve the equivalent system yields the unique solution (3, −1, 5). 34. (a). Let ⎡ a11 ⎢ a21 ⎢ A# = ⎢ ⎢ a31 ⎣ ... an1 0 a22 a32 ... an2 0 0 a33 ... an3 ... ... ... ... ... 0 0 0 ... ann ⎤ b1 b2 ⎥ ⎥ b3 ⎥ ⎥ ... ⎦ bn represent the corresponding augmented matrix of the given system. Since a11 x1 = b1 , we can solve for x1 easily: b1 x1 = , (a11 = 0). a11 Now since a21 x1 + a22 x2 = b2 , by using the expression for x1 we just obtained, we can solve for x2 : x2 = a11 b2 − a21 b1 . a11 a22 In a similar manner, we can solve for x3 , x4 , . . . , xn . (b). We solve instantly for x1 from the first equation: x1 = 2. Substituting this into the middle equation, we obtain 2 · 2 − 3 · x2 = 1, from which it quickly follows that x2 = 1. Substituting for x1 and x2 in the bottom equation yields 3 · 2 + 1 − x3 = 8, from which it quickly follows that x3 = −1. Consequently, the solution of the given system is (2, 1, −1). 35. This system of equations is not linear in x1 , x2 , and x3 ; however, the system is linear in x31 , x22 , and x3 , so we can first solve for x31 , x22 , and x3 . Converting the given system of equations to an augmented matrix and using Gauss-Jordan elimination we obtain the following equivalent matrices: ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 4 2 3 12 2 1 −1 1 2 1 −1 1 1 2 ⎣ 1 −1 1 2 ⎦∼⎣ 4 2 3 12 ⎦ ∼ ⎣ 0 6 −1 4 ⎦ 3 1 −1 2 3 1 −1 2 0 4 −4 −4 (c)2017 Pearson Education. Inc. 161 ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 2 2 1 1 −1 1 1 −1 1 1 0 0 4 5 4 −4 −4 ⎦ ∼ ⎣ 0 1 −1 −1 ⎦ ∼ ⎣ 0 1 −1 −1 ⎦ ∼⎣ 0 0 6 −1 0 6 −1 0 0 5 10 4 4 ⎤ ⎡ ⎤ ⎡ 1 1 0 0 1 0 0 1 6 7 ∼ ⎣ 0 1 −1 −1 ⎦ ∼ ⎣ 0 1 0 1 ⎦ . 0 0 1 0 0 1 2 2 3 1. P12 2. A12 (−4), A13 (−3) 5. A21 (1), A23 (−6) 3. P23 6. M2 (1/5) 4. M2 (1/4) 7. A32 (1) Thus, taking only real solutions, we have x31 = 1, x22 = 1, and x3 = 2. Therefore, x1 = 1, x2 = ±1, and x3 = 2, leading to the two solutions (1, 1, 2) and (1, −1, 2) to the original system of equations. There is no contradiction of Theorem 2.5.9 here since, as mentioned above, this system is not linear in x1 , x2 , and x3 . 36. Reduce the augmented matrix of the system: ⎤ ⎡ ⎤ ⎡ ⎡ 1 1 −2 0 1 1 3 2 −1 0 1 2 ⎣ 2 1 1 0 ⎦ ∼ ⎣ 0 −1 5 0 ⎦∼⎣ 0 1 5 −4 1 0 0 −9 11 0 0 −9 ⎡ ⎤ ⎡ 1 0 3 0 1 0 4 5 ∼ ⎣ 0 1 −5 0 ⎦ ∼ ⎣ 0 1 0 0 1 0 0 0 1. A21 (−1), A12 (−2), A13 (−5) 4. M3 (−1/34) ⎤ ⎡ ⎤ −2 0 1 0 3 0 3 −5 0 ⎦ ∼ ⎣ 0 1 −5 0 ⎦ 11 0 0 0 −34 0 ⎤ 0 0 0 0 ⎦. 1 0 2. M2 (−1) 3. A21 (−1), A23 (9) 5. A31 (−3), A32 (5) Therefore, the unique solution to this system is x1 = x2 = x3 = 0: (0, 0, 0). 37. Reduce the augmented matrix of the system: ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ 2 1 −1 0 1 −1 −1 0 1 −1 −1 0 1 −1 −1 0 ⎢ 3 −1 ⎢ 2 0 ⎥ 2 0 ⎥ 2 5 0 ⎥ 3 1 0 ⎥ 2 ⎢ 3 ⎢ ⎥ 1 ⎢ 3 −1 ⎥∼ ⎥∼ ⎥ ⎢ ⎢ 0 ⎢ 0 ⎣ 1 −1 −1 0 ⎦ ∼ ⎣ 2 1 −1 0 ⎦ ⎣ 0 3 1 0 ⎦ ⎣ 0 2 5 0 ⎦ 5 2 −2 0 5 2 −2 0 0 7 3 0 0 7 3 0 ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 1 −1 −1 0 1 0 −5 0 1 0 −5 0 1 0 0 0 ⎥ 6 ⎢ 0 1 −4 0 ⎥ 7 ⎢ 0 1 0 0 ⎥ 1 −4 0 ⎥ 4 ⎢ 0 5 ⎢ ⎢ 0 1 −4 0 ⎥ ∼ ⎢ ⎥∼ ⎥∼⎢ ⎥. ∼⎢ ⎣ 0 ⎣ ⎦ 2 5 0 0 0 13 0 ⎦ ⎣ 0 0 1 0 ⎦ ⎣ 0 0 1 0 ⎦ 0 7 3 0 0 0 31 0 0 0 31 0 0 0 0 0 1. P13 2. A12 (−3), A13 (−2), A14 (−5) 5. A21 (1), A23 (−2), A24 (−7) 6. M3 (1/13) 3. P23 4. A32 (−1) 7. A31 (5), A32 (4), A34 (−31) Therefore, the unique solution to this system is x1 = x2 = x3 = 0: (0, 0, 0). 38. Reduce the augmented matrix of the system: ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ 1 1 4 0 1 1 4 0 2 −1 −1 0 1 2 ⎣ 5 −1 2 0 ⎦ ∼ ⎣ 5 −1 2 0 ⎦ ∼ ⎣ 0 −6 −18 0 ⎦ 1 1 4 0 2 −1 −1 0 0 −3 −9 0 (c)2017 Pearson Education. Inc. 162 ⎡ 1 ∼⎣ 0 0 3 1. P13 ⎤ ⎡ ⎤ 1 4 0 1 0 1 0 4 1 3 0 ⎦ ∼ ⎣ 0 1 3 0 ⎦. −3 −9 0 0 0 0 0 2. A12 (−5), A13 (−2) 3. M2 (−1/6) 4. A21 (−1), A23 (3) It follows that x1 + x3 = 0 and x2 + 3x3 = 0. Setting x3 = t, where t is a free variable, we get x2 = −3t and x1 = −t. Thus we have that the solution set of the system is {(−t, −3t, t) : t ∈ R}. 39. Reduce the augmented matrix of the system: ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 1 + 2i 1 − i 1 0 0 0 i 1+i −i 1 1−i −1 1 2 ⎣ i 1+i −i 1 1 0 ⎦ ∼ ⎣ 1 + 2i 1 − i 0 ⎦ ∼ ⎣ 1 + 2i 1 − i 0 ⎦ 2i 1 1 + 3i 0 2i 1 1 + 3i 0 2i 1 1 + 3i 0 ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 0 0 0 1 1−i −1 1 1−i −1 1 1−i −1 3 4 5 0 −5 + 8i 0 ⎦ ∼ ⎣ 0 −2 − 2i 1 + 2i 0 ⎦ ∼ ⎣ 0 −2 − 2i 1 + 2i 0 ⎦ ∼ ⎣ 0 0 −1 − 2i 1 + 5i 0 0 1 3i 0 1 3i 0 0 ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 0 1 1−i −1 1 1 − i −1 0 1 0 0 0 6 7 8 1 3i 1 3i 0 ⎦ ∼ ⎣ 0 1 0 0 ⎦ . 0 ⎦∼⎣ 0 ∼⎣ 0 0 0 −5 + 8i 0 0 0 1 0 0 0 1 0 1. P12 2. M1 (−i) 6. P23 3. A12 (−1 − 2i), A13 (−2i) 1 7. M3 ( −5+8i ) 4. A23 (−1) 5. A32 (2 + 2i) 8. A21 (−1 + i), A31 (1), A32 (−3i) Therefore, the unique solution to this system is x1 = x2 = x3 = 0: (0, 0, 0). 40. Reduce the augmented matrix of the system: ⎤ ⎡ ⎤ ⎡ ⎡ 1 2 2 3 2 1 0 0 1 1 3 3 3 1 2 ⎣ 0 −5 ⎣ 6 −1 2 0 ⎦ ∼ ⎣ 6 −1 2 0 ⎦ ∼ 12 6 4 0 12 6 4 0 0 −2 ⎡ ⎡ ⎤ ⎤ 1 2 0 1 1 0 13 0 3 3 3 4 1 0 0 ⎦ ∼ ⎣ 0 1 0 0 ⎦. ∼⎣ 0 0 −2 0 0 0 0 0 0 1. M1 (1/3) 2. A12 (−6), A13 (−12) 3. M2 (−1/5) 1 3 0 0 ⎤ 0 0 ⎦ 0 4. A21 (−2/3), A23 (2) From the last augmented matrix, we have x1 + 13 x3 = 0 and x2 = 0. Since x3 is a free variable, we let x3 = t, where t is a real number. It follows that the solution set for the given system is given by {(t, 0, −3t) : t ∈ R}. 41. Reduce the augmented matrix of the system: ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 2 1 −8 0 3 −2 −5 0 1 −3 3 0 ⎢ 3 −2 −5 0 ⎥ 1 ⎢ 2 ⎢ 1 −8 0 ⎥ 1 −8 0 ⎥ ⎢ ⎥ ⎢ ⎥ 2 ⎢ 2 ⎥ ⎣ 5 −6 −3 0 ⎦ ∼ ⎣ 5 −6 −3 0 ⎦ ∼ ⎣ 5 −6 −3 0 ⎦ 3 −5 1 0 3 −5 1 0 3 −5 1 0 ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 1 −3 3 0 1 −3 3 0 1 0 −3 0 ⎥ 7 −14 0 ⎥ 1 −2 0 ⎥ 3 ⎢ 0 4 ⎢ 5 ⎢ ⎢ 0 ⎢ 0 1 −2 0 ⎥ . ⎥∼ ⎥∼ ∼⎢ ⎣ 0 ⎣ ⎣ ⎦ ⎦ 9 −18 0 0 9 −18 0 0 0 0 0 ⎦ 0 4 −8 0 0 4 −8 0 0 0 0 0 (c)2017 Pearson Education. Inc. 163 1. P12 2. A21 (−1) 3. A12 (−2), A13 (−5), A14 (−3) 4. M2 (1/7) 5. A21 (3), A23 (−9), A24 (−4) From the last augmented matrix we have: x1 − 3x3 = 0 and x2 − 2x3 = 0. Since x3 is a free variable, we let x3 = t, where t is a real number. It follows that x2 = 2t and x1 = 3t. Thus, the solution set for the given system is given by {(3t, 2t, t) : t ∈ R}. 42. Reduce the augmented matrix of the system: ⎡ ⎤ ⎡ 1 1+i 1−i 0 1 1+i 1 ⎣ i 1 i 2−i 0 ⎦∼⎣ 0 1 − 2i −1 + i 1 − 3i 0 0 −4 + 2i ⎡ 1 1+i 3 1 ∼⎣ 0 0 0 1−i −2−i 5 0 1. A12 (−i), A13 (−1 + 2i) 1−i −1 2 ⎤ ⎡ 0 1 0 4 0 ⎦∼⎣ 0 1 0 0 0 ⎤ ⎡ 0 1 1+i 2 0 ⎦∼⎣ 0 2−i 0 0 0 6−2i 5 −2−i 5 0 ⎤ 0 0 ⎦ 0 ⎤ 0 0 ⎦. 0 1 3. M2 ( 2−i ) 2. A23 (2) 1−i −1 0 4. A21 (−1 − i) From the last augmented matrix we see that x3 is a free variable. We set x3 = 5s, where s ∈ C. Then x1 = 2(i − 3)s and x2 = (2 + i)s. Thus, the solution set of the system is {(2(i − 3)s, (2 + i)s, 5s) : s ∈ C}. 43. Reduce the augmented matrix of the system: ⎤ ⎡ ⎤ ⎡ ⎡ 1 −1 1 0 1 −1 1 −1 1 0 ⎥ 1 ⎢ 0 ⎥ 2 ⎢ 0 ⎢ 0 3 2 3 2 1 0 0 ⎥∼⎢ ⎥∼⎢ ⎢ ⎣ 3 0 −1 0 ⎦ ⎣ 0 3 −4 0 ⎦ ⎣ 0 3 5 1 −1 0 0 6 −6 0 0 6 ⎡ 1 0 5/3 0 1 2/3 3 ⎢ ∼⎢ ⎣ 0 0 −6 0 0 −10 ⎤ ⎡ 1 0 5/3 0 ⎢ 0 1 2/3 0 ⎥ 4 ⎥∼⎢ 1 0 ⎦ ⎣ 0 0 0 0 −10 0 1. A13 (−3), A14 (−5) 4. M3 (−1/6) 2. M2 (1/3) 1 2/3 −4 −6 ⎤ 0 0 ⎥ ⎥ 0 ⎦ 0 ⎤ ⎡ ⎤ 1 0 0 0 0 ⎥ 0 ⎥ 5 ⎢ ⎢ 0 1 0 0 ⎥. ⎥∼ ⎣ ⎦ 0 0 1 0 ⎦ 0 0 0 0 0 0 3. A21 (1), A23 (−3), A24 (−6) 5. A31 (−5/3), A32 (−2/3), A34 (10) Therefore, the unique solution to this system is x1 = x2 = x3 = 0: (0, 0, 0). 44. Reduce the augmented matrix of the system: ⎤ ⎡ ⎡ 1 −2 3 2 −4 6 0 ⎢ 3 −6 9 0 ⎥ 1 ⎢ 3 −6 9 ⎥ ⎢ ⎢ ⎣ 1 −2 3 0 ⎦ ∼ ⎣ 2 −4 6 5 −10 15 0 5 −10 15 1. M1 (1/2) ⎤ ⎡ ⎤ 0 1 −2 3 0 0 0 0 ⎥ 0 ⎥ 2 ⎢ ⎢ 0 ⎥∼ ⎥. ⎣ ⎦ 0 0 0 0 ⎦ 0 0 0 0 0 0 2. A12 (−3), A13 (−2), A14 (−5) From the last matrix we have that x1 − 2x3 + 3x3 = 0. Since x2 and x3 are free variables, let x2 = s and let x3 = t, where s and t are real numbers. The solution set of the given system is therefore {(2s − 3t, s, t) : s, t ∈ R}. (c)2017 Pearson Education. Inc. 164 45. Reduce the augmented matrix of the system: ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 4 −2 −1 −1 0 1 −3 1 −4 0 1 −3 1 −4 0 1 2 ⎣ 3 1 −2 3 0 ⎦∼⎣ 3 1 −2 3 0 ⎦ ∼ ⎣ 0 10 −5 15 0 ⎦ 5 −1 −2 1 0 5 −1 −2 1 0 0 14 −7 21 0 ⎡ ⎤ ⎡ ⎤ ⎡ 1 −3 1 −4 0 1 −3 1 −4 0 1 −3 1 4 5 2 −1 3 0 ⎦∼⎣ 0 2 −1 3 0 ⎦∼⎣ 0 1 −1/2 ∼⎣ 0 0 2 −1 3 0 0 0 0 0 0 0 0 0 3 1. A21 (−1) 2. A12 (−3), A13 (−5) 4. A23 (−1) ⎤ −4 0 3/2 0 ⎦ . 0 0 3. M2 (1/5), M3 (1/7) 5. M2 (1/2) From the last augmented matrix above we have that x2 − 12 x3 + 32 x4 = 0 and x1 −3x2 +x3 −4x4 = 0. Since x3 and x4 are free variables, we can set x3 = 2s and x4 = 2t, where s and t are real numbers. Then x2 = s − 3t and x1 = s − t. It follows that the solution set of the given system is {(s − t, s − 3t, 2s, 2t) : s, t ∈ R}. 46. Reduce the augmented matrix of the system: ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ 1 1 1 −1 0 1 1 1 −1 0 2 1 −1 1 0 ⎢ 1 1 1 −1 0 ⎥ 1 −1 1 0 ⎥ 3 0 ⎥ 1 ⎢ 2 ⎢ ⎢ 2 ⎢ 0 −1 −3 ⎥∼ ⎥∼ ⎥ ⎢ ⎣ ⎣ ⎦ ⎦ ⎣ 3 −1 1 −2 0 3 −1 1 −2 0 0 −4 −2 1 0 ⎦ 4 2 −1 1 0 4 2 −1 1 0 0 −2 −5 5 0 ⎡ 1 0 3 ⎢ ∼⎢ ⎣ 0 0 ⎡ 1 0 −2 ⎢ 0 1 3 6 ∼⎢ ⎣ 0 0 1 0 0 10 ⎤ ⎡ 1 1 −1 0 1 0 −2 2 ⎢ 0 1 1 3 −3 0 ⎥ 3 −3 4 ⎥∼⎢ −4 −2 1 0 ⎦ ⎣ 0 0 10 −11 −2 −5 5 0 0 0 −3 3 ⎤ ⎡ ⎤ ⎡ 2 0 1 0 0 0 0 1 ⎢ ⎢ ⎥ −3 0 ⎥ 0 1 0 0 0 7 ⎢ 8 ⎢ 0 ⎥∼ ⎥∼ −1 0 ⎦ ⎣ 0 0 1 −1 0 ⎦ ⎣ 0 −11 0 0 0 0 −1 0 0 1. P12 2. A12 (−2), A13 (−3), A14 (−4) 5. P34 6. M3 (−1/3) ⎤ ⎡ ⎤ 1 0 −2 2 0 0 3 −3 0 ⎥ 0 ⎥ 5 ⎢ ⎢ 0 1 ⎥∼ ⎥ 3 0 ⎦ 0 ⎦ ⎣ 0 0 −3 0 0 10 −11 0 0 ⎤ ⎡ 0 0 0 0 1 0 0 0 ⎢ 1 0 0 0 ⎥ 9 ⎢ 0 1 0 0 ⎥∼ 0 1 −1 0 ⎦ ⎣ 0 0 1 0 0 0 1 0 0 0 0 1 3. M2 (−1) ⎤ 0 0 ⎥ ⎥. 0 ⎦ 0 4. A21 (−1), A23 (4), A24 (2) 7. A31 (2), A32 (−3), A34 (−10) 8. M4 (−1) 9. A43 (1) From the last augmented matrix, it follows that the solution set to the system is given by {(0, 0, 0, 0)}. 47. The equation Ax = 0 is 2 −1 3 4 x1 x2 Reduce the augmented matrix of the system: 2 −1 0 1 1 − 12 0 2 1 − 12 ∼ ∼ 3 4 0 3 4 0 0 11 2 1. M1 (1/2) 2. A12 (−3) = 0 0 0 0 3 . ∼ 1 − 12 0 1 3. M2 (2/11) (c)2017 Pearson Education. Inc. 0 0 4 ∼ 4. A21 (1/2) 1 0 0 0 1 0 . 165 From the last augmented matrix, we see that x1 = x2 = 0. Hence, the solution set is {(0, 0)}. 48. The equation Ax = 0 is 1−i 1+i 2i −2 Reduce the augmented matrix of the system: 1 − i 2i 0 1 1 ∼ 1 + i −2 0 1+i x1 x2 0 0 0 0 = −1 + i −2 1. M1 ( 1+i 2 ) . 2 −1 + i 0 1 0 ∼ 0 0 . 2. A12 (−1 − i) It follows that x1 + (−1 + i)x2 = 0. Since x2 is a free variable, we can let x2 = t, where t is a complex number. The solution set to the system is then given by {(t(1 − i), t) : t ∈ C}. 49. The equation Ax = 0 is 1+i −1 + i 1 − 2i 2+i Reduce the augmented matrix of the system: 1 + i 1 − 2i 0 1 1 ∼ −1 + i 2 + i 0 −1 + i 1. M1 ( 1−i 2 ) x1 x2 = − 1+3i 2 2+i 0 0 0 0 2 . ∼ 1 − 1+3i 2 0 0 0 0 . 2. A12 (1 − i) It follows that x1 − 1+3i 2 x2 = 0. Since x2 is a free variable, we can let x2 = r, where r is any complex number. Thus, the solution set to the given system is {( 1+3i 2 r, r) : r ∈ C}. 50. The equation Ax = 0 is ⎡ ⎤⎡ ⎤ ⎡ ⎤ 1 2 3 x1 0 ⎣ 2 −1 0 ⎦ ⎣ x2 ⎦ = ⎣ 0 ⎦ . 1 1 1 x3 0 Reduce the augmented matrix of the system: ⎤ ⎡ ⎤ ⎡ ⎡ 1 2 3 0 1 1 2 3 0 1 2 ⎣ 0 ⎣ 2 −1 0 0 ⎦ ∼ ⎣ 0 −5 −6 0 ⎦ ∼ 1 1 1 0 0 −1 −2 0 0 ⎡ ⎤ ⎡ 1 0 −1 0 1 0 −1 4 5 2 0 ⎦∼⎣ 0 1 2 ∼⎣ 0 1 0 0 4 0 0 0 1 1. A12 (−2), A13 (−1) 2. P23 3. M2 (−1) ⎤ ⎡ ⎤ 2 3 0 1 2 3 0 3 −1 −2 0 ⎦ ∼ ⎣ 0 1 2 0 ⎦ −5 −6 0 0 −5 −6 0 ⎤ ⎡ ⎤ 0 1 0 0 0 6 0 ⎦ ∼ ⎣ 0 1 0 0 ⎦. 0 0 1 0 0 4. A21 (−2), A23 (5) 5. M3 (1/4) 6. A31 (1), A32 (−2) From the last augmented matrix, we see that the only solution to the given system is x1 = x2 = x3 = 0: {(0, 0, 0)}. 51. The equation Ax = 0 is ⎡ 1 ⎣ −1 1 ⎤ ⎤ ⎡ ⎡ ⎤ x 0 1 1 −1 ⎢ 1 ⎥ ⎢ x2 ⎥ ⎢ 0 ⎥ ⎥ 0 −1 2 ⎦⎢ ⎣ x3 ⎦ = ⎣ 0 ⎦ . 3 2 2 0 x4 (c)2017 Pearson Education. Inc. 166 Reduce the augmented matrix of the system: ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 1 1 1 −1 0 1 1 1 −1 0 1 0 1 −2 0 1 0 0 −3 0 1 2 3 ⎣ −1 0 −1 2 0 ⎦∼⎣ 0 1 0 1 0 ⎦∼⎣ 0 1 0 1 0 ⎦∼⎣ 0 1 0 1 0 ⎦. 1 3 2 2 0 0 2 1 3 0 0 0 1 1 0 0 0 1 1 0 1. A12 (1), A13 (−1) 2. A21 (−1), A23 (−2) 3. A31 (−1) From the last augmented matrix, we see that x4 is a free variable. We set x4 = t, where t is a real number. The last row of the reduced row echelon form above corresponds to the equation x3 + x4 = 0. Therefore, x3 = −t. The second row corresponds to the equation x2 + x4 = 0, so we likewise find that x2 = −t. Finally, from the first equation we have x1 − 3x4 = 0, so that x1 = 3t. Consequently, the solution set of the original system is given by {(3t, −t, −t, t) : t ∈ R}. 52. The equation Ax = 0 is ⎡ 2 − 3i ⎣ 3 + 2i 5−i ⎤⎡ ⎤ ⎡ ⎤ x1 i−1 0 −1 − i ⎦ ⎣ x2 ⎦ = ⎣ 0 ⎦ . −2 x3 0 1+i −1 + i 2i Reduce the augmented matrix of this system: ⎤ ⎡ ⎡ 2 − 3i 1 + i i−1 0 1 1 ⎣ 3 + 2i ⎣ 3 + 2i −1 + i −1 − i 0 ⎦ ∼ 5−i 2i −2 0 5−i 1. M1 ( 2+3i 13 ) −1+5i 13 −1 + i 2i −5−i 13 −1 − i −2 ⎤ ⎡ 0 1 2 0 ⎦∼⎣ 0 0 0 −1+5i 13 0 0 −5−i 13 0 0 ⎤ 0 0 ⎦. 0 2. A12 (−3 − 2i), A13 (−5 + i) −5−i From the last augmented matrix, we see that x1 + −1+5i 13 x2 + 13 x3 = 0. Since x2 and x3 are free variables, we can let x2 = 13r and x3 = 13s, where r and s are complex numbers. It follows that the solution set of the system is {(r(1 − 5i) + s(5 + i), 13r, 13s) : r, s ∈ C}. 53. The equation Ax = 0 is ⎡ ⎤⎡ ⎤ ⎡ ⎤ 1 3 0 x1 0 ⎣ −2 −3 0 ⎦ ⎣ x2 ⎦ = ⎣ 0 ⎦ . 1 4 0 x3 0 Reduce the augmented matrix of the system: ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ 1 3 0 0 1 3 0 0 1 0 0 0 1 3 0 0 1 2 3 ⎣ 0 3 0 0 ⎦∼ ⎣ 0 1 0 0 ⎦∼ ⎣ 0 1 0 0 ⎦. ⎣ −2 −3 0 0 ⎦ ∼ 1 4 0 0 0 1 0 0 0 3 0 0 0 0 0 0 1. A12 (2), A13 (−1) 2. P23 3. A21 (−3), A23 (−3) From the last augmented matrix we see that the solution set of the system is {(0, 0, t) : t ∈ R}. 54. The equation Ax = 0 is ⎤ 1 0 3 ⎡ ⎤ ⎡ ⎤ ⎢ 3 −1 7 ⎥ 0 ⎥ x1 ⎢ ⎢ 2 ⎣ ⎦ ⎣ ⎦ 1 8 ⎥ ⎥ x2 = 0 . ⎢ ⎦ ⎣ 1 1 5 x3 0 −1 1 −1 ⎡ (c)2017 Pearson Education. Inc. 167 Reduce the augmented matrix of the system: ⎡ ⎤ ⎡ ⎤ ⎡ 1 0 3 0 1 1 0 3 0 ⎢ ⎢ 3 −1 ⎥ ⎢ 7 0 ⎥ ⎢ ⎥ 1 ⎢ 0 −1 −2 0 ⎥ 2 ⎢ 0 ⎢ ⎢ ⎢ 2 1 8 0 ⎥ 1 2 0 ⎥ ⎢ ⎥∼⎢ 0 ⎥∼⎢ 0 ⎣ ⎣ 0 ⎣ 1 ⎦ ⎦ 1 5 0 0 1 2 0 −1 1 −1 0 0 1 2 0 0 1. A12 (−3), A13 (−2), A14 (−1), A15 (1) 0 1 1 1 1 2. M2 (−1) 3 2 2 2 2 ⎤ ⎡ ⎤ 1 0 3 0 0 ⎢ ⎥ 0 ⎥ ⎥ 3 ⎢ 0 1 2 0 ⎥ ⎢ ⎥ 0 ⎥∼⎢ 0 0 0 0 ⎥ ⎥. 0 ⎦ ⎣ 0 0 0 0 ⎦ 0 0 0 0 0 3. A23 (−1), A24 (−1), A25 (−1) From the last augmented matrix, we obtain the equations x1 + 3x3 = 0 and x2 + 2x3 = 0. Since x3 is a free variable, we let x3 = t, where t is a real number. The solution set for the given system is then given by {(−3t, −2t, t) : t ∈ R}. 55. The equation Ax = 0 is ⎡ ⎤ ⎡ ⎤ x1 1 −1 0 1 ⎢ 0 ⎥ x ⎣ 3 −2 0 5 ⎦ ⎢ 2 ⎥ = ⎣ 0 ⎦ . ⎣ x3 ⎦ −1 2 0 1 0 x4 ⎡ ⎤ Reduce the augmented matrix of the system: ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ 1 −1 0 1 0 1 0 0 3 0 1 −1 0 1 0 1 2 ⎣ 0 ⎣ 3 −2 0 5 0 ⎦ ∼ 1 0 2 0 ⎦ ∼ ⎣ 0 1 0 2 0 ⎦. −1 2 0 1 0 0 1 0 2 0 0 0 0 0 0 1. A12 (−3), A13 (1) 2. A21 (1), A23 (−1) From the last augmented matrix we obtain the equations x1 + 3x4 = 0 and x2 + 2x4 = 0. Because x3 and x4 are free, we let x3 = t and x4 = s, where s and t are real numbers. It follows that the solution set of the system is {(−3s, −2s, t, s) : s, t ∈ R}. 56. The equation Ax = 0 is ⎡ ⎤ ⎤ x ⎡ ⎤ 1 0 −3 0 ⎢ 1 ⎥ 0 ⎣ 3 0 −9 0 ⎦ ⎢ x2 ⎥ = ⎣ 0 ⎦ . ⎣ x3 ⎦ −2 0 6 0 0 x4 ⎡ Reduce the augmented matrix of the system: ⎤ ⎡ ⎤ ⎡ 1 0 −3 0 0 1 0 −3 0 0 1 ⎣ 0 0 ⎣ 3 0 −9 0 0 ⎦ ∼ 0 0 0 ⎦. −2 0 6 0 0 0 0 0 0 0 1. A12 (−3), A13 (2) From the last augmented matrix we obtain x1 − 3x3 = 0. Therefore, x2 , x3 , and x4 are free variables, so we let x2 = r, x3 = s, and x4 = t, where r, s, t are real numbers. The solution set of the given system is therefore {(3s, r, s, t) : r, s, t ∈ R}. (c)2017 Pearson Education. Inc. 168 57. The equation Ax = 0 is ⎡ 2+i ⎣ i 3−i Reduce the augmented matrix of the system: ⎡ ⎤ ⎡ 2+i i 3 − 2i 0 i 1 ⎣ i 1 − i 4 + 3i 0 ⎦ ∼ ⎣ 2 + i 3 − i 1 + i 1 + 5i 0 3−i ⎡ 1 −1 − i 3 ∼ ⎣ 0 1 + 4i 0 5 + 3i 1. P12 2. M1 (−i) ⎤⎡ ⎤ ⎡ ⎤ 3 − 2i x1 0 4 + 3i ⎦ ⎣ x2 ⎦ = ⎣ 0 ⎦ . 1 + 5i 0 x3 i 1−i 1+i 1−i i 1+i 4 + 3i 3 − 2i 1 + 5i ⎤ ⎡ 0 1 2 0 ⎦∼⎣ 2+i 3−i 0 −1 − i i 1+i ⎤ ⎡ ⎤ ⎡ 0 1 −1 − i 3 − 4i 0 3 − 4i 1 0 4 5 5+31i ⎦∼ ⎣ 0 1 −7 + 3i 0 ⎦ ∼ ⎣ 0 0 1 17 −4 + 20i 0 0 5 + 3i −4 + 20i 0 0 0 ⎤ ⎡ ⎡ ⎤ 0 1 0 25−32i 1 0 0 0 17 6 7 ⎦∼ ⎣ 0 1 0 0 ⎦. ∼ ⎣ 0 1 5+31i 0 17 0 0 1 0 0 0 1 0 3. A12 (−2 − i), A13 (−3 + i) 6. M3 (−i/10) 4. M2 ( 1−4i 17 ) 3 − 3i 3 − 2i 1 + 5i 25−32i 17 5+31i 17 10i ⎤ 0 0 ⎦ 0 ⎤ 0 0 ⎦ 0 5. A21 (1 + i), A23 (−5 − 3i) 7. A31 ( −25+32i ), A32 ( −5−31i ) 17 17 From the last augmented matrix above, we see that the only solution to this system is the trivial solution. Solutions to Section 2.6 True-False Review: (a): FALSE. An invertible matrix is also known as a nonsingular matrix. 1 1 (b): FALSE. For instance, the matrix does not contain a row of zeros, but fails to be invertible. 2 2 (c): TRUE. If A is invertible, then the unique solution to Ax = b is x = A−1 b. ⎤ ⎡ 1 0 1 0 0 (d): FALSE. For instance, if A = and B = ⎣ 0 0 ⎦, then AB = I2 , but A is not even a 0 0 1 0 1 square matrix, hence certainly not invertible. (e): FALSE. For instance, if A = In and B = −In , then A and B are both invertible, but A + B = 0n is not invertible. (f ): TRUE. We have (AB)B −1 A−1 = In and B −1 A−1 (AB) = In , and therefore, AB is invertible, with inverse B −1 A−1 . (g): TRUE. From A2 = A, we subtract to obtain A(A−I) = 0. Left multiplying both sides of this equation by A−1 (since A is invertible, A−1 exists), we have A − I = A−1 0 = 0. Therefore, A = I, the identity matrix. (h): TRUE. From AB = AC, we left-multiply both sides by A−1 (since A is invertible, A−1 exists) to obtain A−1 AB = A−1 AC. Since A−1 A = I, we obtain IB = IC, or B = C. (c)2017 Pearson Education. Inc. 169 (i): TRUE. Any 5 × 5 invertible matrix must have rank 5, not rank 4 (Theorem 2.6.6). (j): TRUE. Any 6 × 6 matrix of rank 6 is invertible (Theorem 2.6.6). Problems: 1. We have −1 = AA 2. We have −1 AA = 4 3 9 7 2 −1 3 −1 3. We have a c b d 7 −9 −3 4 −1 −3 1 2 = (4)(−9) + (9)(4) (3)(−9) + (7)(4) (2)(−1) + (−1)(−3) (3)(−1) + (−1)(−3) = 1 ad − bc (4)(7) + (9)(−3) (3)(7) + (7)(−3) d −c −b a (2)(1) + (−1)(2) (3)(1) + (−1)(2) = 1 0 = 0 1 1 0 = I2 . 0 1 = I2 . 1 a b d −b = −c a ad − bc c d 1 ad − bc 0 = 0 ad − bc ad − bc 1 0 = 0 1 = I2 , and 1 ad − bc d −c −b a a c b d 1 d −b a b a c d ad − bc −c 1 ad − bc 0 = 0 ad − bc ad − bc 1 0 = 0 1 = = I2 . 4. We have ⎡ ⎤⎡ ⎤ 3 5 1 8 −29 3 19 −2 ⎦ AA−1 = ⎣ 1 2 1 ⎦ ⎣ −5 2 6 7 2 −8 1 ⎡ (3)(8) + (5)(−5) + (1)(2) (3)(−29) + (5)(19) + (1)(−8) = ⎣ (1)(8) + (2)(−5) + (1)(2) (1)(−29) + (2)(19) + (1)(−8) (2)(8) + (6)(−5) + (7)(2) (2)(−29) + (6)(19) + (7)(−8) ⎡ ⎤ 1 0 0 = ⎣ 0 1 0 ⎦ = I3 . 0 0 1 5. We have [A|I2 ] = 1 2 1 1 3 0 0 1 1 ∼ 1 0 1 2 0 1 −1 1 2 ∼ ⎤ (3)(3) + (5)(−2) + (1)(1) (1)(3) + (2)(−2) + (1)(1) ⎦ (2)(3) + (6)(−2) + (7)(1) 3 −2 1 0 0 1 −1 1 (c)2017 Pearson Education. Inc. = [I2 |A−1 ]. 170 Therefore, A−1 = 3 −2 −1 1 1. A12 (−1) . 2. A21 (−2) 6. We have [A|I2 ] = 1 1−i 1+i 1 1 0 3 0 1 ∼ 1 1 1+i 0 −1 ∼ 1 0 −1 0 1 1−i Thus, A −1 = 1. A12 (−1 + i) 1 −1 + i 1+i −1 −1 1−i 0 1 2 ∼ 1 1−i 1 1+i 0 1 0 −1 = [I2 |A−1 ]. 1+i −1 . 3. A21 (−1 − i) 2. M2 (−1) 7. We have [A|I2 ] = 1 −i i−1 2 1 0 3 ∼ 1 0 −i 1−i 1 0 1+i 0 1 1 −1+i 2 1+i 2 0 1 1 ∼ Thus, A−1 = 1. A12 (1 − i) 1+i 1 1 1−i 0 1 2 ∼ 1 −i 0 1 1 1 0 1+i 2 = [I2 |A−1 ]. −1+i 2 1+i 2 . 2. M2 (1/(1 − i)) 3. A21 (i) 8. Note that AB = 02 for all 2 × 2 matrices B. Therefore, A is not invertible. 9. We have ⎡ 1 −1 2 1 1 11 0 [A|I3 ] = ⎣ 2 4 −3 10 0 0 1 0 ⎤ ⎡ ⎤ ⎡ ⎤ 0 1 0 0 1 0 0 1 −1 2 1 −1 2 1 2 3 7 −2 1 0 ⎦ ∼ ⎣ 0 1 2 −4 0 1 ⎦ 0 ⎦∼⎣ 0 0 1 2 −4 0 1 0 3 7 −2 1 0 1 ⎡ ⎤ ⎡ ⎤ 1 1 0 4 −3 0 1 0 0 −43 −4 13 3 4 1 ⎦ ∼ ⎣ 0 1 0 −24 −2 7 ⎦ = [I3 |A−1 ]. ∼ ⎣ 0 1 2 −4 0 0 0 1 10 1 −3 0 0 1 10 1 −3 Thus, ⎡ ⎤ −43 −4 13 7 ⎦. A−1 = ⎣ −24 −2 10 1 −3 (c)2017 Pearson Education. Inc. 171 1. A12 (−2), A13 (−4) 2. P23 3. A21 (1), A23 (−3) 4. A31 (−4), A32 (−2) 10. We have ⎡ 3 5 1 1 [A|I3 ] = ⎣ 1 2 1 0 2 6 7 0 ⎤ ⎡ 0 1 2 1 0 1 0 ⎦∼⎣ 3 5 1 1 2 6 7 0 1 0 1 0 1 0 0 ⎤ ⎡ 0 1 2 1 0 2 0 ⎦ ∼ ⎣ 0 −1 −2 1 0 2 5 0 1 ⎤ 1 0 −3 0 ⎦ −2 1 ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 0 1 0 2 −5 0 8 −29 3 1 2 1 1 0 −3 1 0 0 3 4 5 2 −1 3 0 ⎦∼⎣ 0 1 3 0 ⎦ ∼ ⎣ 0 1 0 −5 19 −2 ⎦ = [I3 |A−1 ]. ∼ ⎣ 0 1 2 −1 0 2 5 0 0 1 0 0 1 0 −2 1 2 −8 1 2 −8 1 Thus, ⎡ ⎤ 8 −29 3 19 −2 ⎦ . A−1 = ⎣ −5 2 −8 1 1. P12 2. A12 (−3), A13 (−2) 3. M2 (−1) 4. A21 (−2), A23 (−2) 5. A31 (3), A32 (−2) 11. This matrix is not invertible, because the column of zeros guarantees that the rank of the matrix is less than three. 12. We have ⎡ 4 2 −13 1 [A|I3 ] = ⎣ 2 1 −7 0 3 2 4 0 ⎡ 1 3 ∼⎣ 0 0 1 11 0 −1 −29 0 −2 −57 1 ⎤ ⎡ 3 2 4 0 0 1 −7 0 0 ⎦∼⎣ 2 1 4 2 −13 1 1 0 1 0 0 1 0 ⎤ ⎡ 1 1 11 0 1 2 −7 0 0 ⎦∼⎣ 2 1 4 2 −13 1 0 ⎤ −1 1 1 0 ⎦ 0 0 ⎤ ⎡ ⎤ ⎡ ⎤ −1 1 1 2 −1 1 1 11 0 −1 1 0 −18 0 4 5 1 29 0 −3 29 0 −3 3 −2 ⎦ ∼ ⎣ 0 2 ⎦∼⎣ 0 1 2 ⎦ 0 −2 −57 1 0 0 1 1 −2 4 −4 4 −4 0 ⎡ ⎤ 1 0 18 −34 −1 6 ∼ ⎣ 0 1 0 −29 55 2 ⎦ = [I3 |A−1 ]. 0 0 1 1 −2 0 Thus, ⎡ ⎤ 18 −34 −1 55 2 ⎦. A−1 = ⎣ −29 1 −2 0 1. P13 2. A21 (−1) 3. A12 (−2), A13 (−4) 5. A21 (−1), A23 (2) 4. M2 (−1) 6. A31 (18), A32 (−29) 13. We have ⎡ 1 2 −3 1 [A|I3 ] = ⎣ 2 6 −2 0 −1 1 4 0 0 1 0 ⎤ ⎡ ⎤ ⎡ 1 1 2 −3 0 1 0 0 1 2 −3 1 2 4 −2 1 0 ⎦ ∼ ⎣ 0 1 0 ⎦∼⎣ 0 2 2 −1 0 3 1 1 1 0 1 0 3 1 1 (c)2017 Pearson Education. Inc. 0 1 2 0 ⎤ 0 0 ⎦ 1 172 ⎤ ⎡ ⎤ 3 −1 0 3 −1 0 1 0 −7 1 0 −7 4 1 1 0 ⎦∼⎣ 0 1 0 ⎦ 2 −1 2 −1 ∼⎣ 0 1 2 2 3 3 4 1 4 −2 1 0 0 −5 0 0 1 − 5 10 − 5 ⎤ ⎡ 11 13 7 −5 1 0 0 −5 10 5 2 ⎦ 3 1 − ∼⎣ 0 1 0 = [I3 |A−1 ]. 5 10 5 3 4 1 0 0 1 −5 −5 10 ⎡ 3 Thus, ⎡ A−1 = ⎣ 1. A12 (−2), A13 (1) 14. We have ⎡ 2. M2 ( 12 ) − 13 5 3 5 4 −5 11 10 1 − 10 3 10 − 75 2 5 − 15 ⎤ ⎦. 4. M3 (− 15 ) 3. A21 (−2), A23 (−3) 5. A31 (7), A32 (−2) ⎤ ⎡ ⎤ ⎡ i 2 1 0 0 1 0 0 1 1 i 2 1 i 2 1 2 −1 2i 0 1 0 ⎦ ∼ ⎣ 0 −i −2 −1 − i 1 0 ⎦ ∼ ⎣ 0 1 −2i 1 − i 2i 5 0 0 1 0 0 1 0 0 1 −2 −2 0 1 ⎡ ⎤ ⎡ ⎤ −i 1 0 1 0 0 −i 1 0 1 0 0 3 4 ∼ ⎣ 0 1 −2i 1 − i i 0 ⎦ ∼ ⎣ 0 1 0 1 − 5i i 2i ⎦ = [I3 |A−1 ]. 0 0 1 −2 0 1 0 0 1 −2 0 1 1 [A|I3 ] = ⎣ 1 + i 2 Thus, ⎡ −i A−1 = ⎣ 1 − 5i −2 1. A12 (−1 − i), A13 (−2) 15. We have ⎡ 2 1 3 1 [A|I3 ] = ⎣ 1 −1 2 0 3 3 4 0 0 1 0 ⎤ 0 0 i 0 ⎦ 0 1 ⎤ 1 0 i 2i ⎦ . 0 1 2. M2 (i) 3. A21 (−i) 4. A32 (2i) ⎤ ⎡ ⎤ ⎡ 0 1 −1 2 0 1 0 1 −1 2 0 1 2 1 3 1 0 0 ⎦∼⎣ 0 3 −1 1 0 ⎦∼⎣ 2 3 3 4 0 0 1 0 6 −2 0 1 ⎡ ⎤ 0 1 0 1 −1 2 3 3 −1 1 −2 0 ⎦ ∼⎣ 0 0 0 0 −2 1 1 ⎤ 1 0 −2 0 ⎦ −3 1 Since 2 = rank(A) < rank(A# ) = 3, we know that A−1 does not exist (we have obtained a row of zeros in the block matrix on the left. 1. P12 16. We have ⎡ 1 −1 2 3 ⎢ 2 0 3 −4 [A|I4 ] = ⎢ ⎣ 3 −1 7 8 1 0 3 5 2. A12 (−2), A13 (−3) 1 0 0 0 0 1 0 0 0 0 1 0 3. A23 (−2) ⎤ ⎡ ⎤ 0 1 0 0 0 1 −1 2 3 2 −1 −10 −2 1 0 0 ⎥ 0 ⎥ 1 ⎢ ⎢ 0 ⎥∼ ⎥ ⎣ ⎦ 0 2 1 −1 −3 0 1 0 ⎦ 0 0 1 1 2 −1 0 0 1 1 (c)2017 Pearson Education. Inc. 173 ⎡ ⎤ ⎡ ⎤ 1 0 0 0 0 0 0 1 1 −1 2 3 1 0 3 5 1 1 2 −1 0 0 1 ⎥ 1 2 −1 0 0 1 ⎥ 2 ⎢ 0 3 ⎢ ⎢ 0 1 ⎥∼ ⎥ ∼⎢ ⎣ 0 ⎣ ⎦ 2 1 −1 −3 0 1 0 0 0 −1 −5 −1 0 1 −2 ⎦ 0 2 −1 −10 −2 1 0 0 0 0 −3 −14 0 1 0 −2 ⎡ ⎤ ⎡ ⎤ 1 0 3 5 1 0 0 −10 −3 0 0 0 0 1 3 −5 1 2 −1 0 0 1 ⎥ 1 −1 ⎥ 4 ⎢ 0 1 5 ⎢ ⎢ 0 1 0 −3 −2 0 ⎥∼ ⎥ ∼⎢ ⎣ 0 0 ⎣ ⎦ 1 5 0 0 1 5 1 0 −1 2 1 0 −1 2 ⎦ 0 0 −3 −14 0 0 0 1 0 1 0 −2 3 1 −3 4 ⎡ ⎤ 1 0 0 0 27 10 −27 35 ⎥ 0 1 0 0 7 3 −8 11 6 ⎢ ⎥ = [I4 |A−1 ]. ∼⎢ ⎣ 0 0 1 0 −14 −5 14 −18 ⎦ 0 0 0 1 3 1 −3 4 Thus, ⎤ 27 10 −27 35 ⎢ 7 3 −8 11 ⎥ ⎥. A−1 = ⎢ ⎣ −14 −5 14 −18 ⎦ 3 1 −3 4 ⎡ 1. A12 (−2), A13 (−3), A14 (−1) 4. M3 (−1) 2. P13 5. A31 (−3), A32 (−1), A34 (3) 3. A21 (1), A23 (−2), A24 (−2) 6. A41 (10), A42 (3), A43 (5) 17. We have ⎡ 0 −2 −1 −3 1 ⎢ 2 0 2 1 0 [A|I4 ] = ⎢ ⎣ 1 −2 0 2 0 3 −1 −2 0 0 0 1 0 0 0 0 1 0 ⎤ ⎡ 0 1 −2 ⎢ 2 0 0 ⎥ 1 ⎥∼⎢ 0 ⎦ ⎣ 0 −2 3 −1 1 ⎡ 0 2 0 2 1 0 −1 −3 1 −2 0 0 0 1 0 0 1 0 0 0 ⎤ 0 0 ⎥ ⎥ 0 ⎦ 1 ⎤ ⎡ ⎤ 1 0 1 −2 0 2 0 0 1 0 1 −2 0 2 0 0 1 4 2 −3 0 1 −2 0 ⎥ − 34 0 14 − 12 0 ⎥ 1 2 ⎢ 0 3 ⎢ 2 ⎢ 0 ⎥∼ ⎥ ∼⎢ ⎣ 0 −2 −1 −3 1 0 0 0 ⎦ ⎣ 0 −2 −1 −3 1 0 0 0 ⎦ 0 5 −2 −6 0 0 −3 1 0 5 −2 −6 0 0 −3 1 ⎡ ⎡ ⎤ ⎤ 1 1 1 1 1 0 1 1 0 1 0 0 0 0 0 0 2 2 2 2 1 1 1 1 − 34 0 − 12 0 ⎥ − 34 0 − 12 0 ⎥ 4 ⎢ 0 1 5 ⎢ 2 4 2 4 ⎢ 0 1 ⎥∼ ⎥ ∼⎢ 9 1 9 9 5 ⎣ 0 0 −1 0 ⎦ ⎣ 0 0 − 2 − 4 0 − 4 − 12 1 ⎦ 0 −2 1 2 1 −1 0 0 0 − 92 − 94 0 − 54 − 12 1 0 0 0 − 92 1 2 ⎡ ⎤ ⎡ ⎤ 2 1 1 0 1 1 0 0 0 0 29 − 19 0 12 0 0 2 9 1 ⎥ 1 ⎢ − 34 0 14 − 12 0 ⎥ 0 19 − 59 6 ⎢ 0 1 2 9 ⎥ ⎥ 7 ⎢ 0 1 0 −1 ∼⎢ 1 5 2 ⎦∼⎣ 1 5 2 ⎦ 1 1 ⎣ 0 0 1 0 18 −9 0 0 1 0 18 −9 2 9 2 9 0 0 0 − 92 1 12 −1 0 0 0 0 − 92 1 12 −1 0 ⎡ ⎡ ⎤ ⎤ 2 2 2 2 1 0 0 0 1 0 0 0 0 − 19 0 − 19 9 9 9 9 1 ⎥ 1 ⎥ 1 5 2 1 ⎢ ⎢ 0 1 0 −1 0 −9 0 −3 8 9 ⎢ 0 1 0 0 −9 9 9 ⎥∼ 9 ⎥ = [I |A−1 ]. ∼⎢ 4 1 1 1 5 2 ⎦ 1 2 ⎦ ⎣ 0 0 1 ⎣ 0 18 −9 0 0 1 0 0 −9 2 9 9 3 2 2 2 1 2 1 0 0 0 1 −9 −9 0 0 0 0 1 −9 −9 0 9 9 (c)2017 Pearson Education. Inc. 174 Thus, ⎡ 0 ⎢ −2 9 A−1 = ⎢ ⎣ 1 9 − 29 1. P13 5. P34 2. A12 (−2), A14 (−3) 6. M3 (− 29 ) 2 9 − 19 0 − 13 1 0 3 2 9 − 19 3. M2 ( 14 ) 2 9 1 9 2 −9 ⎤ ⎥ ⎥. ⎦ 0 4. A21 (2), A23 (2), A24 (−5) 7. A31 (−1), A32 (− 12 ) 8. M4 (− 29 ) 9. A42 (1), A43 (− 12 ) 18. We have ⎡ 1 ⎢ 3 =⎢ ⎣ 0 0 ⎡ 1 0 2 ⎢ ⎢ ∼⎣ 0 0 ⎡ 1 0 4 ⎢ ⎢ ∼⎣ 0 0 ⎤ ⎡ 1 2 0 0 0 ⎢ 0 −2 0 0 0 ⎥ 1 ⎥∼⎢ 0 ⎦ ⎣ 0 0 1 65 1 0 0 7 8 ⎤ ⎡ 2 0 0 1 0 0 0 1 0 −2 0 0 −3 1 0 0 ⎥ 0 −2 3 ⎢ ⎢ ⎥ ∼ 6 1 0 0 0 0 ⎦ ⎣ 0 0 1 5 5 0 0 0 0 − 75 1 0 0 − 25 ⎤ 0 0 0 −2 1 0 0 3 − 12 0 0 ⎥ 1 0 0 2 ⎥ = [I4 |A−1 ]. 0 1 0 0 0 −4 −3 ⎦ 7 0 0 − 52 0 0 1 2 2 0 0 1 4 0 0 0 0 5 6 0 0 7 8 0 Thus, 0 1 0 0 0 0 1 0 ⎡ ⎢ A−1 = ⎢ ⎣ 1. A12 (−3), M3 ( 15 ) −2 3 2 1 − 12 0 0 0 0 2. A34 (−7) 0 0 −4 7 2 1 0 −3 1 0 0 0 0 0 0 1 5 0 ⎤ 0 0 ⎥ ⎥ 0 ⎦ 1 ⎤ 0 0 0 0 −2 1 0 0 −3 1 0 0 ⎥ ⎥ 1 0 0 0 −4 3 ⎦ 0 0 − 75 1 0 − 25 ⎤ 0 0 ⎥ ⎥. 3 ⎦ − 52 3. A21 (1), A13 (3) 4. M2 (− 12 ), M4 (− 52 ) −1 19. To⎡determine the third without determining the whole ⎡inverse, we solve the ⎤ ⎤ ⎡ column ⎤ ⎡vector ⎤ of A −1 −2 3 x 0 −1 −2 3 0 1 1 0 ⎦ 1 1 ⎦ ⎣ y ⎦ = ⎣ 0 ⎦. The corresponding augmented matrix ⎣ −1 system ⎣ −1 −1 −2 −1 1 −1 −2 −1 z 1 ⎤ ⎡ 0 1 2 −3 5 0 ⎦. Thus, back substitution yields z = − 14 , y = − 16 , and x = − 12 . can be row-reduced to ⎣ 0 1 − 23 1 0 0 1 −4 ⎡ ⎤ −5/12 Thus, the third column vector of A−1 is ⎣ −1/6 ⎦. −1/4 −1 20. To determine the second vector ⎤ ⎡ column ⎤ ⎡ ⎤ of A without determining the whole⎡ inverse, we solve ⎤ the ⎡ x 2 −1 4 0 2 −1 4 0 1 2 1 ⎦ can 1 2 ⎦ ⎣ y ⎦ = ⎣ 1 ⎦. The corresponding augmented matrix ⎣ 5 linear system ⎣ 5 1 −1 3 z 0 1 −1 3 0 (c)2017 Pearson Education. Inc. 175 ⎤ 0 −1 3 1 −2 0 ⎦. Thus, back-substitution yields z = −1, y = −2, and x = 1. Thus, 0 1 −1 ⎡ ⎤ 1 the second column vector of A−1 is ⎣ −2 ⎦. −1 7 6 20 −8 −10 2 . 21. We have A = , b = , and the Gauss-Jordan method yields A−1 = 2 7 2 −1 3 Therefore, we have 7 −8 −48 −10 −1 2 x=A b= = . 2 14 −1 3 ⎡ 1 be row-reduced to ⎣ 0 0 Hence, we have x1 = −48 and x2 = 14. 1 3 1 −5 3 22. We have A = , b = , and the Gauss-Jordan method yields A−1 = . 2 5 3 2 −1 Therefore, we have −5 3 1 4 −1 = . x=A b= 2 −1 3 −1 So we have x1 = 4 and x2 = −1. ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 7 5 −3 1 1 −2 −2 1 ⎦. 1 ⎦, b = ⎣ 3 ⎦, and the Gauss-Jordan method yields A−1 = ⎣ −2 −1 23. We have A = ⎣ 0 1 2 2 −1 2 4 −3 1 Therefore, we have ⎡ ⎤⎡ ⎤ ⎡ ⎤ 7 5 −3 −2 −2 1 ⎦⎣ 3 ⎦ = ⎣ 2 ⎦. x = A−1 b = ⎣ −2 −1 2 2 −1 1 1 Hence, we have x1 = −2, x2 = 2, and x3 = 1. 1 −2i 2 4i 1 24. We have A = ,b= , and the Gauss-Jordan method yields A−1 = 2+8i 2−i 4i −i −2 + i Therefore, we have −1 x=A 1 b= 2 + 8i 4i −2 + i 2i 1 2 −i 1 = 2 + 8i 2 + 8i −4 + i (c)2017 Pearson Education. Inc. . . Hence, we have x1 = 1 and x2 = −4+i 2+8i . ⎡ ⎤ ⎡ ⎡ ⎤ 3 4 5 −79 1 25. We have A = ⎣ 2 10 1 ⎦, b = ⎣ 1 ⎦, and the Gauss-Jordan method yields A−1 = ⎣ 12 4 1 8 38 1 Therefore, we have ⎡ ⎤⎡ ⎤ ⎡ ⎤ −79 27 46 1 −6 x = A−1 b = ⎣ 12 −4 −7 ⎦ ⎣ 1 ⎦ = ⎣ 1 ⎦ . 38 −13 −22 1 3 Hence, we have x1 = −6, x2 = 1, and x3 = 3. 2i 1 ⎤ 27 46 −4 −7 ⎦. −13 −22 176 ⎡ ⎤ ⎤ ⎡ ⎤ −1 3 5 1 2 12 1 ⎣ 3 3 −3 ⎦. 2 −1 ⎦, b = ⎣ 24 ⎦, and the Gauss-Jordan method yields A−1 = 12 5 −3 −1 −1 1 −36 ⎡ 1 26. We have A = ⎣ 1 2 Therefore, we have −1 x=A ⎡ ⎤⎡ ⎤ ⎡ ⎤ −1 3 5 12 −10 1 ⎣ 3 3 −3 ⎦ ⎣ 24 ⎦ = ⎣ 18 ⎦ . b= 12 5 −3 −1 −36 2 Hence, x1 = −10, x2 = 18, and x3 = 2. 27. We have 0 −1 1 0 0 −1 1 0 (0)(0) + (1)(1) (−1)(0) + (0)(1) (0)(−1) + (1)(0) (−1)(−1) + (0)(0) 1 0 0 1 √ 3/2 √1/2 3/2 √ −1/2 3/2 3/2 −1/2 1/2 √ √ √ √ 1 (1/2)(1/2) ( 3/2)(−1/2) + √ (1/2)( √3/2) ( 3/2)(√ 3/2) + √ = = 0 (−1/2)( 3/2) + ( 3/2)(1/2) (−1/2)(−1/2) + ( 3/2)( 3/2) 0 1 T AA = = = = I2 , so AT = A−1 . 28. We have √ AAT = = I2 , so AT = A−1 . 29. We have AAT = = cos α − sin α sin α cos α cos α sin α − sin α cos α cos2 α + sin2 α (− sin α)(cos α) + (cos α)(sin α) (cos α)(− sin α) + (sin α)(cos α) (− sin α)2 + cos2 α = 1 0 0 1 so AT = A−1 . 30. We have ⎡ ⎤ ⎤ 1 −2x 2x2 1 2x 2x2 1 ⎣ 2x 1 − 2x2 −2x ⎦ ⎣ −2x 1 − 2x2 2x ⎦ AAT = 1 + 2x2 2 2x 2x2 2x 1 −2x 1 ⎤ ⎡ 1 + 4x2 + 4x4 0 0 1 2 4 ⎦ = I3 , ⎣ 0 1 + 4x + 4x 0 = 1 + 4x2 + 4x4 2 4 0 0 1 + 4x + 4x 1 1 + 2x2 ⎡ so AT = A−1 . 31. For part 2, we have (B −1 A−1 )(AB) = B −1 (A−1 A)B = B −1 In B = B −1 B = In , and for part 3, we have (A−1 )T AT = (AA−1 )T = InT = In . (c)2017 Pearson Education. Inc. = I2 , 177 32. We prove this by induction on k, with k = 1 trivial and k = 2 proven in part 2 of Theorem 2.6.10. Assuming the statement is true for a product involving k − 1 matrices, we may proceed as follows: −1 (A1 A2 · · · Ak )−1 = ((A1 A2 · · · Ak−1 )Ak )−1 = A−1 k (A1 A2 · · · Ak−1 ) −1 −1 −1 −1 −1 −1 −1 = A−1 k (Ak−1 · · · A2 A1 ) = Ak Ak−1 · · · A2 A1 . In the second equality, we have applied part 2 of Theorem 2.6.10 to the two matrices A1 A2 · · · Ak−1 and Ak , and in the third equality, we have assumed that the desired property is true for products of k − 1 matrices. 33. Since A is skew-symmetric, we know that AT = −A. We wish to show that (A−1 )T = −A−1 . We have (A−1 )T = (AT )−1 = (−A)−1 = −(A−1 ), which shows that A−1 is skew-symmetric. The first equality follows from part 3 of Theorem 2.6.10, and the second equality results from the assumption that A−1 is skew-symmetric. 34. Since A is symmetric, we know that AT = A. We wish to show that (A−1 )T = A−1 . We have (A−1 )T = (AT )−1 = A−1 , which shows that A−1 is symmetric. The first equality follows from part 3 of Theorem 2.6.10, and the second equality results from the assumption that A is symmetric. 35. We have (In − A3 )(In + A3 + A6 + A9 ) = In (In + A3 + A6 + A9 ) − A3 (In + A3 + A6 + A9 ) = In + A3 + A6 + A9 − A3 − A6 − A9 − A12 = In − A12 = In , where the last equality uses the assumption that A12 = 0. This calculation shows that In − A3 and In + A3 + A6 + A9 are inverses of one another. 36. We have (In − A)(In + A + A2 + A3 ) = In (In + A + A2 + A3 ) − A(In + A + A2 + A3 ) = In + A + A2 + A3 − A − A2 − A3 − A4 = In − A4 = In , where the last equality uses the assumption that A4 = 0. This calculation shows that In − A and In + A + A2 + A3 are inverses of one another. 37. We claim that the inverse of A15 is B 9 . To verify this, use the fact that A5 B 3 = I to observe that A15 B 9 = A5 (A5 (A5 B 3 )B 3 )B 3 = A5 (A5 IB 3 )B 3 = A5 (A5 B 3 )B 3 = A5 IB 3 = A5 B 3 = I. This calculation shows that the inverse of A15 is B 9 . 38. We claim that the inverse of A9 is B −3 . To verify this, use the fact that A3 B −1 = I to observe that A9 B −3 = A3 (A3 (A3 B −1 )B −1 )B −1 = A3 (A3 IB −1 )B −1 = A3 (A3 B −1 )B −1 = A3 IB −1 = A3 B −1 = I. This calculation shows that the inverse of A9 is B −3 . 39. We have B = BIn = B(AC) = (BA)C = In C = C. 40. YES. Since BA = In , we know that A−1 = B (see Theorem 2.6.12). Likewise, since CA = In , A−1 = C. Since the inverse of A is unique, it must follow that B = C. (c)2017 Pearson Education. Inc. 178 41. We can simply compute 1 a22 −a12 a11 a11 a21 Δ −a21 Therefore, a12 a22 a11 a21 1 a22 a11 − a12 a21 a22 a12 − a12 a22 = Δ −a21 a11 + a11 a21 −a21 a12 + a11 a22 1 a11 a22 − a12 a21 0 1 = = 0 a11 a22 − a12 a21 0 Δ a12 a22 −1 1 = Δ a22 −a21 −a12 a11 0 1 = I2 . . 42. Assume that A is an invertible matrix and that Axi = bi for i = 1, 2, . . . , p (where each bi is given). Use elementary row operations on the augmented matrix of the system to obtain the equivalence [A|b1 b2 b3 . . . bp ] ∼ [In |c1 c2 c3 . . . cp ]. The solutions to the system can be read from the last matrix: xi = ci for each i = 1, 2, . . . , p. 43. We have ⎡ 1 −1 ⎣ 2 −1 1 1 ⎡ 1 0 2 ⎣ ∼ 0 1 0 0 ⎤ ⎡ 1 −1 2 1 1 −1 1 4 1 1 2 3 ⎦∼⎣ 0 6 −1 0 2 5 2 ⎤ ⎡ 0 3 1 3 1 0 3 2 −1 4 −1 ⎦ ∼ ⎣ 0 1 1 0 0 0 −2 2 ⎤ 1 −1 2 1 2 −1 4 −1 ⎦ 5 −2 6 0 ⎤ 0 9 −5 0 0 −1 8 −5 ⎦ . 1 0 −2 2 Hence, x1 = (0, −1, 0), x2 = (9, 8, −2), 1. A12 (−2), A13 (−1) x3 = (−5, −5, 2). 2. A21 (1), A23 (−2) 3. A31 (−3), A32 (−2) 44. (a). Let ei denote the ith column vector of the identity matrix Im , and consider the m linear systems of equations Axi = ei for i = 1, 2, . . . , m. Since rank(A) = m and each ei is a column m-vector, it follows that rank(A# ) = m = rank(A) and so each of the systems Axi = ei above has a solution (Note that if m < n, then there will be an infinite number of solutions). If we let B = [x1 , x2 , . . . , xm ], then AB = A [x1 , x2 , . . . , xm ] = [Ax1 , Ax2 , . . . , Axm ] = [e1 , e2 , . . . , em ] = In . ⎡ ⎤ a d (b). A right inverse for A in this case is a 3 × 2 matrix ⎣ b e ⎦ such that c f a + 3b + c d + 3e + f 1 0 = . 2a + 7b + 4c 2d + 7e + 4f 0 1 (c)2017 Pearson Education. Inc. 179 Thus, we must have a + 3b + c = 1, d + 3e + f = 0, 2a + 7b + 4c = 0, 2d + 7e + 4f = 1. 1 3 1 1 The first and third equation comprise a linear system with augmented matrix for a, b, and 2 7 4 0 1 1 3 1 c. The row-echelon form of this augmented matrix is . Setting c = t, we have b = −2 − 2t 0 1 2 −2 and a = 7 + 5t. Next, the second and fourth equation above comprise a linear system with augmented matrix 1 3 1 0 1 3 1 0 for d, e, and f . The row-echelon form of this augmented matrix is . Setting 2 7 4 1 0 1 2 1 f = s, we have e = 1 − ⎡ ⎤ 2s and d = −3 + 5s. Thus, right inverses of A are precisely the matrices of the form 7 + 5t −3 + 5s ⎣ −2 − 2t 1 − 2s ⎦. t s Solutions to Section 2.7 True-False Review: (a): TRUE. Since every elementary matrix corresponds to a (reversible) elementary row operation, the reverse elementary row operation will correspond to an elementary matrix that is the inverse of the original elementary matrix. 2 0 1 0 (b): FALSE. For instance, the matrices and are both elementary matrices, but their 0 1 0 2 2 0 product, , is not. 0 2 (c): FALSE. Every invertible matrix can be expressed as a product of elementary matrices. Since every elementary matrix is invertible and products of invertible matrices are invertible, any product of elementary matrices must be an invertible matrix. (d): TRUE. Performing an elementary row operation on a matrix does not alter its rank, and the matrix EA is obtained from A by performing the elementary row operation associated with the elementary matrix E. Therefore, A and EA have the same rank. (e): FALSE. If Pij is a permutation matrix, then Pij2 = In , since permuting the ith and jth rows of In twice yields In . Alternatively, we can observe that Pij2 = In from the fact that Pij−1 = Pij . 1 0 1 1 (f ): FALSE. For example, consider the elementary matrices E1 = and E2 = . Then we 0 7 0 1 1 1 1 7 have E1 E2 = and E2 E1 = . 0 7 0 7 ⎤ ⎡ ⎤ ⎡ 1 3 0 1 0 0 (g): FALSE. For example, consider the elementary matrices E1 = ⎣ 0 1 0 ⎦ and E2 = ⎣ 0 1 2 ⎦. 0 0 1 0 0 1 ⎡ ⎤ ⎡ ⎤ 1 3 6 1 3 0 Then we have E1 E2 = ⎣ 0 1 2 ⎦ and E2 E1 = ⎣ 0 1 2 ⎦. 0 0 1 0 0 1 (h): FALSE. The only matrices we perform an LU factorization for are invertible matrices for which the reduction to upper triangular form can be accomplished without permuting rows. (c)2017 Pearson Education. Inc. 180 (i): FALSE. The matrix U need not be a unit upper triangular matrix. (j): FALSE. As can be seen in Example 2.7.8, a 4 × 4 matrix with LU factorization will have 6 multipliers, not 10 multipliers. Problems: 1. ⎡ ⎤ ⎡ 0 1 0 0 Permutation Matrices: P12 = ⎣ 1 0 0 ⎦ , P13 = ⎣ 0 0 0 1 1 ⎡ ⎤ ⎡ k 0 0 1 Scaling Matrices: M1 (k) = ⎣ 0 1 0 ⎦ , M2 (k) = ⎣ 0 0 0 1 0 Row Combinations: ⎡ ⎤ 0 0 ⎦, 1 ⎤ 0 0 ⎦, 1 1 0 A12 (k) = ⎣ k 1 0 0 ⎡ 1 k A21 (k) = ⎣ 0 1 0 0 ⎤ ⎡ ⎤ 1 1 0 0 0 ⎦ , P23 = ⎣ 0 0 1 ⎦ . 0 0 1 0 ⎤ ⎡ ⎤ 0 0 1 0 0 k 0 ⎦ , M3 (k) = ⎣ 0 1 0 ⎦. 0 1 0 0 k 0 1 0 ⎡ ⎤ 0 0 ⎦, 1 ⎤ k 0 ⎦, 1 1 0 A13 (k) = ⎣ 0 1 k 0 ⎡ 1 0 A31 (k) = ⎣ 0 1 0 0 2. We have ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ −1 −8 1 8 1 −4 −1 1 2 3 ⎣ 0 3 ⎦∼⎣ 0 3 ⎦∼⎣ 0 3 ⎦∼⎣ 0 −3 7 −3 7 0 −3 7 1. A31 (−1) 2. M1 (−1) 3. A13 (3) ⎡ ⎤ 1 0 0 A23 (k) = ⎣ 0 1 0 ⎦ , 0 k 1 ⎡ ⎤ 1 0 0 A32 (k) = ⎣ 0 1 k ⎦ . 0 0 1 ⎤ ⎡ 8 1 4 3 ⎦∼⎣ 0 31 0 4. M2 ( 13 ) ⎤ ⎡ 8 1 5 1 ⎦∼⎣ 0 31 0 ⎤ 8 1 ⎦. 0 5. A23 (−31) Elementary Matrices: A23 (31), M2 ( 13 ), A13 (3), M1 (−1), A31 (−1). 3. We have 3 1 5 −2 1 ∼ 1 −2 3 5 1. P12 2 ∼ 1 −2 0 11 2. A12 (−3) 3 ∼ 1 −2 0 1 . 1 3. M2 ( 11 ) 1 Elementary Matrices: M2 ( 11 ), A12 (−3), P12 . 4. We have 5 1 8 2 3 −1 1 ∼ 1 5 3 −1 8 2 1. P12 2 ∼ 1 0 2. A12 (−5) 3 −1 −7 7 3 ∼ 1 0 3 −1 1 −1 . 3. M2 (− 17 ) Elementary Matrices: M2 (− 17 ), A12 (−5), P12 . 5. We have ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ 1 3 3 −1 4 1 3 2 1 3 2 1 3 2 1 2 3 4 ⎣ 2 1 3 ⎦∼⎣ 2 1 3 ⎦ ∼ ⎣ 0 −5 −1 ⎦ ∼ ⎣ 0 −5 −1 ⎦ ∼ ⎣ 0 1 1 3 2 3 −1 4 0 −10 −2 0 0 0 0 0 (c)2017 Pearson Education. Inc. 2 1 5 0 ⎤ ⎦. 181 1. P13 2. A12 (−2), A13 (−3) 4. M2 (− 15 ) 3. A23 (−2) Elementary Matrices: M2 (− 15 ), A23 (−2), A13 (−3), A12 (−2), P13 . 6. We have ⎡ 1 ⎣ 2 3 2 3 4 3 4 5 ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ 4 1 2 3 4 1 2 3 4 1 1 2 3 5 ⎦ ∼ ⎣ 0 −1 −2 −3 ⎦ ∼ ⎣ 0 1 2 3 ⎦∼⎣ 0 6 0 −2 −4 −6 0 −2 −4 −6 0 1. A12 (−2), A13 (−3) 2. M2 (−1) 2 1 0 ⎤ 4 3 ⎦. 0 3 2 0 3. A23 (2) Elementary Matrices: A23 (2), M2 (−1), A13 (−3), A12 (−2). 7. We reduce A to the identity matrix: 1 1 2 3 1 ∼ 1 0 2 1 1. A12 (−1) 2 ∼ 1 0 0 1 . 2. A21 (−2) 1 −1 The elementary matrices corresponding to these row operations are E1 = We have E2 E1 A = I2 , so that A = E1−1 E2−1 = 1 1 0 1 1 0 2 1 0 1 and E2 = 1 −2 0 1 , which is the desired expression since E1−1 and E2−1 are elementary matrices. 8. We reduce A to the identity matrix: −2 −3 1 −2 −3 2 1 1 3 1 1 4 1 ∼ ∼ ∼ ∼ 5 7 1 1 −2 −3 0 −1 0 1. A12 (2) 2. P12 3. A12 (2) 4. A21 (1) 0 1 0 1 1 0 1 −2 0 1 5 ∼ 1 0 0 1 1 1 , 1 −1 0 1 E5 = 1 0 1 0 0 −1 9. We reduce A to the identity matrix: 3 −4 1 −1 2 2 1 −2 3 1 −2 4 1 −2 5 1 ∼ ∼ ∼ ∼ ∼ −1 2 3 −4 3 −4 0 2 0 1 0 0 1 A = E1−1 E2−1 E3−1 E4−1 E5−1 = 1 −2 . 5. M2 (−1) The elementary matrices corresponding to these row operations are 1 0 0 1 1 0 1 E1 = , E2 = , E3 = , E4 = 2 1 1 0 2 1 0 We have E5 E4 E3 E2 E1 A = I2 , so 0 −1 0 −1 , which is the desired expression since each Ei−1 is an elementary matrix. (c)2017 Pearson Education. Inc. . . . 182 1. P12 2. M1 (−1) 4. M2 ( 12 ) 3. A12 (−3) 5. A21 (2) The elementary matrices corresponding to these row operations are 1 0 1 −1 0 1 0 , E2 = , E3 = , E4 = E1 = 1 0 0 1 −3 1 0 We have E5 E4 E3 E2 E1 A = I2 , so A = E1−1 E2−1 E3−1 E4−1 E5−1 = 0 1 1 0 −1 0 0 1 1 3 0 1 0 1 2 1 0 , 0 2 1 0 E5 = 1 −2 0 1 , which is the desired expression since each Ei−1 is an elementary matrix. 10. We reduce A to the identity matrix: 4 2 1 4 −5 1 1 ∼ ∼ 4 −5 0 1 4 1. P12 4 −21 3 ∼ 1 0 1 3. M2 (− 21 ) 2. A12 (−4) 4 1 A = E1−1 E2−1 E3−1 E4−1 = 0 1 1 0 1 4 0 1 ∼ 1 0 0 1 . 4. A21 (−4) The elementary matrices corresponding to these row operations are 1 0 0 1 1 0 , , E2 = , E3 = E1 = 1 1 0 −4 1 0 − 21 We have E4 E3 E2 E1 A = I2 , so 4 1 0 0 −21 E4 = 1 0 1 −4 0 1 4 1 . , which is the desired expression since each Ei−1 is an elementary matrix. 11. We reduce A to the identity matrix: ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ 1 −1 0 1 −1 0 1 −1 0 1 −1 0 1 2 3 ⎣ 2 4 2 ⎦∼⎣ 0 4 2 ⎦∼⎣ 0 4 2 ⎦ 2 2 ⎦∼⎣ 0 3 1 3 0 4 3 0 0 1 3 1 3 ⎤ ⎡ ⎡ ⎤ ⎡ ⎤ 1 −1 0 1 −1 0 1 0 0 4 5 6 1 0 ⎦ ∼ ⎣ 0 1 0 ⎦. 1 12 ⎦ ∼ ⎣ 0 ∼⎣ 0 0 0 1 0 0 1 0 0 1 1. A12 (−2) 2. A13 (−3) 3. A23 (−1) 4. M2 ( 41 ) 5. A32 (− 12 ) 6. A21 (1) The elementary matrices corresponding to these row operations are ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 1 0 0 1 0 0 1 0 0 1 0 ⎦, E1 = ⎣ −2 1 0 ⎦ , E2 = ⎣ 0 1 0 ⎦ , E3 = ⎣ 0 0 0 1 −3 0 1 0 −1 1 ⎤ ⎡ ⎡ ⎡ ⎤ ⎤ 1 0 0 1 0 0 1 1 0 E4 = ⎣ 0 14 0 ⎦ , E5 = ⎣ 0 1 − 12 ⎦ , E6 = ⎣ 0 1 0 ⎦ . 0 0 1 0 0 1 0 0 1 (c)2017 Pearson Education. Inc. 2 1 . 183 We have E6 E5 E4 E3 E2 E1 A = I3 , so A = E1−1 E2−1 E3−1 E4−1 E5−1 E6−1 ⎡ ⎤⎡ ⎤⎡ 1 0 0 1 0 0 1 = ⎣ 2 1 0 ⎦⎣ 0 1 0 ⎦⎣ 0 0 0 1 3 0 1 0 0 1 1 ⎤⎡ 0 1 0 ⎦⎣ 0 1 0 0 4 0 ⎤⎡ 1 0 0 0 ⎦⎣ 0 1 1 0 0 ⎤ 1 −1 0 1 ⎦⎣ 0 1 0 ⎦, 2 0 0 1 1 0 ⎤⎡ which is the desired expression since each Ei−1 is an elementary matrix. 12. We reduce A to the identity matrix: ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 0 −4 −2 1 −1 3 1 −1 3 1 −1 3 1 2 3 ⎣ 1 −1 3 ⎦ ∼ ⎣ 0 −4 −2 ⎦ ∼ ⎣ 0 −4 −2 ⎦ ∼ ⎣ 0 −4 −2 ⎦ −2 2 2 −2 2 2 0 0 8 0 0 1 ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ 1 −1 3 1 −1 0 1 −1 0 1 0 0 4 5 6 7 1 0 ⎦ ∼ ⎣ 0 1 0 ⎦. ∼ ⎣ 0 −4 0 ⎦ ∼ ⎣ 0 −4 0 ⎦ ∼ ⎣ 0 0 0 1 0 0 1 0 0 1 0 0 1 1. P12 2. A13 (2) 3. M3 ( 18 ) 4. A32 (2) 5. A31 (−3) 6. M2 (− 14 ) 7. A21 (1) The elementary matrices corresponding to these row operations are ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ ⎡ 1 0 0 0 1 0 1 0 0 1 0 E1 = ⎣ 1 0 0 ⎦ , E2 = ⎣ 0 1 0 ⎦ , E3 = ⎣ 0 1 0 ⎦ , E4 = ⎣ 0 1 0 0 1 2 0 1 0 0 0 0 18 ⎤ ⎤ ⎡ ⎡ ⎡ ⎤ 1 0 0 1 0 −3 1 1 0 0 ⎦ , E6 = ⎣ 0 − 14 0 ⎦ , E7 = ⎣ 0 1 0 ⎦ . E5 = ⎣ 0 1 0 0 1 0 0 1 0 0 1 ⎤ 0 2 ⎦, 1 We have E7 E6 E5 E4 E3 E2 E1 A = I3 , so A = E1−1 E2−1 E3−1 E4−1 E5−1 E6−1 E7−1 ⎡ ⎤⎡ ⎤⎡ 0 1 0 1 0 0 1 = ⎣ 1 0 0 ⎦⎣ 0 1 0 ⎦⎣ 0 0 0 1 −2 0 1 0 0 1 0 ⎤⎡ 0 1 0 ⎦⎣ 0 8 0 ⎤⎡ 0 0 1 1 −2 ⎦ ⎣ 0 0 1 0 0 1 0 ⎤⎡ ⎤⎡ ⎤ 3 1 0 0 1 −1 0 0 ⎦ ⎣ 0 −4 0 ⎦ ⎣ 0 1 0 ⎦, 1 0 0 1 0 0 1 which is the desired expression since each Ei−1 is an elementary matrix. 13. We reduce A to the identity matrix: ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 1 2 3 1 2 3 1 2 3 1 0 3 1 2 3 ⎣ 0 8 0 ⎦∼ ⎣ 0 1 0 ⎦∼ ⎣ 0 1 0 ⎦∼⎣ 0 1 0 ⎦ 3 4 5 3 4 5 0 −2 −4 0 −2 −4 ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 1 0 3 1 0 3 1 0 0 4 5 6 0 ⎦ ∼ ⎣ 0 1 0 ⎦ ∼ ⎣ 0 1 0 ⎦. ∼⎣ 0 1 0 0 −4 0 0 1 0 0 1 1. M2 ( 18 ) 2. A13 (−3) 3. A21 (−2) 4. A23 (2) (c)2017 Pearson Education. Inc. 5. M3 (− 14 ) 6. A31 (−3) 184 The elementary matrices corresponding to these row operations are ⎤ ⎡ ⎡ ⎤ ⎡ 1 0 0 1 0 0 1 E1 = ⎣ 0 18 0 ⎦ , E2 = ⎣ 0 1 0 ⎦ , E3 = ⎣ 0 −3 0 1 0 0 0 1 ⎤ ⎤ ⎡ ⎡ ⎡ 1 0 0 1 0 0 1 0 ⎦ , E6 = ⎣ 0 E4 = ⎣ 0 1 0 ⎦ , E5 = ⎣ 0 1 0 2 1 0 0 0 − 14 ⎤ −2 0 1 0 ⎦, 0 1 ⎤ 0 −3 1 0 ⎦. 0 1 We have E6 E5 E4 E3 E2 E1 A = I3 , so A = E1−1 E2−1 E3−1 E4−1 E5−1 E6−1 ⎡ ⎤⎡ ⎤⎡ 1 0 0 1 0 0 1 = ⎣ 0 8 0 ⎦⎣ 0 1 0 ⎦⎣ 0 0 0 1 3 0 1 0 2 1 0 ⎤⎡ ⎤⎡ 0 1 0 0 1 0 ⎦⎣ 0 1 0 ⎦⎣ 0 1 0 −2 1 0 ⎤⎡ 0 0 1 1 0 ⎦⎣ 0 0 −4 0 0 1 0 ⎤ 3 0 ⎦, 1 which is the desired expression since each Ei−1 is an elementary matrix. 14. We reduce A to the identity matrix: 2 −1 1 1 3 2 1 ∼ ∼ 1 3 2 −1 0 1. P12 3 −7 3 1 0 ∼ 3. M2 (− 17 ) 2. A12 (−2) 3 1 4 ∼ 1 0 0 1 . 4. A21 (−3) The elementary matrices corresponding to these row operations are 1 0 0 1 1 0 , , E2 = , E3 = E1 = 1 0 −2 1 0 − 17 E4 = 1 −3 0 1 . Direct multiplication verifies that E4 E3 E2 E1 A = I2 . 15. We have 3 −2 −1 5 1 ∼ 3 −2 0 13 3 = U. 1. A12 ( 13 ) Hence, E1 = A12 ( 13 ). Then Equation (2.7.3) reads L = E1−1 = A12 (− 13 ) = (2.7.2): LU = 16. We have 2 5 3 1 1 1 − 13 ∼ 2 0 Then LU = 0 1 3 − 13 2 1 5 2 0 1 3 −2 0 13 3 3 −2 −1 5 = = U =⇒ m21 = 2 3 0 − 13 2 = (c)2017 Pearson Education. Inc. 3 1 1 − 13 0 1 . Verifying Equation = A. 5 =⇒ L = 2 2 5 = A. 1 5 2 0 1 . 185 1. A12 (− 52 ) 17. We have 3 5 1 2 1 ∼ 3 0 Then LU = 1 1 3 1 5 3 0 1 5 = U =⇒ m21 = =⇒ L = 3 3 0 1 = 1 3 3 5 1 2 1 5 3 0 1 . = A. 1. A12 (− 53 ) 18. We have ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 3 −1 2 3 −1 2 3 −1 2 1 2 ⎣ 6 −1 1 ⎦ ∼ ⎣ 0 1 −3 ⎦ ∼ ⎣ 0 1 −3 ⎦ = U =⇒ m21 = 2, m31 = −1, m32 = 4. −3 5 2 0 4 4 0 0 16 Hence, ⎡ 1 L=⎣ 2 −1 0 1 4 ⎤ 0 0 ⎦ 1 ⎡ and 1 LU = ⎣ 2 −1 0 1 4 ⎤⎡ ⎤ ⎡ ⎤ 0 3 −1 2 3 −1 2 0 ⎦⎣ 0 1 −3 ⎦ = ⎣ 6 −1 1 ⎦ = A. 1 0 0 16 −3 5 2 1. A12 (−2), A13 (1) 19. We have ⎡ ⎤ ⎡ 5 2 1 5 1 ⎣ −10 −2 3 ⎦∼⎣ 0 15 2 −3 0 ⎤ ⎡ 2 1 5 2 2 5 ⎦∼⎣ 0 −4 −6 0 2 2 0 2. A23 (−4) ⎤ 1 5 ⎦ = U =⇒ m21 = −2, m31 = 3, m32 = −2. 4 Hence, ⎡ 1 L = ⎣ −2 3 ⎤ 0 0 1 0 ⎦ −2 1 ⎡ and 1 LU = ⎣ −2 3 ⎤⎡ 0 0 5 1 0 ⎦⎣ 0 −2 1 0 1. A12 (2), A13 (−3) 2 2 0 ⎤ ⎡ ⎤ 1 5 2 1 5 ⎦ = ⎣ −10 −2 3 ⎦ = A. 4 15 2 −3 2. A23 (2) 20. We have ⎤ ⎡ ⎤ ⎡ ⎤ ⎤ ⎡ ⎡ 1 −1 2 3 1 −1 2 3 1 −1 2 3 1 −1 2 3 ⎢ 2 2 −1 −10 ⎥ 2 −1 −10 ⎥ 2 −1 −10 ⎥ 0 3 −4 ⎥ 1 ⎢ 2 ⎢ 3 ⎢ ⎥∼ ⎥∼ ⎥ = U. ⎥∼ ⎢ 0 ⎢ 0 ⎢ 0 ⎢ ⎣ 3 −1 7 2 1 −1 ⎦ ⎣ 0 0 2 9 ⎦ ⎣ 0 0 2 9 ⎦ 8 ⎦ ⎣ 0 0 4 2 2 0 0 4 22 0 0 0 4 1 3 4 5 1. A12 (−2), A13 (−3), A14 (−1) 2. A23 (−1), A24 (−2) (c)2017 Pearson Education. Inc. 3. A34 (−2) 186 Hence, m21 = 2, m31 = 3, m41 = 1, m32 = 1, m42 = 2, m43 = 2. Hence, ⎡ 1 ⎢ 2 L=⎢ ⎣ 3 1 0 1 1 2 21. We have ⎡ 2 −3 ⎢ 4 −1 ⎢ ⎣ −8 2 6 1 0 0 1 2 ⎤ 0 0 ⎥ ⎥ 0 ⎦ 1 ⎡ 1 ⎢ 2 LU = ⎢ ⎣ 3 1 and 0 1 1 2 ⎤ ⎤ ⎡ ⎤⎡ 1 −1 2 3 1 −1 2 3 0 ⎢ ⎢ 0 3 −4 ⎥ 2 −1 −10 ⎥ 0 ⎥ ⎥ = A. ⎥=⎢ 2 ⎥⎢ 0 8 ⎦ 0 2 9 ⎦ ⎣ 3 −1 7 0 ⎦⎣ 0 1 3 4 5 0 0 0 4 1 0 0 1 2 ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 1 2 2 −3 1 2 2 −3 1 2 2 −3 1 2 1 1 ⎥ 5 −1 −3 ⎥ 5 −1 −3 ⎥ 5 −1 −3 ⎥ 1 ⎢ 2 ⎢ 3 ⎢ ⎥∼ ⎥∼ ⎥∼ ⎥ = U. ⎢ 0 ⎢ 0 ⎢ 0 ⎦ ⎦ ⎦ ⎣ ⎣ ⎣ 2 −5 0 −10 6 3 0 0 4 −3 0 0 4 −3 ⎦ 5 2 0 10 2 −4 0 0 4 2 0 0 0 5 1. A12 (−2), A13 (4), A14 (−3) 2. A23 (2), A24 (−2) 3. A34 (−1) Hence, m31 = −4, m21 = 2, m41 = 3, m32 = −2, m42 = 2, m43 = 1. Hence, ⎤ ⎤ ⎡ ⎤⎡ ⎤ ⎡ 2 −3 1 2 2 −3 1 2 1 0 0 0 1 0 0 0 ⎢ ⎢ ⎢ ⎢ 2 1 ⎥ 5 −1 −3 ⎥ 1 0 0 ⎥ 1 0 0 ⎥ ⎥ = A. ⎥ = ⎢ 4 −1 1 ⎥⎢ 0 ⎥ and LU = ⎢ 2 L=⎢ ⎣ −4 −2 1 0 ⎦ ⎣ 0 ⎣ −4 −2 1 0 ⎦ 2 2 −5 ⎦ 0 4 −3 ⎦ ⎣ −8 6 1 5 2 0 0 0 5 3 2 1 1 3 2 1 1 ⎡ 22. We have 1 2 2 3 1 ∼ 1 0 2 −1 = U =⇒ m21 = 2 =⇒ L = 1 2 0 1 . 1. A12 (−2) We now solve the triangular systems Ly = b and U x = y. From Ly = b, we obtain y = −11 U x = y yields x = . 7 3 −7 . Then 23. We have ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 1 −3 5 1 −3 5 1 −3 5 1 2 ⎣ 3 2 2 ⎦ ∼ ⎣ 0 11 −13 ⎦ ∼ ⎣ 0 11 −13 ⎦ = U =⇒ m21 = 3, m31 = 2, m32 = 1. 2 5 2 0 11 −8 0 0 5 1. A12 (−3), A13 (−2) 2. A23 (−1) (c)2017 Pearson Education. Inc. 187 ⎡ ⎤ 1 0 0 Hence, L = ⎣ 3 1 0 ⎦. We now solve the triangular systems Ly = b and U x = y. From Ly = b, we 2 1 1 ⎡ ⎤ ⎡ ⎤ 1 3 obtain y = ⎣ 2 ⎦. Then U x = y yields x = ⎣ −1 ⎦. −5 −1 24. We have ⎡ ⎤ ⎡ 2 2 2 1 1 ⎣ 0 ⎣ 6 3 −1 ⎦ ∼ 0 −4 2 2 ⎤ ⎡ ⎤ 2 1 2 2 1 2 −3 −4 ⎦ ∼ ⎣ 0 −3 −4 ⎦ = U =⇒ m21 = 3, m31 = −2, m32 = −2. 0 −4 0 0 −4 1. A12 (−3), A13 (2) 2. A23 (2) ⎡ ⎤ 1 0 0 1 0 ⎦. We now solve the triangular systems Ly = b and U x = y. From Ly = b, we Hence, L = ⎣ 3 −2 −2 1 ⎡ ⎤ ⎡ ⎤ 1 −1/12 1/3 ⎦. obtain y = ⎣ −3 ⎦. Then U x = y yields x = ⎣ −2 1/2 25. We have ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ ⎡ 4 3 0 0 4 3 0 0 4 3 0 4 3 0 0 ⎥ 2 ⎢ 0 −5 ⎥ 3 ⎢ 0 −5 2 ⎥ 1 ⎢ 0 −5 ⎢ 8 1 2 0 2 0 2 0 ⎥∼⎢ ⎥∼⎢ ⎥∼⎢ ⎢ ⎣ 0 5 5 3 6 ⎦ ⎣ 0 0 5 6 ⎦ ⎣ 0 0 5 3 6 ⎦ ⎣ 0 0 0 −5 7 0 0 −5 7 0 0 0 0 0 −5 7 1. A12 (−2) 2. A23 (1) ⎤ 0 0 ⎥ ⎥ = U. 6 ⎦ 13 3. A34 (1) ⎤ 1 0 0 0 ⎢ 2 1 0 0 ⎥ ⎥. We The only nonzero multipliers are m21 = 2, m32 = −1, and m43 = −1. Hence, L = ⎢ ⎣ 0 −1 1 0 ⎦ 0⎤ −1 1 ⎡0 2 ⎢ −1 ⎥ ⎥ now solve the triangular systems Ly = b and U x = y. From Ly = b, we obtain y = ⎢ ⎣ −1 ⎦. Then U x = y 4 ⎤ ⎡ 677/1300 ⎢ −9/325 ⎥ ⎥ yields x = ⎢ ⎣ −37/65 ⎦. 4/13 ⎡ 26. We have 2 −1 −8 3 1 ∼ 2 −1 0 −1 = U =⇒ m21 = −4 =⇒ L = 1. A12 (4) (c)2017 Pearson Education. Inc. 1 −4 0 1 . 188 We now solve the triangular systems Lyi = bi , U xi = yi for i = 1, 2, 3. We have 3 −4 . Then U x1 = y1 =⇒ x1 = ; Ly1 = b1 =⇒ y1 = 11 −11 2 −6.5 . Then U x2 = y2 =⇒ x2 = ; Ly2 = b2 =⇒ y2 = 15 −15 5 −3 Ly3 = b3 =⇒ y3 = . Then U x3 = y3 =⇒ x3 = . 11 −11 27. We have ⎡ −1 ⎣ 3 5 ⎤ ⎡ 4 2 −1 4 1 1 4 ⎦ ∼ ⎣ 0 13 −7 1 0 13 ⎤ ⎡ 2 −1 2 10 ⎦ ∼ ⎣ 0 11 0 1. A12 (3), A13 (5) 4 13 0 ⎤ 2 10 ⎦ = U. 1 2. A23 (−1) Thus, m21 = −3, m31 = −5, and m32 = 1. We now solve the triangular systems Lyi = bi , U xi = yi for i = 1, 2, 3. We have ⎡ ⎤ ⎡ ⎤ 1 −29/13 Ly1 = e1 =⇒ y1 = ⎣ 3 ⎦. Then U x1 = y1 =⇒ x1 = ⎣ −17/13 ⎦; 2 2 ⎡ ⎤ ⎡ ⎤ 0 18/13 Ly2 = e2 =⇒ y2 = ⎣ 1 ⎦. Then U x2 = y2 =⇒ x2 = ⎣ 11/13 ⎦; −1 −1 ⎡ ⎤ ⎡ ⎤ 0 −14/13 Ly3 = e3 =⇒ y3 = ⎣ 0 ⎦. Then U x3 = y3 =⇒ x3 = ⎣ −10/13 ⎦. 1 1 28. Observe that if Pi is an elementary permutation matrix, then Pi−1 = Pi = PiT . Therefore, we have −1 T . . . P2−1 P1−1 = PkT Pk−1 . . . P2T . . . P1T = (P1 P2 . . . Pk )T = P T . P −1 = (P1 P2 . . . Pk )−1 = Pk−1 Pk−1 29. (a). Let A be an invertible upper triangular matrix with inverse B. Therefore, we have AB = In . Write A = [aij ] and B = [bij ]. We will show that bij = 0 for all i > j, which shows that B is upper triangular. We have n , aik bkj = δij . k=1 Since A is upper triangular, aik = 0 whenever i > k. Therefore, we can reduce the above summation to n , aik bij = δij . k=i Let i = n. Then the above summation reduces to ann bnj = δnj . If j = n, we have ann bnn = 1, so ann = 0. For j < n, we have ann bnj = 0, and therefore bnj = 0 for all j < n. (c)2017 Pearson Education. Inc. 189 Next let i = n − 1. Then we have an−1,n−1 bn−1,j + an−1,n bnj = δn−1,j . Setting j = n−1 and using the fact that bn,n−1 = 0 by the above calculation, we obtain an−1,n−1 bn−1,n−1 = 1, so an−1,n−1 = 0. For j < n − 1, we have an−1,n−1 bn−1,j = 0 so that bn−1,j = 0. Next let i = n − 2. Then we have an−2,n−2 bn−2,j + an−2,n−1 bn−1,j + an−2,n bnj = δn−2,j . Setting j = n − 2 and using the fact that bn−1,n−2 = 0 and bn,n−2 = 0, we have an−2,n−2 bn−2,n−2 = 1, so that an−2,n−2 = 0. For j < n − 2, we have an−2,n−2 bn−2,j = 0 so that bn−2,j = 0. Proceeding in this way, we eventually show that bij = 0 for all i > j. For an invertible lower triangular matrix A with inverse B, we can either modify the preceding argument, or we can proceed more briefly as follows: Note that AT is an invertible upper triangular matrix with inverse B T . By the preceding argument, B T is upper triangular. Therefore, B is lower triangular, as required. (b). Let A be an invertible unit upper triangular matrix with inverse B. Use the notations from (a). By (a), we know that B is upper triangular. We simply must show that bjj = 0 for all j. From ann bnn = 1 (see proof of (a)), we see that if ann = 1, then bnn = 1. Moreover, from an−1,n−1 bn−1,n−1 = 1, the fact that an−1,n−1 = 1 proves that bn−1,n−1 = 1. Likewise, the fact that an−2,n−2 bn−2,n−2 = 1 implies that if an−2,n−2 = 1, then bn−2,n−2 = 1. Continuing in this fashion, we prove that bjj = 1 for all j. For the last part, if A is an invertible unit lower triangular matrix with inverse B, then AT is an invertible unit upper triangular matrix with inverse B T , and by the preceding argument, B T is a unit upper triangular matrix. This implies that B is a unit lower triangular matrix, as desired. 30. (a). Since A is invertible, Corollary 2.6.13 implies that both L2 and U1 are invertible. Since L1 U1 = L2 U2 , −1 −1 we can left-multiply by L−1 to obtain L−1 2 and right-multiply by U1 2 L1 = U 2 U 1 . is a unit lower triangular matrix and U1−1 is an upper triangular (b). By Problem 29, we know that L−1 2 −1 matrix. Therefore, L2 L1 is a unit lower triangular matrix and U2 U1−1 is an upper triangular matrix. Since −1 these two matrices are equal, we must have L−1 = In . Therefore, L1 = L2 and U1 = U2 . 2 L1 = In and U2 U1 31. The system Ax = b can be written as QRx = b. If we can solve Qy = b for y and then solve Rx = y for x, then QRx = b as desired. Multiplying Qy = b by QT and using the fact that QT Q = In , we obtain y = QT b. Therefore, Rx = y can be replaced by Rx = QT b. Therefore, to solve Ax = b, we first determine y = QT b and then solve the upper triangular system Rx = QT b by back-substitution. Solutions to Section 2.8 True-False Review: (a): FALSE. According to the given information, part (c) of the Invertible Matrix Theorem fails, while part (e) holds. This is impossible. (b): TRUE. This holds by the equivalence of parts (d) and (f) of the Invertible Matrix Theorem. (c): FALSE. Part (d) of the Invertible Matrix Theorem fails according to the given information, and therefore part (b) also fails. Hence, the equation Ax = b does not have a unique solution. But it is not valid⎡to conclude⎤that the equation ⎡ ⎤ has infinitely many solutions; it could have no solutions. For instance, if 1 0 0 0 A = ⎣ 0 1 0 ⎦ and b = ⎣ 0 ⎦, there are no solutions to Ax = b, although rank(A) = 2. 0 0 0 1 (d): FALSE. An easy counterexample is the matrix 0n , which fails to be invertible even though it is upper triangular. Since it fails to be invertible, it cannot e row-equivalent to In , by the Invertible Matrix Theorem. (c)2017 Pearson Education. Inc. 190 Problems: 1. Since A is an invertible matrix, the only solution to Ax = 0 is x = 0. However, if we assume that AB = AC, then A(B − C) = 0. If xi denotes the ith column of B − C, then xi = 0 for each i. That is, B − C = 0, or B = C, as required. 2. If rank(A) = n, then the augmented matrix A# for the system Ax = 0 can be reduced to REF such that each column contains a pivot except for the right-most column of all-zeros. Solving the system by back-substitution, we find that x = 0, as claimed. 3. Since Ax = 0 has only the trivial solution, REF(A) contains a pivot in every column. Therefore, the linear system Ax = b can be solved by back-substitution for every b in Rn . Therefore, Ax = b does have a solution. Now suppose there are two solutions y and z to the system Ax = b. That is, Ay = b and Az = b. Subtracting, we find A(y − z) = 0, and so by assumption, y − z = 0. That is, y = z. Therefore, there is only one solution to the linear system Ax = b. 4. If A and B are each invertible matrices, then A and B can each be expressed as a product of elementary matrices, say A = E1 E2 . . . Ek and B = E1 E2 . . . El . Then AB = E1 E2 . . . Ek E1 E2 . . . El , so AB can be expressed as a product of elementary matrices. Thus, by the equivalence of (a) and (e) in the Invertible Matrix Theorem, AB is invertible. 5. We are assuming that the equations Ax = 0 and Bx = 0 each have only the trivial solution x = 0. Now consider the linear system (AB)x = 0. Viewing this equation as A(Bx) = 0, we conclude that Bx = 0. Thus, x = 0. Hence, the linear equation (AB)x = 0 has only the trivial solution. Solutions to Section 2.9 Problems: ⎤ ⎤ ⎡ ⎤ ⎡ 13 −1 −15 0 −2 −1 ⎥ ⎢ ⎢ 4 −1 ⎥ ⎢ 10 10 ⎥ ⎥ = ⎢ −6 −11 ⎥. ⎥−⎢ 1. AT − 5B = ⎢ ⎦ ⎣ ⎦ ⎣ ⎣ 2 −3 20 ⎦ 5 −15 5 6 −5 0 5 6 0 ⎡ ⎤ −3 0 ⎢ 2 2 ⎥ ⎥ 2. C T B = −5 −6 3 1 ⎢ ⎣ 1 −3 ⎦ = 6 −20 . 0 1 ⎡ 3. Since A is not a square matrix, it is not possible to compute A2 . 8 −16 −8 −24 −3 2 1 0 11 −18 −9 −24 4. −4A − B T = − = . 4 4 −20 0 0 2 −3 1 4 2 −17 −1 (c)2017 Pearson Education. Inc. 191 5. We have −2 4 2 6 −1 −1 5 0 AB = ⎤ −3 0 ⎢ 2 2 ⎥ ⎥ = 16 ⎢ ⎣ 1 −3 ⎦ 6 0 1 ⎡ 8 −17 . Moreover, tr(AB) = −1. 6. We have (AC)(AC)T = −2 26 −2 26 4 −52 −52 676 = . ⎤ ⎡ ⎤ −24 48 24 72 12 0 ⎢ 24 −24 −56 −48 ⎥ ⎢ −8 −8 ⎥ −2 4 2 6 ⎥. ⎢ ⎥ 7. (−4B)A = ⎢ ⎣ −4 12 ⎦ −1 −1 5 0 = ⎣ −4 −28 52 −24 ⎦ 4 4 −20 0 0 −4 ⎡ 8. Using Problem 5, we find that (AB) −1 = −1 16 8 6 −17 1 =− 320 −17 −8 −6 16 . ⎤ −5 ⎢ −6 ⎥ ⎥ ⎢ ⎣ 3 ⎦ = [71], 1 ⎡ 9. We have −5 −6 CT C = 3 1 and tr(C T C) = 71. 10. (a). We have AB = 1 2 2 5 3 7 ⎡ ⎤ 3 b 3a − 5 ⎣ −4 a ⎦ = 7a − 14 a b 2a + 4b 5a + 9b . In order for this product to equal I2 , we require 3a − 5 = 1, 7a − 14 = 0, 2a + 4b = 0, 5a + 9b = 1. We quickly solve this for the unique solution: a = 2 and b = −1. (b). We have ⎡ ⎤ 3 −1 1 2 ⎦ BA = ⎣ −4 2 2 −1 2 5 3 7 ⎡ ⎤ 1 1 2 2 2 ⎦. =⎣ 0 0 −1 −1 11. We compute the (i, j)-entry of each side of the equation. We will denote the entries of AT by aTij , which equals aji . On the left side, note that the (i, j)-entry of (AB T )T is the same as the (j, i)-entry of AB T , and (j, i)-entry of AB T = n , k=0 ajk bTki = n , ajk bik = k=0 (c)2017 Pearson Education. Inc. n , k=0 bik aTkj , 192 and the latter expression is the (i, j)-entry of BAT . Therefore, the (i, j)-entries of (AB T )T and BAT are the same, as required. 12. (a). The (i, j)-entry of A2 is n , aik akj . k=1 (b). Assume that A is symmetric. That means that AT = A. We claim that A2 is symmetric. To see this, note that (A2 )T = (AA)T = AT AT = AA = A2 . Thus, (A2 )T = A2 , and so A2 is symmetric. 13. We are assuming that A is skew-symmetric, so AT = −A. To show that B T AB is skew-symmetric, we observe that (B T AB)T = B T AT (B T )T = B T AT B = B T (−A)B = −(B T AB), as required. 14. We have A2 = 3 9 −1 −3 2 = 0 0 0 0 , so A is nilpotent. 15. We have ⎡ 0 A2 = ⎣ 0 0 and ⎡ 0 A3 = A2 A = ⎣ 0 0 0 0 0 ⎤ 1 0 ⎦ 0 0 0 0 ⎤⎡ 1 0 0 ⎦⎣ 0 0 0 1 0 0 ⎤ ⎡ 1 0 1 ⎦=⎣ 0 0 0 0 0 0 ⎤ 0 0 ⎦, 0 so A is nilpotent. 16. We have 17. We have ⎡ −3e−3t ⎣ 6t2 A (t) = 6/t ⎡ −7t ⎢ 6t − t2 /2 B(t) dt = ⎢ ⎣ t + t2 /2 0 et 1 ⎤ −2 sec2 t tan t ⎦. − sin t −5 ⎡ ⎤ t3 /3 −7 1 ⎢ 11/2 3t4 /4 + 2t3 ⎥ ⎥ = ⎢ 2 ⎣ 3/2 ⎦ π sin(πt/2) 0 4 e−1 t − t /4 ⎤ 1/3 11/4 ⎥ ⎥. 2/π ⎦ 3/4 18. Since A(t) is 3 × 2 and B(t) is 4 × 2, it is impossible to perform the indicated subtraction. 19. Since A(t) is 3 × 2 and B(t) is 4 × 2, it is impossible to perform the indicated subtraction. (c)2017 Pearson Education. Inc. 193 20. From the last equation, we see that x3 = 0. Substituting this into the middle equation, we find that x2 = 0.5. Finally, putting the values of x2 and x3 into the first equation, we find x1 = −6 − 2.5 = −8.5. Thus, there is a unique solution to the linear system, and the solution set is {(−8.5, 0.5, 0)}. 21. To solve this system, we need to reduce the corresponding augmented matrix for the linear system to row-echelon form. This gives us ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ 5 −1 2 7 7 1 11 20 1 11 20 7 1 11 1 2 3 ⎣ −2 6 9 9 0 ⎦ ∼ ⎣ −2 6 0 ⎦ ∼ ⎣ 0 28 49 14 ⎦ ∼ ⎣ 0 1 −7 5 −3 −7 −7 5 −3 −7 0 82 137 42 0 82 ⎡ ⎤ ⎡ ⎤ 7 1 11 20 7 1 11 20 4 5 7/4 1/2 ⎦ ∼ ⎣ 0 1 7/4 1/2 ⎦ . ∼⎣ 0 1 0 0 −13/2 1 0 0 1 −2/13 ⎤ 7 20 7/4 1/2 ⎦ 137 42 From the we conclude that x3 = −2/13, and using the middle row, we can solve for x2 : we have row, 2last 10 10 2 x2 + 74 · − 13 = 12 , so x2 = 20 26 = 13 . Finally, from the first row we can get x1 : we have x1 +11· 13 +20· − 13 = 7, and so x1 = 21 13 . So there is a unique solution: 21 10 2 . , ,− 13 13 13 1. A21 (2) 2. A12 (2), A13 (7) 3. M2 (1/28) 4. A23 (−82) 5. M3 (−2/13) 22. To solve this system, we need to reduce the corresponding augmented matrix for the linear system to row-echelon form. This gives us ⎡ ⎤ ⎡ ⎤ ⎡ 1 2 −1 1 1 2 −1 1 1 1 2 ⎣ 1 0 1 5 ⎦ ∼ ⎣ 0 −2 2 4 ⎦∼⎣ 0 4 4 0 12 0 −4 4 8 0 ⎤ ⎡ ⎤ 1 1 2 −1 1 2 −1 3 1 −1 −2 ⎦ ∼ ⎣ 0 1 −1 −2 ⎦ . −4 4 0 0 0 8 0 From this row-echelon form, we see that z is a free variable. Set z = t. Then from the middle row of the matrix, y = t − 2, and from the top row, x + 2(t − 2) − t = 1 or x = −t + 5. So the solution set is {(−t + 5, t − 2, t) : t ∈ R} = {(5, −2, 0) + t(−1, 1, 1) : t ∈ R}. 1. A12 (−1), A13 (−4) 2. M2 (−1/2) 3. A23 (4) 23. To solve this system, we need to reduce the corresponding augmented matrix for the linear system to row-echelon form. This gives us ⎡ ⎤ ⎡ 1 −2 −1 3 0 1 −2 −1 1 ⎣ −2 4 5 −5 3 ⎦ ∼ ⎣ 0 0 3 3 −6 −6 8 2 0 0 −3 ⎤ ⎡ 3 0 1 2 1 3 ⎦∼⎣ 0 −1 2 0 ⎤ ⎡ −2 −1 3 0 1 3 0 3 1 3 ⎦∼⎣ 0 0 0 0 5 0 The bottom row of this matrix shows that this system has no solutions. (c)2017 Pearson Education. Inc. −2 −1 0 1 0 0 ⎤ 3 0 1/3 1 ⎦ . 0 1 194 1. A12 (2), A13 (−3) 2. A23 (1) 3. M2 (1/3), M3 (1/3) 24. To solve this system, we need to reduce the corresponding augmented matrix for the linear system to row-echelon form. This gives us ⎤ ⎡ 1 3 3 0 −1 2 −1 1 ⎢ 1 ⎥ 1 ⎢ 3 3 1 −3 2 0 −1 ⎢ ⎥∼⎢ ⎣ 4 −2 −3 6 −1 5 ⎦ ⎣ 4 −2 0 0 0 1 4 −2 0 0 ⎡ 1 3 1 −3 2 0 −27 −12 33 −21 3 ⎢ ∼⎢ ⎣ 0 28 14 −36 18 0 0 0 1 4 ⎡ ⎤ ⎡ 1 3 1 −3 2 −1 1 ⎥ 6 ⎢ 0 0 1 2 −3 −3 −6 5 ⎢ ⎥ ⎢ ∼⎢ ⎣ 0 −27 −12 33 −21 12 ⎦ ∼ ⎣ 0 0 0 0 1 4 −2 0 ⎡ ⎤ ⎡ ⎤ 1 −3 2 −1 1 3 1 −3 2 −1 −1 2 −1 −9 −4 11 −7 1 ⎥ 4 ⎥ 2 ⎢ ⎢ 0 ⎥∼ ⎥ ⎣ ⎦ −3 6 −1 0 −14 −7 18 −9 5 9 ⎦ 0 1 4 −2 0 0 0 1 4 −2 ⎤ ⎡ ⎤ 1 3 1 −3 2 −1 −1 ⎢ 0 −27 −12 33 −21 12 ⎥ 12 ⎥ 4 ⎥∼⎢ ⎥ 1 2 −3 −3 −6 ⎦ −18 ⎦ ⎣ 0 0 0 0 1 4 −2 −2 ⎤ ⎡ ⎤ −1 1 3 1 −3 2 3 1 −3 2 −1 −3 −6 ⎥ 1 2 −3 −3 −6 ⎥ 7 ⎢ ⎢ 0 1 2 −3 ⎥∼ ⎥ 25 ⎦ . 0 42 −48 −102 −150 ⎦ ⎣ 0 0 1 − 87 − 17 − 7 7 0 0 1 4 −2 0 0 0 1 4 −2 We see that x5 = t is the only free variable. Back substitution yields the remaining values: x5 = t, So the solution set is 1. P12 x4 = −4t − 2, x3 = − 41 15 − t, 7 7 2 33 x2 = − − t, 7 7 2 16 x1 = − + t. 7 7 2 16 2 33 41 15 − + t, − − t, − − t, −4t − 2, t : t ∈ R 7 7 7 7 7 7 2 2 41 16 33 15 , − , − , −4, 1 + − , − , − , −2, 0 : t ∈ R . = t 7 7 7 7 7 7 2. A12 (−3), A13 (−4) 3. M2 (3), M3 (−2) 4. A23 (1) 5. P23 6. A23 (27) 7. M3 (1/42) 25. To solve this system, we need to reduce the corresponding augmented matrix for the linear system to row-echelon form. This gives us ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ 1 1 1 1 −3 1 1 1 1 −3 6 6 1 1 1 1 −3 6 ⎢ 1 1 1 2 −5 8 ⎥ 1 ⎢ 0 0 0 1 −2 0 1 −2 2 ⎥ 2 ⎥ 2 ⎢ ⎢ 0 0 ⎥∼⎢ ⎥∼ ⎥ ⎢ ⎣ 2 3 1 4 −9 17 ⎦ ⎣ 0 1 −1 2 −3 5 ⎦ ⎣ 0 1 −1 2 −3 5 ⎦ 2 2 2 3 −8 14 0 0 0 −1 2 −2 0 0 0 0 0 0 ⎤ ⎡ 1 1 1 1 −3 6 ⎢ 0 1 −1 2 −3 5 ⎥ 3 ⎥. ∼⎢ ⎣ 0 0 0 1 −2 2 ⎦ 0 0 0 0 0 0 From this row-echelon form, we see that x5 = t and x3 = s are free variables. Furthermore, solving this system by back-substitution, we see that x5 = t, x4 = 2t + 2, x3 = s, x2 = s − t + 1, (c)2017 Pearson Education. Inc. x1 = 2t − 2s + 3. 195 So the solution set is {(2t − 2s + 3, s − t + 1, s, 2t + 2, t) : s, t ∈ R} = {t(2, −1, 0, 2, 1) + s(−2, 1, 1, 0, 0) + (3, 1, 0, 2, 0) : s, t ∈ R}. 1. A12 (−1), A13 (−2), A14 (−2) 2. A24 (1) 3. P23 26. To solve this system, we need to reduce the corresponding augmented matrix for the linear system to row-echelon form. This gives us 1 −3 2i 1 1 −3 2i 1 1 1 1 −3 2i 2 . ∼ ∼ −2i 6 2 −2 0 6 − 6i −2 −2 + 2i 0 1 − 16 (1 + i) − 13 1. A12 (2i) 1 2. M2 ( 6−6i ) From the last augmented matrix above, we see that x3 is a free variable. Let us set x3 = t, where t is a complex number. Then we can solve for x2 using the equation corresponding to the second row of the row-echelon form: x2 = − 13 + 16 (1+i)t. Finally, using the first row of the row-echelon form, we can determine that x1 = 12 t(1 − 3i). Therefore, the solution set for this linear system of equations is 1 1 1 {( t(1 − 3i), − + (1 + i)t, t) : t ∈ C}. 2 3 6 27. We reduce the corresponding linear system as follows: 1 −k 6 1 −k 1 ∼ 2 3 k 0 3 + 2k 6 k − 12 . If k = − 32 , then each column of the row-reduced coefficient matrix will contain a pivot, and hence, the linear system will have a unique solution. If, on the other hand, k = − 32 , then the system is inconsistent, because the last row of the row-echelon form will have a pivot in the right-most column. Under no circumstances will the linear system have infinitely many solutions. 28. First observe that if k = 0, then the second equation requires that x3 = 2, and then the first equation requires x2 = 2. However, x1 is a free variable in this case, so there are infinitely many solutions. Now suppose that k = 0. Then multiplying each row of the corresponding augmented matrix for the linear system by 1/k yields a row-echelon form with pivots in the first two columns only. Therefore, the third variable, x3 , is free in this case. So once again, there are infinitely many solutions to the system. We conclude that the system has infinitely many solutions for all values of k. 29. Since this linear system is homogeneous, it already has at least one solution: (0, 0, 0). Therefore, it only remains to determine the values of k for which this will be the only solution. We reduce the corresponding matrix as follows: ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 10 k −1 0 −k 0 10k k 2 1 1/2 −1/2 0 1 2 ⎣ k 1 −1 0 ⎦ ∼ ⎣ 10k 10 −10 0 ⎦ ∼ ⎣ 10k 10 −10 0 ⎦ 2 2 1 −1 0 1 1/2 −1/2 0 10k k −k 0 ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 1 1/2 −1/2 0 1 1/2 −1/2 0 1 1/2 −1/2 0 3 4 5 1 −1 1 −1 0 ⎦∼⎣ 0 0 ⎦. ∼ ⎣ 0 10 − 5k 5k − 10 0 ⎦ ∼ ⎣ 0 2 2 2 0 k − 5k 0 k − 5k 0 0 k −k 0 4k 0 4k 0 (c)2017 Pearson Education. Inc. 196 1. M1 (k), M2 (10), M3 (1/2) 2. P13 3. A12 (−10k), A13 (−10k) 1 4. M2 ( 10−5k ) 5. A23 (5k − k 2 ) Note that the steps above are not valid if k = 0 or k = 2 (because Step 1 is not valid with k = 0 and Step 4 is not valid if k = 2). We will discuss those special cases individually in a moment. However if k = 0, 2, then the steps are valid, and we see from the last row of the last matrix that if k = 1, we have infinitely many solutions. Otherwise, if k = 0, 1, 2, then the matrix has full rank, and so there is a unique solution to the linear system. If k = 2, then the last two rows of the original matrix are the same, and so the matrix of coefficients of the linear system is not invertible. Therefore, the linear system must have infinitely many solutions. If k = 0, we reduce the original linear system as follows: ⎡ ⎤ ⎡ 10 0 −1 0 1 0 −1/10 1 ⎣ 0 1 −1 0 ⎦ ∼ ⎣ 0 1 −1 2 1 −1 0 2 1 −1 ⎤ ⎡ ⎤ ⎡ 0 1 0 −1/10 0 1 0 −1/10 2 3 −1 −1 0 ⎦∼⎣ 0 1 0 ⎦∼⎣ 0 1 0 1 −4/5 0 0 0 1/5 0 ⎤ 0 0 ⎦. 0 The last matrix has full rank, so there will be a unique solution in this case. 1. M1 (1/10) 2. A13 (−2) 3. A23 (−1) To summarize: The linear system has infinitely many solutions if and only if k = 1 or k = 2. Otherwise, the system has a unique solution. 30. To solve this system, we need to reduce the corresponding augmented matrix for the linear system to row-echelon form. This gives us ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 0 0 0 1 −k k 2 0 1 −k k2 1 −k k2 1 −k k2 1 2 3 ⎣ 1 0 k 0 ⎦∼⎣ 0 k k − k2 0 ⎦ ∼ ⎣ 0 1 −1 1 −1 1 ⎦∼⎣ 0 1 ⎦. 2 2 0 1 −1 1 0 1 −1 0 k k−k 0 0 0 2k − k −k 1 1. A12 (−1) 2. P23 3. A23 (−k) Now provided that 2k − k 2 = 0, the system can be solved without free variables via back-substitution, and therefore, there is a unique solution. Consider now what happens if 2k − k 2 = 0. Then either k = 0 or k = 2. If k = 0, then only the first two columns of the last augmented matrix above are pivoted, and we have a free variable corresponding to x3 . Therefore, there are infinitely many solutions in this case. On the other hand, if k = 2, then the last row of the last matrix above reflects an inconsistency in the linear system, and there are no solutions. To summarize, the system has no solutions if k = 2, a unique solution if k = 0 and k = 2, and infinitely many solutions if k = 0. 31. No, there are no common points of intersection. A common point of intersection would be indicated by a solution to the linear system consisting of the equations of the three planes. However, the corresponding augmented matrix can be row-reduced as follows: ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 1 2 1 4 4 4 1 2 1 1 2 1 1 2 ⎣ 0 1 −1 1 ⎦ ∼ ⎣ 0 1 −1 1 ⎦ ∼ ⎣ 0 1 −1 1 ⎦. 1 3 0 0 0 1 −1 −4 0 0 0 −5 The last row of this matrix shows that the linear system is inconsistent, and so there are no points common to all three planes. (c)2017 Pearson Education. Inc. 197 1. A13 (−1) 2. A23 (−1) 32. (a). We have 4 7 −2 5 1 ∼ 1 7/4 −2 5 1. M1 (1/4) 2 ∼ 1 7/4 0 17/2 2. A12 (2) 3 1 0 ∼ 7/4 1 . 3. M2 (2/17) (b). We have: rank(A) = 2, since the row-echelon form of A in (a) consists two nonzero rows. (c). We have 4 7 1 −2 5 0 0 1 1 ∼ 1 7/4 1/4 0 −2 5 0 1 4 ∼ 1. M1 (1/4) 2 ∼ 7/4 1/4 17/2 1/2 1 0 1 0 5/34 −7/34 0 1 1/17 2/17 2. A12 (2) Thus, A −1 = 7 − 34 3 ∼ 1 7/4 1/4 0 1 1/17 0 2/17 3. M2 (2/17) 5 34 1 17 0 1 . 4. A21 (−7/4) . 2 17 33. (a). We have 2 −7 −4 14 1 ∼ 2 −7 0 0 1. A12 (2) 2 ∼ 1 −7/2 0 0 . 2. M1 (1/2) (b). We have: rank(A) = 1, since the row-echelon form of A in (a) has one nonzero row. (c). Since rank(A) < 2, A is not invertible. 34. (a). We have ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ 3 −1 6 1 −1/3 2 1 −1/3 2 1 −1/3 1 2 3 ⎣ 0 2 3 ⎦∼⎣ 0 2 3 ⎦∼⎣ 0 2 3 ⎦∼⎣ 0 2 3 −5 0 1 −5/3 0 0 −4/3 −2 0 0 1. M1 (1/3), M3 (1/3) 2. A13 (−1) 3. A23 (2/3) (c)2017 Pearson Education. Inc. ⎤ ⎡ 2 1 −1/3 4 3 ⎦∼⎣ 0 1 0 0 0 4. M2 (1/2) ⎤ 2 3/2 ⎦ . 0 198 (b). We have: rank(A) = 2, since the row-echelon form of A in (a) consists of two nonzero rows. (c). Since rank(A) < 3, A is not invertible. 35. (a). We have ⎡ 2 ⎢ 1 ⎢ ⎣ 0 0 1 2 0 0 0 0 3 4 ⎤ ⎡ 1 0 ⎢ 2 0 ⎥ 1 ⎥∼⎢ 4 ⎦ ⎣ 0 0 3 ⎡ 2 1 0 0 0 0 3 4 1 2 0 1 ∼⎣ 0 0 0 0 4 ⎢ ⎢ 1. P12 ⎤ ⎡ ⎤ ⎡ ⎤ 0 1 2 0 0 1 2 0 0 0 ⎥ 0 ⎥ 0 ⎥ 2 ⎢ 3 ⎢ ⎥∼ ⎥∼ ⎥ ⎢ 0 −3 0 ⎢ 0 −3 0 4 ⎦ ⎣ 0 0 3 4 ⎦ ⎣ 0 0 1 −1 ⎦ 3 0 0 1 −1 0 0 3 4 ⎤ ⎡ ⎤ 0 0 1 2 0 0 ⎢ 0 1 0 0 0 ⎥ 0 ⎥ 5 ⎥∼⎢ ⎥. 1 −1 ⎦ ⎣ 0 0 1 −1 ⎦ 0 7 0 0 0 1 2. A12 (−2), A34 (−1) 3. P34 4. M2 (−1/3), A34 (−3) 5. M4 (1/7) (b). We have: rank(A) = 4, since the row-echelon form of A in (a) consists of four nonzero rows. (c). We have ⎡ 2 1 0 0 ⎢ 1 2 0 0 ⎢ ⎣ 0 0 3 4 0 0 4 3 ⎡ 1 0 0 0 0 1 0 0 1 2 0 −3 ∼⎣ 0 0 0 0 3 ⎢ ⎢ ⎡ 1 0 0 0 0 1 0 0 ∼⎣ 0 0 1 −1 0 0 0 1 5 ⎢ ⎢ 1. P12 ⎤ ⎡ ⎤ ⎡ 0 1 2 0 0 0 1 0 0 1 2 ⎢ 2 1 0 0 1 0 0 0 ⎥ 2 ⎢ 0 −3 0 ⎥ 1 ⎥∼⎢ ⎥∼⎢ 0 0 ⎦ ⎣ 0 0 3 4 0 0 1 0 ⎦ ⎣ 0 0 0 4 3 0 0 0 1 0 0 1 ⎤ ⎡ 0 0 0 1 2 0 0 1 0 0 0 0 0 1 −2 0 −1/3 0 0 ⎥ 4 ⎢ ⎢ 0 1 0 ⎥∼ 1 −1 0 0 −1 1 ⎦ ⎣ 0 0 1 −1 0 3 4 0 0 0 0 7 0 1 0 0 ⎤ ⎡ 2/3 −1/3 0 0 1 0 0 0 2/3 ⎢ 0 1 0 0 −1/3 −1/3 2/3 0 0 ⎥ 6 ⎥∼⎢ 0 0 −1 1 ⎦ ⎣ 0 0 1 0 0 0 0 0 1 0 0 4/7 −3/7 0 0 0 1 0 2. A12 (−2), A34 (−1) Thus, 3. P34 ⎡ 4. A34 (−3), M2 (−1/3) 2/3 ⎢ −1/3 −1 A =⎢ ⎣ 0 0 −1/3 0 2/3 0 0 −3/7 0 4/7 ⎤ 1 0 0 −2 0 0 ⎥ ⎥ 0 1 0 ⎦ 0 −1 1 ⎤ 1 0 0 2/3 0 0 ⎥ ⎥ 0 −1 1 ⎦ 0 4 −3 ⎤ −1/3 0 0 2/3 0 0 ⎥ ⎥. 0 −3/7 4/7 ⎦ 0 4/7 −3/7 0 0 0 0 3 4 1 −1 0 1 0 0 5. M4 (1/7), A21 (−2) 6. A43 (1) ⎤ 0 0 ⎥ ⎥. 4/7 ⎦ −3/7 36. (a). We have ⎡ ⎤ ⎡ 3 0 0 1 1 ⎣ 0 2 −1 ⎦ ∼ ⎣ 0 1 −1 2 1 ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ 0 0 1 0 0 1 0 0 1 0 0 1 2 3 4 5 2 −1 ⎦ ∼ ⎣ 0 2 −1 ⎦ ∼ ⎣ 0 −1 2 ⎦ ∼ ⎣ 0 −1 2 ⎦ ∼ ⎣ 0 −1 2 0 −1 2 0 2 −1 0 0 3 0 (c)2017 Pearson Education. Inc. ⎤ 0 0 1 −2 ⎦ . 0 1 199 1. M1 (1/3) 2. A13 (−1) 3. P23 4. A23 (2) 5. M2 (−1), M3 (1/3) (b). We have: rank(A) = 3, since the row-echelon form of A in (a) has 3 nonzero rows. (c). We have ⎡ 3 0 0 1 ⎣ 0 2 −1 0 1 −1 2 0 0 1 0 ⎤ ⎡ ⎤ ⎡ ⎤ 0 1 0 0 1/3 0 0 1 0 0 1/3 0 0 1 2 2 −1 0 2 −1 0 ⎦∼⎣ 0 1 0 ⎦∼⎣ 0 0 1 0 ⎦ 1 −1 2 0 0 −1 2 −1/3 0 1 1 0 1 ⎡ ⎤ ⎡ 1 0 0 1/3 0 0 1 4 2 −1/3 0 1 ⎦ ∼ ⎣ 0 ∼ ⎣ 0 −1 0 2 −1 0 0 1 0 ⎡ ⎤ ⎡ 0 0 1 0 0 1/3 1 5 6 0 −1 ⎦ ∼ ⎣ 0 ∼ ⎣ 0 1 −2 1/3 0 0 1 −2/9 1/3 2/3 0 3 1. M1 (1/3) 2. A13 (−1) 3. P23 Hence, 4. A23 (2) ⎡ 1/3 0 A−1 = ⎣ −1/9 2/3 −2/9 1/3 ⎤ 0 0 0 1 ⎦ 1 2 0 0 1/3 −1 2 −1/3 0 3 −2/3 0 0 1/3 1 0 −1/9 0 1 −2/9 0 2/3 1/3 ⎤ 0 1/3 ⎦ . 2/3 5. M2 (−1), M3 (1/3) 6. A32 (2) ⎤ 0 1/3 ⎦ . 2/3 37. (a). We have ⎤ ⎡ ⎡ ⎤ ⎡ 1 4 2 1 −2 −3 1 1 2 ⎣ 1 4 2 ⎦ ∼ ⎣ −2 −3 1 ⎦ ∼ ⎣ 0 0 5 3 0 0 5 3 1. P12 2. A12 (2) 4 5 5 ⎤ ⎡ 2 1 3 5 ⎦∼⎣ 0 3 0 3. A23 (−1) ⎤ ⎡ 4 2 1 4 5 5 ⎦∼⎣ 0 0 −2 0 4 1 0 ⎤ 2 1 ⎦. 1 4. M2 (1/5), M3 (−1/2) (b). We have: rank(A) = 3, since the row-echelon form of A in (a) consists of 3 nonzero rows. (c). We have ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ −2 −3 1 1 0 0 1 4 2 0 1 0 1 4 2 0 1 0 1 2 ⎣ 1 4 2 0 1 0 ⎦ ∼ ⎣ −2 −3 1 1 0 0 ⎦ ∼ ⎣ 0 5 5 1 2 0 ⎦ 0 5 3 0 0 1 0 5 3 0 0 1 0 5 3 0 0 1 ⎤ ⎡ ⎤ ⎡ 0 1 0 1 0 1 4 2 1 4 2 0 3 4 5 1 2 0 ⎦ ∼ ⎣ 0 1 1 1/5 2/5 0 ⎦ ∼⎣ 0 5 0 0 −2 −1 −2 1 0 0 1 1/2 1 −1/2 ⎡ ⎤ ⎡ ⎤ 0 1/5 7/5 −1 1 0 −2 −4/5 −3/5 1 0 0 5 6 1 1/5 2/5 0 ⎦ ∼ ⎣ 0 1 0 −3/10 −3/5 1/2 ⎦ . ∼⎣ 0 1 0 0 1 1/2 0 0 1 1 −1/2 1/2 1 −1/2 (c)2017 Pearson Education. Inc. 200 1. P12 2. A12 (2) 3. A23 (−1) Thus, 4. M2 (1/5), M3 (−1/2) 5. A21 (−4) 6. A31 (2), A32 (−1) ⎡ ⎤ 1/5 7/5 −1 A−1 = ⎣ −3/10 −3/5 1/2 ⎦ . 1/2 1 −1/2 38. We use the Gauss-Jordan method to find A−1 : ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 1 0 0 1 0 0 1 −1 3 1 0 0 1 −1 3 1 −1 3 1 2 ⎣ 4 −3 13 0 1 0 ⎦ ∼ ⎣ 0 1 1 −4 1 0 ⎦ ∼ ⎣ 0 1 1 −4 1 0 ⎦ 1 1 4 0 0 1 0 2 1 −1 0 1 0 0 −1 7 −2 1 ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 1 0 0 0 4 1 −1 3 1 0 4 −3 1 1 0 0 25 −7 3 4 5 1 1 −4 1 0 ⎦ ∼ ⎣ 0 1 1 −4 1 0 ⎦∼⎣ 0 1 0 3 −1 1 ⎦. ∼⎣ 0 0 0 1 −7 2 −1 0 0 1 −7 2 −1 0 0 1 −7 2 −1 1. A12 (−4), A13 (−1) 2. A23 (−2) Thus, 3. M3 (−1) 4. A21 (1) 5. A31 (−4), A32 (−1) ⎡ ⎤ 25 −7 4 1 ⎦. A−1 = ⎣ 3 −1 −7 2 −1 Now xi = A−1 ei for each i. So ⎡ ⎤ 25 x1 = A−1 e1 = ⎣ 3 ⎦ , −7 ⎡ ⎤ −7 x2 = A−1 e2 = ⎣ −1 ⎦ , 2 39. We have xi = A−1 bi , where A−1 = − 1 39 −2 −5 −7 2 ⎡ ⎤ 4 x3 = A−1 e3 = ⎣ 1 ⎦ . −1 . Therefore, 1 1 1 −2 −5 1 −12 12 4 =− = = , −7 2 2 −3 3 39 39 13 1 1 1 1 −2 −5 4 −23 23 −1 =− = , x2 = A b2 = − 2 3 39 −7 39 −22 39 22 x1 = A−1 b1 = − and x3 = A−1 b3 = − 1 39 1 39 −2 −5 −7 2 −2 5 =− 1 39 −21 24 = 1 39 21 −24 40. (a). We have and (A−1 B)(B −1 A) = A−1 (BB −1 )A = A−1 In A = A−1 A = In (B −1 A)(A−1 B) = B −1 (AA−1 )B = B −1 In B = B −1 B = In . (c)2017 Pearson Education. Inc. = 1 13 7 −8 . 201 Therefore, (B −1 A)−1 = A−1 B. (b). We have (A−1 B)−1 = B −1 (A−1 )−1 = B −1 A, as required. 41(a). We have B 4 = (S −1 AS)(S −1 AS)(S −1 AS)(S −1 AS) = S −1 A(SS −1 )A(SS −1 )A(SS −1 )AS = S −1 AIAIAIAS = S −1 A4 S, as required. 41(b). We can prove this by induction on k. For k = 1, the result is B = S −1 AS, which was already given. Now assume that B k = S −1 Ak S. Then B k+1 = BB k = S −1 AS(S −1 Ak S) = S −1 A(SS −1 )Ak S = S −1 AIAk S = S −1 Ak+1 S, which completes the induction step. 42. (a). We reduce A to the identity matrix: 4 7 −2 5 1 ∼ 7 4 1 −2 7 4 17 2 1 0 ∼ 5 1. M1 ( 14 ) 2 3 ∼ 1 0 2 3. M2 ( 17 ) 2. A12 (2) 7 4 4 ∼ 1 1 0 0 1 . 4. A21 (− 74 ) The elementary matrices corresponding to these row operations are 1 1 0 1 0 0 4 , E2 = E1 = , , E3 = 2 2 1 0 1 0 17 E4 = 1 − 74 0 1 . We have E4 E3 E2 E1 A = I2 , so that A = E1−1 E2−1 E3−1 E4−1 = 4 0 0 1 1 −2 0 1 1 0 0 17 2 1 0 7 4 1 , which is the desired expression since Ei−1 is an elementary matrix for each i. (b). We can reduce A to upper triangular form by the following elementary row operation: 4 −2 7 5 1 ∼ 4 0 7 . 17 2 1. A12 ( 12 ) Therefore we have the multiplier m12 = − 12 . Hence, setting L= 1 − 12 0 1 and U= 4 0 7 17 2 , we have the LU factorization A = LU , which can be easily verified by direct multiplication. 43. (c)2017 Pearson Education. Inc. 202 (a). We reduce A to the identity matrix: ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ ⎡ 1 2 0 0 1 2 0 0 1 2 1 0 0 ⎢ 1 2 0 0 ⎥ 1 ⎢ 2 1 0 0 ⎥ 2 ⎢ 0 −3 0 0 ⎥ 3 ⎢ 0 ⎥ ⎢ ⎥ ⎢ ⎥∼⎢ ⎢ ⎣ 0 0 3 4 ⎦∼⎣ 0 0 3 4 ⎦∼⎣ 0 0 3 4 ⎦ ⎣ 0 0 0 4 3 0 0 0 4 3 0 0 4 3 ⎡ 1 0 0 0 1 0 5 ⎢ ∼⎢ ⎣ 0 0 1 0 0 4 ⎤ ⎡ ⎤ ⎡ 1 0 0 0 0 1 0 0 ⎢ 0 1 0 ⎥ 7 ⎢ 0 1 0 0 0 ⎥ 6 ⎥ ⎢ ⎥ ⎢ 4 ⎦∼⎣ 4 ⎦∼⎣ 0 0 1 0 0 1 3 3 3 0 0 0 0 0 0 − 73 1. P12 2. A12 (−2) 6. A34 (−4) 3. M2 (− 13 ) 2 1 0 0 ⎤ ⎡ 0 1 ⎢ 0 ⎥ 4 ⎢ 0 ⎥∼ 4 ⎦ ⎣ 0 3 0 ⎤ ⎡ 0 1 ⎢ 0 0 ⎥ 8 ⎥ ⎢ 4 ⎦∼⎣ 0 3 0 1 4. A21 (−2) 7. M4 (− 37 ) 0 0 3 4 8. A43 (− 43 ) 1 0 ⎢ 0 1 E5 = ⎢ ⎣ 0 0 0 0 0 0 1 3 0 ⎤ 0 0 ⎥ ⎥, 0 ⎦ 1 ⎡ 1 ⎢ 0 E6 = ⎢ ⎣ 0 0 ⎤ 0 0 0 1 0 0 ⎥ ⎥, 0 1 0 ⎦ 0 −4 1 ⎡ 1 ⎢ 0 E7 = ⎢ ⎣ 0 0 0 1 0 0 0 0 3 4 ⎤ 0 0 ⎥ ⎥. 0 ⎦ 1 0 0 1 0 ⎤ 0 0 0 0 ⎥ ⎥, 1 0 ⎦ 0 − 37 ⎤ 1 −2 0 0 ⎢ 0 1 0 0 ⎥ ⎥ E4 = ⎢ ⎣ 0 0 1 0 ⎦ 0 0 0 1 ⎡ ⎡ ⎤ 1 0 0 0 ⎢ 0 1 0 0 ⎥ ⎥ E8 = ⎢ ⎣ 0 0 1 −4 ⎦ . 3 0 0 0 1 We have E8 E7 E 6 E 5 E4 E3 E 2 E1 A = I 4 so that A = E1−1 E2−1 E3−1 E4−1 E5−1 E6−1 E7−1 E8−1 ⎤ ⎤⎡ ⎤⎡ ⎤⎡ ⎡ 1 2 0 0 1 0 0 0 1 0 0 0 0 1 0 0 ⎢ 1 0 0 0 ⎥ ⎢ 2 1 0 0 ⎥ ⎢ 0 −3 0 0 ⎥ ⎢ 0 1 0 0 ⎥ ⎥··· ⎥⎢ ⎥⎢ ⎥⎢ =⎢ ⎣ 0 0 1 0 ⎦⎣ 0 0 1 0 ⎦⎣ 0 0 1 0 ⎦⎣ 0 0 1 0 ⎦ 0 0 0 1 0 0 0 1 0 0 0 1 0 0 0 1 ⎤⎡ ⎤⎡ ⎡ ⎤⎡ 1 0 0 0 1 0 1 0 0 0 1 0 0 0 ⎢ 0 1 0 0 ⎥⎢ 0 1 0 0 ⎥⎢ 0 1 0 ⎥⎢ 0 1 0 ⎥⎢ ⎥⎢ ⎥⎢ ···⎢ ⎣ 0 0 3 0 ⎦⎣ 0 0 1 0 ⎦⎣ 0 0 1 0 ⎦⎣ 0 0 0 0 4 1 0 0 0 1 0 0 0 − 73 0 0 0 0 1 0 ⎤ 0 0 ⎥ ⎥ 4 ⎦, 3 1 which is the desired expression since Ei−1 is an elementary matrix for each i. (b). We can reduce A to upper triangular form by the following elementary row operations: ⎤ ⎡ ⎤ ⎡ ⎡ ⎤ 2 1 0 0 2 1 0 0 2 1 0 0 ⎢ 1 2 0 0 ⎥ 1 ⎢ 0 3 0 0 ⎥ 2 ⎢ 0 3 0 0 ⎥ 2 2 ⎥ ⎢ ⎥ ⎢ ⎢ ⎥. ⎣ 0 0 3 4 ⎦∼⎣ 0 0 3 4 ⎦∼⎣ 0 0 3 4 ⎦ 0 0 4 3 0 0 4 3 0 0 0 − 73 (c)2017 Pearson Education. Inc. ⎤ 0 0 ⎥ ⎥ 4 ⎦ 3 5. M3 ( 13 ) The elementary matrices corresponding to these row operations are ⎤ ⎤ ⎤ ⎡ ⎡ ⎡ 1 0 0 0 0 1 0 0 1 0 0 0 1 ⎥ ⎥ ⎢ 1 0 0 0 ⎥ ⎢ ⎢ ⎥ , E2 = ⎢ −2 1 0 0 ⎥ , E3 = ⎢ 0 − 3 0 0 ⎥ , E1 = ⎢ ⎣ 0 0 1 0 ⎦ ⎣ 0 0 1 0 ⎦ ⎣ 0 0 1 0 ⎦ 0 0 0 1 0 0 0 1 0 0 0 1 ⎡ 0 1 0 0 0 1 0 0 203 1. A12 (− 12 ) 2. A34 (− 43 ) Therefore, the nonzero multipliers are m12 = 12 and m34 = 43 . Hence, setting ⎡ ⎤ ⎤ ⎡ 1 0 0 0 2 1 0 0 ⎢ 1 1 0 0 ⎥ ⎢ 0 3 0 0 ⎥ 2 2 ⎥, ⎥ L=⎢ and U = ⎢ ⎣ 0 0 1 0 ⎦ ⎣ 0 0 3 4 ⎦ 0 0 43 1 0 0 0 − 73 we have the LU factorization A = LU , which can be easily verified by direct multiplication. 44. (a). We reduce A to the identity matrix: ⎤ ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ ⎡ ⎤ ⎡ 1 −1 2 1 −1 2 1 −1 2 1 −1 2 3 0 0 4 1 2 3 ⎣ 0 2 −1 ⎦ ∼ ⎣ 0 2 −1 ⎦ ∼ ⎣ 0 2 −1 ⎦ ∼ ⎣ 0 1 − 12 ⎦ 1 − 12 ⎦ ∼ ⎣ 0 3 0 0 0 3 −6 1 −1 2 0 3 −6 0 0 − 92 ⎤ ⎡ ⎡ ⎤ ⎤ ⎡ ⎤ ⎡ 3 1 −1 2 1 0 0 1 0 1 0 0 2 5 6 7 8 1 − 12 ⎦ ∼ ⎣ 0 1 − 12 ⎦ ∼ ⎣ 0 1 − 12 ⎦ ∼ ⎣ 0 1 0 ⎦ . ∼⎣ 0 0 0 1 0 0 1 0 0 1 0 0 1 1. P13 2. A13 (−3) 6. A21 (1) 3. M2 ( 12 ) 7. A31 (− 32 ) 4. A23 (−3) 8. A32 ( 12 ) 5. M3 (− 29 ) The elementary matrices corresponding to these row operations are ⎤ ⎤ ⎤ ⎡ ⎡ ⎤ ⎡ ⎡ 1 0 0 0 0 1 1 0 0 1 0 0 1 0 ⎦ E1 = ⎣ 0 1 0 ⎦ , E2 = ⎣ 0 1 0 ⎦ , E3 = ⎣ 0 12 0 ⎦ , E4 = ⎣ 0 1 0 0 0 −3 1 −3 0 1 0 0 1 ⎤ ⎤ ⎡ ⎡ ⎤ ⎡ ⎡ ⎤ 3 1 0 0 1 0 0 1 1 0 1 0 −2 0 ⎦ , E6 = ⎣ 0 1 0 ⎦ , E7 = ⎣ 0 1 0 ⎦ , E8 = ⎣ 0 1 12 ⎦ . E5 = ⎣ 0 1 2 0 0 1 0 0 1 0 0 1 0 0 −9 We have E 8 E7 E6 E 5 E4 E3 E2 E 1 A = I 3 so that A = E1−1 E2−1 E3−1 E4−1 E5−1 E6−1 E7−1 E8−1 ⎡ ⎤⎡ ⎤⎡ ⎤⎡ ⎤ 0 0 1 1 0 0 1 0 0 1 0 0 = ⎣ 0 1 0 ⎦⎣ 0 1 0 ⎦⎣ 0 2 0 ⎦⎣ 0 1 0 ⎦··· 1 0 0 3 0 1 0 0 1 0 3 1 ⎤⎡ ⎤⎡ ⎤⎡ ⎤ ⎡ 1 0 0 1 0 0 1 −1 0 1 0 32 0 ⎦⎣ 0 1 0 ⎦ ⎣ 0 1 0 ⎦ ⎣ 0 1 − 12 ⎦ , ···⎣ 0 1 9 0 0 1 0 0 1 0 0 1 0 0 −2 which is the desired expression since Ei−1 is an elementary matrix for each i. (b). We can reduce A to upper triangular form by the following elementary row operations: ⎤ ⎡ ⎡ ⎤ ⎤ ⎡ 3 0 0 3 0 0 3 0 0 1 2 ⎣ 0 2 −1 ⎦ ∼ ⎣ 0 2 −1 ⎦ ∼ ⎣ 0 2 −1 ⎦ . 3 1 −1 2 0 −1 2 0 0 2 (c)2017 Pearson Education. Inc. 204 1. A13 (− 13 ) 2. A23 ( 12 ) Therefore, the nonzero multipliers are m13 = 13 and m23 = − 12 . Hence, setting ⎡ 1 L=⎣ 0 1 3 0 1 − 12 ⎤ 0 0 ⎦ 1 ⎤ 0 0 2 −1 ⎦ , 3 0 2 ⎡ and 3 U =⎣ 0 0 we have the LU factorization A = LU , which can be verified by direct multiplication. 45. (a). We reduce A to the identity matrix: ⎡ ⎤ ⎡ ⎤ ⎡ −2 −3 1 1 4 2 1 1 2 ⎣ 1 4 2 ⎦ ∼ ⎣ −2 −3 1 ⎦ ∼ ⎣ 0 0 5 3 0 5 3 0 ⎡ 1 ∼⎣ 0 0 5 ⎤ ⎡ 4 2 1 6 1 −8 ⎦ ∼ ⎣ 0 0 45 0 1. P12 ⎤ ⎡ 4 2 1 7 1 −8 ⎦ ∼ ⎣ 0 0 1 0 2. A12 (2) 1 6. M3 ( 45 ) ⎤ ⎡ 4 2 1 3 5 5 ⎦∼⎣ 0 5 −3 0 ⎤ ⎡ 4 2 1 4 5 5 ⎦∼⎣ 0 1 −8 0 ⎤ ⎡ 0 34 1 8 1 −8 ⎦ ∼ ⎣ 0 0 1 0 0 1 0 3. A23 (−1) 7. A21 (−4) 4. P23 8. A32 (8) ⎤ 4 2 1 −8 ⎦ 5 5 ⎤ ⎡ 34 1 9 0 ⎦∼⎣ 0 1 0 0 1 0 ⎤ 0 0 ⎦. 1 5. A23 (−5) 9. A31 (−34) The elementary matrices corresponding to these row operations are ⎡ ⎤ ⎡ ⎤ ⎡ 0 1 0 1 0 0 1 E1 = ⎣ 1 0 0 ⎦ , E2 = ⎣ 2 1 0 ⎦ , E3 = ⎣ 0 0 0 0 1 0 0 1 ⎤ 0 0 1 0 ⎦, −1 1 ⎤ ⎤ ⎡ ⎤ ⎡ 1 0 0 1 0 0 1 0 0 1 0 ⎦ , E6 = ⎣ 0 1 0 ⎦ , E4 = ⎣ 0 0 1 ⎦ , E5 = ⎣ 0 1 0 1 0 0 −5 1 0 0 45 ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 1 −4 0 1 0 0 1 0 −34 1 0 ⎦ , E8 = ⎣ 0 1 8 ⎦ , E9 = ⎣ 0 1 0 ⎦. E7 = ⎣ 0 0 0 1 0 0 1 0 0 1 ⎡ We have E9 E8 E7 E 6 E5 E4 E3 E 2 E1 A = I 3 so that A = E1−1 E2−1 E3−1 E4−1 E5−1 E6−1 E7−1 E8−1 E9−1 ⎡ ⎤⎡ ⎤⎡ ⎤⎡ ⎤ 0 1 0 1 0 0 1 0 0 1 0 0 = ⎣ 1 0 0 ⎦ ⎣ −2 1 0 ⎦ ⎣ 0 1 0 ⎦ ⎣ 0 0 1 ⎦ · · · 0 0 1 0 0 1 0 1 1 0 1 0 ⎤⎡ ⎤⎡ ⎤⎡ ⎤⎡ ⎡ 1 0 0 1 1 4 0 1 0 0 1 0 0 · · · ⎣ 0 1 0 ⎦ ⎣ 0 1 0 ⎦ ⎣ 0 1 0 ⎦ ⎣ 0 1 −8 ⎦ ⎣ 0 0 0 0 1 0 0 1 0 0 45 0 5 1 (c)2017 Pearson Education. Inc. 0 1 0 ⎤ 34 0 ⎦, 1 205 which is the desired expression since Ei−1 is an elementary matrix for each i. (b). We can reduce A to upper triangular form by the following elementary row operations: ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ −2 −3 1 −2 −3 1 −2 −3 1 1 5 ⎦ 2 ⎣ 5 ⎦ 5 5 ⎣ 1 4 2 ⎦∼⎣ 0 0 ∼ . 2 2 2 2 0 5 3 0 5 3 0 0 −2 Therefore, the nonzero multipliers are m12 = − 12 and m23 = 2. Hence, setting ⎤ ⎤ ⎡ ⎡ −2 −3 1 1 0 0 5 5 ⎦ and U = ⎣ 0 L = ⎣ − 12 1 0 ⎦ , 2 2 0 2 1 0 0 −2 we have the LU factorization A = LU , which can be verified by direct multiplication. 46(a). Using the distributive laws of matrix multiplication, first note that (A+2B)2 = (A+2B)(A+2B) = A(A+2B)+2B(A+2B) = A2 +A(2B)+(2B)A+(2B)2 = A2 +2AB+2BA+4B 2 . Thus, we have (A + 2B)3 = (A + 2B)(A + 2B)2 = A(A + 2B)2 + 2B(A + 2B)2 = A(A2 + 2AB + 2BA + 4B 2 ) + 2B(A2 + 2AB + 2BA + 4B 2 ) = A3 + 2A2 B + 2ABA + 4AB 2 + 2BA2 + 4BAB + 4B 2 A + 8B 3 , as needed. 46(b). Each occurrence of B in the answer to part (a) must now be accompanied by a minus sign. Therefore, all terms containing an odd number of Bs will experience a sign change. The answer is (A − 2B)3 = A3 − 2A2 B − 2ABA − 2BA2 + 4AB 2 + 4BAB + 4B 2 A − 8B 3 . 47.The answer is 2k , because each term in the expansion of (A + B)k consists of a string of k matrices, each of which is either A or B (2 possibilities for each matrix in the string). Multiplying the possibilities for each position in the string of length k, we get 2k different strings, and hence 2k different terms in the expansion of (A + B)k . So, for instance, if k = 4, we expect 16 terms, corresponding to the 16 strings AAAA, AAAB, AABA, ABAA, BAAA, AABB, ABAB, ABBA, BAAB, BABA, BBAA, ABBB, BABB, BBAB, BBBA, and BBBB. Indeed, one can verify that the expansion of (A + B)4 is precisely the sum of the 16 terms we just wrote down. 48. We claim that To see this, simply note that and A 0 0 B −1 A−1 0 0 B A 0 0 B −1 A−1 0 −1 = 0 B A 0 0 B −1 A−1 0 = = 0 B In 0 0 Im In 0 0 Im (c)2017 Pearson Education. Inc. . = In+m = In+m . 206 49. For a 2 × 4 matrix, the leading ones can occur in 6 different positions: 1 ∗ ∗ 0 1 ∗ ∗ ∗ 1 , 0 ∗ 0 ∗ ∗ 1 ∗ 1 ∗ ∗ ∗ 0 , , 0 0 0 1 0 1 ∗ ∗ 0 1 ∗ 0 , 0 For a 3 × 4 matrix, the leading ones can occur in 4 different positions: ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ 1 ∗ ∗ ∗ 1 ∗ ∗ ∗ 1 ∗ ∗ ∗ 0 ⎣ 0 1 ∗ ∗ ⎦,⎣ 0 1 ∗ ∗ ⎦,⎣ 0 0 1 ∗ ⎦,⎣ 0 0 0 1 ∗ 0 0 0 1 0 0 0 1 0 1 ∗ ∗ 0 0 1 0 , 0 0 0 1 ∗ 0 1 ⎤ 1 ∗ ∗ 0 1 ∗ ⎦ 0 0 1 For a 4 × 6 matrix, the leading ones can occur in 15 different positions: ⎡ 1 ⎢ 0 ⎢ ⎣ 0 0 ⎡ 1 ⎢ 0 ⎢ ⎣ 0 0 ⎡ 1 ⎢ 0 ⎢ ⎣ 0 0 ∗ ∗ 1 ∗ 0 1 0 0 ∗ ∗ ∗ 1 ∗ ∗ ∗ 1 ∗ ∗ 0 0 1 0 0 0 ∗ ∗ ∗ 0 1 ∗ 0 0 0 0 0 0 ⎡ ⎤ ⎡ ⎤ ⎡ 1 ∗ ∗ ∗ 1 ∗ ∗ ∗ ∗ ∗ ∗ ∗ ∗ ∗ ∗ ⎢ 0 1 ∗ ∗ ∗ ∗ ⎥ ⎢ 0 1 ∗ ∗ 1 ∗ ∗ ∗ ∗ ⎥ ⎥,⎢ ⎥,⎢ 0 1 ∗ ∗ ∗ ⎦ ⎣ 0 0 1 ∗ ∗ ∗ ⎦ ⎣ 0 0 0 1 0 0 0 0 0 0 0 0 0 1 0 0 0 1 ∗ ⎤ ⎡ ⎤ ⎡ 1 ∗ ∗ ∗ ∗ ∗ ∗ ∗ ∗ ∗ ∗ 1 ∗ ∗ ∗ ⎢ 0 0 1 ∗ ∗ ∗ ⎥ ⎢ 0 0 1 ∗ 1 ∗ ∗ ∗ ∗ ⎥ ⎥,⎢ ⎥,⎢ 0 0 0 1 ∗ ⎦ ⎣ 0 0 0 1 ∗ ∗ ⎦ ⎣ 0 0 0 1 0 0 0 0 1 ∗ 0 0 0 0 1 0 0 0 0 ⎤ ⎡ ⎤ ⎡ 0 1 ∗ ∗ 0 1 ∗ ∗ ∗ ∗ ∗ ∗ ∗ ∗ ∗ ⎢ 0 0 1 ∗ ∗ ∗ ⎥ ⎢ 0 0 1 ∗ 0 0 1 ∗ ∗ ⎥ ⎥,⎢ ⎥,⎢ 0 0 0 1 ∗ ⎦ ⎣ 0 0 0 1 ∗ ∗ ⎦ ⎣ 0 0 0 1 0 0 0 0 0 0 0 0 1 ∗ 0 0 0 0 1 ⎤ ⎡ ⎤ ⎤ ⎡ ∗ ∗ ∗ ∗ 0 0 1 ∗ ∗ ∗ 0 1 ∗ ∗ ∗ ∗ ⎥ ⎥ ⎢ ⎢ 1 ∗ ∗ ∗ ⎥ ⎥,⎢ 0 0 0 1 ∗ ∗ ⎥,⎢ 0 0 0 1 ∗ ∗ ⎥ 0 0 1 ∗ ⎦ ⎣ 0 0 0 0 1 ∗ ⎦ ⎣ 0 0 0 0 1 ∗ ⎦ 0 0 0 1 0 0 0 0 0 1 0 0 0 0 0 1 ⎤ ⎡ 1 ∗ ∗ ⎢ 0 ∗ ∗ ⎥ ⎥,⎢ ∗ ∗ ⎦ ⎣ 0 0 ∗ ∗ ⎤ ⎡ 1 ∗ ∗ ⎢ 0 ∗ ∗ ⎥ ⎥,⎢ ∗ ∗ ⎦ ⎣ 0 0 0 1 ⎤ ⎡ 1 ∗ ∗ ⎢ 0 ∗ ∗ ⎥ ⎥,⎢ 1 ∗ ⎦ ⎣ 0 0 0 1 0 ⎢ 0 ⎢ ⎣ 0 0 1 0 0 0 ∗ ∗ ∗ 1 ∗ ∗ ∗ 0 ∗ ∗ ∗ 0 ⎤ ∗ ∗ ⎥ ⎥, ∗ ⎦ ∗ ⎤ ∗ ∗ ⎥ ⎥, ∗ ⎦ 1 ⎤ ∗ ∗ ⎥ ⎥, ∗ ⎦ 1 For an m × n matrix with m ≤ n, the answer is the binomial coefficient n! n C(n, m) = = . m m!(n − m)! This represents n “choose” m, which is the number of ways to choose m columns from the n columns of the matrix in which to put the leading ones. This choice then determines the structure of the matrix. 50. We claim that the inverse of A10 is B 5 . To prove this, use the fact that A2 B = I to observe that A10 B 5 = A2 A2 A2 A2 (A2 B)BBBB = A2 A2 A2 A2 IBBBB = A2 A2 A2 (A2 B)BBB = A2 A2 A2 IBBB = A2 A2 (A2 B)BB = A2 A2 IBB = A2 (A2 B)B = A2 IB = A2 B = I, as required. 51. We claim that the inverse of A9 is B 6 . To prove this, use the fact that A3 B 2 = I to observe that A9 B 6 = A3 A3 (A3 B 2 )B 2 B 2 = A3 A3 IB 2 B 2 = A3 (A3 B 2 )B 2 = A3 IB 2 = A3 B 2 = I, as required. (c)2017 Pearson Education. Inc. 207 Chapter 3 Solutions Solutions to Section 3.1 True-False Review: (a): TRUE. Let A = a b 0 c . Then det(A) = ac − b0 = ac, which is the product of the elements on the main diagonal of A. ⎡ ⎤ a b c (b): TRUE. Let A = ⎣ 0 d e ⎦. Using the schematic of Figure 3.1.1, we have 0 0 f det(A) = adf + be0 + c00 − 0dc − 0ea − f 0b = adf, which is the product of the elements on the main diagonal of A. (c): FALSE. The volume of this parallelepiped is determined by the absolute value of det(A), since det(A) could very well be negative. (d): TRUE. There are 12 of each. The ones of even parity are (1, 2, 3, 4), (1, 3, 4, 2), (1, 4, 2, 3), (2, 1, 4, 3), (2, 4, 3, 1), (2, 3, 1, 4), (3, 1, 2, 4), (3, 2, 4, 1), (3, 4, 1, 2), (4, 1, 3, 2), (4, 2, 1, 3), (4, 3, 2, 1), and the others are all of odd parity. 1 0 0 0 (e): FALSE. Many examples are possible here. If we take A = and B = , then det(A) = 0 0 0 1 det(B) = 0, but A + B = I2 , and det(I2 ) = 1 = 0. 1 2 (f ): FALSE. Many examples are possible here. If we take A = , for example, then det(A) = 3 4 1 · 4 − 2 · 3 = −2 < 0, even though all elements of A are positive. (g): TRUE. In the summation that defines the determinant, each term in the sum is a product consisting of one element from each row and each column of the matrix. But that means one of the factors in each term will be zero, since it must come from the row containing all zeros. Hence, each term is zero, and the summation is zero. (h): TRUE. If the determinant of the 3 × 3 matrix [v1 , v2 , v3 ] is zero, then the volume of the parallelepiped determined by the three vectors is zero, and this means precisely that the three vectors all lie in the same plane. (c)2017 Pearson Education. Inc. 208 (i): TRUE. Applying cofactor expansion to the first row of the given 4 × 4 matrix, we see that a1 a3 0 0 a2 a4 0 0 0 0 b1 b3 0 a4 0 = a1 0 b2 0 b4 0 b1 b3 b1 b3 = a1 a4 0 a3 b2 − a 2 0 b4 0 0 b1 b3 b2 b − a2 a3 1 b4 b3 b2 b4 = (a1 a4 − a2 a3 ) = a1 a3 a2 b · 1 a4 b3 b1 b3 0 b2 b4 b2 b4 b2 . b4 (j): TRUE. Applying cofactor expansion to the first column of the given 4 × 4 matrix, we see that 0 0 b1 b3 0 0 b2 b4 a1 a3 0 0 a2 0 a4 = b1 0 0 b4 0 = b1 b4 a1 a3 0 a1 a3 a2 0 a 4 − b3 0 0 b2 a1 a3 0 a2 a − b3 b2 1 a4 a3 a2 a4 = (b1 b4 − b3 b2 ) a1 a3 a2 a4 0 a2 = (b1 b4 − b3 b2 )(a1 a4 − a2 a3 ) a4 = b1 b3 b2 a · 1 b4 a3 a2 a4 = a1 a3 a2 b · 1 a4 b3 b2 . b4 Problems: 1. σ(3, 1, 4, 2) = (−1)N (3,1,4,2) = (−1)3 = −1, odd. 2. σ(2, 4, 3, 1) = (−1)N (2,4,3,1) = (−1)4 = 1, even. 3. σ(5, 4, 3, 2, 1) = (−1)N (5,4,3,2,1) = (−1)10 = 1, even. 4. σ(2, 4, 1, 5, 3) = (−1)N (2,4,1,5,3) = (−1)4 = 1, even. 5. σ(6, 1, 4, 2, 5, 3) = (−1)N (6,1,4,2,5,3) = (−1)8 = 1, even. 6. σ(6, 5, 4, 3, 2, 1) = (−1)N (6,5,4,3,2,1) = (−1)15 = −1, odd. a a12 7. det(A) = 11 = σ(1, 2)a11 a22 + σ(2, 1)a12 a21 = a11 a22 − a12 a21 . a21 a22 8. a11 a23 a34 a43 a52 . This is not a possible term of an order 5 determinant, since the column indices are not distinct. 9. a11 a25 a33 a42 a54 . All row and column indices are distinct, so this is a term of an order 5 determinant. Further, N (1, 5, 3, 2, 4) = 4, so that σ(1, 5, 3, 2, 4) = (−1)4 = +1. (c)2017 Pearson Education. Inc. 209 10. a11 a32 a24 a43 a55 . This is a possible term of an order 5 determinant. N (1, 4, 2, 3, 5) = 2 =⇒ σ(1, 4, 2, 3, 5) = (−1)2 = +1. 11. a13 a25 a31 a44 a42 . This is not a possible term of an order 5 determinant, since the row indices are not distinct. 12. a21 a3q ap2 a43 = ap2 a21 a3q a43 . We must choose p = 1 and q = 4. N (2, 1, 4, 3) = 2 and σ(2, 1, 4, 3) = (−1)2 = +1. 13. a13 ap4 a32 a2q = a13 a2q a32 ap4 . We must choose p = 4 and q = 1 in order for the row and column indices to be distinct. N (3, 1, 2, 4) = 2 so that σ(3, 1, 2, 4) = (−1)2 = +1. 14. apq a34 a13 a42 . We must choose p = 2 and q = 1. N (3, 1, 4, 2) = 3 and σ(3, 1, 4, 2) = (−1)3 = −1. 15. a3q ap4 a13 a42 . We must choose p = 2 and q = 1. N (3, 4, 1, 2) = 4 and σ(3, 4, 1, 2) = (−1)4 = +1. 0 −2 16. det = (0)(1) − (−2)(5) = 10. 5 1 6 −3 17. det = (6)(−1) − (−3)(−5) = −21. −5 −1 −4 7 18. det = (−4)(7) − (7)(1) = −35. 1 7 2 −3 19. det = (2)(5) − (−3)(1) = 13. 1 5 9 −8 20. det = (9)(−3) − (−8)(−7) = −83. −7 −3 2 −4 21. det = (2)(0) − (−4)(−1) = −4. −1 0 1 4 22. det = (1)(3) − (4)(−4) = 19. −4 3 −3 3e10 e = (e−3 )(6e8 ) − (3e10 )(2e−5 ) = 0. 23. det 2e−5 6e8 π π2 √ √ = 2π 2 − 2π 2 = (2 − 2)π 2 . √ 24. 2 2π 25: Using the schematic in Figure 3.1.1, we have ⎡ ⎤ 6 −1 2 7 1 ⎦ = (6)(7)(1) + (−1)(1)(0) + (2)(−4)(3) − (0)(7)(2) − (3)(1)(6) − (1)(−4)(−1) det ⎣ −4 0 3 1 = 42 + 0 + (−24) − 0 − 18 − 4 = −4. (c)2017 Pearson Education. Inc. 210 26: Using the schematic in Figure 3.1.1, we have ⎡ ⎤ 5 −3 0 4 −1 ⎦ = (5)(4)(−2) + (−3)(−1)(−8) + (0)(1)(2) − (−8)(4)(0) − (2)(−1)(5) − (−2)(1)(−3) det ⎣ 1 −8 2 −2 = −40 − 24 + 0 − 0 + 10 − 6 = −60. 27: Using the schematic in Figure 3.1.1, we have ⎡ ⎤ −2 −4 1 1 1 ⎦ = (−2)(1)(3) + (−4)(1)(−2) + (1)(6)(−1) − (−2)(1)(1) − (−1)(1)(−2) − (3)(6)(−4) det ⎣ 6 −2 −1 3 = −6 + 8 − 6 + 2 − 2 + 72 = 68. 28: Using the schematic in Figure 3.1.1, we have ⎡ ⎤ 0 0 −3 3 ⎦ = (0)(4)(5) + (0)(3)(−2) + (−3)(0)(1) − (−2)(4)(−3) − (1)(3)(0) − (5)(0)(0) det ⎣ 0 4 −2 1 5 = 0 + 0 + 0 − 24 − 0 − 0 = −24. 29: Using the schematic in Figure 3.1.1, we have ⎡ ⎤ 9 1 −7 1 ⎦ = (9)(2)(−2) + (1)(1)(−4) + (−7)(6)(0) − (−4)(2)(−7) − (0)(1)(9) − (−2)(6)(1) det ⎣ 6 2 −4 0 −2 = −36 − 4 + 0 − 56 − 0 + 12 = −84. 30: Using the schematic in Figure 3.1.1, we have ⎡ ⎤ 2 −10 3 1 1 ⎦ = (2)(1)(−3) + (−10)(1)(0) + (3)(1)(8) − (0)(1)(3) − (8)(1)(2) − (−3)(1)(−10) det ⎣ 1 0 8 −3 = −6 + 0 + 24 − 0 − 16 − 30 = −28. 31: Using the schematic in Figure 3.1.1, we have ⎡ ⎤ 5 4 3 9 12 ⎦ = (5)(9)(0) + (4)(12)(1) + (3)(−2)(−1) − (1)(9)(3) − (−1)(12)(5) − (0)(−2)(4) det ⎣ −2 1 −1 0 = 0 + 48 + 6 − 27 + 60 − 0 = 87. (c)2017 Pearson Education. Inc. 211 32: Using the schematic in Figure 3.1.1, we have ⎡ 5 det ⎣ 0 2 0 3 0 ⎤ 4 0 ⎦ = (5)(3)(1) + (0)(0)(2) + (4)(0)(0) − (2)(3)(4) − (0)(0)(5) − (1)(0)(0) 1 = 15 + 0 + 0 − 24 − 0 − 0 = −9. √ e2 e−1 √π √ √ √ 33. 67 1/30 2001 = ( π)(1/30)(π 3 )+(e2 )(2001)(π)+(e−1 )( 67)(π 2 )−(π)(1/30)(e−1 )−(π 2 )(2001)( π)− π π2 π3 √ 3 2 (π )( 67)(e ) ≈ 9601.882. 2 3 −1 1 = (2)(4)(6) + (3)(1)(3) + (−1)(1)(1) − (3)(4)(−1) − (1)(1)(2) − (6)(1)(3) = 48. 34. 1 4 3 1 6 35. The definition of the determinant for this 4 × 4 matrix includes 24 terms, but based on the positions occupied by zeros in this matrix, only one of these 24 terms will be nonzero; namely, det(A) = σ(4, 3, 2, 1)a14 a23 a32 a41 = (−1)6 (−3)(−7)(2)(1) = 42. 36. The definition of the determinant for this 4 × 4 matrix includes 24 terms, but based on the positions occupied by zeros in this matrix, only one of these 24 terms will be nonzero; namely, det(A) = σ(1, 2, 3, 4)a11 a22 a33 a44 = (−1)0 (4)(−2)(−6)(−3) = −144. 37. The definition of the determinant for this 4 × 4 matrix includes 24 terms, but based on the positions occupied by zeros in this matrix, many of these terms will be zero. The nonzero terms will correspond to the permutations that we have enumerated in the table below, along with the number of inversions, the sign, and the value of the term before adopting the sign: Permutation (1, 2, 4, 3) (1, 3, 2, 4) (1, 3, 4, 2) (1, 4, 2, 3) (3, 1, 2, 4) (3, 1, 4, 2) (3, 2, 1, 4) (3, 4, 1, 2) (4, 1, 2, 3) (4, 2, 1, 3) (4, 3, 1, 2) Number of Inversions 1 1 2 2 2 3 3 4 3 4 5 Sign (+/−) − − + + + − − + − + − Value of Term 72 −4 30 −32 2 −15 −12 −40 24 −144 −60 Thus, det(A) = −72 − (−4) + 30 + (−32) + 2 − (−15) − (−12) + (−40) − 24 + (−144) − (−60) = −189. 38. The definition of the determinant for this 4 × 4 matrix includes 24 terms, but based on the positions occupied by zeros in this matrix, many of these terms will be zero. The nonzero terms will correspond to the permutations that we have enumerated in the table below, along with the number of inversions, the sign, and the value of the term before adopting the sign: (c)2017 Pearson Education. Inc. 212 Permutation (1, 2, 3, 4) (1, 2, 4, 3) (1, 4, 2, 3) (1, 4, 3, 2) Number of Inversions 0 1 2 3 Sign (+/−) + − + − Value of Term −6 −20 120 −96 Thus, det(A) = −6 − (−20) + 120 − (−96) = 230. 39. The definition of the determinant for this 4 × 4 matrix includes 24 terms, but based on the positions occupied by zeros in this matrix, many of these terms will be zero. The nonzero terms will correspond to the permutations that we have enumerated in the table below, along with the number of inversions, the sign, and the value of the term before adopting the sign: Permutation (1, 2, 3, 4) (1, 2, 4, 3) (2, 1, 3, 4) (2, 1, 4, 3) Number of Inversions 0 1 1 2 Sign (+/−) + − − + Value of Term −16 −24 −8 −12 Thus, det(A) = (−16) − (−24) − (−8) + (−12) = 4. 40. The definition of the determinant for this 5×5 matrix includes 5! = 120 terms, but based on the positions occupied by zeros in this matrix, many of these terms will be zero. The nonzero terms will correspond to the permutations that we have enumerated in the table below, along with the number of inversions, the sign, and the value of the term before adopting the sign: Permutation (1, 2, 3, 4, 5) (1, 2, 3, 5, 4) (1, 3, 2, 4, 5) (1, 3, 2, 5, 4) (2, 1, 3, 4, 5) (2, 1, 3, 5, 4) (2, 3, 1, 4, 5) (2, 3, 1, 5, 4) (3, 1, 2, 4, 5) (3, 1, 2, 5, 4) (3, 2, 1, 4, 5) (3, 2, 1, 5, 4) Number of Inversions 0 1 1 2 1 2 2 3 2 3 3 4 Sign (+/−) + − − + − + + − + − − + Value of Term 8 3 −32 −12 −32 −12 −384 −144 −48 −18 144 54 Thus, det(A) = 8−3−(−32)+(−12)−(−32)+(−12)+(−384)−(−144)+(−48)−(−18)−144+54 = −315. 41. The definition of the determinant for this 5×5 matrix includes 5! = 120 terms, but based on the positions occupied by zeros in this matrix, many of these terms will be zero. The nonzero terms will correspond to the permutations that we have enumerated in the table below, along with the number of inversions, the sign, and the value of the term before adopting the sign: Permutation (1, 2, 3, 4, 5) (1, 2, 3, 5, 4) (2, 1, 3, 4, 5) (2, 1, 3, 5, 4) Number of Inversions 0 1 1 2 Sign (+/−) + − − + (c)2017 Pearson Education. Inc. Value of Term 1080 1120 1620 1680 213 Thus, det(A) = 1080 − 1120 − 1620 + 1680 = 20. 42. The definition of the determinant for this 5×5 matrix includes 5! = 120 terms, but based on the positions occupied by zeros in this matrix, many of these terms will be zero. The nonzero terms will correspond to the permutations that we have enumerated in the table below, along with the number of inversions, the sign, and the value of the term before adopting the sign: Permutation (4, 5, 3, 1, 2) (4, 5, 3, 2, 1) (5, 4, 3, 1, 2) (5, 4, 3, 2, 1) Number of Inversions 8 9 9 10 Sign (+/−) + − − + Value of Term −64 −192 32 96 Thus, det(A) = (−64) − (−192) − 32 + 96 = 192. 6t e4t e 43. We have det = (e6t )(4e4t ) − (e4t )(6e6t ) = −2e10t . 6e6t 4e4t 44. Using the schematic in Figure 3.1.1, we have ⎡ ⎤ sin t cos t 1 det ⎣ cos t − sin t 0 ⎦ = (sin t)(− sin t)(0) + (cos t)(0)(sin t) + (1)(cos t)(− cos t) sin t − cos t 0 − (sin t)(− sin t)(1) − (− cos t)(0)(sin t) − (0)(cos t)(cos t) = − cos2 t + sin2 t. 2t 3t e e−4t 2t e 3t 3e −4e−4t = (e2t )(3e3t )(16e−4t )+(e3t )(−4e−4t )(4e2t )+(e−4t )(2e2t )(9e3t )−(4e2t )(3e3t )(e−4t )− 45. 2e 4e2t 9e3t 16e−4t (9e3t )(−4e−4t )(e2t ) − (16e−4t )(2e2t )(e3t ) = 42et . 46. Using the schematic in Figure 3.1.1, we have ⎤ ⎡ −t e−5t e2t e det ⎣ −e−t −5e−5t 2e2t ⎦ = (e−t )(−5e−5t )(4e2t ) + (e−5t )(2e2t )(e−t ) + (e2t )(−e−t )(25e−5t ) e−t 25e−5t 4e2t − (e−t )(−5e−5t )(e2t ) − (25e−5t )(2e2t )(e−t ) − (4e2t )(−e−t )(e−5t ) = −20e−4t + 2e−4t − 25e−4t + 5e−4t − 50e−4t + 4e−4t = −84e−4t . 47. y1 − y1 + 4y1 − 4y1 = 8 sin 2x + 4 cos 2x − 8 sin 2x − 4 cos 2x = 0, y2 − y2 + 4y2 − 4y2 = −8 cos 2x + 4 sin 2x + 8 cos 2x − 4 sin 2x = 0, y3 − y3 + 4y3 − 4y3 = ex − ex + 4ex − 4ex = 0. y1 y 2 y 3 cos 2x sin 2x ex y1 y2 y3 = −2 sin 2x 2 cos 2x ex y1 y2 y3 −4 cos 2x −4 sin 2x ex = 2ex cos2 2x − 4ex sin 2x cos 2x + 8ex sin2 2x + 8ex cos2 2x + 4ex sin 2x cos 2x + 2ex sin2 2x = 10ex . (c)2017 Pearson Education. Inc. 214 48. (a). y1 − y1 − y1 + y1 = ex − ex − ex + ex = 0, y2 − y2 − y2 + y2 = sinh x − cosh x − sinh x + cosh x = 0, y3 − y3 − y3 + y3 = cosh x − sinh x − cosh x + sinh x = 0. y1 y1 y1 y2 y2 y2 y3 ex y3 = ex y3 ex cosh x sinh x cosh x sinh x cosh x sinh x = ex sinh2 x + ex cosh2 x + ex sinh x cosh x − ex sinh2 x − ex cosh2 x − ex sinh x cosh x = 0. (b). The formulas we need are cosh x = ex + e−x 2 and sinh x = ex − e−x . 2 Adding the two equations, we find that cosh x + sinh x = ex , so that −ex + cosh x + sinh x = 0. Therefore, we may take d1 = −1, d2 = 1, and d3 = 1. 49. (a). S4 = {1, 2, 3, 4} (1, 2, 3, 4) (1, 2, 4, 3) (1, 3, 2, 4) (1, 3, 4, 2) (1, 4, 2, 3) (1, 4, 3, 2) (2, 1, 3, 4) (2, 1, 4, 3) (2, 3, 1, 4) (2, 3, 4, 1) (2, 4, 1, 3) (2, 4, 3, 1) (3, 1, 2, 4) (3, 1, 4, 2) (3, 2, 1, 4) (3, 2, 4, 1) (3, 4, 1, 2) (3, 4, 2, 1) (4, 1, 2, 3) (4, 1, 3, 2) (4, 2, 1, 3) (4, 2, 3, 1) (4, 3, 1, 2) (4, 3, 2, 1) (b). N (1, 2, 3, 4) = 0, σ(1, 2, 3, 4) = 1, even; N (1, 3, 2, 4) = 1, σ(1, 3, 2, 4) = −1, odd; N (1, 4, 2, 3) = 2, σ(1, 4, 2, 3) = 1, even; N (2, 1, 3, 4) = 1, σ(2, 1, 3, 4) = −1, odd; N (2, 3, 1, 4) = 2, σ(2, 3, 1, 4) = 1, even; N (2, 4, 1, 3) = 3, σ(2, 4, 1, 3) = −1, odd; N (3, 1, 2, 4) = 2, σ(3, 1, 2, 4) = 1, even; N (3, 2, 1, 4) = 3, σ(3, 2, 1, 4) = −1, odd; N (3, 4, 1, 2) = 4, σ(3, 4, 1, 2) = 1, even; N (4, 1, 2, 3) = 3, σ(4, 1, 2, 3) = −1, odd; N (4, 2, 1, 3) = 4, σ(4, 2, 1, 3) = 1, even; N (4, 3, 1, 2) = 5, σ(4, 3, 1, 2) = −1, odd; (c). N (1, 2, 4, 3) = 1, σ(1, 2, 4, 3) = −1, odd; N (1, 3, 4, 2) = 2, σ(1, 3, 4, 2) = 1, even; N (1, 4, 3, 2) = 3, σ(1, 4, 3, 2) = −1, odd; N (2, 1, 4, 3) = 2, σ(2, 1, 4, 3) = 1, even; N (2, 3, 4, 1) = 3, σ(2, 3, 4, 1) = −1, odd; N (2, 4, 3, 1) = 4, σ(2, 4, 3, 1) = 1, even; N (3, 1, 4, 2) = 3, σ(3, 1, 4, 2) = −1, odd; N (3, 2, 4, 1) = 4, σ(3, 2, 4, 1) = 1, even; N (3, 4, 2, 1) = 5, σ(3, 4, 2, 1) = −1, odd; N (4, 1, 3, 2) = 4, σ(4, 1, 3, 2) = 1, even; N (4, 2, 3, 1) = 5, σ(4, 2, 3, 1) = −1, odd; N (4, 3, 2, 1) = 6, σ(4, 3, 2, 1) = 1, even. (c)2017 Pearson Education. Inc. 215 det(A) = a11 a22 a33 a44 − a11 a22 a34 a43 − a11 a23 a32 a44 + a11 a23 a34 a42 + a11 a24 a32 a43 − a11 a24 a33 a42 − a12 a21 a33 a44 + a12 a21 a34 a43 + a12 a23 a31 a44 − a12 a23 a34 a41 − a12 a24 a31 a43 + a12 a24 a33 a41 + a13 a21 a32 a44 − a13 a21 a34 a42 − a13 a22 a31 a44 + a13 a22 a34 a41 + a13 a24 a31 a42 − a13 a24 a32 a41 − a14 a21 a32 a43 + a14 a21 a33 a42 + a14 a22 a31 a43 − a14 a22 a33 a41 − a14 a23 a31 a42 + a14 a23 a32 a41 50. The definition of the determinant for this 4 × 4 matrix includes 24 terms, but based on the positions occupied by zeros in this matrix, many of these terms will be zero. The nonzero terms will correspond to the permutations that we have enumerated in the table below, along with the number of inversions, the sign, and the value of the term before adopting the sign: Permutation (1, 2, 3, 4) (1, 3, 2, 4) (1, 3, 4, 2) (1, 4, 3, 2) (3, 2, 1, 4) (3, 4, 1, 2) (4, 3, 1, 2) Number of Inversions 0 1 2 3 3 4 5 Sign (+/−) + − + − − + − Value of Term 96 8 12 −16 24 −4 10 Thus, det(A) = 96 − 8 + 12 − (−16) − 24 + (−4) − 10 = 78. 51. The definition of the determinant for this 4 × 4 matrix includes 24 terms, but based on the positions occupied by zeros in this matrix, many of these terms will be zero. The nonzero terms will correspond to the permutations that we have enumerated in the table below, along with the number of inversions, the sign, and the value of the term before adopting the sign: Permutation (1, 3, 2, 4) (1, 3, 4, 2) (1, 4, 2, 3) (1, 4, 3, 2) (2, 1, 3, 4) (2, 1, 4, 3) (2, 3, 1, 4) (2, 4, 1, 3) (3, 1, 2, 4) (3, 1, 4, 2) (3, 4, 1, 2) Number of Inversions 1 2 2 3 1 2 2 3 2 3 4 Sign (+/−) − + + − − + + − + − + Value of Term 4 −6 −2 −6 144 −72 −32 16 −84 126 −28 Thus, det(A) = −4 + (−6) + (−2) − (−6) − 144 + (−72) + (−32) − 16 + (−84) − 126 + (−28) = −508. 52. The definition of the determinant for this 5 × 5 matrix includes 5! = 120 terms, but based on the positions occupied by zeros in this matrix, many of these terms will be zero. In fact, the only nonzero terms will consist of permutations whose first entry is 1. Thus, a11 = 2 will be a factor in each of the nonzero terms. The 24 permutations of {2, 3, 4, 5} will yield the same terms that we already found in Problem 50, since the 4 × 4 submatrix consisting of the last four rows and columns is the same as the matrix in Problem 50. Therefore, the answer here is simply 2 · 78 = 156. 53. The definition of the determinant for this 5 × 5 matrix includes 5! = 120 terms, but based on the positions occupied by zeros in this matrix, many of these terms will be zero. In fact, the only nonzero terms (c)2017 Pearson Education. Inc. 216 will consist of permutations whose last entry is 5. Thus, a55 = −3 will be a factor in each of the nonzero terms. The 24 permutations of {1, 2, 3, 4} will yield the same terms that we already found in Problem 51, since the 4 × 4 submatrix consisting of the first four rows and columns is the same as the matrix in Problem 51. Therefore, the answer here is simply −3 · (−508) = 1524. 54. (a). det(cA) = ca11 ca21 = c2 a11 a22 − c2 a12 a21 = c2 (a11 a22 − a12 a21 ) = c2 det(A). ca12 = (ca11 )(ca22 ) − (ca12 )(ca21 ) ca22 (b). det(cA) , = σ(p1 , p2 , p3 , . . . , pn )ca1p1 ca2p2 ca3p3 · · · canpn , σ(p1 , p2 , p3 , . . . , pn )a1p1 a2p2 a3p3 · · · anpn cn = cn det(A), = where each summation above runs over all permutations σ of {1, 2, 3, . . . , n}. 55. (a): 123 = 1, (b): Consider 132 = −1, 3 , 3 3 , , 213 = −1, 231 = 1, 312 = 1, 321 = −1. ijk a1i a2j a3k . The only nonzero terms arise when i, j, and k are distinct. Conse- i=1 j=1 k=1 quently, 3 , 3 3 , , ijk a1i a2j a3k = 123 a11 a22 a33 + 132 a11 a23 a32 + 213 a12 a21 a33 = = + 231 a12 a23 a31 + 312 a13 a21 a32 + 321 a13 a22 a31 a11 a22 a33 + a12 a23 a31 + a13 a21 a32 − a11 a23 a32 − a12 a21 a33 − a13 a22 a31 det(A). i=1 j=1 k=1 56. From the given term, we have N (n, n − 1, n − 2, . . . , 1) = 1 + 2 + 3 + · · · + (n − 1) = n(n − 1) , 2 because the series of (n−1) terms is just an arithmetic series which has a first term of one, common difference of one, and last term (n − 1). Thus, σ(n, n − 1, n − 2, . . . , 1) = (−1)n(n−1)/2 . Solutions to Section 3.2 True-False Review: n (a): FALSE. The determinant of the matrix will increase by a factor of 2 . For instance, if A = I2 , then 2 0 det(A) = 1. However, det(2A) = det = 4, so the determinant in this case increases by a factor of 0 2 four. (c)2017 Pearson Education. Inc. 217 (b): TRUE. In both cases, the determinant of the matrix is multiplied by a factor of c. (c): TRUE. This follows by repeated application of Property (P8): det(A5 ) = det(AAAAA) = (detA)(detA)(detA)(detA)(detA) = (detA)5 . (d): TRUE. Since det(A2 ) = (detA)2 and since det(A) is a real number, det(A2 ) must be nonnegative. (e): FALSE. The matrix is not invertible if and only if its determinant, x2 y − xy 2 = xy(x − y), is zero. For example, if x = y = 1, the matrix is not invertible since x = y; however, neither x nor y is zero in this case. We conclude that the statement is false. (f ): TRUE. We have det(AB) = (detA)(detB) = (detB)(detA) = det(BA). Problems: From this point on, P2 will denote an application of the property of determinants that states that if every element in any row (column) of a matrix A is multiplied by a number c, then the determinant of the resulting matrix is c det(A). 1. −2 5 1 −2 5 2 1 6 3 1 = =− =− 5 −4 1 6 −2 5 0 6 = −17. 17 1. A12 (2) 2. 2. P12 1 1 1 2 3 2 3 2 3 1 1 2 3 2 2 −2 = 2 0 1 −1 = 2 0 6 4 = 0 0 −11 −7 0 3 −5 2 0 −11 −7 1. A12 (−2), A13 (−3) 3. 2. P2 2 3 1 −1 = 2(−18) = −36. 0 −18 3. A23 (11) 2 −1 4 2 −1 4 2 2 1 −1/2 2 3 1 −1/2 1 3 2 1 = 2 3 2 1 = 2 0 7/2 −5 = 2 · 28 = 56. 2 1 = 3 −2 0 0 8 0 0 8 0 0 8 1 4 1. A13 (1) 4. 3. A12 (2) 2 −1 4 1 2 1 −1 3 1 6 = − 2 4 12 2. P2 3. A12 (−3) −1 2 6 −1 2 6 6 2 3 3 = − 0 5 15 = −5 · 9 0 1 3 0 9 36 0 1 4 12 −1 2 6 4 = −45 0 1 3 = (−45)(−1) = 45. 0 0 1 1. P12 2 1 1 2. A12 (2), A13 (4) 3. P2 (c)2017 Pearson Education. Inc. 4. A23 (−1) 218 5. 0 1 −2 1 1 −1 0 3 = 0 2 −3 0 2 0 −3 1 0 −3 1 2 3 1 −2 = 0 1 −2 = 0 −3 0 0 −3 6 0 1. P12 , P2 6. 3 5 2 7 1 1 1 9 −6 = 5 2 1 3 1. A31 (−1) 7. 2. A13 (−2) 0 −3 1 −2 = 0. 0 0 3. A23 (3) 6 −2 1 6 −2 1 6 −2 2 3 9 −6 = 0 −21 4 = 0 1 −10 1 3 0 −11 7 0 −11 7 1 6 −2 4 = −103. 0 1 −10 = 0 0 −103 2. A12 (−5), A13 (−2) 3. A32 (−2) 4. A24 (11) 1 −1 2 4 1 −1 1 −1 1 −1 2 4 2 4 2 4 3 0 0 1 2 4 1 0 4 −4 −8 2 1 −1 −2 3 1 −1 −2 = 0 = 3 · 4 0 = −12 0 −1 1 3 2 0 5 6 0 5 6 1 0 −2 2 0 0 1 4 2 0 3 0 −6 1 0 −2 0 5 6 1 −1 1 −1 2 4 2 4 0 0 1 −1 −2 1 −1 −2 5 4 = −12 = −12 · 6 = −72. = −12 0 1 0 0 1 0 0 0 0 0 0 0 6 0 5 6 1. A12 (−3), A13 (1), A14 (−2) 2. P2 3. P34 4. A23 (−1) 5. A34 (−5) 8. Note that in the first step below, we extract a factor of 13 from the second row of the matrix, and we also extract a factor of 8 from the second column of the matrix. 2 32 1 2 4 1 5 −1 1 4 4 5 26 104 26 −13 1 2 1 2 −1 2 2 −1 = 13 · 8 2 1 = −104 2 2 7 2 7 56 2 7 2 7 2 7 1 1 5 −1 5 2 4 1 4 40 1 5 1 1 5 1 5 5 1 5 0 −9 0 0 −11 4 1 0 1 3 = −104 = −104(−3) 0 −11 0 −3 0 −3 0 −9 0 −6 −1 −6 0 −6 −1 −6 1 5 1 5 1 5 1 5 0 1 0 1 0 1 6 0 1 5 = 312 = 312 0 −2 0 0 0 0 0 −1 0 0 −1 0 0 0 0 −2 = 312(−1)(−2) = 624. (c)2017 Pearson Education. Inc. 219 1. P2 2. P14 3. A12 (−2), A13 (−2), A14 (−2) 0 1 −1 −1 0 1 1 −1 0 −1 −1 −1 1 0 3 = − 0 0 9. 1. P13 2 3 4 5 10. 1 0 1 2 1 −1 1 0 1 1 −1 0 1 = − 1 0 1 −1 −1 −1 −1 0 1 −1 0 1 −1 1 2 4 0 = − 0 0 0 3 0 0 −3 −3 2. A12 (1), A14 (1) 4. P2, P23 5. A23 (9), A24 (6) 6. P34 1 −1 1 0 1 1 2 0 −1 1 2 = − 1 1 −1 1 0 0 0 −2 −1 1 −1 0 1 −1 1 2 = 9. 0 −3 −3 0 0 3 3. A23 (1), A24 (−2) 4. P34 1 2 3 5 1 2 5 3 5 2 1 0 3 1 2 2 0 3 1 2 = − = − 0 2 3 1 4 4 3 1 −2 2 5 5 3 0 1 −1 −7 3 1 2 3 4 1 2 3 4 0 1 −1 −7 4 0 1 −1 −7 3 = = − 3 12 1 −2 0 0 0 2 0 3 4 23 1 2 0 0 1 2 1 2 3 4 3 4 0 1 −1 −7 6 0 1 −1 −7 5 = 3 = 3 0 0 1 4 1 4 0 0 0 0 0 0 4 23 0 7 3 1 4 5 = 3 · 1 · 1 · 1 · 7 = 21. 1. CP12 11. 2. A13 (−1), A14 (−2) 3. P24 4. A23 (−2), A24 (−3) 5. P2 6. A34 (−4) 2 −1 3 −1 1 −2 −3 −4 4 2 3 4 7 1 1 2 3 1 7 2 3 2 1 7 2 3 = −6 = 6 −2 4 −2 8 4 −2 4 8 6 6 8 6 6 −6 18 −24 −1 1 3 −4 −1 1 3 −4 1 −2 −3 −4 1 −2 −3 −4 1 −2 −3 −4 0 0 −1 0 −1 9 5 7 4 0 −8 5 0 −8 3 = 6 = −6 = −6 0 0 6 20 22 6 20 22 0 20 −26 0 0 −1 0 −8 0 0 5 −65 0 9 5 7 1 −2 −3 −4 1 −2 −3 −4 0 −1 0 −8 7 0 −1 0 −8 6 = 6 · 1(−1) · 5 · 234 = −7020. = 6 = 6 0 0 5 −65 0 5 −65 0 0 0 0 20 −26 0 0 234 (c)2017 Pearson Education. Inc. 220 1. CP12 , P2 2. M1 (−1) 5. A23 (6), A24 (9) 6. P34 3. A12 (−1), A13 (−4), A14 (1) 4. P24 7. A34 (−4) 12. 7 −1 3 4 1 −1 3 2 1 −1 3 2 14 2 0 2 4 6 1 2 4 3 2 4 −2 −1 = 7 · 2 3 21 = 14 0 1 3 4 1 3 2 4 −6 −4 −7 −1 0 4 5 8 4 5 4 3 8 6 1 −1 1 −1 1 −1 3 2 3 2 3 2 0 0 0 4 −2 −1 4 1 −10 −7 5 1 −10 −7 3 = 14 = 14 = 14 0 −4 −3 0 −4 −3 0 −4 −3 0 0 0 0 3 8 6 0 3 8 6 0 0 38 27 1 −1 3 0 1 −10 6 = −56 0 1 0 0 0 38 1. P2 1 −1 2 3 2 0 7 7 1 −10 −7 = −56 · (−3/2) = 84. = −56 3/4 0 1 3/4 0 27 0 0 0 −3/2 2. A12 (−2), A13 (−3), A14 (1) 5. A24 (−3) 6. P2 3. A23 (−1) 4. A42 (−1) 7. A34 (−38) 13. 3 7 1 1 1 −1 4 8 −1 3 7 0 8 16 −1 1 0 3 = − 0 0 0 1. P12 1 4 0 0 0 −1 4 −1 −1 −1 1 2 3 3 0 1 1 6 6 = − 4 3 9 4 8 8 12 0 2 4 7 4 1 1 0 0 4 2 = − 0 0 1 0 4 1 1 −1 0 1 0 7 1 2 3 2 8 −1 6 6 = − 0 0 7 0 9 4 0 16 −1 8 12 1 −1 0 1 4 4 2 0 4 3 6 2 4 3 9 1 8 7 8 4 1 −1 0 1 4 4 2 0 0 −1 4 2 = −(1)(4)(−1)(3)(2) = 24. 0 0 3 −1 0 0 0 2 2. A12 (−3), A13 (−4), A14 (−3), A15 (−8) 3. A23 (−1), A24 (−1), A25 (−2) (c)2017 Pearson Education. Inc. 4. A34 (−1), A35 (−1) 221 14. 1 5 2 6 0 2 4 3 5 2 3 3 4 4 4 4 2 5 3 6 5 1 2 3 4 5 1 0 −6 −12 −18 −24 1 6 = 0 −1 −2 −3 −4 2 0 −7 −14 −21 −28 8 0 2 4 6 8 1 2 3 4 5 0 1 2 3 4 2 = −6 0 −1 −2 −3 −4 0 −7 −14 −21 −28 0 2 4 6 8 1 0 3 = −6 0 0 0 2 1 0 0 0 3 2 0 0 0 4 3 0 0 0 5 4 0 0 0 = −6 · 0 = 0. 1. A12 (−5), A13 (−2), A14 (−6) 2. M2 (− 16 ) 3. A23 (1), A24 (7), A25 (−2) −1 1 15. = 0; not invertible. 1 −1 2 1 = 1; invertible. 16. 3 2 −1 2 3 17. 5 −2 1 = −8; invertible. 8 −2 5 2 6 −1 1 = 14; invertible. 18. 3 5 2 0 1 19. 1 1 1 1 1 −1 1 −1 1 0 1 1 −1 −1 = 0 −1 1 1 −1 0 1 1 1 2 0 2 2 0 2 = 0 −2 −2 = 16; invertible. 0 −2 −2 2 2 0 2 2 0 1 0 2 −1 1 0 2 −1 3 −2 1 0 −2 −5 −2 −5 7 4 7 = = 1 2 4 = 133; invertible. 2 1 6 2 0 1 2 4 −3 2 1 1 −3 4 0 0 −3 2 1 1 1 2 −3 5 2 −3 5 −1 3 −6 2 2 −3 6 1 1 −2 1. M2 (−1) 2. P7 = − = 0; not invertible. 21. 3 −1 4 3 −1 4 2 2 1 −2 1 −2 3 −6 3 −6 20. (c)2017 Pearson Education. Inc. 222 1 2 k 1 2 k = (3k −1)(k +4). Consequently, the system has an infinite 1 22. det(A) = 2 −k 1 = 2 −k 3 6 1 0 0 1 − 3k number of solutions if and only if k = −4 or k = 1/3 (Corollary 3.2.6). b1 1 k x1 , and b = . According to Corollary 23. The system is Ax = b, where A = , x= x2 b2 k 4 3.2.6, the system has unique solution if and only if det(A) = 0. But, det(A) = 4 − k 2 , so that the system has a unique solution if and only if k = ±2. 1 k 0 24. det(A) = k 1 1 = 1 + k − 1 − k 2 = k(1 − k). Consequently, the system has a unique solution if 1 1 1 and only if k = 0, 1. 25. The given system is (1 − k)x1 + 2x1 + x1 + 2x2 + (1 − k)x2 + x2 + x3 x3 (2 − k)x3 = 0, = 0, = 0. The determinant of the coefficients as a function of k is given by 1−k 2 1 1−k 1 = −(1 + k)(k − 1)(k − 4). det(A) = 2 1 1 2−k Consequently, the system has an infinite number of solutions if and only if k = −1, k = 1, or ,k = 4. 26. From the schematic provided in Figure 3.1.1, we find that det(A) = (1)(1)(3) + (−1)(4)(0) + (2)(3)(1) − (0)(1)(2) − (1)(4)(1) − (3)(3)(−1) = 3 + 0 + 6 − 0 − 4 + 9 = 14. Thus, by Property P10, we have 1 det(A−1 ) = , and by Property P4, we have det(−3A) = (−3)3 · 14 = −378. 14 1 −1 1 2 27. A = and B = . det(A) det(B) = [3 · 1 − (−1)2][1 · 4 − 2(−2)] = 5 · 8 = 40. 2 3 −2 4 3 −2 = 3 · 16 − (−2)(−4) = 40. Hence, det(AB) = det(A) det(B). det(AB) = −4 16 28. Using identities for hyperbolic trignometric functions, we have det(A) = (cosh x)(cosh x)−(sinh x)(sinh x) = cosh2 x − sinh2 x = 1 and det(B) = (cosh y)(cosh y) − (sinh y)(sinh y) = cosh2 y − sinh2 y = 1. Furthermore, we have cosh x cosh y + sinh x sinh y cosh x sinh y + sinh x cosh y cosh(x + y) sinh(x + y) AB = = , sinh x cosh y + cosh x sinh y sinh x sinh y + cosh x cosh y sinh(x + y) cosh(x + y) so that det(AB) = cosh2 (x + y) − sinh2 (x + y) = 1. Thus, det(AB) = det(A)det(B), and Property P9 is verified. 29. B is obtained from AT by the following elementary row operations: (1) A13 (3), (2) M1 (−2). Since det(AT ) = det(A) = 1, and (1) leaves the determinant unchanged, we have det(B) = −2. 30. B is obtained from A by the following elementary row operations: (1) M1 (4), (2) M2 (3), (3) P12 . Thus, det(B) = det(A) · 4 · 3 · (−1) = −12. (c)2017 Pearson Education. Inc. 223 31. B is obtained from A via the following sequence of four elementary row and column operations: (1) P12 , (2) CP12 , (3) M1 (−6), (4) M2 (3). Thus, det(B) = det(A) · (−1) · (−1) · (−6) · (3) = −18. 32. B is obtained from A by the following operations: (1) Interchange the two columns, (2) M1 (−1), (3) A12 (−4). Now (3) leaves the determinant unchanged, and (1) and (2) each change the sign of the determinant. Therefore, det(B) = det(A) = 1. 33. B is obtained from A via the following sequence of elementary row operations: (1) M1 (−1), (2) M2 (−2), (3) P13 . Thus, det(B) = det(A) · (−1) · (−2) · (−1) = 12. 34. B is obtained from A by the following row operations: (1) A13 (5), (2) M2 (−4), (3) P12 , (4) P23 . Thus, det(B) = det(A) · (−4) · (−1) · (−1) = (−6) · (−4) = 24. 35. B is obtained from A by the following operations: (1) M1 (−3), (2) A23 (−4), (3) P12 . Thus, det(B) = det(A) · (−3) · (−1) = (−6) · (3) = −18. 36. B is obtained from AT by the following row operations: (1) M1 (2), (2) A32 (−1), (3) A13 (−1). Thus, det(B) = det(AT ) · 2 = (−6) · (2) = −12. 37. We have det(AB T ) = det(A)det(B T ) = 5 · 3 = 15. 38. We have det(A2 B 5 ) = (detA)2 (detB)5 = 52 · 35 = 6075. −1 39. We have det((A 2 3 B ) ) = (det(A −1 2 3 −1 B )) = (detA )(detB) 40. We have det((2B) −1 T (AB) ) = (det((2B) −1 T ))(det(AB) ) = 2 3 = 1 2 ·3 5 det(A)det(B) det(2B) 3 3 729 9 = = = 5.832. 5 125 = 5·3 3 · 24 = 5 . 16 41. We have det((5A)(2B)) = (5 · 54 )(3 · 24 ) = 150, 000. 1 . 42. We have det(B −1 A−1 ) = det(B −1 )det(A−1 ) = 13 · 15 = 15 43. We have det(2A) = 24 · 5 = 80, so that det(B −1 (2A)B) = det(B −1 )det(2A)det(B) = det(2A) = 80. 44. We have det(4B) = 44 · 3 = 768, so that det((4B)3 ) = 7683 = 452, 984, 832. 1 1 2 0 , and A = 45. It is not true that A and B must be equal. For instance, if we take S = 2 3 4 0 1 1 4 then it is easy to verify that B = S −1 AS = 3 = A. However, using the properties of determinants, 4 2 we see that det(B) = det(S −1 AS) = (det S −1 )(det A)(det S) = (det A)(det S) = det(A). det S 46. (a). The volume of the parallelepiped is given by |det(A)|. In this case, we have |det(A)| = |2 + 12k + 36 − 4k − 18 − 12| = |8 + 8k|. (b). NO. The answer does not change because the determinants of A and AT are the same, so the volume determined by the columns of A, which is the same as the volume determined by the rows of AT , is |det(AT )| = |det(A)|. (c)2017 Pearson Education. Inc. 224 (c). The matrix A is invertible if and only if det(A) = 0. That is, A is invertible if and only if 8 + 8k = 0, if and only if k = −1. 47. 1 −1 x 1 −1 1 −1 x x 2 2 2 0 2 = 0 3 x 3 x 1 x − 2x − 2x = 4 −1 x3 0 3 x3 − 4x 0 0 x3 − x2 − 2x 1 −1 x . 3 x(x − 2) = 0 0 0 x(x − 2)(x + 1) We see directly that the determinant will be zero if and only if x ∈ {0, −1, 2}. 48. αx − βy βx + αy βx − αy αx βx − αy −βy βx − αy = + αx + βy βx αx + βy αy αx + βy αx βx αx −αy −βy βx −βy −αy + = + + βy αy αx αy βx αx βx βy α −α β −β 2 α β 2 α β =x + xy − xy +y β α β α β β α α α −α β β 2 2 α β + xy + xy = (x + y ) β α β β α −α α −α α −α 2 2 α β − xy + xy = (x + y ) β β β β β α 2 2 α β = (x + y ) . β α 49. a1 + βb1 a2 + βb2 a3 + βb3 b1 + γc1 b2 + γc2 b3 + γc3 c1 + αa1 a1 c2 + αa2 = a2 c3 + αa3 a3 a1 = a2 a3 b1 + γc1 c1 + αa1 βb1 b1 + γc1 b2 + γc2 c2 + αa2 + βb2 b2 + γc2 b3 + γc3 c3 + αa3 βb3 b3 + γc3 b1 c 1 a 1 b1 c1 b2 c2 + αβγ b2 c2 a2 b3 c 3 a 3 b3 c 3 b1 c 1 a 1 = (1 + αβγ) b2 c2 a2 . b3 c 3 a 3 c1 + αa1 c2 + αa2 c3 + αa3 Now if the last expression is to be zero for all ai , bi , and ci , then it must be the case that 1 + αβγ = 0; hence, αβγ = −1. 50. Suppose A is a matrix with a row of zeros. We will use (P3) and (P7) to justify the fact that det(A) = 0. If A has more than one row of zeros, then by (P7), since two rows of A are the same, det(A) = 0. Assume instead that only one row of A consists entirely of zeros. Adding a nonzero row of A to the row of zeros yields a new matrix B with two equal rows. Thus, by (P7), det(B) = 0. However, B was obtained from A by adding a multiple of one row to another row, and by (P3), det(B) = det(A). Hence, det(A) = 0, as required. (c)2017 Pearson Education. Inc. 225 51. A is orthogonal, so AT = A−1 . Using the properties of determinants, it follows that 1 = det(In ) = det(AA−1 ) = det(A) det(A−1 ) = det(A) det(AT ) = det(A) det(A). Therefore, det(A) = ±1. 52. (a). From the definition of determinant we have: det(A) = , σ(p1 , p2 , p3 . . . , pn )a1p1 a2p2 a3p3 · · · anpn . (52.1) n! If A is lower triangular, then aij = 0 whenever i < j, and therefore the only nonzero terms in (47.1) are those with pi ≤ i for all i. Since all the pi must be distinct, the only possibility is pi = i for all i with 1 ≤ i ≤ n, and so (52.1) reduces to the single term: det(A) = σ(1, 2, 3, . . . , n)a11 a22 a33 · · · ann . (b). 2 −1 3 1 2 2 det(A) = 3 0 1 1 2 0 2 16 0 13 3 = − −1 −8 1 2 5 −3 −11 0 1 1 0 = 4 −1 −8 2 1 1 2 0 0 0 0 4 0 = − 1 0 −1 1 0 1 1. A41 (−5), A42 (−1), A43 (−4) 3 0 0 13 0 0 2 0 2 2 16 0 0 = 1 0 −1 −8 1 0 0 1 1 2 0 1 0 0 0 13 0 0 = −26. −8 1 0 2 0 1 2. A31 (−3), A32 (−2) 3. P12 4. A21 (−16/13) 53. To prove this, take the determinant of both sides of AB = AC to get det(AB) = det(AC). Thus, by Property P8, we have det(A)det(B) = det(A)det(C). Since A is invertible, det(A) = 0. Thus, we can cancel det(A) from the last equality to obtain det(B) = det(C). 54. Note that (S −1 AS)2 = S −1 ASS −1 AS = S −1 AIAS = S −1 A2 S. Therefore, det((S −1 AS)2 ) = det(S −1 A2 S) = det(S −1 det(A2 )det(S) = det(A2 ) = (det(A))2 , as required. 55. No. If A were invertible, then det(A) = 0, so that det(A3 ) = det(A) det(A) det(A) = 0. 56. Let E be an elementary matrix. There are three different possibilities for E. (a). E permutes two rows: Then E is obtained from In by interchanging two rows of In . Since det(In ) = 1 and using Property P1, we obtain det(E) = −1. (b). E adds a multiple of one row to another: Then E is obtained from In by adding a multiple of one row of In to another. Since det(In ) = 1 and using Property P3, we obtain det(E) = +1. (c). E scales a row by k: Then E is obtained from In by multiplying a row of In by k. Since det(In ) = 1 and using Property P2, we obtain det(E) = k. (c)2017 Pearson Education. Inc. 226 57. We have x 0 = x1 x2 y y1 y2 1 1 = xy1 + yx2 + x1 y2 − x2 y1 − xy2 − yx1 , 1 which can be rewritten as x(y1 − y2 ) + y(x2 − x1 ) = x2 y1 − x1 y2 . Setting a = y1 − y2 , b = x2 − x1 , and c = x2 y1 − x1 y2 , we can express this equation as ax + by = c, the equation of a line. Moreover, if we substitute x1 for x and y1 for y, we obtain a valid identity. Likewise, if we substitute x2 for x and y2 for y, we obtain a valid identity. Therefore, we have a straight line that includes the points (x1 , y1 ) and (x2 , y2 ). 58. 1 1 1 x y z 1 x x2 x2 2 (y − x)(y + x) = (y − x)(z − x) 0 1 y + x 0 1 z+x (z + x)(z − x) 1 x x2 3 = (y − x)(z − x) 0 1 y + x = (y − x)(z − x)(z − y) = (y − z)(z − x)(x − y). 0 0 z−y x x2 1 1 y 2 = 0 y − x z2 0 z − x 1. A12 (−1)A13 (−1) 2. P2 3. A23 (−1) 59. Since A is an n × n skew-symmetric matrix, AT = −A; thus, det(AT ) = det(−A) = (−1)n det(A) = − det(A) since n is given as odd. But by P4, det(AT ) = det(A), so det(A) = − det(A) or det(A) = 0. 60. Solving b = c1 a1 + c2 a2 + · · · + cn an for c1 a1 yields c1 a1 = b − c2 a2 − · · · − cn an . Consequently, det(Bk ) can be written as det(Bk ) = det([a1 , a2 , . . . , ak−1 , b, ak+1 , . . . , an ]) = det([a1 , a2 , . . . , ak−1 , (c1 a1 + c2 a2 + · · · + cn an ), ak+1 , . . . , an ]) = c1 det([a1 , . . . , ak−1 , a1 , ak+1 , . . . , an ]) + c2 det([a1 , . . . , ak−1 , a2 , ak+1 , . . . , an ]) + · · · + ck det([a1 , . . . , ak−1 , ak , ak+1 , . . . , an ]) + · · · + cn det([a1 , . . . , ak−1 , an , ak+1 , . . . , an ]). Now by P7, all except the k th determinant are zero since they have two equal columns, so that we are left with det(Bk ) = ck det(A). 63. Using technology we find that: 1 2 3 4 a 2 1 2 3 4 3 2 1 2 3 = −192 + 88a − 8a2 = −8(a − 3)(a − 8). 4 3 2 1 2 a 4 3 2 1 Consequently, the matrix is invertible provided a = 3, 8. (c)2017 Pearson Education. Inc. 227 64. Using technology we find that: 1−k 3 3 4 2−k 4 1 = −(k − 6)(k + 2)2 . 1 −1 − k Consequently, the system has an infinite number of solutions ⎡ if and only if ⎤k = 6, −2. −5 4 1 1 ⎦. This system has solution set k = 6: In this case, the system is Bx = 0, where B = ⎣ 3 −4 3 4 −7 {t(1, 1, 1) : t ∈ R}. ⎡ ⎤ 3 4 1 k = −2: In this case, the system is Cx = 0, where C = ⎣ 3 4 1 ⎦. This system has solution set 3 4 1 {(r, s, −3r − 4s) : r, s ∈ R}. 65. Using technology we find that: ⎤ ⎤ ⎡ −2/5 1/2 0 1/10 19/10 ⎢ 1/2 −1 1/2 ⎢ 0 ⎥ −5 ⎥ ⎥ ; x = A−1 b = ⎢ ⎥. det(A) = −20; A−1 = ⎢ ⎦ ⎣ ⎣ 0 1/2 −1 1/2 1/2 ⎦ 1/10 0 1/2 −2/5 12/5 ⎡ Solutions to Section 3.3 True-False Review: (a): FALSE. Because 2 + 3 = 5 is odd, the (2, 3)-cofactor is the negative of the (2, 3)-minor of the matrix. (b): TRUE. Examining the submatrix obtained by deleting row i and column j from the matrix A, note by examining the remaining entries from column i or from row j of A, they now comprise a column (or row) of zeros, so this submatrix has a determinant of zero. That is, Mij = 0. (c): TRUE. The submatrix obtained by removing the ith row and ith column of A is the matrix diag(a1 , . . . , ai−1 , ai+1 , . . . , an whose determinant is a1 · · · ai−1 ai+1 · · · an . ⎡ ⎤ 1 2 3 (d): FALSE. Here is a counterexample. If we let A = ⎣ 4 5 6 ⎦, then M11 = (5)(9) − (6)(8) = −3. 7 8 9 However, in the matrix 2A, we have M11 = (10)(18) − (12)(16) = −12, which is not doubled the value of the (1, 1)-minor of A. (e): TRUE. This just requires a slight modification of the proof of Theorem 3.3.17. We compute (A · adj(A))ij = n , k=1 aik adj(A)kj = n , aik Cjk = δij · det(A). k=1 Therefore, A · adj(A) = det(A) · In . (f ): TRUE. The Cofactor Expansion Theorem allows for expansion along any row or any column of the matrix, and in all cases, the result is the determinant of the matrix. (c)2017 Pearson Education. Inc. 228 ⎡ ⎤ 1 2 3 (g): FALSE. For example, let A = ⎣ 4 5 6 ⎦, and let c = 2. It is easy to see that the (1, 1)-entry of 7 8 9 adj(A) is −3. But the (1, 1)-entry of adj(2A) is −12, not −6. Therefore, the equality posed in this review item does not generally hold. Many other counterexamples could be given as well. ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 1 2 3 9 8 7 10 10 10 (h): FALSE. For example, let A = ⎣ 4 5 6 ⎦ and B = ⎣ 6 5 4 ⎦. Then A + B = ⎣ 10 10 10 ⎦. 7 8 9 3 2 1 10 10 10 The (1, 1)-entry of adj(A + B) is therefore 0. However, the (1, 1)-entry of adj(A) is −3, and the (1, 1)-entry of adj(B) is also −3. But (−3) + (−3) = 0. Many other examples abound, of course. ae + bg af + bh a b e f . Then AB = . We compute (i): FALSE. Let A = and let B = ce + dg cf + dh c d g h cf + dh −(af + bh) d −b h −f adj(AB) = = = adj(A)adj(B). −(ce + dg) ae + bg −c a −g e (j): TRUE. This can be immediately deduced by substituting In for A in Theorem 3.3.17. Problems: From this point on, CET(col#n) will mean that the Cofactor Expansion Theorem has been applied to column n of the determinant, and CET(row#n) will mean that the Cofactor Expansion Theorem has been applied to row n of the determinant. 1. We have M11 = 5, M12 = 0, M21 = 2, M22 = −9 C11 = 5, C12 = 0, C21 = −2, C22 = −9. and 2. Minors: M11 = 4, M21 = −3, M12 = 2, M22 = 1; Cofactors: C11 = 4, C21 = 3, C12 = −2, C22 = 1. 3. Minors: M11 = −9, M21 = −7, M31 = −2, M12 = 7, M22 = 1, M32 = −2, M13 = 5, M23 = 3, M33 = 2; Cofactors: C11 = −9, C21 = 7, C31 = −2, C12 = −7, C22 = 1, C32 = 2, C13 = 5, C23 = −3, C33 = 2. 4. Minors: M11 = −5, M21 = 47, M31 = 3, M12 = 0, M22 = −2, M32 = 0, M13 = 4, M23 = −38, M33 = −2; Cofactors: C11 = −5, C21 = −47, C31 = 3, C12 = 0, C22 = −2, C32 = 0, C13 = 4, C23 = 38, C33 = −2. 5. 3 M12 = 7 5 1 M23 = 7 5 1 4 1 3 1 0 2 6 = −4, 2 2 6 = 40, 2 3 −1 2 1 2 = 16, M31 = 4 0 1 2 1 −1 2 1 2 = 12, M42 = 3 7 4 6 C12 = 4, C31 = 16, C23 = −40, C42 = 12. 6. We have 9 M41 = −6 −1 0 −1 8 8 = 478, −3 4 (c)2017 Pearson Education. Inc. 229 M22 = −2 0 −1 0 −3 4 = 9, 7 3 1 M23 = −2 9 −1 0 −1 4 = 191, 7 −7 1 M43 = −2 9 −1 4 −6 8 = −108. 0 −1 4 Therefore, C41 = −478, C22 = 9, C23 = −191, C43 = 108. 7. We have det(A) = 6 · C12 + 9 · C22 = 6 · 4 + 9 · (−8) = −48. 1 −2 = 1 · |3| + 2 · |1| = 5. 8. 1 3 −1 2 3 −1 2 −1 2 1 4 = 3(−11) + 2(−7) + 4(−6) = −71. +4· +2· 9. 1 4 −2 = 3 · 1 4 3 1 3 1 3 1 4 2 1 −4 2 1 1 −4 2 −4 = −7 · 18 + 0 − 3 · 9 = −153. 3 = −7 · 10. 7 1 −3· + 1 5 5 −2 1 −2 1 5 −2 3 1 4 1 4 1 4 2 1 2 = 3 · + 2 · − 11. 7 1 1 2 = 3(−11) − 7(−17) + 2(−2) = 82. 3 −5 3 −5 2 3 −5 0 2 −3 0 5 = 3 · 10 + 5(−6) + 0 · 4 = 0. 12. −2 3 −5 0 1 −2 ⎡ ⎤ 3 0 1 −2 1 −2 3 3 1 4 0 7 −2 = −2 · 0 0 7 ⎦ = −2 · (−26) − 4 · (−5) = 72. 1 3 − 4 · ⎣ 4 13. 1 3 4 0 1 1 5 −2 5 −2 1 5 −2 0 1. CET(col#4) 14. We have det(A) = −3 · C11 + 1 · C31 = −3 · 16 + 1 · (−4) = −52. 15. We can choose to apply the Cofactor Expansion Theorem to any row or column of our choice. Choosing row 1, we find that det(A) = −4 · C11 + 2 · C12 + (−1) · C13 = −4 · (−18) + 2 · (−26) + (−1) · (24) = −4. (c)2017 Pearson Education. Inc. 230 1 0 −2 3 1 1 1 −1 = 7 − 2(−1) = 9. 1. CET(row#1) −2· 16. 3 1 −1 = 7 2 2 5 7 2 5 −1 2 3 2 3 1 4 1 = −7 + 10 = 3. 1. CET(col#1) 1 4 = −1 · +2· 17. 0 1 4 −1 3 2 −1 3 2 −1 3 1 2 1 − 5 · −1 3 + 3 · −1 3 = 2 · 17 − 5 · 2 + 3(−7) = 3. 2 1 = 2 · 18. 5 −3 7 −3 7 2 1 3 −3 7 1. CET(col#1) 0 −2 1 −2 −2 1 0 −3 1 0 −3 = 0 · 19. 2 −2· 3 0 −1· 0 3 0 −1 3 0 1 = 0 + 6 − 6 = 0. −3 1. CET(col#1) 20. 1 0 −1 0 0 −1 1 0 0 −1 1 0 1 0 −1 = 1 · 0 −1 0 −1 = −2 − 2 = −4. 0 − 1 · 1 −1 0 −1 0 1 1 0 1 0 1 0 1 0 1 1. CET(col#1) 21. 2 −1 5 3 1 3 1 2 5 1 2 2 1 1 1 0 −2 3 4 −2 3 = − 1 = 0 3 0 0 −1 0 2 −1 0 0 1 −1 4 1 3 −2 4 1 −1 −2 4 2 5 1 5 = 21 − 21 = 0. +3· = 1 −1 −1 4 1. CA32 (2) 2. CET(row#3) 22. There are two zeros in the last row of this matrix, so we will apply the Cofactor Expansion Theorem to the last row: det(A) = −5 · C43 + 2 · C44 = −5 · −4 1 −2 −4 1 −3 1 2 −6 + 2 · 1 2 −1 = −5 · 33 + 2 · (−3) = −171. 3 0 3 3 0 2 23. By applying elementary row operations A12 (4) and A13 (2) to the matrix (which does not change the determinant), we have −1 3 3 −1 3 3 4 −6 −3 = 0 6 9 , 2 −1 4 0 5 10 (c)2017 Pearson Education. Inc. 231 and we evaluate the latter determinant by cofactor expansion along the first column to obtain −1 3 3 4 −6 −3 = −1 · (6 · 10 − 9 · 5) = −1 · 15 = −15. 2 −1 4 24. 3 2 −3 11 1 2 −3 11 5 2 6 1 2 3 5 −5 2 0 −1 11 −27 3 5 −5 1 2 = 7 = 7 5 −3 −16 0 −9 18 −93 5 −3 −16 9 −6 27 −12 9 −6 27 −12 0 −24 54 −111 −1 11 −27 −1 11 −27 −81 150 3 4 5 = 11997. −81 150 = − = −9 18 −93 = 0 −210 537 −24 54 −111 0 −210 537 1. A21 (−1) 2. A12 (−2), A13 (−7), A14 (−9) 4. A12 (−9), A13 (−24) 25. 3. CET(col#1) 5. CET(col#1) 2 −7 2 −1 4 3 0 −13 4 3 2 −7 −13 2 −1 5 3 1 2 3 3 1 2 2 1 5 −3 7 1 1 0 −9 = − −16 = 0 −16 = 6 6 0 −9 2 6 3 2 6 3 −10 −8 −3 4 2 −4 5 0 −10 −8 −3 2 −4 5 4 13 −2 −2 1 1 13 5 6 101 −18 4 0 −9 = 101 −18 0 = = −892. = −16 29 14 −10 −8 −3 29 −14 0 1. A42 (−1) 2. A21 (−2), A23 (−6), A24 (−4) 5. A12 (9), A13 (3) 3. CET(col#1) 4. P2 6. CET(col#3) 26. By applying elementary row operations A21 (2), A23 (4), and A24 (1) to the matrix (which does not change the determinant), we have −2 0 1 1 0 4 5 1 1 2 2 0 1 2 2 0 = , −4 4 6 1 0 12 14 1 −1 1 0 5 0 3 2 5 and we evaluate the latter determinant by cofactor expansion along the first column to obtain −2 0 1 1 1 2 2 0 = C21 = 31. −4 4 6 1 −1 1 0 5 27. We proceed as follows: −3 2 0 −1 −4 0 1 3 1 3 5 1 1 0 2 1 −3 −1 = 0 2 0 4 −2 −3 0 −1 2 6 0 −1 0 11 3 8 11 1 3 1 3 5 0 2 2 1 −3 −1 = − 0 −6 2 −8 −13 0 5 7 3 4 0 3 11 2 −6 5 1 3 5 1 3 8 11 0 3 1 −3 −1 = 0 2 −8 −13 0 7 3 4 0 (c)2017 Pearson Education. Inc. 3 2 11 −6 5 1 3 5 1 −3 −1 3 8 11 2 −8 −13 7 3 4 232 1 0 4 =2· 0 0 0 3 1 11 −6 5 1 1 2 3 2 7 3 5 − 32 − 12 5 8 11 = 2 · −8 −13 3 4 1. A21 (3), A24 (−2), A25 (1) 2. P12 1 0 0 0 0 3 1 1 1 2 0 − 52 0 5 9 0 2 3. P23 3 − 32 5 − 12 1 0 6 1 49 33 = · 0 2 2 2 0 −17 −16 21 13 0 2 2 4. P 2 3 1 3 5 1 1 −3 −1 0 −5 49 33 0 5 −17 −16 0 9 21 13 5. A23 (−11), A24 (6), A25 (−5) 6. P 2 The 5 × 5 determinant remaining in the last step can be computed via cofactor expansion along the first column, followed by cofactor expansion along the second column, which then leads to a 3 × 3 determinant that can be computed by using the schematic in Figure 3.1.1: −3 2 0 −1 −4 1 3 1 3 5 1 0 2 1 −3 −1 = · 2 2 0 4 −2 −3 −1 2 6 0 −1 28. 1 0 0 0 0 3 1 3 5 1 1 −3 −1 −5 49 33 1 2302 0 −5 49 33 = · 5 −17 −16 = − = −1151. 2 2 0 5 −17 −16 9 21 13 0 9 21 13 5 0 0 −1 3 0 0 0 −3 0 −3 5 0 0 1 2 3 0 1 2 0 3 3 0 1 2 1 2 3 0 4 = − 1 3 0 4 = 0 1 3 0 4 1 1 −1 0 0 1 −1 0 1 0 0 −1 3 5 3 5 0 2 0 5 0 0 −1 0 −3 5 0 −3 5 0 −3 5 0 0 −9 1 −10 4 3 5 1 −10 = − = − −9 1 −10 = −9 1 3 0 4 −1 3 5 1 −3 −5 0 −1 3 5 42 0 50 42 50 6 7 = −9 1 −10 = = −170. −26 −35 −26 0 −35 2 0 0 1 3 1. A41 (−2),A45 (−3) 5. P2 29. 2. CET(col#1) 6. A21 (−5), A23 (3) 3. A12 (−3) 4. CET(col#1) 7. CET(col#2) 0 xz 0 x + y z x y z xz x + y z 2 −x 1 −x 0 1 −1 0 1 −1 = = −x 1 −1 −y −1 −1 1 0 1 −y − z 0 −y − z −1 1 −z −z 1 −1 0 1 −1 0 xz − yz + z 2 x + y + z 0 4 xz − yz + z 2 x + y + z 3 = (x + y + z)2 . 0 0 = = −x + y − z −x − y − z 0 −y − z −1 1 1. A41 (−x), A43 (1) 2. CET(col#2) 3. A32 (1), A31 (−z) (c)2017 Pearson Education. Inc. 4. CET(col#3) 233 30. (a). 1 1 1 1 0 0 1 V (r1 , r2 , r3 ) = r1 r2 r3 = r1 r2 − r1 r3 − r1 r12 r22 r32 r12 r22 − r12 r32 − r12 r3 − r1 2 r2 − r1 = 2 = (r2 − r1 )(r32 − r12 ) − (r3 − r1 )(r22 − r12 ) r2 − r12 r32 − r12 = (r3 − r1 )(r2 − r1 )[(r3 + r1 ) − (r2 + r1 )] = (r2 − r1 )(r3 − r1 )(r3 − r2 ). 1. CA12 (−1), CA13 (−1) 2. CET(row#1) (b). We use mathematical induction. The result is vacuously true when n = 1 and quickly verified when n = 2. Suppose that the result is true when n = k − 1, for k ≥ 3, and consider 1 1 1 ··· 1 r1 r2 r3 ··· rk 2 2 r12 r2 r3 ··· rk2 . V (r1 , r2 , . . . , rk ) = .. .. .. .. . . . . k−1 k−1 k−1 k−1 r r r ··· r 1 2 3 k The determinant vanishes when rk = r1 , r2 , . . . , rk−1 , so we can write V (r1 , r2 , . . . , rk ) = a(r1 , r2 , . . . , rk ) n−1 - (rk − ri ), i=1 where a(r1 , r2 , . . . , rk ) is the coefficient of rkk−1 in the expansion of V (r1 , r2 , . . . , rk ). However, using the Cofactor Expansion Theorem along column k, we see that this coefficient is just V (r1 , r2 , . . . , rk−1 ), so by hypothesis, a(r1 , r2 , . . . , rk ) = V (r1 , r2 , . . . , rk−1 ) = (rm − ri ). 1≤i<m≤n−1 Thus, - V (r1 , r2 , . . . , rk ) = (rm − ri ) 1≤i<m≤n−1 n−1 - (rk − ri ) = i=1 - (rm − ri ). 1≤i<m≤n Hence the result is true for n = k, so, by induction, is true for all non-negative integers n. 31. We have det(A − λI) = det 2−λ 2 −1 4−λ = (2 − λ)(4 − λ) + 2 = λ2 − 6λ + 10. The quadratic formula supplies the two roots of the equation λ2 − 6λ + 10 = 0; namely, λ = 3 ± i. 32. We have det(A − λI) = det 2−λ 3 4 13 − λ = (2 − λ)(13 − λ) − 12 = λ2 − 15λ + 14 = (λ − 1)(λ − 14). (c)2017 Pearson Education. Inc. 234 The two roots are λ = 1 and λ = 14. 33. We have det(A − λI) = det −1 − λ −4 2 7−λ = (−1 − λ)(7 − λ) + 8 = λ2 − 6λ + 1. √ The quadratic formula supplies the two roots of the equation λ2 − 6λ + 1 = 0; namely, λ = 3 ± 2 2. 34. We have ⎡ 2−λ det(A − λI) = det ⎣ −1 3 0 −6 − λ 3 ⎤ 0 0 ⎦ = (2 − λ)(−6 − λ)(7 − λ), 7−λ since the matrix A − λI is lower triangular. Thus, the three roots of the equation (2 − λ)(−6 − λ)(7 − λ) = 0 are λ = 2, λ = −6, and λ = 7. 35. We have ⎡ 2−λ det(A − λI) = det ⎣ 7 7 ⎤ 0 7 ⎦, 7−λ 0 7−λ 7 and using cofactor expansion along the first row, we obtain det(A − λI) = (2 − λ) 7−λ 7 7 = (2 − λ)(λ2 − 14λ) = λ(λ − 14)(2 − λ). 7−λ The roots of λ(λ − 14)(2 − λ) = 0 are λ = 0, λ = 14, and λ = 2. 36. We have ⎡ 6−λ det(A − λI) = det ⎣ 0 −5 0 7−λ 0 ⎤ −2 ⎦, 0 −3 − λ and using cofactor expansion along the second row, we obtain det(A − λI) = (7 − λ) 6−λ −5 −2 = (7 − λ)(λ2 − 3λ − 28) = (7 − λ)(λ − 7)(λ + 4). −3 − λ The roots of (7 − λ)(λ − 7)(λ + 4) = 0 are λ = 7 (repeated twice) and λ = −4. 37. We have ⎡ −5 − λ det(A − λI) = det ⎣ −8 −5 −5 1−λ 3 ⎤ 0 0 ⎦, 7−λ and using cofactor expansion along the third column, we obtain det(A − λI) = (7 − λ) −5 − λ −8 −5 = (7 − λ)(λ2 + 4λ − 45) = (7 − λ)(λ + 9)(λ − 5). 1−λ The roots of (7 − λ)(λ + 9)(λ − 5) = 0 are λ = 7, λ = −9, and λ = 5. 38. We have ⎡ −5 − λ det(A − λI) = det ⎣ −2 −5 −1 −1 − λ 2 (c)2017 Pearson Education. Inc. ⎤ 1 0 ⎦, 3−λ 235 and using cofactor expansion along the third column, we obtain det(A−λ) = (−9−5−λ)+(3−λ)[(−5−λ)(−1−λ)−2] = −λ3 −3λ2 +10λ = −λ(λ2 +3λ−10) = −λ(λ+5)(λ−2). The roots of −λ(λ + 5)(λ − 2) = 0 are λ = 0, λ = −5, and λ = 2. 39. (a). det(A) = 11; 5 −4 (b). MC = ; −1 3 5 −1 (c). adj(A) = ; −4 3 1 5 −1 −1 (d). A = . 3 11 −4 40. (a). det(A) = 7; 1 −4 ; 2 −1 1 2 (c). adj(A) = ; −4 −1 1 1 2 . (d). A−1 = 7 −4 −1 (b). MC = 41. (a). det(A) = 0; −6 15 ; −2 5 −6 −2 (c). adj(A) = ; 15 5 (b). MC = (d). A−1 does not exist because det(A) = 0. 2 −3 0 2 −3 0 4 5 = 2 · 13 = 26; 1 5 = 0 (a). det(A) = 2 0 −1 2 0 −1 2 ⎡ ⎤ 7 −4 −2 6 4 2 ⎦; (b). MC = ⎣ −15 −10 8 ⎡ ⎤ 7 6 −15 (c). adj(A) = ⎣ −4 4 −10 ⎦; −2 2 8 42. (c)2017 Pearson Education. Inc. 236 ⎡ ⎤ 7 6 −15 1 1 ⎣ −4 4 −10 ⎦. (d). A−1 = adj(A) = det(A) 26 −2 2 8 −2 3 −1 −2 3 −1 4 = −2 · 4 = −8; 5 = 0 4 (a). det(A) = 2 1 0 2 3 3 0 2 ⎡ ⎤ −7 −6 4 4 ⎦; (b). MC = ⎣ −11 −6 16 8 −8 ⎡ ⎤ −7 −11 16 8 ⎦; (c). adj(A) = ⎣ −6 −6 4 4 −8 ⎡ ⎤ −7 −11 16 1 1 8 ⎦. adj(A) = − ⎣ −6 −6 (d). A−1 = det(A) 8 4 4 −8 43. 1 −1 2 6 0 9 (a). det(A) = 3 −1 4 = 8 0 11 = 6; 5 1 7 5 1 7 ⎡ ⎤ −11 −1 8 9 −3 −6 ⎦; (b). MC = ⎣ −2 2 2 ⎡ ⎤ −11 9 −2 2 ⎦; (c). adj(A) = ⎣ −1 −3 8 −6 2 ⎡ ⎤ ⎡ ⎤ −11 9 −2 −11/6 3/2 −1/3 1 1 2 ⎦ = ⎣ −1/6 −1/2 1/3 ⎦. adj(A) = ⎣ −1 −3 (d). A−1 = det(A) 6 8 −6 2 4/3 −1 1/3 44. 0 1 2 0 (a). det(A) = −1 −1 3 = 0 1 −2 1 1 ⎡ ⎤ 5 4 3 (b). MC = ⎣ −5 −2 1 ⎦; 5 −2 1 ⎡ ⎤ 5 −5 5 (c). adj(A) = ⎣ 4 −2 −2 ⎦; 3 1 1 ⎡ 5 1 1 ⎣ 4 adj(A) = (d). A−1 = det(A) 10 3 45. 1 2 −3 4 = 10; −2 1 ⎤ −5 5 −2 −2 ⎦. 1 1 (c)2017 Pearson Education. Inc. 237 2 −3 5 0 −7 3 2 1 = 1 2 1 = 14; (a). det(A) = 1 0 7 −1 0 7 −1 ⎡ ⎤ −9 1 7 (b). MC = ⎣ 32 −2 −14 ⎦; −13 3 7 ⎡ ⎤ −9 32 −13 3 ⎦; (c). adj(A) = ⎣ 1 −2 7 −14 7 ⎡ ⎤ ⎡ ⎤ −9 32 −13 −9/14 16/7 −13/14 1 1 ⎣ 1 −2 3 ⎦ = ⎣ 1/14 −1/7 3/14 ⎦. adj(A) = (d). A−1 = det(A) 14 7 −14 7 1/2 −1 1/2 46. 1 1 1 1 1 −1 1 −1 1 0 (a). det(A) = = 1 1 −1 −1 0 −1 1 1 −1 0 ⎤ ⎡ 4 4 4 4 ⎢ −4 4 −4 4 ⎥ ⎥ (b). MC = ⎢ ⎣ 4 4 −4 −4 ⎦; −4 4 4 −4 ⎡ ⎤ 4 −4 4 −4 ⎢ 4 4 4 4 ⎥ ⎥; (c). adj(A) = ⎢ ⎣ 4 −4 −4 4 ⎦ 4 4 −4 −4 ⎡ 4 −4 ⎢ 1 1 4 ⎢ 4 (d). A−1 = adj(A) = ⎣ 4 −4 det(A) 16 4 4 47. 1 1 1 2 0 2 2 0 2 0 = 16; = 2 −2 2 −2 0 2 0 −2 2 0 −2 ⎤ 4 −4 4 4 ⎥ ⎥. −4 4 ⎦ −4 −4 1 0 3 5 11 0 −2 1 1 3 4 1 = (a). det(A) = 2 7 9 3 9 0 2 0 3 −1 0 0 ⎤ ⎡ 84 −46 −29 81 ⎢ −162 60 99 −27 ⎥ ⎥; (b). MC = ⎢ ⎣ 18 38 −11 3 ⎦ −30 26 130 −72 ⎤ ⎡ 84 −162 18 −30 ⎢ −46 60 38 26 ⎥ ⎥; (c). adj(A) = ⎢ ⎣ −29 99 −11 130 ⎦ 81 −27 3 −72 48. 18 5 11 10 3 4 = − 6 2 7 0 −1 (c)2017 Pearson Education. Inc. 0 1 9 11 0 18 18 4 1 10 = 402; 10 = − −29 0 −84 6 238 (d). ⎤ 84 −162 18 −30 1 1 ⎢ 60 38 26 ⎥ ⎥ ⎢ −46 adj(A) = A−1 = ⎦ ⎣ −29 99 −11 130 det(A) 402 81 −27 3 −72 ⎤ ⎡ 14/67 −27/67 3/67 −5/67 ⎢ −23/201 10/67 19/201 13/201 ⎥ ⎥ =⎢ ⎣ −29/402 33/134 −11/402 65/201 ⎦ . 27/134 −9/134 1/134 −12/67 ⎡ 49. (a). Using the Cofactor Expansion Theorem along row 1 yields det(A) = (1 + 2x2 ) + 2x(2x + 4x3 ) + 2x2 (2x2 + 4x4 ) = 8x6 + 12x4 + 6x2 + 1 = (1 + 2x2 )3 . (b). ⎤ −(2x + 4x3 ) 2x2 + 4x4 1 + 2x2 1 − 4x4 −(2x + 4x3 ) ⎦ MC = ⎣ 2x + 4x3 2 4 3 2x + 4x 2x + 4x 1 + 2x2 ⎤ ⎡ −2x(1 + 2x2 ) 2x2 (1 + 2x2 ) 1 + 2x2 = ⎣ 2x(1 + 2x2 ) (1 + 2x2 )(1 − 2x2 ) −2x(1 + 2x2 ) ⎦ , 2x2 (1 + 2x2 ) 2x(1 + 2x2 ) 1 + 2x2 ⎡ so that ⎡ 1 1 1 ⎣ −2x adj(A) = A−1 = det(A) (1 + 2x2 )2 2x2 50. det(A) = 2x 1 − 2x2 −2x −1 −2 4 −1 −2 4 2 −1 0 2 −1 = − 0 2 −1 = = −18 and −8 13 3 −2 1 0 −8 13 C11 = 2 −1 = 0. −2 1 Thus, 1 51. det(A) = 1 1 and 1 2 2 1 1 2 = 0 3 0 1 1 1 (A−1 )11 = (adj(A))11 0 = = 0. det(A) −18 1 1 1 = 1 2 1 = 2 − 1 = 1, 2 1 C23 = − 1 1 = −(2 − 1) = −1. 2 (c)2017 Pearson Education. Inc. ⎤ 2x2 2x ⎦ . 1 239 Thus, (adj(A))32 C23 = = −1. det(A) det(A) 1 0 4 1 1 1 = − = −(−4 − 3) = 7, 3 −1 −1 0 (A−1 )32 = 2 52. det(A) = 2 3 and 0 −1 4 1 1 = 2 −1 0 3 2 1 C13 = = −2 − 3 = −5. 3 −1 Thus, (A−1 )31 = C13 5 (adj(A))31 = =− . det(A) det(A) 7 1 0 0 0 0 1 0 1 2 −1 −1 3 −1 −1 3 2 −1 1 3 1 −1 2 = = det(A) = 1 −1 2 1 −1 2 0 0 1 3 0 −1 1 3 0 1 2 0 −1 −1 2 3 2 3 = 4 + 12 = 16, = 1 −4 2 = −4 2 1 0 0 53. and 1 1 0 1 C32 = − 2 1 3 = −(−3) −1 −1 2 0 1 = 3(2 + 1) = 9. 2 Thus, (adj(A))23 C32 9 = = . det(A) det(A) 16 −t −et cos(2t) e sin(2t) et cos(2t) , adj(A) = , and −et cos(2t) et sin(2t) et sin(2t) (A−1 )23 = 54. MC = e−t sin(2t) −et cos(2t) det(A) = sin2 (2t) + cos2 (2t) = 1, so −t 1 e sin(2t) et cos(2t) . adj(A) = A = −et cos(2t) et sin(2t) det(A) −2et 2e2t −e2t , adj(A) = , and det(A) = 4e3t , so 3et −2et 3et 1 1 2e2t −e2t . A−1 = adj(A) = 3t −2et 3et det(A) 4e −1 55. MC = 2e2t −e2t (c)2017 Pearson Education. Inc. 240 56. We compute ⎡ −et ⎣ −et MC = t e (9t + 1) and thus, ⎡ −et T adj(A) = MC = ⎣ −tet 0 −tet −tet t −e (1 − t) ⎤ 0 −e6t (9t2 + 1) ⎦ , e6t (9t2 + 1) −et −tet 6t −e (9t2 + 1) ⎤ et (9t + 1) −et (1 − t) ⎦ . e6t (9t2 + 1) We can find det(A) via the schematic in Figure 3.1.1: det(A) = 0 − 9t2 e4t + te4t − te4t − e4t − 0 = −e4t (9t2 + 1). Therefore, ⎡ −et −4t 1 e −1 ⎣ −tet A = · adj(A) = − 2 det(A) 9t + 1 0 ⎤ et (9t + 1) −et (1 − t) ⎦ . e6t (9t2 + 1) −et −tet 6t −e (9t2 + 1) ⎤ ⎤ ⎡ −e−t −te2t 3te−t −te−t −te−t e−t 0 ⎦ , adj(A) = ⎣ −e−t e−t 0 ⎦ , and 2t 2t −te 0 te 0 te2t t e 2tet e−2t et tet e−2t t e−2t = 0 tet 0 = et (te−t ) = t, det(A) = e 2tet t −2t −2t et 0 0 e te 2e ⎡ 3te−t 57. MC = ⎣ −te−t −te−t so ⎡ 3te−t 1 1⎣ −1 −e−t A = adj(A) = det(A) t −te2t ⎤ −te−t −te−t e−t 0 ⎦. 0 te2t ⎡ ⎤ ⎡ ⎤ −1 2 −1 −1 3 −2 3 ⎦ , adj(A) = ⎣ 2 −6 4 ⎦ . Hence, 58. MC = ⎣ 3 −6 −2 4 −2 −1 3 −2 ⎡ 1 A · adj(A) = ⎣ 3 4 2 4 5 ⎤⎡ ⎤ ⎡ 3 −1 3 −2 0 5 ⎦ ⎣ 2 −6 4 ⎦=⎣ 0 6 −1 3 −2 0 0 0 0 ⎤ 0 0 ⎦ = 03 . 0 From Equation (3.3.4) of the text we have that, in general, A · adj(A) = det(A) · In . Since, for the given matrix, A · adj(A) = 03 , we must have det(A) = 0. 2 −3 2 2 2 −3 = 6. Thus, = 16, and det(B2 ) = = 7, det(B1 ) = 59. det(A) = 4 2 1 4 1 2 x1 = det(B1 ) 16 = and det(A) 7 x2 = det(B2 ) 6 = . det(A) 7 Solution: (16/7, 6/7). (c)2017 Pearson Education. Inc. 241 60. det(A) = 1 5 1 5 = 26, = 21, det(B1 ) = −3 6 −4 6 x1 = det(B1 ) 26 = det(A) 21 and 1 −3 and det(B2 ) = x2 = det(B2 ) 1 =− . det(A) 21 1 = −1. Thus, −4 Solution: (26/21, −1/21). 61. det(A) = −2 4 1 3 −2 −1 = −27, 4 −3 2 det(B1 ) = −5 4 1 2 −2 −1 = 11, 1 −3 2 det(B2 ) = −2 −5 1 3 2 −1 = 35, 4 1 2 det(B3 ) = −2 4 −5 3 −2 2 = 17. 4 −3 1 Thus, x1 = det(B1 ) det(B2 ) det(B3 ) 11 35 17 = − , x2 = = , x3 = =− . det(A) 27 det(A) 27 det(A) 27 Solution: (−11/27, 35/27, −17/27). 62. 3 det(A) = 1 1 4 det(B1 ) = 2 1 3 det(B2 ) = 1 1 3 det(B3 ) = 1 1 −2 1 1 −1 = 6, 0 1 −2 1 1 −1 = 9, 0 1 4 1 2 −1 = 0, 1 1 −2 4 1 2 = −3. 0 1 det(B2 ) det(B3 ) 9 3 det(B1 ) = , x2 = = 0, and x3 = =− . det(A) 6 det(A) det(A) 6 Solution: (3/2, 0, −1/2). 1 −3 1 1 1 −3 7 −2 7 −2 = 4 −1 = 0 = −35 + 14 = −21 = 0. 63. det(A) = 1 7 −5 2 7 −5 1 −3 0 Thus, the system has only the trivial solution. Thus, x1 = (c)2017 Pearson Education. Inc. 242 64. 1 −2 3 −1 3 −1 −5 −2 2 0 1 0 0 1 0 0 det(A) = = 1 1 0 −1 1 1 0 −1 0 1 −2 1 1 −2 1 4 −1 −1 0 −5 −2 −1 2 0 = −3, 1 −1 = − 5 = − 1 4 4 1 1 1 1 1 −2 3 −1 1 −3 3 −1 3 3 −5 −3 1 −3 2 2 0 1 0 0 1 0 = 2 = 0 1 = −11, 0 1 = 0 det(B1 ) = 0 0 −1 1 0 −1 0 0 7 2 −2 3 2 −2 3 2 −2 1 1 −2 1 3 1 1 3 0 3 −1 1 1 3 0 3 0 1 1 2 2 2 2 1 2 1 0 = 2 = 5 0 = 17, 1 2 = −4 det(B2 ) = 0 0 0 −1 1 0 1 0 3 −2 1 3 −2 1 0 3 −2 1 0 3 −2 1 1 −2 1 −1 0 −2 1 −1 2 0 2 0 0 2 0 0 det(B3 ) = = 1 0 −1 1 1 0 −1 1 0 1 3 1 1 3 1 −3 −3 −1 0 0 −2 −1 2 0 = −2(−8) = 16, 1 −1 = −2 −2 = −2 1 −3 −3 1 1 1 1 1 −2 3 1 3 −2 3 1 3 3 1 2 2 0 1 2 0 1 2 = = − 2 1 2 det(B4 ) = 1 0 0 0 1 0 0 −1 −2 3 1 0 1 −2 3 −1 1 −2 3 −3 3 −5 −3 −5 = 6. 0 = − = − 0 1 3 7 3 2 7 Therefore, x1 = 17 16 11 , x2 = − , x3 = − , 3 3 3 and x4 = −2. Solution: (11/3, −17/3, −16/3, −2). 65. t e 1 e−2t t −2t 1 −t e = e det(A) = t 1 −2 = −3e , e −2e−2t 3 sin t e−2t 1 −2t 3 sin t −2t = e det(B1 ) = 4 cos t −2 = −2e (3 sin t + 2 cos t), 4 cos t −2e−2t t e 3 sin t = et 1 3 sin t = et (4 cos t − 3 sin t). det(B2 ) = t 1 4 cos t e 4 cos t (c)2017 Pearson Education. Inc. 243 det(B1 ) det(B2 ) 2(3 sin t + 2 cos t) e2t (3 sin t − 4 cos t) Thus, x1 = and x2 = = = . t 3e det(A) 3 det(A) 1 2 −t e (3 sin t + 2 cos t), e2t (3 sin t − 4 cos t) . Solution: 3 3 66. 1 4 −2 1 1 −1 −2 2 2 9 −3 −2 2 −1 −3 0 −1 −2 2 1 −16 0 det(A) = = = −1 −3 0 = −1 −3 0 = −19, 0 −1 1 0 0 0 1 5 −1 7 1 −1 7 1 3 14 7 −2 3 −1 7 1 1 2 −2 1 1 −1 −2 2 2 5 −3 −2 2 −1 −3 0 −1 −2 2 5 −16 0 = −1 −3 0 = −1 −3 0 = det(B2 ) = 0 0 0 0 −1 1 1 3 7 1 −3 7 1 −3 3 6 7 1 7 −2 3 −3 5 −16 = −31. Therefore x2 = det(B2 ) = 31 . = −1 −3 det(A) 19 67. The coefficient matrix for this linear system is ⎤ −3 1 −3 −9 ⎢ 1 −2 0 4 ⎥ ⎥. A=⎢ ⎣ 0 0 2 1 ⎦ 1 1 0 1 ⎡ By applying the cofactor expansion theorem to the third row of A, we obtain det(A) = 2 · C33 + 1 · C34 = 2 · (−6) + 1 · 9 = −3. To find x4 , we form ⎤ −3 1 −3 −3 ⎢ 1 −2 0 1 ⎥ ⎥, B4 = ⎢ ⎣ 0 0 2 −1 ⎦ 1 1 0 0 ⎡ and using the cofactor expansion theorem on the last row, we have det(B4 ) = 1 · C41 + 1 · C42 = 1 · (−16) + 1 · (−3) = −19. Thus, by Cramer’s Rule we conclude that x4 = det(B4 ) −19 19 = = . det(A) −3 3 (c)2017 Pearson Education. Inc. 244 68. b+c a a −a + b + c a a a+c b = −a + b − c a + c b det(A) = b c c a+b 0 c a+b −a + b + c a −a + b + c a = 4abc, = −c + (a + b) −a + b − c b −a + b − c a + c a 0 a a a a b b = b a − b + c det(B1 ) = b a + c c 0 a+b c a+b c a a = a(a − b + c)(a + b − c), = (a − b + c) c a+b b+c a a b + c a − b − c a b b = b 0 b det(B2 ) = b c c a+b c 0 a+b b b = −b(a − b − c)(a + b − c), = −(a − b − c) c a+b b+c a a a a −a + b + c 0 a + c b a + c b = det(B3 ) = b c 0 c c c c a+c b = −c(a − b − c)(a − b + c). = (−a + b + c) c c Case 1: If a = 0, b = 0, and c = 0, then det(A) = 0, so it follows by Theorem 3.2.5 that the given system of equations has a unique solution. The solution in this case is given by: det(B1 ) (a − b + c)(a + b − c) = , det(A) 4bc det(B2 ) −(a − b − c)(a + b − c) x2 = = , det(A) 4ac det(B3 ) −(a − b − c)(a − b + c) x3 = = . det(A) 4ab x1 = Case 2: If a = b = 0 and c = 0, then it is easy to see from the system that the system is inconsistent. By the symmetry of the equations, the same will be true if a = c = 0 with b = 0, or b = c = 0 with a = 0. Case 3: Suppose a = 0 where b = 0 and c = 0, and consider the reduced row echelon form of the augmented matrix: ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 0 b+c 0 0 0 1 0 0 1 0 0 0 ⎣ c b b b ⎦ ∼ ⎣ 0 b − c b − c b − c ⎦ ∼ ⎣ 0 1 1 1 ⎦. b c c c 0 c−b c−b c−b 0 0 0 0 From the last matrix we see that the system has an infinite number of solutions of the form {(0, 1 − r, r) : r ∈ R}. By the symmetry of the three equations, it follows that the other two cases: b = 0 with a = 0 and c = 0, and c = 0 with a = 0 and b = 0 have similar forms of solutions. 69. Let B be the matrix obtained from A by adding column i to column j (i = j) in the matrix A. By the property for columns corresponding to Property P3, we have det(B) = det(A). Cofactor expansion of B (c)2017 Pearson Education. Inc. 245 along column j gives det(A) = det(B) = n , (akj + aki )Ckj = k=1 That is, det(A) = det(A) + n , akj Ckj + k=1 n , n , aki Ckj . k=1 aki Ckj , k=1 since by the Cofactor Expansion Theorem the first summation on the right-hand side is simply det(A). It follows immediately that n , aki Ckj = 0, i = j. k=1 73. ⎡ 1.21 A = ⎣ 5.41 21.63 ⎡ 1.21 B2 = ⎣ 5.41 21.63 ⎤ ⎡ ⎤ 3.42 2.15 3.25 3.42 2.15 2.32 7.15 ⎦ , B1 = ⎣ 4.61 2.32 7.15 ⎦ , 3.51 9.22 9.93 3.51 9.22 ⎤ ⎡ ⎤ 3.25 2.15 1.21 3.42 3.25 4.61 7.15 ⎦ , B3 = ⎣ 5.41 2.32 4.61 ⎦ . 9.93 9.22 21.63 3.51 9.93 From Cramer’s Rule, x1 = det(B1 ) det(B2 ) det(B3 ) ≈ 0.25, x2 = ≈ 0.72, x3 = ≈ 0.22. det(A) det(A) det(A) 74. det(A) = 32, det(B1 ) = −3218, det(B2 ) = 3207, det(B3 ) = 2896, det(B4 ) = −9682, det(B5 ) = 2414. So, x1 = − 1609 3207 181 4841 1207 , x2 = , x3 = , x4 = − , x5 = . 16 32 2 32 16 75. We have (BA)ji = n , bjk aki = k=1 n , k=1 , 1 1 Ckj aki = δij , · adj(A)jk · aki = det(A) det(A) where we have used Equation (3.3.4) in the last step. Solutions to Section 3.4 Problems: 1. | − 3| = −3. 5 −1 = 5 · 7 − 3(−1) = 38. 2. 3 7 (c)2017 Pearson Education. Inc. n k=1 246 3 5 3. −1 2 6 3 5 1 4. 6 1 14 2 7 4 = −43. −2 4 3 = −3. 7 2.3 1.5 7.9 4.2 3.3 5.1 = 2.3 3.3 3.6 6.8 3.6 5.7 5. 4.2 5.1 4.2 3.3 5.1 + 7.9 − 1.5 6.8 5.7 5.7 6.8 3.6 = 1.035 + 16.11 − 57.828 = −40.683. a b c 6. b c a c c a = a a b b a − b c b b a + c c b c a = a(bc − a2 ) − b(b2 − ac) + c(ab − c2 ) = 3abc − a3 − b3 − c3 . 8. 8 −7 −1 5 −1 2 0 8 −7 −1 3 1 0 1 5 2 0 1 0 = −1 3 = 5 −1 4 2 5 7 0 5 −1 4 2 1 −1 2 1 1 −1 29 −7 −1 29 −1 = −124. 1 0 = −1 = −1 0 8 4 8 −1 4 3 2 3 1 7. 7 1 2 2 −2 4 3 −1 5 18 9 27 9. det(A) = 11. MC = 1 10. det(A) = 2 3 2 3 1 7 1 2 3 7 3 1 2 3 16 16 0 8 12 2 −2 4 6 6 = 9 = 9 10 0 7 7 = −9 10 3 −1 5 4 4 −5 −5 0 1 3 2 1 3 6 54 14 0 −3 4 2 3 = −36 10 7 7 = −36 45 0 −14 −5 1 −5 1 3 3 14 −3 = −2196. = 36 45 −14 7 −5 =⇒ adj(A) = , so that −2 3 1 7/11 −5/11 A−1 = adj(A) = . −2/11 3/11 det(A) 7 −2 −5 3 3 1 1 = 0 2 0 2 3 −1 −5 = −18. −5 −7 (c)2017 Pearson Education. Inc. 8 7 1 12 7 3 247 ⎡ ⎤ ⎡ ⎤ 5 −1 −7 5 −1 −7 5 ⎦ =⇒ adj(A) = ⎣ −1 −7 5 ⎦, so that MC = ⎣ −1 −7 −7 5 −1 −7 5 −1 ⎡ ⎤ −5/18 1/18 7/18 1 7/18 −5/18 ⎦ . adj(A) = ⎣ 1/18 A−1 = det(A) 7/18 −5/18 1/18 3 4 7 7 −11 −38 −11 −38 = 30. 0 0 1 = − 1 = 11. det(A) = 2 6 5 20 3 14 −1 5 20 −1 ⎡ ⎤ ⎡ ⎤ −20 5 10 −20 102 −38 5 −24 11 ⎦, so that MC = ⎣ 102 −24 −30 ⎦ =⇒ adj(A) = ⎣ −38 11 10 10 −30 10 ⎡ ⎤ −2/3 17/5 −19/15 1 11/30 ⎦ . adj(A) = ⎣ 1/6 −4/5 A−1 = det(A) 1/3 −1 1/3 2 5 7 12. det(A) = 4 −3 2 = 116. 6 9 11 ⎡ ⎤ ⎡ ⎤ −51 −32 54 −51 8 31 8 −20 12 ⎦ =⇒ adj(A) = ⎣ −32 −20 24 ⎦, so that MC = ⎣ 31 24 −26 54 12 −26 ⎡ ⎤ −51/116 2/29 31/116 1 −8/29 −5/29 6/29 ⎦ . adj(A) = ⎣ A−1 = det(A) 27/58 3/29 −13/58 5 −1 2 5 3 −1 4 5 = −152. 13. det(A) = 2 1 1 −1 5 9 −3 2 ⎤ ⎤ ⎡ ⎡ −38 0 38 0 −38 32 34 2 ⎢ 32 ⎢ 28 −124 −24 ⎥ 0 28 44 −60 ⎥ ⎥, so that ⎥ ⎢ MC = ⎢ ⎣ 38 −124 −222 130 ⎦ =⇒ adj(A) = ⎣ 34 44 −222 −16 ⎦ 2 −60 130 8 0 −24 −16 8 ⎤ ⎡ 1/4 0 −1/4 0 ⎢ −4/19 −7/38 1 31/38 3/19 ⎥ ⎥. A−1 = adj(A) = ⎢ ⎣ −17/76 −11/38 111/76 2/19 ⎦ det(A) −1/76 15/38 −65/76 −1/19 14. Applying cofactor expansion to the first row of A, we find that det(A) = (−1) · C11 + 4 · C14 = (−1) · 2 + 4 · 14 = 54. To apply the adjoint method, we must first compute the matrix of cofactors: ⎤ ⎡ 2 26 −16 14 ⎢ −4 56 −22 −1 ⎥ ⎥. MC = ⎢ ⎣ 12 −6 12 3 ⎦ 4 −2 22 1 (c)2017 Pearson Education. Inc. 248 Thus, ⎤ 2 −4 12 4 ⎢ 26 56 −6 −2 ⎥ ⎥ adj(A) = MCT = ⎢ ⎣ −16 −22 12 22 ⎦ . 14 −1 3 1 ⎡ Therefore, we have 15. det(A) = ⎤ 2 −4 12 4 1 ⎢ 56 −6 −2 ⎥ ⎥. ⎢ 26 A−1 = 54 ⎣ −16 −22 12 22 ⎦ 14 −1 3 1 ⎡ 2 8 0 8 2 0 = 24, det(B1 ) = = 24, det(B2 ) = = −6. Thus, −2 4 −3 4 −2 −3 x1 = det(B1 ) 24 = =1 det(A) 24 Solution: (1, − 14 ). 4 3 5 16. A = , B1 = 9 6 2 17. A = cos t sin t x2 = det(B2 ) 6 1 =− =− . det(A) 24 4 3 4 , B2 = , so that 6 9 det(B2 ) det(B1 ) 37 1 = , x2 = =− . x1 = det(A) 24 det(A) 8 5 2 and −t sin t e cos t e−t sin t , B , so that = , B1 = 2 3e−t − cos t sin t 3e−t − cos t det(B1 ) det(B2 ) x1 = = e−t [cos t + 3 sin t], x2 = = e−t [sin t − 3 cos t]. det(A) det(A) ⎤ ⎡ ⎤ ⎡ ⎡ ⎤ ⎤ 4 1 3 5 1 3 4 5 3 4 1 5 18. A = ⎣ 2 −1 5 ⎦ , B1 = ⎣ 7 −1 5 ⎦ , B2 = ⎣ 2 7 5 ⎦ , B3 = ⎣ 2 −1 7 ⎦, so that 2 3 1 2 3 1 2 3 2 2 2 1 det(B1 ) det(B2 ) det(B3 ) 1 1 21 = , x2 = = , x3 = = . x1 = det(A) 4 det(A) 16 det(A) 16 ⎡ ⎡ 5 19. A = ⎣ 2 2 ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ ⎤ 3 3 6 5 3 6 5 3 3 3 6 4 −7 ⎦ , B1 = ⎣ −1 4 −7 ⎦ , B2 = ⎣ 2 −1 −7 ⎦ , B3 = ⎣ 2 4 −1 ⎦, so that 4 5 9 2 4 9 2 5 4 5 9 det(B1 ) det(B2 ) det(B3 ) 30 59 81 = , x2 = = , x3 = = . x1 = det(A) 271 det(A) 271 det(A) 271 ⎡ ⎤ 3.6 3.5 3.1 3.5 7.1 20. A = ⎣ 2.2 5.2 6.3 ⎦ , B1 = ⎣ 2.5 5.2 9.3 8.1 1.4 8.1 0.9 so that ⎡ x1 = ⎤ ⎡ 7.1 3.1 6.3 ⎦ , B2 = ⎣ 2.2 0.9 1.4 3.6 2.5 9.3 ⎤ ⎡ ⎤ 7.1 3.1 3.5 3.6 6.3 ⎦ , B3 = ⎣ 2.2 5.2 2.5 ⎦, 0.9 1.4 8.1 9.3 det(B1 ) det(B2 ) det(B3 ) = 3.77, x2 = = 0.66, x3 = = −1.46. det(A) det(A) det(A) (c)2017 Pearson Education. Inc. 249 21. Since A is invertible, A−1 exists. Then, AA−1 = In =⇒ det(AA−1 ) = det(In ) = 1 =⇒ det(A) det(A−1 ) = 1, so that det(A−1 ) = ⎡ a 22. Write A = ⎣ d g b e h 1 . det(A) ⎤ c f ⎦. By cofactor expansion along the first row, we have i det(A) = a(ei − f h) − b(di − f g) + c(dh − eg) = aei + bf g + cdh − af h − bdi − ceg. ⎡ ⎤ a d g Next, we have AT = ⎣ b e h ⎦. By cofactor expansion along the first row, we have c f i det(AT ) = a(ei − f h) − d(bi − ch) + g(bf − ce) = aei + bf g + cdh − af h − bdi − ceg. Therefore, we see that det(A) = det(AT ). 23. det(2A) = 23 det(A) = 24. 24. det(A−1 ) = 1 1 = . det(A) 3 25. det(AT B) = det(AT ) det(B) = det(A) det(B) = −12. 26. det(B 5 ) = [det(B)]5 = −1024. 27. det(B −1 AB) = det(B −1 ) det(A) det(B) = 32 = 9. 1 det(A) det(B) = det(A) = 3, so that det(B −1 AB)2 = det(B) 28. Interchanging two columns multiplies the determinant by −1 and multiplying a column by 4 multiplies the determinant by 4. Therefore, det(C) = −1 · 4 · det(B) = 16. 29. Multiplying a row by 3 multiplies the determinant by 3 and adding the first row to the last row does not change the determinant. Thus, det(C) = 3 · det(A) = 9. Solutions to Section 3.5 Additional Problems: 1. (a). We have det(A) = (6)(1) − (−2)(6) = 18. (b). We have 1 A∼ 1 −2 1 1 2 ∼ 1. M1 (1/6) 1 0 1 3 = B. 2. A12 (2) Now, det(A) = 6 det(B) = 6 · 3 = 18. (c)2017 Pearson Education. Inc. 250 (c). Let us use the Cofactor Expansion Theorem along the first row: det(A) = a11 C11 + a12 C12 = (6)(1) + (6)(2) = 18. 2. (a). We have det(A) = (−7)(−5) − (1)(−2) = 37. (b). We have 1 1 −5 −7 −2 A∼ 2 ∼ 1. P12 1 −5 0 −37 = B. 2. A12 (7) Now, det(A) = −det(B) = −(−37) = 37. (c). Let us use the Cofactor Expansion Theorem along the first row: det(A) = a11 C11 + a12 C12 = (−7)(−5) + (−2)(−1) = 37. 3. (a). Using Equation (3.1.2), we have det(A) = (2)(0)(3)+(3)(2)(6)+(−5)(−4)(−3)−(−6)(0)(−5)−(−3)(2)(2)−(3)(−4)(3) = 0+36−60+−0+12+36 = 24. (b). We have ⎡ ⎤ ⎡ 2 3 −5 2 2 6 −8 ⎦ ∼ ⎣ 0 A∼⎣ 0 0 −12 18 0 1 1. A12 (2), A13 (−3) ⎤ 3 −5 6 −8 ⎦ = B. 0 2 2. A23 (2) Now, det(A) = det(B) = (2)(6)(2) = 24. (c). Let us use the Cofactor Expansion Theorem along the second row: det(A) = (−4)(6) + (0)(36) + (2)(24) = −24 + 0 + 48 = 24. 4. (a). Using Equation (3.1.2), we have det(A) = (−1)(2)(−3)+(4)(2)(2)+(1)(0)(2)−(−1)(2)(2)−(4)(0)(−3)−(1)(2)(2) = 6+16+0−(−4)−0−4 = 22. (b). We have ⎡ −1 A∼⎣ 0 0 1 ⎤ ⎡ 4 1 −1 2 2 2 ⎦∼⎣ 0 10 −1 0 1. A13 (2) ⎤ 4 1 2 2 ⎦ = B. 0 −11 2. A23 (−5) (c)2017 Pearson Education. Inc. 251 Now, det(A) = det(B) = (−1)(2)(−11) = 22. (c). Let us use the Cofactor Expansion Theorem along the first column: det(A) = a11 C11 + a21 C21 + a31 C31 = (−1)(−10) + (0)(14) + (2)(6) = 22. 5. (a). Of the 24 terms in the determinant expression (3.1.3), the only nonzero term is a41 a32 a23 a14 , which has a positive sign: σ(4, 3, 2, 1) = +1. Therefore, det(A) = (−3)(1)(−5)(−2) = −30. (b). By permuting the first and last rows, and by permuting the middle two rows, we bring A to upper triangular form. The two permutations introduce two sign changes, so the resulting upper triangular matrix has the same determinant as A, and it is (−3)(1)(−5)(−2) = −30. 0 0 −2 1 . This latter (c). Cofactor expansion along the first column yields: det(A) = −(−3) · 0 −5 1 −4 1 determinant can be found by cofactor expansion along the first column: (1)[(0)(1) − (−5)(−2)] = −10. Hence, det(A) = 3 · (−10) = −30. 6. (a). Of the 24 terms appearing in the determinant expression (3.1.3), only terms containing the factors a11 and a44 will be nonzero (all other entries in the first column and fourth row of A are zero). Looking at entries in the second and third rows and columns of A, we see that only the product a23 a32 is nonzero. Therefore, the only nonzero term in the summation (3.1.3) is a11 a23 a32 a44 = (3)(1)(2)(−4) = −24. The permutation associated with this term is (1, 3, 2, 4) which contains one inversion. Therefore, σ(1, 3, 2, 4) = −1, and so the determinant is (−24)(−1) = 24. (b). We have ⎡ ⎤ 3 −1 −2 1 2 1 −1 ⎥ 1 ⎢ 0 ⎥ = B. A∼⎢ ⎣ 0 0 1 4 ⎦ 0 0 0 −4 1. P23 Now, det(A) = −det(B) = −(3)(2)(1)(−4) = 24. 0 1 4 (c). Cofactor expansion along the first column yields: det(A) = 3· 2 1 −1 . This latter determinant can 0 0 −4 be found by cofactor expansion along the last column: (−4)[(0)(1) − (2)(1)] = 8. Thus, det(A) = 3 · 8 = 24. 7. To obtain the given matrix from A, we add 5 times the middle row to the top row, we multiply the last row by 3, we multiply the middle row by −1, we add a multiple of the last row to the middle row, and we perform a row permutation. The combined effect of these operations is to multiply the determinant of A by (1) · (3) · (−1) · (−1) = 3. Hence, the given matrix has determinant det(A) · (3) = 4 · (3) = 12. 8. To obtain the given matrix from A, we perform two row permutations, multiply a row through by −4, and multiply a row through by 2. The combined effect of these operations is to multiply the determinant of A by (−1)2 · (−4) · (2) = −8. Hence, the given matrix has determinant det(A) · (−8) = 4 · (−8) = −32. (c)2017 Pearson Education. Inc. 252 9. To obtain the given matrix from A, we permute the bottom two rows, multiply a row by 2, multiply a row by −1, add −1 times one row to another, and then multiply each row of the matrix by 3. The combined effect of these operations is to multiply the determinant of A by (−1)(2)(−1)(1)(3)3 = 54. Hence, the given matrix has determinant det(A) · 54 = (4)(54) = 216. 10. To obtain the given matrix from AT , we do two row permutations, multiply a row by −1, multiply a row by 3, and add 2 times one row to another. The combined effect of these operations is to multiply the determinant of A by (−1)(−1)(−1)(3)(1) = −3. Hence, the given matrix has determinant det(A) · (−3) = (4) · (−3) = −12. 2 3 11. We have det(B 2 A−1 ) = (det(B))2 det(A−1 ) = −2 = −4.5. 12. We have det(AB) = det(A)det(B) = (−2) · 3 = −6. 13. We have det((−A)3 (2B 2 )) = det(−A)3 det(2B 2 ) = (−1)3 (detA)3 · 24 · (detB)2 = (−1)3 (−2)3 · 24 · 32 = 1152. 14. We have det((A−1 B)T (2B −1 )) = det(A−1 B) · det(2B −1 ) = − 12 · 3 · 24 · 13 = −8. 15. det(A) cannot be computed. 16. det(B) cannot be computed. 17. det(C) = −18. 18. det(C T ) = −18. 19. We have AB = ⎡ 8 −10 25 28 4 20. We have BA = ⎣ 1 18 , and so det(AB) = 474. ⎤ 5 2 8 −13 ⎦, we have det(BA) = 0. 15 24 21. We have det(B T AT ) = det((AB)T ) = det(AB) = 474. 22. We have det(BAC) = det(BA)det(C) = 0 · (−18) = 0. 38 54 23. We have det(ACB) = det = 4104. 133 297 ⎡ ⎤ 1 2 1 2−1 6 0 ⎣ 2 1 ⎦ = 24. Note that AAT = , so that det(AAT ) = 6 · 21 = 126. 2 1 4 0 21 −1 4 Therefore, we conclude that det((AAT )2 ) = [det(AAT )]2 = 1262 = 15876. ⎡ ⎤ ⎡ ⎤ 2 1 4 5 2 1 2 −1 8 −13 ⎦, and using the schematic in Figure 25. Note that BA = ⎣ 5 −2 ⎦ =⎣ 1 2 1 4 4 7 18 15 24 1 · 0 = 0. 3.1.1., for example, we find that det(BA) = 0. Therefore, det(C −1 BA) = det(C −1 )det(BA) = − 18 26. Since B is a 3 × 2 matrix and C is a 3 × 3 matrix, we cannot compute BCC T . 1 −1 1 −4 , and so adj(B) = . Since det(B) = 1, 27. The matrix of cofactors of B is MC = −4 5 −1 5 (c)2017 Pearson Education. Inc. 253 we have B −1 = 1 −4 −1 5 . Therefore, (A−1 B T )−1 = (B T )−1 A = (B −1 )T A = 28. ⎤ 16 −1 −5 5 −3 ⎦; MC = ⎣ 4 −4 2 10 ⎡ ⎤ 16 4 −4 5 2 ⎦; adj(A) = ⎣ −1 −5 −3 10 det(A) ⎡ = 28; ⎤ 4/7 1/7 −1/7 1/14 ⎦. A−1 = ⎣ −1/28 5/28 −5/28 −3/28 5/14 29. ⎤ 5 12 −11 −1 ⎢ 65 −24 −23 −13 ⎥ ⎥; MC = ⎢ ⎣ −25 0 −5 5 ⎦ −35 0 5 −5 ⎤ ⎡ 5 65 −25 −35 ⎢ 12 −24 0 0 ⎥ ⎥; adj(A) = ⎢ ⎣ −11 −23 −5 5 ⎦ −1 −13 5 −5 det(A) = −60; ⎤ ⎡ 5 65 −25 −35 1 ⎢ 12 −24 0 0 ⎥ ⎥. A−1 = − ⎢ ⎦ ⎣ −11 −23 −5 5 60 −1 −13 5 −5 30. ⎤ 88 −24 −40 −48 ⎢ 32 12 −20 0 ⎥ ⎥; MC = ⎢ ⎣ 16 12 −4 0 ⎦ −4 6 −2 0 ⎤ ⎡ 88 32 16 −4 ⎢ −24 12 12 6 ⎥ ⎥ adj(A) = ⎢ ⎣ −40 −20 −4 −2 ⎦; −48 0 0 0 det(A) = −48; ⎤ ⎡ 88 32 16 −4 1 ⎢ −24 12 12 6 ⎥ ⎥. A−1 = − ⎢ 48 ⎣ −40 −20 −4 −2 ⎦ −48 0 0 0 1 −1 −4 5 ⎡ ⎡ ⎡ 31. (c)2017 Pearson Education. Inc. 1 3 2 4 = −2 −2 11 12 . 254 ⎡ ⎤ 12 −12 9 −12 ⎦; 24 −27 ⎤ 21 24 48 9 24 ⎦; adj(A) = ⎣ 12 −12 −12 −27 det(A) ⎡ = 9; ⎤ 7/3 8/3 16/3 1 8/3 ⎦. A−1 = ⎣ 4/3 −4/3 −4/3 −3 21 MC = ⎣ 24 48 ⎡ 32. ⎡ ⎤ 7 −2 0 2 −2 ⎦; MC = ⎣ 0 −6 0 2 ⎡ ⎤ 7 0 −6 2 0 ⎦; adj(A) = ⎣ −2 0 −2 2 det(A) ⎡ = 2; ⎤ 3.5 0 −3 1 0 ⎦. A−1 = ⎣ −1 0 −1 1 ⎤ 4 −1 0 1 4 ⎦. 33. There are many ways to do this. One choice is to let B = ⎣ 5 0 0 10 9 ⎡ ⎤ 1 2 3 34. FALSE. For instance, if one entry of the matrix A = ⎣ 1 2 3 ⎦ is changed, two of the rows of A 1 2 3 will still be identical, and therefore, the determinant of the resulting matrix must be zero. It is not possible to force the determinant to equal r. ⎡ 35. Note that det(A) = 2 + 12k + 36 − 4k − 18 − 12 = 8k + 8 = 8(k + 1). (a). Based on the calculation above, we see that A fails to be invertible if and only if k = −1. (b). The volume of the parallelepiped determined by the row vectors of A is precisely |det(A)| = |8k + 8|. The volume of the parallelepiped determined by the column vectors of A is the same as the volume of the parallelepiped determined by the row vectors of AT , which is |det(AT )| = |det(A)| = |8k + 8|. Thus, the volume is the same. 36. Note that det(A) = 3(k + 1) + 2k + 0 − 3 − k(k + 1) − 0 = −k 2 + 4k = k(4 − k). (a). Based on the calculation above, we see that A fails to be invertible if and only if k = 0 or k = 4. (b). The volume of the parallelepiped determined by the row vectors of A is precisely |det(A)| = | − k 2 + 4k|. The volume of the parallelepiped determined by the column vectors of A is the same as the volume of the parallelepiped determined by the row vectors of AT , which is |det(AT )| = |det(A)| = | − k 2 + 4k|. Thus, the volume is the same. (c)2017 Pearson Education. Inc. 255 37. Note that det(A) = 0 + 4(k − 3) + 2k 3 − k 2 − 8k − 0 = 2k 3 − k 2 − 4k − 12. (a). Based on the calculation above, we see that A fails to be invertible if and only if 2k 3 − k 2 − 4k − 12 = 0, and the only real solution to this equation for k is k ≈ 2.39 (from calculator). (b). The volume of the parallelepiped determined by the row vectors of A is precisely |det(A)| = |2k 3 − k 2 − 4k−12|. The volume of the parallelepiped determined by the column vectors of A is the same as the volume of the parallelepiped determined by the row vectors of AT , which is |det(AT )| = |det(A)| = |2k 3 − k 2 − 4k − 12|. Thus, the volume is the same. 38. From the assumption that AB = −BA, we can take the determinant of each side: det(AB) = det(−BA). Hence, det(A)det(B) = (−1)n det(B)det(A) = −det(A)det(B). From this, it follows that det(A)det(B) = 0, and therefore, either det(A) = 0 or det(B) = 0. Thus, either A or B (or both) fails to be invertible. 39. Since AAT = In , we can take the determinant of both sides to get det(AAT ) = det(In ) = 1. Hence, det(A)det(AT ) = 1. Therefore, we have (det(A))2 = 1. We conclude that det(A) = ±1. −3 1 40. The coefficient matrix of this linear system is A = . We have 1 2 −3 1 = −7, det(B1 ) = 3 1 = 5, and det(B2 ) = −3 3 = −6. det(A) = 1 2 1 2 1 1 Thus, x1 = Solution: (−5/7, 6/7). det(B1 ) det(B2 ) 5 6 =− and x2 = = . det(A) 7 det(A) 7 ⎤ 2 −1 1 5 3 ⎦. We have 41. The coefficient matrix of this linear system is A = ⎣ 4 4 −3 3 2 −1 1 2 −1 1 5 3 = 32, 5 3 = 16, det(B1 ) = 0 det(A) = 4 2 −3 3 4 −3 3 2 2 1 2 −1 2 5 0 = −36. det(B2 ) = 4 0 3 = −4, det(B3 ) = 4 4 2 3 4 −3 2 ⎡ Thus, x1 = det(B1 ) det(B2 ) 1 = 2, x2 = =− , det(A) det(A) 4 Solution: (2, −1/4, −9/4). and x3 = ⎡ det(B3 ) 9 =− . det(A) 4 ⎤ 3 1 2 42. The coefficient matrix of this linear system is A = ⎣ 2 −1 1 ⎦. We have 0 5 5 −1 3 1 2 1 2 det(A) = 2 −1 1 = −20, det(B1 ) = −1 −1 1 = −10, −5 0 5 5 5 5 (c)2017 Pearson Education. Inc. 256 3 −1 2 3 1 −1 det(B2 ) = 2 −1 1 = −10, det(B3 ) = 2 −1 −1 = 30. 0 −5 5 0 5 −5 Thus, x1 = det(B1 ) det(B2 ) 1 1 = , x2 = = , det(A) 2 det(A) 2 and x3 = det(B3 ) 3 =− . det(A) 2 Solution: (1/2, 1/2, −3/2). Chapter 4 Solutions Solutions to Section 4.1 True-False Review: (a): FALSE. The vectors (x, y) and (x, y, 0) do not belong to the same set, so they are not even comparable, let alone equal to one another. (b): TRUE. The unique additive inverse of (x, y, z) is (−x, −y, −z). (c): TRUE. The solution set refers to collections of the unknowns that solve the linear system. Since this system has 6 unknowns, the solution set will consist of vectors belonging to R6 . (d): TRUE. The vector (−1) · (x1 , x2 , . . . , xn ) is precisely (−x1 , −x2 , . . . , −xn ), and this is the additive inverse of (x1 , x2 , . . . , xn ). (e): FALSE. There is no such name for a vector whose components are all positive. (f ): FALSE. The correct result is (s + t)(x + y) = (s + t)x + (s + t)y = sx + tx + sy + ty, and in this item, only the first and last of these four terms appear. (g): TRUE. When the vector x is scalar multiplied by zero, each component becomes zero: 0x = 0. This is the zero vector in Rn . (h): TRUE. This is seen geometrically from addition and subtraction of geometric vectors. (i): FALSE. If k < 0, then kx is a vector in the third quadrant. For instance, (1, 1) lies in the first quadrant, but (−2)(1, 1) = (−2, −2) lies in the third quadrant. (j): TRUE. Recalling that i = (1, 0, 0), j = (0, 1, 0), and k = (0, 0, 1), we have √ √ √ 5i − 6j + 2k = 5(1, 0, 0) − 6(0, 1, 0) + 2(0, 0, 1) = (5, −6, 2), as stated. (k): FALSE. If the three vectors lie on the same line or the same plane, the resulting object may determine a one-dimensional segment or two-dimensional area. For instance, if x = y = z = (1, 0, 0), then the vectors x,y, and z rest on the segment from (0, 0, 0) to (1, 0, 0), and do not determine a three-dimensional solid region. (c)2017 Pearson Education. Inc. 257 (l): FALSE. The components of kx only remain even integers if k is an integer. But, for example, if k = π, then the components of kx are not even integers at all, let alone even integers. Problems: 1. v1 = (−3, −12), v2 = (20, −4), v3 = (−3, −12) + (20, −4) = (17, −16). y x v2 v1 (20, -4) v3 (-3, -12) (17, -16) Figure 0.0.60: Figure for Problem 1 2. v1 = (6, 2), v2 = (−3, 6), v3 = (6, 2) + (−3, 6) = (3, 8). y (-3, 6) (3, 8) v3 v2 (6, 2) v1 x Figure 0.0.61: Figure for Problem 2 3. We have v = −2x + 10y = −2(5, −2, 9) + 10(−1, 6, 4) = (−20, 64, 22), and therefore, the additive inverse of v is −v = (20, −64, −22). 4. v = 5(3, −1, 2, 5) − 7(−1, 2, 9, −2) = (15, −5, 10, 25) − (−7, 14, 63, −14) = (22, −19, −53, 39). Additive inverse: −v = (−1)v = (−22, 19, 53, −39). (c)2017 Pearson Education. Inc. 258 5. We have 2 1 2 1 y = x + z = (1, 2, 3, 4, 5) + (−1, 0, −4, 1, 2) = 3 3 3 3 1 4 2 , , , 3, 4 . 3 3 3 6. Solving for z, we have 1 y − 2x 2 1 = (3, 0, −5) − 2(−3, 9, 9) 2 15 41 = ( , −18, − ). 2 2 z= 7. Solving for z, we have 1 1+i x− y 2i i 1−i (−2 + i, 3i) + i(5, 2 − 2i) = 2 1 = (−1 + 3i, 3 + 3i) + (5i, 2 + 2i) 2 1 13 7 7 = (− + i, + i). 2 2 2 2 z= 8. Solving for z, we have 2 3i x+ y 1+i 1+i 3 3 = (1 − i)x + ( + i)y 2 2 z= 3 3 = (1 − i)(5 + i, 0, −1 − 2i, 1 + 8i) + ( + i)(−3, i, i, 3) 2 2 9 9 3 3 3 3 9 9 = (6 − 4i, 0, −3 − i, 9 + 7i) + (− − i, − + i, − + i, + i) 2 2 2 2 2 2 2 2 3 3 9 1 27 23 3 17 − i, − + i, − + i, + i . = 2 2 2 2 2 2 2 2 9. Let x = (x1 , x2 , x3 , x4 ), y = (y1 , y2 , y3 , y4 ) be arbitrary vectors in R4 . Then x+y = (x1 , x2 , x3 , x4 ) + (y1 , y2 , y3 , y4 ) = (x1 + y1 , x2 + y2 , x3 + y3 , x4 + y4 ) = = (y1 + x1 , y2 + x2 , y3 + x3 , y4 + x4 ) = (y1 , y2 , y3 , y4 ) + (x1 , x2 , x3 , x4 ) y + x. 10. Let x = (x1 , x2 , x3 , x4 ), y = (y1 , y2 , y3 , y4 ), z = (z1 , z2 , z3 , z4 ) be arbitrary vectors in R4 . Then x + (y + z) = = = (x1 , x2 , x3 , x4 ) + (y1 + z1 , y2 + z2 , y3 + z3 , y4 + z4 ) (x1 + (y1 + z1 ), x2 + (y2 + z2 ), x3 + (y3 + z3 ), x4 + (y4 + z4 )) ((x1 + y1 ) + z1 , (x2 + y2 ) + z2 , (x3 + y3 ) + z3 , (x4 + y4 ) + z4 ) = = (x1 + y1 , x2 + y2 , x3 + y3 , x4 + y4 ) + (z1 , z2 , z3 , z4 ) (x + y) + z. (c)2017 Pearson Education. Inc. 259 11. Let x = (x1 , x2 , x3 ) and y = (y1 , y2 , y3 ) be arbitrary vectors in R3 , and let r, s, t be arbitrary real numbers. Then: 1x = 1(x1 , x2 , x3 ) = (1x1 , 1x2 , 1x3 ) = (x1 , x2 , x3 ) = x. (st)x = (st)(x1 , x2 , x3 ) = ((st)x1 , (st)x2 , (st)x3 ) = (s(tx1 ), s(tx2 ), s(tx3 )) = s(tx1 , tx2 , tx3 ) = s(tx). r(x + y) = r(x1 + y1 , x2 + y2 , x3 + y3 ) = (r(x1 + y1 ), r(x2 + y2 ), r(x3 + y3 )) = (rx1 + ry1 , rx2 + ry2 , rx3 + ry3 ) = (rx1 , rx2 , rx3 ) + (ry1 , ry2 , ry3 ) = rx + ry. (s + t)x = (s + t)(x1 , x2 , x3 ) = ((s + t)x1 , (s + t)x2 , (s + t)x3 ) = (sx1 + tx1 , sx2 + tx2 , sx3 + tx3 ) = (sx1 , sx2 , sx3 ) + (tx1 , tx2 , tx3 ) = sx + tx. 12. Let x = (x1 , x2 , x3 , x4 , x5 ) and y = (y1 , y2 , y3 , y4 , y5 ) be arbitrary vectors in R5 , and let r, s, t be arbitrary real numbers. Then: 1x = 1(x1 , x2 , x3 , x4 , x5 ) = (1x1 , 1x2 , 1x3 , 1x4 , 1x5 ) = (x1 , x2 , x3 , x4 , x5 ) = x. (st)x = (st)(x1 , x2 , x3 , x4 , x5 ) = ((st)x1 , (st)x2 , (st)x3 , (st)x4 , (st)x5 ) = (s(tx1 ), s(tx2 ), s(tx3 ), s(tx4 ), s(tx5 )) = s(tx1 , tx2 , tx3 , tx4 , tx5 ) = s(tx). r(x + y) = r(x1 + y1 , x2 + y2 , x3 + y3 , x4 + y4 , x5 + y5 ) = (r(x1 + y1 ), r(x2 + y2 ), r(x3 + y3 ), r(x4 + y4 ), r(x5 + y5 )) = (rx1 + ry1 , rx2 + ry2 , rx3 + ry3 , rx4 + ry4 , rx5 + ry5 ) = (rx1 , rx2 , rx3 , rx4 , rx5 ) + (ry1 , ry2 , ry3 , ry4 , ry5 ) = rx + ry. (s + t)x = (s + t)(x1 , x2 , x3 , x4 , x5 ) = ((s + t)x1 , (s + t)x2 , (s + t)x3 , (s + t)x4 , (s + t)x5 ) = (sx1 + tx1 , sx2 + tx2 , sx3 + tx3 , sx4 + tx4 , sx5 + tx5 ) = (sx1 , sx2 , sx3 , sx4 , sx5 ) + (tx1 , tx2 , tx3 , tx4 , tx5 ) = sx + tx. 13. For example, if x = (2, 2) and y = (−1, −1), then x + y = (1, 1) lies in the first quadrant. If x = (1, 2) and y = (−5, 1), then x + y = (−4, 3) lies in the second quadrant. If x = (1, 1) and y = (−2, −2), then x + y = (−1, −1) lies in the third quadrant. If x = (2, 1) and y = (−1, −5), then x + y = (1, −4) lies in the fourth quadrant. Solutions to Section 4.2 True-False Review: (a): TRUE. This is part 1 of Theorem 4.2.7. (b): FALSE. The statement would be true if it was required that v be nonzero. However, if v = 0, then rv = sv = 0 for all values of r and s, and r and s need not be equal. We conclude that the statement is not true. (c)2017 Pearson Education. Inc. 260 (c): TRUE. From rv = sv, we subtract to obtain 0 = rv − sv = (r − s)v. By part (6) of Theorem 4.2.7, we must have either r − s = 0 or v = 0. However, v = 0 is given. Thus, we must conclude that r − s = 0, or r = s. (d): FALSE. This set is not closed under scalar multiplication. In particular, if k is an irrational number such as k = π and v is an integer, then kv is not an integer. (e): TRUE. We have (x + y) + ((−x) + (−y)) = (x + (−x)) + (y + (−y)) = 0 + 0 = 0, where we have used the vector space axioms in these steps. Therefore, the additive inverse of x + y is (−x) + (−y). (f ): TRUE. This is part 1 of Theorem 4.2.7. (g): TRUE. This is called the trivial vector space. Since 0 + 0 = 0 and k · 0 = 0, it is closed under addition and scalar multiplication. Both sides of the remaining axioms yield 0, and 0 is the zero vector, and it is its own additive inverse. (h): FALSE. This set is not closed under addition, since 1 + 1 ∈ {0, 1}. Therefore, (A1) fails, and hence, this set does not form a vector space. (It is worth noting that the set is also not closed under scalar multiplication.) (i): FALSE. This set is not closed under scalar multiplication. If k < 0 and x is a positive real number, the result kx is a negative real number, and therefore no longer belongs to the set of positive real numbers. (j): FALSE. Any positive element of the given set fails to have an additive inverse in the set, so Axiom (A6) fails. Problems: 1. If x = p/q and y = r/s, where p, q, r, s are integers (q = 0, s = 0), then x + y = (ps + qr)/(qs), which is a rational number. Consequently, the set of all rational numbers is closed under addition. The set is not closed under scalar multiplication since, if we multiply a rational number by an irrational number, the result is an irrational number. 2. Let A = [aij ] and B = [bij ] be upper triangular matrices, and let k be an arbitrary real number. Then whenever i > j, it follows that aij = 0 and bij = 0. Consequently, aij + bij = 0 whenever i > j, and kaij = 0 whenever i > j. Therefore, A + B and kA are upper triangular matrices, so the set of all upper triangular matrices with real elements is closed under both addition and scalar multiplication. 3. In this case, S fails to be closed under addition, but it is closed under scalar multiplication. (A1) fails: S is not closed under addition. For instance, using the case n = 2, we can give this specific 0 1 0 0 example (many such examples are acceptable): Let A = and B = . Since A is upper 0 0 1 0 0 1 triangular and B is lower triangular, both A and B belong to S. However, A + B = is neither 1 0 upper nor lower triangular. Thus, A + B ∈ S. (A2) holds: S is closed under scalar multiplication. To see this, suppose A ∈ S. This means A is either upper triangular, or A is lower triangular. If A is upper triangular, then for any real constant r, rA is still upper triangular. Thus, rA ∈ S. A similar argument holds if A is lower triangular. 4. In this case, S is closed under both addition and scalar multiplication. (c)2017 Pearson Education. Inc. 261 (A1) holds: S is closed under addition. To see this, suppose p(x) = a0 + a1 x + a2 x2 and q(x) = b0 + b1 x + b2 x2 belong to S.This means that a0 + a1 + a2 = 0 and b0 + b1 + b2 = 0. Then p(x) + q(x) = (a0 + a1 x + a2 x2 ) + (b0 + b1 x + b2 x2 ) = (a0 + b0 ) + (a1 + b1 )x + (a2 + b2 )x2 , and since (a0 + b0 ) + (a1 + b1 ) + (a2 + b2 ) = (a0 + a1 + a2 ) + (b0 + b1 + b2 ) = 0 + 0 = 0, we conclude that p(x) + q(x) ∈ S. Hence, S is closed under addition. (A2) holds: S is closed under scalar multiplication. To see this, suppose p(x) is as above, and let r be a constant. Then r · p(x) = (ra0 ) + (ra1 )x + (ra2 )x2 , and since ra0 + ra1 + ra2 = r(a0 + a1 + a2 ) = r · 0 = 0, we conclude that r · p(x) ∈ S. Hence, S is closed under scalar multiplication. 5. In this case, S is not closed under addition and not closed under scalar multiplication. (A1) fails: S is not closed under addition. We can demonstrate with an example (many such examples can be given). Let p(x) = 1 and let q(x) = x. Note that p(x) ∈ S and q(x) ∈ S. However, p(x)+q(x) = 1+x ∈ S. Thus, S fails to be closed under scalar multiplication. (A2) fails: S is not closed under scalar multiplication. We can demonstrate with an example (many such examples can be given). With p(x) = 1 and r = 2, we have p(x) ∈ S, but r · p(x) = 2 ∈ S. Thus, S fails to be closed under scalar multiplication. 6. In this case, S is not closed under addition and not closed under scalar multiplication. (A1) fails: S is not closed under addition. To see this, suppose y1 and y2 are both solutions to the given differential equation. That is, y1 + 3y1 = 6x3 + 5 and y2 + 3y2 = 6x3 + 5. Then (y1 + y2 ) + 3(y1 + y2 ) = y1 + y2 + 3y1 + 3y2 = (y1 + 3y1 ) + (y2 + 3y2 ) = (6x3 + 5) + (6x3 + 5) = 6x3 + 5, so that y1 + y2 ∈ S. Thus, S is not closed under addition. (A2) fails: S is not closed under scalar multiplication. To see this, let y1 be as above and let r be a scalar. Then (ry1 ) + 3(ry1 ) = r(y1 + 3y1 ) = r(6x3 + 5) = 6x3 + 5 for r = 1. Thus, unless r = 1, we have that ry1 ∈ S. Thus, S is not closed under scalar multiplication. 7. In this case, S is closed under both addition and scalar multiplication. (A1) holds: S is closed under addition. To see this, suppose y1 and y2 are both solutions to the given differential equation. That is, y1 + 3y1 = 0 and y2 + 3y2 = 0. Then (y1 + y2 ) + 3(y1 + y2 ) = y1 + y2 + 3y1 + 3y2 = (y1 + 3y1 ) + (y2 + 3y2 ) = 0 + 0 = 0, so that y1 + y2 ∈ S. Thus, S is closed under addition. (c)2017 Pearson Education. Inc. 262 (A2) holds: S is closed under scalar multiplication. To see this, let y1 be as above and let r be a scalar. Then (ry1 ) + 3(ry1 ) = r(y1 + 3y1 ) = r(0) = 0. Thus, we have that ry1 ∈ S. Thus, S is closed under scalar multiplication. 8. In this case, S is closed under both addition and scalar multiplication. (A1) holds: S is closed under addition. To see this, suppose x1 and x2 belong to S. Thus, Ax1 = 0 and Ax2 = 0. Then, since A(x1 + x2 ) = Ax1 + Ax2 = 0 + 0 = 0, we conclude that x1 + x2 ∈ S. Thus, S is closed under addition. (A2) holds: S is closed under scalar multiplication. To see this, let x1 be as above and let r be a scalar. Then A(rx1 ) = r(Ax1 ) = r(0) = 0, so that rx1 ∈ S. Thus, S is closed under scalar multiplication. 9. In this case, S is not closed under addition, but S is closed under scalar multiplication. 1 0 0 0 (A1) fails: S is not closed under addition. As a specific example to show this, note that A = ∈S 0 0 and B = ∈ S. However, A + B = I2 , and since det(I2 ) = 1 = 0, we have det(A + B) = 0, so 0 1 that A + B ∈ S. Thus, S is not closed under addition. (Many other examples could have been used to demonstrate this.) (A2) holds: S is closed under scalar multiplication. Suppose that A ∈ S and r is a scalar. This means that det(A) = 0. Using Property (P4) of determinants, we then have det(rA) = r2 det(A) = r2 · 0 = 0, so that rA ∈ S. Thus, S is closed under scalar multiplication. 10. In this case, S is closed under neither addition nor scalar multiplication. (A1) fails: S is not closed under addition. As a specific example to show this, note that v = (−1, 1) ∈ S and w = (1, 1) ∈ S. However, v + w = (−1, 1) + (1, 1) = (0, 2) ∈ S. Thus, S is not closed under addition. (Many other examples could have been used to demonstrate this.) (A2) fails: S is not closed under scalar multiplication. As a specific example to show this, note that w = (1, 1) ∈ S. However, 2 · w = (2, 2) ∈ S. Thus, S is not closed under scalar multiplication. (Many other examples could have been used to demonstrate this.) 11. In this case, S is closed under neither addition nor scalar multiplication. (A1) fails: S is not closed under addition. As a specific example to show this, note that v = (0, 1) ∈ S and w = (1, 2) ∈ S. However, v + w = (0, 1) + (1, 2) = (1, 3) ∈ S. Thus, S is not closed under addition. (Many other examples could have been used to demonstrate this.) (A2) fails: S is not closed under scalar multiplication. As a specific example to show this, note that v = (0, 1) ∈ S, while 2 · v = (0, 2) ∈ S. Thus, S is not closed under scalar multiplication. (Many other examples could have been used to demonstrate this.) 12. In this case, S is closed under addition, but S is not closed under scalar multiplication. (A1) holds: N is closed under addition. To see this, note that if m, n ∈ N, then m + n ∈ N since the sum of two positive integers is another positive integer. Thus, S is closed under addition. (A2) fails: N is not closed under scalar multiplication. To see this, note that if m ∈ N, then −1 · m = −m ∈ N, since −m is a negative integer. Thus, S is not closed under scalar multiplication. (Many other examples could have been used to demonstrate this.) 13. In this case, S is closed under neither addition nor scalar multiplication. (c)2017 Pearson Education. Inc. 263 (A1) fails: S is not closed under addition. For a specific example of this, note that p(x) = x2 ∈ S and q(x) = x − x2 ∈ S. However, p(x) + q(x) = x2 + (x − x2 ) = x ∈ S, which shows that S not closed under addition. (Many other examples could have been used to demonstrate this.) (A2) fails: S is not closed under scalar multiplication. For a specific example of this, note that p(x) = x2 ∈ S, but 0 · p(x) = 0 ∈ S. Thus, S is not closed under scalar multiplication. (Many other examples could have been used to demonstrate this.) 14. In this case, S is closed under both addition and scalar multiplication. (A1) holds: S is closed under addition. To see this, suppose p(x) = a1 + b1 x3 + c1 x4 and q(x) = a2 + b2 x3 + c2 x4 belong to S. Then p(x) + q(x) = (a1 + b1 x3 + c1 x4 ) + (a2 + b2 x3 + c2 x4 ) = (a1 + a2 ) + (b1 + b2 )x3 + (c1 + c2 )x4 ∈ S, since the form of the final expression fits the form for elements of S. Thus, S is closed under addition. (A2) holds: S is closed under scalar multiplication. To see this, suppose p(x) is as above, and let r be a scalar. Then r · p(x) = r(a1 + b1 x3 + c1 x4 ) = (ra1 ) + (rb1 )x3 + (rc1 )x4 ∈ S, since the form of the final expression fits the form for elements of S. Thus, S is closed under scalar multiplication. 15. Let V be R2 , i.e. {(x, y) : x, y ∈ R}. With addition and scalar multiplication as defined in the text, this set is clearly closed under both operations. A3: u + v = (u1 , u2 ) + (v1 , v2 ) = (u1 + v1 , u2 + v2 ) = (v1 + u1 , v2 + u2 ) = v + u. A4: [u+v]+w = [(u1 +v1 , u2 +v2 )]+(w1 , w2 ) = ([u1 +v1 ]+w1 , [u2 +v2 ]+w2 ) = (u1 +[v1 +w1 ], u2 +[v2 +w2 ]) = (u1 , u2 ) + [(v1 + w1 , v2 + w2 )] = u + [v + w]. A5: 0 = (0, 0) since (x, y) + (0, 0) = (x + 0, y + 0) = (x, y). A6: If u = (a, b), then −u = (−a, −b), a, b ∈ R, since (a, b) + (−a, −b) = (a − a, b − b) = (0, 0). Now, let u = (u1 , u2 ), v = (v1 , v2 ), where u1 , u2 , v1 , v2 ∈ R, and let r, s, t ∈ R. A7: 1 · v = 1(v1 , v2 ) = (1 · v1 , 1 · v2 ) = (v1 , v2 ) = v. A8: (rs)v = (rs)(v1 , v2 ) = ((rs)v1 , (rs)v2 ) = (rsv1 , rsv2 ) = (r(sv1 ), r(sv2 )) = r(sv1 , sv2 ) = r(sv). A9: r(u + v) = r((u1 , u2 ) + (v1 , v2 )) = r(u1 + v1 , u2 + v2 ) = (r(u1 + v1 ), r(u2 + v2 )) = (ru1 + rv1 , ru2 + rv2 ) = (ru1 , ru2 ) + (rv1 , rv2 ) = r(u1 , u2 ) + r(v1 , v2 ) = ru + rv. A10: (r + s)u = (r + s)(u1 , u2 ) = ((r + s)u1 , (r + s)u2 ) = (ru1 + su1 , ru2 + su2 ) = (ru1 , ru2 ) + (su1 , su2 ) = r(u1 , u2 ) + s(u1 , u2 ) = ru + su. Thus, R2 is a vector space. ⎤ ⎡ ⎡ ⎤ 0 0 a b ⎢ 0 0 ⎥ ⎥ ⎣ c d ⎦, so that the additive inverse is 16. The zero vector is ⎢ ⎣ 0 0 ⎦, a general element in V is A = e f 0 0 ⎡ ⎤ −a −b −A = ⎣ −c −d ⎦. −e −f 17. The zero vector of Mm×n (R) is 0m×n , the m × n zero matrix. The additive inverse of the m × n matrix A with (i, j)-element aij is the m × n matrix −A with (i, j)-element −aij . 18. The zero vector is p(x) = 0, a general element in V is p(x) = a0 + a1 x + a2 x2 + a3 x3 , so that the additive inverse is −p(x) = −a0 − a1 x − a2 x2 − a3 x3 . 19. The zero vector is p(x) = 0, a general element in V is p(x) = a0 + a1 x + a2 x2 + · · · + an xn , so that the additive inverse is −p(x) = −a0 − a1 x − a2 x2 − · · · − an xn . (c)2017 Pearson Education. Inc. 264 20. YES. We verify the ten axioms of a vector space. A1: The product of two positive real numbers is a positive real number, so the set is closed under addition. A2: Any power of a positive real number is a positive real number, so the set is closed under scalar multiplication. A3: We have x + y = xy = yx = y + x for all x, y ∈ R+ , so commutativity under addition holds. A4: We have (x + y) + z = (xy) + z = (xy)z = x(yz) = x(y + z) = x + (y + z) for all x, y, z ∈ R+ , so that associativity under addition holds. A5: We claim that the zero vector in this set is the real number 1. To see this, note that 1 + x = 1x = x = x1 = x + 1 for all x ∈ R+ . A6: We claim that the additive inverse of the vector x ∈ R+ is x1 ∈ R+ . To see this, note that x + x1 = x x1 = 1 = x1 x = x1 + x. A7: Note that 1 · x = x1 = x for all x ∈ R+ , so that the unit property holds. A8: For all r, s ∈ R and x ∈ R+ , we have (rs) · x = xrs = xsr = (xs )r = r · xs = r · (s · x), as required for associativity of scalar multiplication. A9: For all r ∈ R and x, y ∈ R+ , we have r · (x + y) = r · (xy) = (xy)r = xr y r = xr + y r = r · x + r · y, as required for distributivity of scalar multiplication over vector addition. A10: For all r, s ∈ R and x ∈ R+ , we have (r + s) · x = xr+s = xr xs = xr + xs = r · x + s · x, as required for distributivity of scalar multiplication over scalar addition. The above verification of axioms A1-A10 shows that we have a vector space structure here. 21: Since the output vectors on the right hand side of the addition operation and the scalar multiplication operation are both in R2 , we conclude that axioms (A1) and (A2) are satisfied. Furthermore, since the addition operation is simply the ordinary addition of vectors in R2 , all of the axioms pertaining only to addition will hold as usual. This includes Axioms (A3), (A4), (A5), and (A6). For (A7), note that for all (x1 , y1 ) ∈ R2 , we have 1 (x1 , y1 ) = (x1 , y1 ), so (A7) holds. For (A8), observe that for (x1 , y1 ) ∈ R2 and scalars r, s ∈ R, we have (rs) (x1 , y1 ) = ((rs)x1 , y1 ) = (r(sx1 ), y1 ) = r (sx1 , y1 ) = r (s (x1 , y1 )), so that (A8) holds. For (A9), observe that for (x1 , y1 ), (x2 , y2 ) ∈ R2 and r ∈ R, we have r ((x1 , y1 ) + (x2 , y2 )) = r (x1 + x2 , y1 + y2 ) = (r(x1 + x2 ), y1 + y2 ), while r (x1 , y1 ) + r (x2 , y2 ) = (rx1 , y1 ) + (rx2 , y2 ) = (rx1 + rx2 , y1 + y2 ) = (r(x1 + x2 ), y1 + y2 ), (c)2017 Pearson Education. Inc. 265 so that (A9) holds. Finally, for (A10), observe that for scalars r, s ∈ R and (x1 , y1 ) ∈ R2 , we have (r + s) (x1 , y1 ) = ((r + s)x1 , y1 ) = (rx1 + sx1 , y1 ), while r (x1 , y1 ) + s (x1 , y1 ) = (rx1 , y1 ) + (sx1 , y1 ) = (rx1 + sx1 , y1 + y1 ) = (rx1 + sx1 , y1 ) for y1 = 0. Therefore, (A10) fails. Thus, this is NOT a vector space structure. 22. Axioms A1 and A2 clearly hold under the given operations. A3: u + v = (u1 , u2 ) + (v1 , v2 ) = (u1 − v1 , u2 − v2 ) = (−(v1 − u1 ), −(v2 − u2 )) = (v1 − u1 , v2 − u2 ) = v + u. Consequently, A3 does not hold. A4: (u+v)+w = (u1 −v1 , u2 −v2 )+(w1 , w2 ) = ((u1 −v1 )−w1 , (u2 −v2 )−w2 ) = (u1 −(v1 +w1 ), u2 −(v2 +w2 )) = (u1 , u2 ) + (v1 + w1 , v2 + w2 ) = (u1 , u2 ) + (v1 − w1 , v2 − w2 ) = u + (v + w). Consequently, A4 does not hold. A5: 0 = (0, 0) since u + 0 = (u1 , u2 ) + (0, 0) = (u1 − 0, u2 − 0) = (u1 , u2 ) = u. A6: If u = (u1 , u2 ), then −u = (u1 , u2 ) since u + (−u) = (u1 , u2 ) + (u1 , u2 ) = (u1 − u1 , u2 − u2 ) = (0, 0) = 0. Each of the remaining axioms do not hold. 23. Axiom A5: The zero vector on R2 with the defined operation of addition is given by (1,1), for if (u1 , u2 ) is any element in R2 , then (u1 , u2 ) + (1, 1) = (u1 · 1, u2 · 1) = (u1 , u2 ). Thus, Axiom A5 holds. Axiom A6: Suppose that (u1 , v1 ) is any element in R2 with additive inverse (a, b). From the first part of the problem, we know that (1, 1) is the zero element, so it must be the case that (u1 , v1 ) + (a, b) = (1, 1) so that (u1 a, v1 b) = (1, 1); hence, it follows that u1 a = 1 and v1 b = 1, but this system is not satisfied for all (u1 , v1 ) ∈ R, namely, (0, 0). Thus, Axiom A6 is not satisfied. 24. Let A, B, C ∈ M2 (R) and r, s, t ∈ R. A3: The addition operation is not commutative since A + B = AB = BA = B + A. A4: Addition is associative since (A + B) + C = AB + C = (AB)C = A(BC) = A(B + C) = A + (B + C). 1 0 A5: I2 = is the zero vector in M2 (R) because A + I2 = AI2 = A for all A ∈ M2 (R). 0 1 A6: We wish to determine whether for each matrix A ∈ M2 (R) we can find a matrix B ∈ M2 (R) such that A+B = I2 (remember that we have shown in A5 that the zero vector is I2 ), equivalently, such that AB = I2 . However, this equation can be satisfied only if A is nonsingular, therefore the axiom fails. A7: 1 · A = A is true for all A ∈ M2 (R). A8: (st)A = s(tA) is true for all A ∈ M2 (R) and s, t ∈ R. A9: sA + tA = (sA)(tA) = st(AA) = (s + t)A for all s, t ∈ R and A ∈ M2 (R). Consequently, the axiom fails. A10: rA + rB = (rA)(rB) = r2 AB = r2 (A + B). Thus, rA + rB = rA + rB for all r ∈ R, so the axiom fails. 25. M2 (R) = {A : A is a 2 × 2 real matrix}. Let A, B ∈ M2 (R) and k ∈ R. A1: A ⊕ B = −(A + B). A2: k · A = −kA. A3 and A4: A ⊕ B = −(A + B) = −(B + A) = B ⊕ A. Hence, the operation ⊕ is commutative. (A ⊕ B) ⊕ C = −((A ⊕ B) + C) = −(−(A + B) + C) = A + B − C, but A ⊕ (B ⊕ C) = −(A + (B ⊕ C)) = −(A + (−(B + C)) = −A + B + C. Thus the operation ⊕ is not associative. A5: An element B is needed such that A ⊕ B = A for all A ∈ M2 (R), but −(A + B) = A =⇒ B = −2A. (c)2017 Pearson Education. Inc. 266 Since this depends on A, there is no zero vector. A6: Since there is no zero vector, we cannot define the additive inverse. Let r, s, t ∈ R. A7: 1 · A = −1A = −A = A. A8: (st) · A = −[(st)A], but s · (t · A) = s · (−tA) = −[s(−tA)] = s(tA) = (st)A, so it follows that (st) · A = s · (t · A). A9: r ·(A⊕B) = −r(A⊕B) = −r(−(A+B)) = r(A+B) = rA+rB = −[(−rA)+(−rB)] = −rA⊕(−rB) = r · A + r · B, whereas rA ⊕ rB = −(rA + rB) = −rA − rB, so this axiom fails to hold. A10: (s+t)·A = −(s+t)A = −sA+(−tA), but s·A⊕t·A = −(sA)⊕(−tA) = −[−(sA)+(−tA)] = sA+tA, hence (s + t) · A = s · A ⊕ t · A. We conclude that only the axioms (A1)-(A3) hold. 26. Let C2 = {(z1 , z2 ) : zi ∈ C} under the usual operations of addition and scalar multiplication. A3 and A4: Follow from the properties of addition in C2 . A5: (0, 0) is the zero vector in C2 since (z1 , z2 ) + (0, 0) = (z1 + 0, z2 + 0) = (z1 , z2 ) for all (z1 , z2 ) ∈ C2 . A6: The additive inverse of the vector (z1 , z2 ) ∈ C2 is the vector (−z1 , −z2 ) for all (z1 , z2 ) ∈ C2 . A7-A10: Follows from properties in C2 . Thus, C2 together with its defined operations, is a complex vector space. 27. Let M2 (C) = {A : A is a 2 × 2 matrix with complex entries} under the usual operations of matrix addition and multiplication. A3 and A4: Follows from properties of matrix addition. A5: The zero vector, 0, for M2 (C) is the 2 × 2 zero matrix, 02 . A6: For each vector A = [aij ] ∈ M2 (C), the vector −A = [−aij ] ∈ M2 (C) satisfies A + (−A) = 02 . A7-A10: Follow from properties of matrix algebra. Hence, M2 (C) together with its defined operations, is a complex vector space. 28. Let u = (u1 , u2 , u3 ) and v = (v1 , v2 , v3 ) be vectors in C3 , and let k ∈ R. A1: u + v = (u1 , u2 , u3 ) + (v1 , v2 , v3 ) = (u1 + v1 , u2 + v2 , u3 + v3 ) ∈ C3 . A2: ku = k(u1 , u2 , u3 ) = (ku1 , ku2 , ku3 ) ∈ C3 . A3 and A4: Satisfied by the properties of addition in C3 . A5: (0, 0, 0) is the zero vector in C3 since (0, 0, 0) + (z1 , z2 , z3 ) = (0 + z1 , 0 + z2 , 0 + z3 ) = (z1 , z2 , z3 ) for all (z1 , z2 , z3 ) ∈ C3 . A6: (−z1 , −z2 , −z3 ) is the additive inverse of (z1 , z2 , z3 ) because (z1 , z2 , z3 ) + (−z1 , −z2 , −z3 ) = (0, 0, 0) for all (z1 , z2 , z3 ) ∈ C3 . Let r, s, t ∈ R. A7: 1 · u = 1 · (u1 , u2 , u3 ) = (1u1 , 1u2 , 1u3 ) = (u1 , u2 , u3 ) = u. A8: (st)u = (st)(u1 , u2 , u3 ) = ((st)u1 , (st)u2 , (st)u3 ) = (s(tu1 ), s(tu2 ), s(tu3 )) = s(tu1 , tu2 , tu3 ) = s(t(u1 , u2 , u3 )) = s(tu). A9: r(u + v) = r(u1 + v1 , u2 + v2 , u3 + v3 ) = (r(u1 + v1 ), r(u2 + v2 ), r(u3 + v3 )) = (ru1 + rv1 , ru2 + rv2 , ru3 + rv3 ) = (ru1 , ru2 , ru3 ) + (rv1 , rv2 , rv3 ) = r(u1 , u2 , u3 ) + r(v1 , v2 , v3 ) = ru + rv. A10: (s + t)u = (s + t)(u1 , u2 , u3 ) = ((s + t)u1 , (s + t)u2 , (s + t)u3 ) = (su1 + tu1 , su2 + tu2 , su3 + tu3 ) = (su1 , su2 , su3 ) + (tu1 , tu2 , tu3 ) = s(u1 , u2 , u3 ) + t(u1 , u2 , u3 ) = su + tu. Thus, C3 is a real vector space. 29. NO. If we scalar multiply a vector (x, y, z) ∈ R3 by a non-real (complex) scalar r, we will obtain the vector (rx, ry, rz) ∈ R3 , since rx, ry, rz ∈ R. 30. Let k be an arbitrary scalar, and let v be an arbitrary vector in V . Then, using property 2 of Theorem 4.2.7, we have k0 = k(0v) = (k0)v = 0v = 0. (c)2017 Pearson Education. Inc. 267 31. Assume that k is a scalar and v ∈ V such that kv = 0. If k = 0, the desired conclusion is already reached. If, on the other hand, k = 0, then we have k1 ∈ R and 1 1 · (kv) = · 0 = 0, k k or 1 · k v = 0. k Hence, 1·v = 0, and the unit property A7 now shows that v = 0, and again, we reach the desired conclusion. 32. We verify the axioms A1-A10 for a vector space. A1: If a0 + a1 x + · · · + an xn and b0 + b1 x + · · · + bn xn belong to Pn (R), then (a0 + a1 x + · · · + an xn ) + (b0 + b1 x + · · · + bn xn ) = (a0 + b0 ) + (a1 + b1 )x + · · · + (an + bn )xn , which again belongs to Pn (R). Therefore, Pn (R) is closed under addition. A2: If a0 + a1 x + · · · + an xn and r is a scalar, then r · (a0 + a1 x + · · · + an xn ) = (ra0 ) + (ra1 )x + · · · + (ran )xn , which again belongs to Pn (R). Therefore, Pn (R) is closed under scalar multiplication. A3: Let p(x) = a0 + a1 x + · · · + an xn and q(x) = b0 + b1 x + · · · + bn xn belong to Pn (R). Then p(x) + q(x) = (a0 + a1 x + · · · + an xn ) + (b0 + b1 x + · · · + bn xn ) = (a0 + b0 ) + (a1 + b1 )x + · · · + (an + bn )xn = (b0 + a0 ) + (b1 + a1 )x + · · · + (bn + an )xn = (b0 + b1 x + · · · + bn xn ) + (a0 + a1 x + · · · + an xn ) = q(x) + p(x), so Pn (R) satisfies commutativity under addition. A4: Let p(x) = a0 + a1 x + · · · + an xn , q(x) = b0 + b1 x + · · · + bn xn , and r(x) = c0 + c1 x + · · · + cn xn belong to Pn (R). Then [p(x) + q(x)] + r(x) = [(a0 + a1 x + · · · + an xn ) + (b0 + b1 x + · · · + bn xn )] + (c0 + c1 x + · · · + cn xn ) = [(a0 + b0 ) + (a1 + b1 )x + · · · + (an + bn )xn ] + (c0 + c1 x + · · · + cn xn ) = [(a0 + b0 ) + c0 ] + [(a1 + b1 ) + c1 ]x + · · · + [(an + bn ) + cn ]xn = [a0 + (b0 + c0 )] + [a1 + (b1 + c1 )]x + · · · + [an + (bn + cn )]xn = (a0 + a1 x + · · · + an xn ) + [(b0 + c0 ) + (b1 + c1 )x + · · · + (bn + cn )xn ] = (a0 + a1 x + · · · + an xn ) + [(b0 + b1 x + · · · + bn xn ) + (c0 + c1 x + · · · + cn xn )] = p(x) + [q(x) + r(x)], so Pn (R) satisfies associativity under addition. A5: The zero vector is the zero polynomial z(x) = 0 + 0 · x + · · · + 0 · xn , and it is readily verified that this polynomial satisfies z(x) + p(x) = p(x) = p(x) + z(x) for all p(x) ∈ Pn (R). A6: The additive inverse of p(x) = a0 + a1 x + · · · + an xn is −p(x) = (−a0 ) + (−a1 )x + · · · + (−an )xn . (c)2017 Pearson Education. Inc. 268 It is readily verified that p(x) + (−p(x)) = z(x), where z(x) is defined in A5. A7: We have 1 · (a0 + a1 x + · · · + an xn ) = a0 + a1 x + · · · + an xn , which demonstrates the unit property in Pn (R). A8: Let r, s ∈ R, and p(x) = a0 + a1 x + · · · + an xn ∈ Pn (R). Then (rs) · p(x) = (rs) · (a0 + a1 x + · · · + an xn ) = [(rs)a0 ] + [(rs)a1 ]x + · · · + [(rs)an ]xn = r[(sa0 ) + (sa1 )x + · · · + (san )xn ] = r[s(a0 + a1 x + · · · + an xn )] = r · (s · p(x)), which verifies the associativity of scalar multiplication. A9: Let r ∈ R, let p(x) = a0 + a1 x + · · · + an xn ∈ Pn (R), and let q(x) = b0 + b1 x + · · · + bn xn ∈ Pn (R). Then r · (p(x) + q(x)) = r · ((a0 + a1 x + · · · + an xn ) + (b0 + b1 x + · · · + bn xn )) = r · [(a0 + b0 ) + (a1 + b1 )x + · · · + (an + bn )xn ] = [r(a0 + b0 )] + [r(a1 + b1 )]x + · · · + [r(an + bn )]xn = [(ra0 ) + (ra1 )x + · · · + (ran )xn ] + [(rb0 ) + (rb1 )x + · · · + (rbn )xn ] = [r(a0 + a1 x + · · · + an xn )] + [r(b0 + b1 x + · · · + bn xn )] = r · p(x) + r · q(x), which verifies the distributivity of scalar multiplication over vector addition. A10: Let r, s ∈ R and let p(x) = a0 + a1 x + · · · + an xn ∈ Pn (R). Then (r + s) · p(x) = (r + s) · (a0 + a1 x + · · · + an xn ) = [(r + s)a0 ] + [(r + s)a1 ]x + · · · + [(r + s)an ]xn = [ra0 + ra1 x + · · · + ran xn ] + [sa0 + sa1 x + · · · + san xn ] = r(a0 + a1 x + · · · + an xn ) + s(a0 + a1 x + · · · + an xn ) = r · p(x) + s · p(x), which verifies the distributivity of scalar multiplication over scalar addition. The above verification of axioms A1-A10 shows that Pn (R) is a vector space. Solutions to Section 4.3 True-False Review: (a): FALSE. The null space of an m × n matrix A is a subspace of Rn , not Rm . (b): FALSE. It is not necessarily the case that 0 belongs to the solution set of the linear system. In fact, 0 belongs to the solution set of the linear system if and only if the system is homogeneous. (c): TRUE. If b = 0, then the line is y = mx, which is a line through the origin of R2 , a one-dimensional subspace of R2 . On the other hand, if b = 0, then the origin does not lie on the given line, and therefore since the line does not contain the zero vector, it cannot form a subspace of R2 in this case. (c)2017 Pearson Education. Inc. 269 (d): FALSE. The spaces Rm and Rn , with m < n, are not comparable. Neither of them is a subset of the other, and therefore, neither of them can form a subspace of the other. (e): TRUE. Choosing any vector v in S, the scalar multiple 0v = 0 still belongs to S. (f ): FALSE. In order for a subset of V to form a subspace, the same operations of addition and scalar multiplication must be used in the subset as used in V . (g): FALSE. This set is not closed under addition. For instance, the point (1, 1, 0) lies in the xy-plane, the point (0, 1, 1) lies in the yz-plane, but (1, 1, 0) + (0, 1, 1) = (1, 2, 1) does not belong to S. Therefore, S is not a subspace of V . (h): FALSE. For instance, if we consider V = R3 , then the xy-plane forms a subspace of V , and the x-axis forms a subspace of V . Both of these subspaces contain in common all points along the x-axis. Other examples abound as well. Problems: 1. S = {x ∈ R3 : x = (r − 2s, 3r + s, s), r, s ∈ R}. (a). S is certainly nonempty. Let x, y ∈ S. Then for some r, s, u, v ∈ R, x = (r − 2s, 3r + s, s) and y = (u − 2v, 3u + v, v). Hence, x + y = (r − 2s, 3r + s, s) + (u − 2v, 3u + v, v) = ((r + u) − 2(s + v), 3(r + u) + (s + v), s + v) = (a − 2b, 3a + b, b), where a = r + u, and b = s + v. Consequently, S is closed under addition. Further, if c ∈ R, then cx = c(r − 2s, 3r + s, s) = (cr − 2cs, 3cr + cs, cs) = (k − 2l, 3k + l, l), where k = cr and l = cs. Therefore S is also closed under scalar multiplication. It follows from Theorem 4.3.2 that S is a subspace of R3 . (b). The coordinates of the points in S are (x, y, z) where x = r − 2s, y = 3r + s, z = s. Eliminating r and s between these three equations yields 3x − y + 7z = 0. 2. S = {x ∈ R2 : x = (2k, −3k), k ∈ R}. (a). S is certainly nonempty. Let x, y ∈ S. Then for some r, s ∈ R, x = (2r, −3r) and y = (2s, −3s). Hence, x + y = (2r, −3r) + (2s, −3s) = (2(r + s), −3(r + s)) = (2k, −3k), (c)2017 Pearson Education. Inc. 270 where k = r + s. Consequently, S is closed under addition. Further, if c ∈ R, then cx = c(2r, −3r) = (2cr, −3cr) = (2t, −3t), where t = cr. Therefore S is also closed under scalar multiplication. It follows from Theorem 4.3.2 that S is a subspace of R2 . (b). The subspace S consists of all points lying along the line in the accompanying figure. y y=-3x/2 x Figure 0.0.62: Figure for Problem 2(b) 3. YES. We can write S = {(x, y, z) ∈ R3 : z = 3x and y = 2x} = {(x, 2x, 3x) : x ∈ R}. Note that S is nonempty since (0, 0, 0) ∈ S. To show that the nonempty set S is a subspace of V , we will verify that S is closed under addition and S is closed under scalar multiplication: Closure under Addition: Let (x, 2x, 3x), (y, 2y, 3y) ∈ S. Then (x, 2x, 3x) + (y, 2y, 3y) = (x + y, 2x + 2y, 3x + 3y) = (x + y, 2(x + y), 3(x + y)) = (t, 2t, 3t) ∈ S, where t = x + y. Closure under Scalar Multiplication: Let (x, 2x, 3x) ∈ S and let r be a scalar. Then r · (x, 2x, 3x) = (rx, 2rx, 3rx) = (t, 2t, 3t) ∈ S, where t = rx. 4. YES. S = {(x, y) ∈ R2 : 3x + 2y = 0}. S = ∅ since (0, 0) ∈ S. Closure under Addition: Let (x1 , x2 ), (y1 , y2 ) ∈ S. Then 3x1 + 2x2 = 0 and 3y1 + 2y2 = 0, so 3(x1 + y1 ) + 2(x2 + y2 ) = 0, which implies that (x1 + y1 , x2 + y2 ) ∈ S. Closure under Scalar Multiplication: Let a ∈ R and (x1 , x2 ) ∈ S. Then 3x1 + 2x2 = 0 =⇒ a(3x1 + 2x2 ) = a · 0 =⇒ 3(ax1 ) + 2(ax2 ) = 0, which shows that (ax1 , ax2 ) ∈ S. Thus, S is a subspace of R2 by Theorem 4.3.2. 5. NO. S = {(x1 , 0, x3 , 2) : x1 , x3 ∈ R} is not a subspace of R4 because it is not closed under addition. To see this, let (a, 0, b, 2), (c, 0, d, 2) ∈ S. Then (a, 0, b, 2) + (c, 0, d, 2) = (a + c, 0, b + d, 4) ∈ / S. Note that S also fails the zero vector check. (c)2017 Pearson Education. Inc. 271 / S since 0 + 0 + 0 = 1. 6. NO. S = {(x, y, z) ∈ R3 : x + y + z = 1} is not a subspace of R3 because (0, 0, 0) ∈ / S. 7. NO. S = {u ∈ R2 : Au = b, A is a fixed m × n matrix} is not a subspace of Rn since 0 ∈ 8. NO. S = {(x, y) ∈ R2 : x2 − y 2 = 0} is not a subspace of R2 , since it is not closed under addition, as we now observe: If (x1 , y1 ), (x2 , y2 ) ∈ S, then (x1 , y1 ) + (x2 , y2 ) = (x1 + x2 , y1 + y2 ). (x1 + x2 )2 − (y1 + y2 )2 = x21 + 2x1 x2 + x22 − (y12 + 2y1 y2 + y22 ) = (x21 − y12 ) + (x22 − y22 ) + 2(x1 x2 − y1 y2 ) = 0 + 0 + 2(x1 x2 − y1 y2 ) = 0, in general. / S. Thus, (x1 , y1 ) + (x2 , y2 ) ∈ 9. NO. S = {A ∈ M2 (R) : det(A) = 1} is not a subspace of M2 (R). To see this, let k ∈ R be a scalar and let A ∈ S. Then det(kA) = k 2 det(A) = k 2 · 1 = k 2 = 1, unless k = ±1. Note also that det(A) = 1 and det(B) = 1 does not imply that det(A + B) = 1. 10. YES. S = {A = [aij ] ∈ Mn (R) : aij = 0 whenever i < j}. Note that S = ∅ since 0n ∈ S. Now let A = [aij ] and B = [bij ] be lower triangular matrices. Then aij = 0 and bij = 0 whenever i < j. Then aij + bij = 0 and caij = 0 whenever i < j. Hence A + B = [aij + bij ] and cA = [caij ] are also lower triangular matrices. Therefore S is closed under addition and scalar multiplication. Consequently, S is a subspace of M2 (R) by Theorem 4.3.2. / S. 11. NO. S = {A ∈ Mn (R) : A is invertible} is not a subspace of Mn (R) because 0n ∈ 12. YES. We can write S= or S= : a+b+c+d=0 a c b d a c b −(a + b + c) ∈V : a, b, c ∈ R . Note that S is nonempty since the zero matrix belongs to S. To show that the nonempty set S is a subspace of V , we will verify that S is closed under addition and S is closed under scalar multiplication. In this case, we will use the first descriptionof S, although either of them can be used. a b e f ∈ S. Therefore, a + b + c + d = 0 Closure under Addition: Let A = ∈ S and B = c d g h and e + f + g + h = 0. Now, a b e f a+e b+f A+B = + = , c d g h c+g d+h and since (a + e) + (b + f ) + (c + g) + (d + h) = (a + b + c + d) + (e + f + g + h) = 0 + 0 = 0, we conclude that A + B ∈ S, as needed. Closure under Scalar Multiplication: Let A be as above and let r be a scalar. Then a b ra rb rA = r = , c d rc rd and since ra + rb + rc + rd = r(a + b + c + d) = r(0) = 0, we conclude that rA ∈ S, as needed. (c)2017 Pearson Education. Inc. 272 13. YES. We can write ⎧⎡ ⎨ a S= ⎣ c ⎩ e ⎤ b d ⎦∈V f ⎧⎡ ⎨ S= ⎣ ⎩ or a c −(a + c) ⎫ ⎬ : a + c + e = 0 and b + d + f = 0 ⎭ ⎫ ⎤ b ⎬ ⎦ : a, b, c, d ∈ R . d ⎭ −(b + d) Note that S is nonempty since the zero matrix belongs to S. To show that the nonempty set S is a subspace of V , we will verify that S is closed under addition and S is closed under scalar multiplication. In this case, we will use the first description ⎡of S, although either of ⎡ them can⎤be used. ⎤ a b u v Closure under Addition: Let A = ⎣ c d ⎦ ∈ S and B = ⎣ w x ⎦ ∈ S. Therefore, y z e f a + c + e = b + d + f = u + w + y = v + x + z = 0. Now, ⎡ a A+B =⎣ c e ⎤ ⎡ b u d ⎦+⎣ w f y ⎤ ⎡ v a+u x ⎦=⎣ c+w z e+y ⎤ b+v d + x ⎦, f +z and since (a + u) + (c + w) + (e + y) = (a + c + e) + (u + w + y) = 0 + 0 = 0 and (b + v) + (d + x) + (f + z) = (b + d + f ) + (v + x + z) = 0 + 0 = 0, we conclude that A + B ∈ S, as needed. Closure under Scalar Multiplication: Let A be as above and let r be a scalar. Then ⎡ ⎤ ⎡ ⎤ a b ra rb rA = r ⎣ c d ⎦ = ⎣ rc rd ⎦ , e f re rf and since ra + rc + re = r(a + c + e) = r(0) = 0 and rb + rd + rf = r(b + d + f ) = r(0) = 0, we conclude that rA ∈ S, as needed. 14. NO. We can write S= a b d e c f ∈V : a + b + c = 10 and d + e + f = 10 . Note that S is not a subspace of V , for example, by the Zero Vector Check: 15. YES. S = {A ∈ M2 (R) : AT = A}. S = ∅ since 02 ∈ S. (c)2017 Pearson Education. Inc. 0 0 0 0 0 0 ∈ S. 273 Closure under Addition: If A, B ∈ S, then (A + B)T = AT + B T = A + B, which shows that A + B ∈ S. Closure under Scalar Multiplication: If r ∈ R and A ∈ S, then (rA)T = rAT = rA, which shows that rA ∈ S. Consequently, S is a subspace of M2 (R) by Theorem 4.3.2. 16. YES. We can write S = {f ∈ V : f (a) = 5 · f (b)}. Note that the zero function f (x) = 0 for all x belongs to S, so that S is nonempty. To show that the nonempty set S is a subspace of V , we will verify that S is closed under addition and S is closed under scalar multiplication. Closure under Addition: Suppose that f and g belong to S. Therefore, f (a) = 5f (b) and g(a) = 5g(b). Hence, (f + g)(a) = f (a) + g(a) = 5f (b) + 5g(b) = 5(f (b) + g(b)) = 5 · (f + g)(b), which shows that f + g ∈ S. Closure under Scalar Multiplication: Suppose that f belongs to S and r is a scalar. Therefore (rf )(a) = rf (a) = r(5(f (b)) = 5(rf (b)) = 5 · (rf )(b), which shows that rf ∈ S. 17. NO. S = {f ∈ V : f (a) = 1}, where V is the vector space of all real-valued functions defined on [a, b]. We claim that S is not a subspace of V . Not Closed under Addition: If f, g ∈ S, then (f + g)(a) = f (a) + g(a) = 1 + 1 = 2 = 1, which shows that f +g ∈ / S. It can also be shown that S is not closed under scalar multiplication. 18. YES. S = {f ∈ V : f (−x) = f (x) for all x ∈ R}. Note that S = ∅ since the zero function O(x) = 0 for all x belongs to S. Let f, g ∈ S. Then (f + g)(−x) = f (−x) + g(−x) = f (x) + g(x) = (f + g)(x) and if c ∈ R, then (cf )(−x) = cf (−x) = cf (x) = (cf )(x), so f + g and c · f belong to S. Therefore, S is closed under addition and scalar multiplication. Therefore, S is a subspace of V by Theorem 4.3.2. 19. YES. S = {p ∈ P2 (R) : p(x) = ax2 + b, a, b ∈ R}. Note that S = ∅ since p(x) = 0 belongs to S. Closure under Addition: Let p, q ∈ S. Then for some a1 , a2 , b1 , b2 ∈ R, p(x) = a1 x2 + b1 and q(x) = a2 x2 + b2 . Hence, (p + q)(x) = p(x) + q(x) = (a1 + a2 )x2 + b1 + b2 = ax2 + b, where a = a1 + a2 and b = b1 + b2 , so that S is closed under addition. Closure under Scalar Multiplication: If k ∈ R, then (kp)(x) = kp(x) = ka1 x2 + kb1 = cx2 + d, (c)2017 Pearson Education. Inc. 274 where c = ka1 and d = kb1 , so that S is also closed under scalar multiplication. It follows from Theorem 4.3.2 that S is a subspace of P2 (R). 20. NO. S = {p ∈ P2 (R) : p(x) = ax2 + 1, a ∈ R}. We claim that S is not closed under addition: Not Closed under Addition: Let p, q ∈ S. Then for some a1 , a2 ∈ R, p(x) = a1 x2 + 1 and q(x) = a2 x2 + 1. Hence, (p + q)(x) = p(x) + q(x) = (a1 + a2 )x2 + 2 = ax2 + 2 where a = a1 + a2 . Consequently, p + q ∈ / S, and therefore S is not closed under addition. It follows that S is not a subspace of P2 (R). 21. YES. S = {y ∈ C 2 (I) : y +2y −y = 0}. Note that S is nonempty since the function y = 0 belongs to S. Closure under Addition: Let y1 , y2 ∈ S. (y1 + y2 ) + 2(y1 + y2 ) − (y1 + y2 ) = y1 + y2 + 2(y1 + y2 ) − y1 − y2 = y1 + 2y1 − y1 + y2 + 2y2 − y2 Thus, y1 + y2 ∈ S. = (y1 + 2y1 − y1 ) + (y2 + 2y2 − y2 ) = 0 + 0 = 0. Closure under Scalar Multiplication: Let k ∈ R and y1 ∈ S. (ky1 ) + 2(ky1 ) − (ky1 ) = ky1 + 2ky1 − ky1 = k(y1 + 2y1 − y1 ) = k · 0 = 0, which shows that ky1 ∈ S. Hence, S is a subspace of V by Theorem 4.3.2. 22. NO. S = {y ∈ C 2 (I) : y + 2y − y = 1}. S is not a subspace of V . We show that S fails to be closed under addition (one can also verify that it is not closed under scalar multiplication, but this is unnecessary if one shows the failure of closure under addition): Not Closed under Addition: Let y1 , y2 ∈ S. (y1 + y2 ) + 2(y1 + y2 ) − (y1 + y2 ) = y1 + y2 + 2(y1 + y2 ) − y1 − y2 = (y1 + 2y1 − y1 ) + (y2 + 2y2 − y2 ) Thus, y1 + y2 ∈ / S. = 1 + 1 = 2 = 1. Or alternatively: Not Closed under Scalar Multiplication: Let k ∈ R and y1 ∈ S. ((ky1 ) + 2(ky1 ) − (ky1 ) = ky1 + 2ky1 − ky1 = k(y1 + 2y1 − y1 ) = k · 1 = k = 1, unless k = 1. Therefore, ky1 ∈ / S unless k = 1. 23. The null space of A is the subspace of R2 consisting of points (x, y) with x + 4y = 0. The matrix A is already in row-echelon form. Setting y = t to be a free variable, then x = −4t. Hence, nullspace(A) = {(−4t, t) : t ∈ R}. 24. The null space of A is the subspace of R3 consisting of points (x, y, z) with x − 3y + 2z = 0. The matrix A is already in row-echelon form. Setting y = t and z = r to be free variables, then x = 3t − 2r. Hence, nullspace(A) = {(3t − 2r, t, r) : r, t ∈ R}. 25. The null space of A is the subspace of R2 consisting of points (x, y) with 2x − 4y = 0, x + 2y = 0, and −3x − 5y = 0. A short series of elementary row operations provides the following row-echelon form for A: (c)2017 Pearson Education. Inc. 275 ⎡ ⎤ 1 2 ⎣ 0 1 ⎦. Since there are no free variables here, we conclude that the null space of A contains no nonzero 0 0 vectors. Therefore, nullspace(A) = {(0, 0)}. 26. The null space of A is the subspace of R4 consisting of points (x, y, z, w) with x + 2y + 3z + 4w = 0 and 5x + 6y+ 7z + 8w = 0. A short series of elementary row operations provides the following row-echelon form 1 2 3 4 for A: . Therefore, we take free variables z = t and w = r. Hence, y = −2t − 3r. Thus, 0 1 2 3 x = −2(−2t − 3r) − 3t − 4r = t + 2r. Therefore, nullspace(A) = {(t + 2r, −2t − 3r, t, r) : r, t ∈ R}. ⎡ ⎤ 1 −2 ⎦. nullspace(A) = {x ∈ R3 : Ax = 0}. The reduced row echelon form of the 4 ⎡ ⎤ 1 0 0 0 augmented matrix of the system Ax = 0 is ⎣ 0 1 0 0 ⎦. Consequently, the linear system Ax = 0 has 0 0 1 0 only the trivial solution (0, 0, 0), so nullspace(A) = {(0, 0, 0)}. ⎡ ⎤ 1 3 −2 1 6 ⎦. nullspace(A) = {x ∈ R3 : Ax = 0}. The REDUCED ROW ECHELON 28. A = ⎣ 3 10 −4 2 5 −6 −1 ⎡ ⎤ 1 0 −8 −8 0 2 3 0 ⎦, so that nullspace(A) = FORM of the augmented matrix of the system Ax = 0 is ⎣ 0 1 0 0 0 0 0 {(8r + 8s, −2r − 3s, r, s) : r, s ∈ R}. ⎡ ⎤ 1 i −2 4i −5 ⎦. nullspace(A) = {x ∈ C3 : Ax = 0}. The REDUCED ROW ECHELON 29. A = ⎣ 3 −1 −3i i ⎡ ⎤ 1 0 0 0 FORM of the augmented matrix of the system Ax = 0 is ⎣ 0 1 0 0 ⎦. Consequently, the linear system 0 0 1 0 Ax = 0 has only the trivial solution (0, 0, 0), so nullspace(A) = {(0, 0, 0)}. 1 −2 27. A = ⎣ 4 −7 −1 3 30. Since the zero function y(x) = 0 for all x ∈ I is not a solution to the differential equation, the set of all solutions does not contain the zero vector from C 2 (I), hence it is not a vector space at all and cannot be a subspace. 31. (a). As an example, we can let V = R2 . If we take S1 to be the set of points lying on the x-axis and S2 to be the set of points lying on the y-axis, then it is readily seen that S1 and S2 are both subspaces of V . However, the union of these subspaces is not closed under addition. For instance, the points (1, 0) and (0, 1) lie in S1 ∪ S2 , but (1, 0) + (0, 1) = (1, 1) ∈ S1 ∪ S2 . Therefore, the union S1 ∪ S2 does not form a subspace of V . (b). Since S1 and S2 are both subspaces of V , they both contain the zero vector. It follows that the zero vector is an element of S1 ∩ S2 , hence this subset is nonempty. Now let u and v be vectors in S1 ∩ S2 , and let (c)2017 Pearson Education. Inc. 276 c be a scalar. Then u and v are both in S1 and both in S2 . Since S1 and S2 are each subspaces of V , u + v and cu are vectors in both S1 and S2 , hence they are in S1 ∩ S2 . This implies that S1 ∩ S2 is a nonempty subset of V , which is closed under both addition and scalar multiplication. Therefore, S1 ∩ S2 is a subspace of V . (c). Note that S1 is a subset of S1 + S2 (every vector v ∈ S1 can be written as v + 0 ∈ S1 + S2 ), so S1 + S2 is nonempty. Closure under Addition: Now let u and v belong to S1 + S2 . We may write u = u1 + u2 and v = v1 + v2 , where u1 , v1 ∈ S1 and u2 , v2 ∈ S2 . Then u + v = (u1 + u2 ) + (v1 + v2 ) = (u1 + v1 ) + (u2 + v2 ). Since S1 and S2 are closed under addition, we have that u1 + v1 ∈ S1 and u2 + v2 ∈ S2 . Therefore, u + v = (u1 + u2 ) + (v1 + v2 ) = (u1 + v1 ) + (u2 + v2 ) ∈ S1 + S2 . Hence, S1 + S2 is closed under addition. Closure under Scalar Multiplication: Next, let u ∈ S1 + S2 and let c be a scalar. We may write u = u1 + u2 , where u1 ∈ S1 and u2 ∈ S2 . Thus, c · u = c · u1 + c · u2 , and since S1 and S2 are closed under scalar multiplication, c · u1 ∈ S1 and c · u2 ∈ S2 . Therefore, c · u = c · u 1 + c · u 2 ∈ S1 + S2 . Hence, S1 + S2 is closed under scalar multiplication. Solutions to Section 4.4 True-False Review: (a): TRUE. By its very definition, when a linear span of a set of vectors is formed, that span becomes closed under addition and under scalar multiplication. Therefore, it is a subspace of V . (b): FALSE. In order to say that S spans V , it must be true that all vectors in V can be expressed as a linear combination of the vectors in S, not simply “some” vector. (c): TRUE. Every vector in V can be expressed as a linear combination of the vectors in S, and therefore, it is also true that every vector in W can be expressed as a linear combination of the vectors in S. Therefore, S spans W , and S is a spanning set for W . (d): FALSE. To illustrate this, consider V = R2 , and consider the spanning set {(1, 0), (0, 1), (1, 1)}. Then the vector v = (2, 2) can be expressed as a linear combination of the vectors in S in more than one way: v = 2(1, 1) and v = 2(1, 0) + 2(0, 1). Many other illustrations, using a variety of different vector spaces, are also possible. (e): TRUE. To say that a set S of vectors in V spans V is to say that every vector in V belongs to span(S). So V is a subset of span(S). But of course, every vector in span(S) belongs to the vector space V , and so span(S) is a subset of V . Therefore, span(S) = V . (c)2017 Pearson Education. Inc. 277 (f ): FALSE. This is not necessarily the case. For example, the linear span of the vectors (1, 1, 1) and (2, 2, 2) is simply a line through the origin, not a plane. (g): FALSE. There are vector spaces that do not contain finite spanning sets. For instance, if V is the vector space consisting of all polynomials with coefficients in R, then since a finite spanning set could not contain polynomials of arbitrarily large degree, no finite spanning set is possible for V . (h): FALSE. To illustrate this, consider V = R2 , and consider the spanning set S = {(1, 0), (0, 1), (1, 1)}. The proper subset S = {(1, 0), (0, 1)} is still a spanning set for V . Therefore, it is possible for a proper subset of a spanning set for V to still be a spanning set for V . ⎡ ⎤ a b c (i): TRUE. The general matrix ⎣ 0 d e ⎦ in this vector space can be written as aE11 + bE12 + cE13 + 0 0 f dE22 + eE23 + f E33 , and therefore the matrices in the set {E11 , E12 , E13 , E22 , E23 , E33 } span the vector space. (j): FALSE. For instance, it is easily verified that {x2 , x2 + x, x2 + 1} is a spanning set for P2 (R), and yet, it contains only polynomials of degree 2. (k): FALSE. For instance, consider m = 2 and n = 3. Then one spanning set for R2 is {(1, 0), (0, 1), (1, 1), (2, 2)}, which consists of four vectors. On the other hand, one spanning set for R3 is {(1, 0, 0), (0, 1, 0), (0, 0, 1)}, which consists of only three vectors. (l): TRUE. This is explained in True-False Review Question (g) above. Problems: 1. NO. This single vector only spans all of its multiples, which is a line in R2 . Any point in R2 that does not lie along the lie through the origin with parallel vector (5, −1) does not lie in the span of this set. 2. YES. {(1, −1), (2, −2), (2, 3)}. Since v1 = (1, −1), and v2 = (2, 3) are noncolinear, the given set of vectors does span R2 . (See the comment immediately preceding Example 4.4.3 in the text.) 3. NO. The given set of vectors does not span R2 since it does not contain two nonzero and non-collinear vectors. 4. NO. The three vectors in the given set are all collinear. Consequently, the set of vectors does not span R2 . 1 2 4 5. YES. Since −1 5 −2 = −23 = 0, the given vectors are not coplanar, and therefore span R3 . 1 3 1 6. YES. We compute 1 −2 1 2 4 3 −1 = −7 = 0, 1 2 which shows that the given vectors are not coplanar. Therefore, they span all of R3 . 2 3 1 7. NO. Since −1 −3 1 = 0, the vectors are coplanar, and therefore the given set does not span R3 . 4 5 3 (c)2017 Pearson Education. Inc. 278 8. NO. We compute 1 2 3 4 5 6 7 8 = 0, 9 which shows that the given vectors are coplanar. Therefore, they only span the plane in which they lie, which is not all of R3 . −4 5 6 9. Since 1 1 0 = 0, the vectors are coplanar, and therefore the given set does not span R3 .The linear 3 6 2 span of the vectors is those points (x, y, z) for which the system c1 (−4, 1, 3) + c2 (5, 1, 6) + c3 (6, 0, 2) = (x, y, z) ⎤ ⎤ ⎡ y −4 5 6 x 1 1 0 ⎦. −3y + z is consistent. Reducing the augmented matrix ⎣ 1 1 0 y ⎦ of this system yields ⎣ 0 3 2 3 6 2 z 0 0 0 x + 13y − 3z This system is consistent if and only if x + 13y − 3z = 0. Consequently, the linear span of the given set of vectors consists of all points lying on the plane with equation x + 13y − 3z = 0. 1 3 4 10. Since 2 4 5 = 0, the vectors are coplanar, and therefore the given set does not span R3 . The 3 5 6 linear span of the vectors is those points (x, y, z) for which the system ⎡ c1 (1, 2, 3) + c2 (3, 4, 5) + c3 (4, 5, 6) = (x, y, z) ⎤ ⎤ ⎡ x 1 3 4 x 1 3 4 2x − y ⎦. is consistent. Reducing the augmented matrix ⎣ 2 4 5 y ⎦ of this system yields ⎣ 0 2 3 3 5 6 z 0 0 0 x − 2y + z This system is consistent if and only if x − 2y + z = 0. Consequently, the linear span of the given set of vectors consists of all points lying on the plane with the equation x − 2y + z = 0. ⎡ 11. Let (x1 , x2 ) ∈ R2 and a, b ∈ R. (x1 , x2 ) = av1 + bv2 = a(2, −1) + b(3, 2) = (2a, −a) + (3b, 2b) = (2a + 3b, −a + 2b). It follows that 2x1 − 3x2 x1 + 2x2 2a + 3b = x1 and −a + 2b = x2 , which implies that a = and b = so (x1 , x2 ) = 7 7 x1 + 2x2 31 9 2x1 − 3x2 v1 + v2 . {v1 , v2 } spans R2 , and in particular, (5, −7) = v1 − v2 . 7 7 7 7 12. Let (x1 , x2 ) ∈ R2 and a, b ∈ R. (x1 , x2 ) = av1 + bv2 = a(1, −5) + b(6, 3) = (a, −5a) + (6b, 3b) = (a + 6b, −5a + 3b). It follows that x1 − 2x2 5x1 + x2 a + 6b = x1 and −5a + 3b = x2 , which implies that a = and b = . Thus, 11 33 x1 − 2x2 5x1 + x2 v1 + v2 . (x1 , x2 ) = 11 33 Hence, {v1 , v2 } spans R2 . 13. Let (x, y, z) ∈ R3 and a, b, c ∈ R. (c)2017 Pearson Education. Inc. 279 (x, y, z) = av1 + bv2 + cv3 = a(1, −3, 2) + b(1, 0, −1) + c(1, 2, −4) = (a, −3a, 2a) + (b, 0, −b) + (c, 2c, −4c) = (a + b + c, −3a + 2c, 2a − b − 4c). Therefore, a + b + c = x, −3a + 2c = y, 2a − b − 4c = z. Substituting x = 9, y = 8, and z = 7, we obtain the system −3a + 2c = 8, 2a − b − 4c = 7. ⎡ ⎤ 1 1 1 9 0 2 8 ⎦. Row-reducing this augmented matrix The augmented matrix for this system is ⎣ −3 2 −1 −4 7 ⎤ ⎡ 9 1 1 1 56 35 ⎦. Back substitution now yields c = −24, b = 155 yields ⎣ 0 3 5 3 and a = − 3 . Therefore, we have 0 0 1 −24 155 (9, 8, 7) = − 56 3 v1 + 3 v2 − 24v3 . a + b + c = 9, 14. Let (x, y, z) ∈ R3 and a, b, c ∈ R. (x, y, z) = av1 + bv2 + cv3 = a(−1, 3, 2) + b(1, −2, 1) + c(2, 1, 1) = (−a, 3a, 2a) + (b, −2b, b) + (2c, c, c) = (−a + b + ⎧ 2c, 3a − 2b + c, 2a + b + c). ⎪ a + b + 2c = x ⎨ These equalities result in the system: 3a − 2b + c = y ⎪ ⎩ 2a + b + c = z Upon solving for a, b, and c we obtain a= −x − 5y + 7z 7x + 3y − z −3x + y + 5z ,b= , and c = . 16 16 16 Consequently, {v1 , v2 , v3 } spans R3 , and −3x + y + 5z −x − 5y + 7z 7x + 3y − z (x, y, z) = v1 + v2 + v3 . 16 16 16 15. Let (x, y) ∈ R2 and a, b, c ∈ R. (x, y) = av1 + bv2 + cv3 = a(1, 1) + b(−1, 2) + c(1, 4) = (a, a) + (−b, 2b) + (c, 4c) = (a − b + c, a + 2b + 4c). a−b+c=x These equalities result in the system: a + 2b + 4c = y Solving the system we find that a= 2x + y − 6c y − x − 3c , and b = 3 3 where c is a free real variable. It follows that y − x − 3c 2x + y − 6c v1 + v2 + cv3 , (x, y) = 3 3 (c)2017 Pearson Education. Inc. 280 which implies that the vectors v1 , v2 , and v3 span R2 . Moreover, if c = 0, then (x, y) = 2x + y 3 v1 + y−x 3 v2 , so R2 = span{v1 , v2 } also. 16. We can write v = (c1 , c2 , c2 − 2c1 ) = c1 (1, 0, −2) + c2 (0, 1, 1), which shows that S is spanned by {(1, 0, −2), (0, 1, 1)}. 17. v = (c1 , c2 , c2 − 2c1 , c1 − 2c2 ) = (c1 , 0, −c1 , c1 ) + (0, c2 , c2 , −2c2 ) = c1 (1, 0, −1, 1) + c2 (0, 1, 1, −2). Thus, {(1, 0, −1, 1), (0, 1, 1, −2)} spans S. 0 α 0 −1 0 −1 18. S = A ∈ M2 (R) : A = , α ∈ R = A ∈ M2 (R) : A = α = span . −α 0 1 0 1 0 19. (a). S = ∅ since 0 0 0 0 ∈ S. Closure under Addition: Let x, y ∈ S. Then, x+y = x12 x22 x11 0 + y12 y22 y11 0 = x11 + y11 0 x12 + y12 x22 + y22 , which implies that x + y ∈ S. Closure under Scalar Multiplication: Let r ∈ R and x ∈ S. Then rx = r x11 0 x12 x22 = rx11 0 rx12 rx22 , which implies thatrx ∈ S. Consequently, S is a subspace of M2 (R) by Theorem 4.3.2. 1 0 0 1 0 0 a11 a12 = a11 + a12 + a22 . (b). A = 0 1 0 a22 0 0 0 0 1 0 0 1 0 0 Therefore, S = span , , . 0 0 0 0 0 1 a b 20. A typical element of S can be written in the form v = . Therefore, we can write c −(a + b + c) v=a 1 0 0 −1 +b 0 0 1 −1 1 −1 0 , 1 +c 0 1 0 −1 which shows that S is spanned by 1 0 0 −1 0 , 0 0 −1 (c)2017 Pearson Education. Inc. . , 281 ⎡ ⎤ a b −(a + b) c d −(c + d) ⎦. There21. A typical element of S can be written in the form v = ⎣ −(a + c) −(b + d) a + b + c + d fore, we can write ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 1 0 −1 0 1 −1 0 0 0 0 0 0 0 ⎦ + b⎣ 0 0 0 ⎦ + c ⎣ 1 0 −1 ⎦ + d ⎣ 0 1 −1 ⎦ , v = a⎣ 0 0 −1 0 1 0 −1 1 −1 0 1 0 −1 1 which shows that S is spanned by ⎧⎡ ⎤ ⎡ ⎤ ⎡ 1 0 −1 0 1 −1 0 ⎨ ⎣ 0 0 0 ⎦,⎣ 0 0 0 ⎦,⎣ 1 ⎩ −1 0 1 0 −1 1 −1 ⎤ ⎡ ⎤⎫ 0 0 0 0 0 ⎬ 0 −1 ⎦ , ⎣ 0 1 −1 ⎦ . ⎭ 0 1 0 −1 1 ⎡ a 22. A typical element of S can be written in the form v = ⎣ b c ⎡ 1 v = a⎣ 0 0 0 0 0 ⎤ ⎡ 0 0 0 ⎦+b⎣ 1 0 0 1 0 0 ⎤ ⎡ 0 0 0 ⎦+c⎣ 0 0 1 which shows that S is spanned by ⎧⎡ ⎤ ⎡ ⎤ ⎡ 0 1 0 0 0 ⎨ 1 0 0 ⎣ 0 0 0 ⎦,⎣ 1 0 0 ⎦,⎣ 0 0 ⎩ 0 0 0 0 0 0 1 0 0 0 0 b d e ⎤ c e ⎦. Therefore, we can write f ⎤ ⎡ 1 0 0 ⎦+d⎣ 0 0 0 0 1 0 ⎤ ⎡ 1 0 0 ⎦,⎣ 0 0 0 ⎤ ⎡ 0 0 0 ⎦,⎣ 0 0 0 0 1 0 ⎤ ⎡ 0 0 0 ⎦+e⎣ 0 0 0 0 0 1 0 0 1 ⎤ ⎡ 0 0 1 ⎦+f ⎣ 0 0 0 ⎤ ⎡ 0 0 1 ⎦,⎣ 0 0 0 0 0 0 0 0 0 ⎤ 0 0 ⎦, 1 ⎤⎫ 0 ⎬ 0 ⎦ . ⎭ 1 23. x − 2y − z = 0 =⇒ x = 2y + z, so v ∈ R3 . =⇒ v = (2y + z, y, z) = (2y, y, 0) + (z, 0, z) = y(2, 1, 0) + z(1, 0, 1). Therefore S = {v ∈ R3 : v = a(2, 1, 0) + b(1, 0, 1), a, b ∈ R}, hence {(2, 1, 0), (1, 0, 1)} spans S. 24. Note that S consists precisely of the constant polynomials: S = {c : c ∈ R}. Thus, S can be spanned by {1}. 25. In Problem 23 in Section 4.3, we found that nullspace(A) = {(−4t, t) : t ∈ R}. Hence, nullspace(A) = {t(−4, 1) : t ∈ R}. Thus, a spanning set for nullspace(A) is {(−4, 1)}. 26. In Problem 24 in Section 4.3, we found that nullspace(A) = {(3t − 2r, t, r) : r, t ∈ R}. Hence, nullspace(A) = {t(3, 1, 0)+r(−2, 0, 1) : r, t ∈ R}. Thus, a spanning set for nullspace(A) is {(3, 1, 0), (−2, 0, 1)}. 27. In Problem 25 in Section 4.3, we found that nullspace(A) = {(0, 0)}, so it can be spanned by {(0, 0)}. 28. In Problem 26 in Section 4.3, we found that nullspace(A) = {(t + 2r, −2t − 3r, t, r) : r, t ∈ R}. Hence, nullspace(A) = {t(1, −2, 1, 0) + r(2, −3, 0, 1) : r, t ∈ R}. Thus, a spanning set for nullspace(A) is {(1, −2, 1, 0), (2, −3, 0, 1)}. 29. In Problem 27 in Section 4.3, we found that nullspace(A) = {(0, 0, 0)}, so it can be spanned by {(0, 0, 0)}. 30. In Problem 28 in Section 4.3, we found that nullspace(A) = {(8r + 8s, −2r − 3s, r, s) : r, s ∈ R}. Hence, nullspace(A) = {r(8, −2, 1, 0) + s(8, −3, 0, 1) : r, s ∈ R}. Thus, a spanning set for nullspace(A) is {(8, −2, 1, 0), (8, −3, 0, 1)}. (c)2017 Pearson Education. Inc. 282 31. In Problem 29 in Section 4.3, we found that nullspace(A = {(0, 0, 0)}, so it can be spanned by {(0, 0, 0)}. ⎡ ⎤ 1 2 3 5 2 ⎦. nullspace(A) = {x ∈ R4 : Ax = 0}. The REDUCED ROW ECHELON FORM 32. A = ⎣ 1 3 4 2 4 6 −1 ⎡ ⎤ 1 0 1 0 0 of the augmented matrix of this system is ⎣ 0 1 1 0 0 ⎦. Consequently, nullspace(A) = {r(1, 1, −1, 0) : 0 0 0 1 0 r ∈ R} = span{(1, 1, −1, 0)}. ⎤ ⎡ ⎡ ⎤⎡ ⎤ 1 2 3 x1 0 33. nullspace(A) = {x ∈ R3 : Ax = 0}. Ax = 0 =⇒ ⎣ 1 3 4 ⎦ ⎣ x2 ⎦ = ⎣ 0 ⎦. 2 4 6 x3 0 Performing Gauss-Jordan elimination on the augmented matrix of the system we obtain: ⎤ ⎡ ⎤ ⎡ ⎤ 1 2 3 0 1 0 1 0 1 2 3 0 ⎣ 1 3 4 0 ⎦ ∼ ⎣ 0 1 1 0 ⎦ ∼ ⎣ 0 1 1 0 ⎦. 2 4 6 0 0 0 0 0 0 0 0 0 ⎡ From the last matrix we find that x1 = −r and x2 = −r where x3 = r, with r ∈ R. Consequently, x = (x1 , x2 , x3 ) = (−r, −r, r) = r(−1, −1, 1). Thus, nullspace(A) = {x ∈ R3 : x = r(−1, −1, 1), r ∈ R} = span{(−1, −1, 1)}. 34 Let v ∈ span{v1 , v2 } and a, b ∈ R. v = av1 + bv2 = a(1, −1, 2) + b(2, −1, 3) = (a, −a, 2a) + (2b, −b, 3b) = (a + 2b, −a − b, 2a + 3b). Thus, span{v1 , v2 } = {v ∈ R3 : v = (a + 2b, −a − b, 2a + 3b), a, b ∈ R}. Geometrically, span{v1 , v2 } is the plane through the origin determined by the two given vectors. The plane has parametric equations x = a + 2b, y = −a − b, and z = 2a + 3b. If a, b, and c are eliminated from the equations, then the resulting Cartesian equation is given by x − y − z = 0. 35. Let v ∈ span{v1 , v2 } and a, b ∈ R. v = av1 + bv2 = a(1, 2, −1) + b(−2, −4, 2) = (a, 2a, −a) + (−2b, −4b, 2b) = (a − 2b, 2a − 4b, −a + 2b) = (a − 2b)(1, 2, −1) = k(1, 2, −1), where k = a − 2b. Thus, span{v1 , v2 } = {v ∈ R3 : v = k(1, 2, −1), k ∈ R}. Geometrically, span{v1 , v2 } is the line through the origin determined by the vector (1, 2, −1). 36. Let v ∈ span{v1 , v2 , v3 } and a, b, c ∈ R. v = av1 + bv2 + cv3 = a(1, 1, −1) + b(2, 1, 3) + c(−2, −2, 2) = (a, a, −a) + (2b, b, 3b) + (−2c, −2c, 2c) = (a + 2b − 2c, a + b − 2c, −a + 3b + 2c). ⎧ ⎪ ⎨ a + 2b − 2c = x a + b − 2c = y −a + 3b + 2c = z Performing Gauss-Jordan elimination on the augmented matrix of the system, we obtain: ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ x 2y − x 1 2 −2 x 1 2 −2 1 0 −2 ⎦. ⎣ 1 1 −2 y ⎦ ∼ ⎣ 0 −1 0 y−x ⎦∼⎣ 0 1 0 x−y −1 3 2 z 0 5 0 x+z 0 0 0 5y − 4x + z Assuming that v = (x, y, z) and using the last ordered triple, we obtain the system: (c)2017 Pearson Education. Inc. ⎪ ⎩ 283 It is clear from the last matrix that the subspace, S, of R3 is a plane through (0, 0, 0) with Cartesian equation 4x − 5y − z = 0. Moreover, {v1 , v2 } also spans the subspace S since v = av1 + bv2 + cv3 = a(1, 1, −1) + b(2, 1, 3) + c(−2, −2, 2) = a(1, 1, −1) + b(2, 1, 3) − 2c(1, 1, −1) = (a − 2c)(1, 1, −1) + b(2, 1, 3) = dv1 + bv2 where d = a − 2c ∈ R. 37. If v ∈ span{v1 , v2 } then there exist a, b ∈ R such that v = av1 + bv2 = a(1, −1, 2) + b(2, 1, 3) = (a, −a, 2a) + (2b, b, 3b) = (a + 2b, −a + b, 2a + 3b). ⎧ ⎪ ⎨ a + 2b = 3 −a + b = 3 Hence, v = (3, 3, 4) is in span{v1 , v2 } provided there exists a, b ∈ R satisfying the system: ⎪ ⎩ 2a + 3b = 4 Solving this system we find that a = −1 and b = 2. Consequently, v = −v1 + 2v2 so that (3, 3, 4) ∈ span{v1 , v2 }. 38. If v ∈ span{v1 , v2 } then there exist a, b ∈ R such that v = av1 + bv2 = a(−1, 1, 2) + b(3, 1, −4) = (−a, a, 2a) + (3b, b, −4b) = (−a + 3b, a + b, 2a − 4b). ⎧ ⎪ ⎨ −a + 3b = 5 a+b= 3 Hence v = (5, 3, −6) is in span{v1 , v2 } provided there exists a, b ∈ R satisfying the system: ⎪ ⎩ 2a − 4b = −6 Solving this system we find that a = 1 and b = 2. Consequently, v = v1 + 2v2 so that (5, 3, −6) ∈ span{v1 , v2 }. 39. If v ∈ span{v1 , v2 } then there exist a, b ∈ R such that v = av1 + bv2 = (3a, a, 2a) + (−2b, −b, b) = (3a − 2b, a − b, 2a + b). ⎧ ⎪ ⎨ 3a − 2b = 1 a−b= 1 Hence v = (1, 1, −2) is in span{v1 , v2 } provided there exists a, b ∈ R satisfying the system: ⎪ ⎩ 2a + b = −2 ⎤ ⎤ ⎡ ⎡ 1 1 3 −2 1 −1 1 −2 ⎦, it follows that the system has no solution. Hence it 1 ⎦ reduces to ⎣ 0 Since ⎣ 1 −1 2 1 −2 0 0 2 must be the case that (1, 1, −2) ∈ / span{v1 , v2 }. 40. If p ∈ span{p1 , p2 } then there exist a, b ∈ R such that p(x) = ap1 (x) + bp2 (x), so p(x) = 2x2 − x + 2 is in span{p1 , p2 } provided there exist a, b ∈ R such that 2x2 − x + 2 = a(x − 4) + b(x2 − x + 3) = ax − 4a + bx2 − bx + 3b = bx2 + (a − b)x + (3b − 4a). Equating like coefficients and solving, we find that a = 1 and b = 2. Thus, 2x2 − x + 2 = 1 · (x − 4) + 2 · (x2 − x + 3) = p1 (x) + 2p2 (x) so p ∈ span{p1 , p2 }. 41. Let A ∈ span{A1 , A2 , A3 } and c1 , c2, c3 ∈ R. 1 −1 0 1 3 0 + c2 + c3 A = c1 A1 + c2 A2 + c3 A3 = c1 2 0 −2 1 1 2 −c1 + c2 0 c2 3c3 0 c1 + 3c3 c1 −c1 + + = . = c3 2c3 2c1 0 −2c2 c2 2c1 − 2c2 + c3 c2 + 2c3 (c)2017 Pearson Education. Inc. 284 Therefore span{A1 , A2 , A3 } = A ∈ M2 (R) : A = 42. Let A ∈ span{A1 , A2 } and a, b ∈ R. 1 2 −2 +b A = aA1 + bA2 = a −1 3 1 c1 + 3c3 2c1 − 2c2 + c3 −c1 + c2 c2 + 2c3 . 1 a 2a −2b b a − 2b 2a + b = + = . −1 −a 3a b −b −a + b 3a − b a − 2b 2a + b . Now, to determine whether B ∈ span{A1 , A2 }, −a + b 3a − b So span{A1 , A2 } = A ∈ M2 (R) : A = a − 2b 2a + b 3 1 let = . This implies that a = 1 and b = −1, thus B ∈ span{A1 , A2 }. −a + b 3a − b −2 4 43. (a). The general vector in span{f, g} is of the form h(x) = c1 cosh x + c2 sinh x where c1 , c2 ∈ R. (b). Let h ∈ S and c1 , c2 ∈ R. Then h(x) = c1 f (x) + c2 g(x) = c1 cosh x + c2 sinh x ex + e−x ex − e−x c1 ex c1 e−x c2 ex c2 e−x + c2 = + + − 2 2 2 2 2 2 c1 + c2 x c1 − c2 −x x −x = e + e = d1 e + d2 e 2 2 c1 + c2 c1 − c2 where d1 = and d2 = . Therefore S =span{ex , e−x }. 2 2 44. The origin in R3 . = c1 45. All points lying on the line through the origin with direction v1 . 46. All points lying on the plane through the origin containing v1 and v2 . 47. If v1 = v2 = 0, then the subspace is the origin in R3 . If at least one of the vectors is nonzero, then the subspace consists of all points lying on the line through the origin in the direction of the nonzero vector. 48. Suppose that S is a subset of S . We must show that every vector in span(S) also belongs to span(S ). Every vector v that lies in span(S) can be expressed as v = c1 v1 + c2 v2 + · · · + ck vk , where v1 , v2 , . . . , vk belong to S. However, since S is a subset of S , v1 , v2 , . . . , vk also belong to S , and therefore, v belongs to span(S ). Thus, we have shown that every vector in span(S) also lies in span(S ). 49. Proof of =⇒: We begin by supposing that span{v1 , v2 , v3 } = span{v1 , v2 }. Since v3 ∈ span{v1 , v2 , v3 }, our supposition implies that v3 ∈ span{v1 , v2 }, which means that v3 can be expressed as a linear combination of the vectors v1 and v2 . Proof of ⇐=: Now suppose that v3 can be expressed as a linear combination of the vectors v1 and v2 . We must show that span{v1 , v2 , v3 } = span{v1 , v2 }, and we do this by showing that each of these subsets is a subset of the other. Since it is clear that span{v1 , v2 } is a subset of span{v1 , v2 , v3 }, we focus our attention on proving that every vector in span{v1 , v2 , v3 } belongs to span{v1 , v2 }. To see this, suppose that v belongs to span{v1 , v2 , v3 }, so that we may write v = c1 v1 + c2 v2 + c3 v3 . By assumption, v3 can be expressed as a linear combination of v1 and v2 , so that we may write v3 = d1 v1 + d2 v2 . Hence, we obtain v = c1 v1 + c2 v2 + c3 v3 = c1 v1 + c2 v2 + c3 (d1 v1 + d2 v2 ) = (c1 + c3 d1 )v1 + (c2 + c3 d2 )v2 ∈ span{v1 , v2 }, as required. Solutions to Section 4.5 (c)2017 Pearson Education. Inc. 285 (a): FALSE. For instance, consider the vector space V = R2 . Here are two different minimal spanning sets for V : {(1, 0), (0, 1)} and {(1, 0), (1, 1)}. Many other examples of this abound. (b): TRUE. We have seven column vectors, and each of them belongs to R5 . Therefore, the number of vectors present exceeds the number of components in those vectors, and hence they must be linearly dependent. (c): FALSE. For instance, the 7 × 5 zero matrix, 07×5 , does not have linearly independent columns. (d): TRUE. Any linear dependencies within the subset also represent linear dependencies within the original, larger set of vectors. Therefore, if the nonempty subset were linearly dependent, then this would require that the original set is also linearly dependent. In other words, if the original set is linearly independent, then so is the nonempty subset. (e): TRUE. This is stated in Theorem 4.5.23. (f ): TRUE. If we can write v = c1 v1 + c2 v2 + · · · + ck vk , then {v, v1 , v2 , . . . , vk } is a linearly dependent set. (g): TRUE. This is a rephrasing of the statement in True-False Review Question (d) above. (h): FALSE. None of the vectors (1, 0), (0, 1), and (1, 1) in R2 are proportional to each other, and yet, they form a linearly dependent set of vectors. (i): FALSE. The illustration given in part (c) of Example 4.5.24 gives an excellent case-in-point here. Problems: 1. This set is linearly independent by Theorem 4.5.5. 2. {(1, −1), (1, 1)}. These vectors are elements of R2 . Since there are two vectors, and the dimension of R2 1 1 = 0. Now is two, Corollary 4.5.17 states that the vectors will be linearly dependent if and only if −1 1 1 1 −1 1 = 2 = 0. Consequently, the given vectors are linearly independent. 3. {(2, −1), (3, 2), (0, 1)}. These vectors are elements of R2 , but since there are three vectors, the vectors are linearly dependent by Corollary 4.5.17. Let v1 = (2, −1), v2 = (3, 2), v3 = (0, 1). We now determine a dependency relationship. The condition c1 v 1 + c2 v2 + c3 v 3 = 0 requires 2c1 + 3c2 = 0 and −c1 + 2c2 + c3 = 0. 1 0 − 37 2 0 1 7 Hence the system has solution c1 = 3r, c2 = −2r, c3 = 7r. Consequently, 3v1 − 2v2 + 7v3 = 0. The REDUCED ROW ECHELON FORM of the augmented matrix of this system is 3 4. 1, −1), (1, 1, 1)}. These vectors are elements of R . {(1, −1, 0), (0, 1 0 1 −1 1 1 = 3 = 0, so by Corollary 4.5.17, the vectors are linearly independent. 0 −1 1 5. {(1, 2, 3), (1, −1, 2), (1, −4, 1)}. These vectors are elements of R3 . (c)2017 Pearson Education. Inc. 0 0 . 286 1 1 1 2 −1 −4 = 0, so by Corollary 4.5.17, the vectors are linearly dependent. Let v1 = (1, 2, 3), v2 = 3 2 1 (1, −1, 2), v3 = (1, −4, 1). We determine a dependency relation. The condition c1 v1 + c2 v2 + c3 v3 = 0 requires c1 + c2 + c3 = 0, 2c1 − c2 − 4c3 = 0, 3c1 + 2c2 + c3 = 0. ⎡ ⎤ 1 0 −1 0 2 0 ⎦. The REDUCED ROW ECHELON FORM of the augmented matrix of this system is ⎣ 0 1 0 0 0 0 Hence c1 = r, c2 = −2r, c3 = r, and so v1 − 2v2 + v3 = 0. 6. Given {(−2, 4, −6), (3, −6, 9)}. The vectors are linearly dependent because 3(−2, 4, −6) + 2(3, −6, 9) = (0, 0, 0), which gives a linear dependency relation. Alternatively, let a, b ∈ R and observe that a(−2, 4, −6) + b(3, −6, 9) = (0, 0, 0) =⇒ (−2a, 4a, −6a) + (3b, −6b, 9b) = (0, 0, 0). ⎧ ⎪ ⎨ −2a + 3b = 0 4a − 6b = 0 The last equality results in the system: ⎪ ⎩ −6a + 9b = 0 ⎤ ⎡ 1 − 32 0 0 0 ⎦, which The REDUCED ROW ECHELON FORM of the augmented matrix of this system is ⎣ 0 0 0 0 3 implies that a = b. Thus, the given set of vectors is linearly dependent. 2 7. {(1, −1, 2), (2, 1, 0)}. Let a, b ∈ R. a(1, −1, 2) + b(2, 1, 0) = (0, 0, 0) =⇒ (a, ⎧ −a, 2a) + (2b, b, 0) = (0, 0, 0) =⇒ (a + 2b, −a + b, 2a) = (0, 0, 0). The ⎪ a + 2b = 0 ⎨ −a + b = 0 . Since the only solution of the system is a = b = 0, it last equality results in the system: ⎪ ⎩ 2a = 0 follows that the vectors are linearly independent. 8. {(−1, 1, 2), (0, 2, −1), (3, 1, 2), (−1, −1, 1)}. These vectors are elements of R3 . Since there are four vectors, it follows from Corollary 4.5.17 that the vectors are linearly dependent. Let v1 = (−1, 1, 2), v2 = (0, 2, −1), v3 = (3, 1, 2), v4 = (−1, −1, 1). Then c1 v 1 + c2 v 2 + c3 v3 + c4 v 4 = 0 requires −c1 + 3c3 − c4 = 0, c1 + 2c2 + c3 − c4 = 0, 2c1 − c2 + 2c3 + c4 = 0. ⎡ 2 1 0 0 5 ⎢ ⎢ 3 The REDUCED ROW ECHELON FORM of the augmented matrix of this system is ⎢ ⎢ 0 1 0 −5 ⎣ 0 0 1 − 15 so that c1 = −2r, c2 = 3r, c3 = r, c4 = 5r. Hence, −2v1 + 3v2 + v3 + 5v4 = 0. (c)2017 Pearson Education. Inc. 0 ⎤ ⎥ ⎥ 0 ⎥ ⎥ ⎦ 0 287 9. {(1, −1, 2, 3), (2, −1, 1, −1), (−1, 1, 1, 1)}. Let a, b, c ∈ R. a(1, −1, 2, 3) + b(2, −1, 1, −1) + c(−1, 1, 1, 1) = (0, 0, 0, 0) =⇒ (a, −a, 2a, 3a) + (2b, −b, b, −b) + (−c, c, c, c) = (0, 0, 0, 0) =⇒ (a + 2b − c, −a − b + c, 2a + b + c, 3a − b + c) = (0, 0, 0, 0). ⎧ a + 2b − c = 0 ⎪ ⎪ ⎪ ⎨ −a − b + c = 0 . The REDUCED ROW ECHELON FORM of the The last equality results in the system: ⎪ 2a + b + c = 0 ⎪ ⎪ ⎩ 3a − b + c = 0 ⎡ ⎤ 1 0 0 0 ⎢ 0 1 0 0 ⎥ ⎥ augmented matrix of this system is ⎢ ⎣ 0 0 1 0 ⎦. Consequently, a = b = c = 0. Thus, the given set of 0 0 0 0 vectors is linearly independent. 10. {(2, −1, 0, 1), (1, 0, −1, 2), (0, 3, 1, 2), (−1, 1, 2, 1)}. These vectors are elements in R4 . By CET (row 1), we obtain: 2 1 0 −1 0 3 1 −1 3 1 −1 0 3 −1 0 3 1 = 2 −1 1 2 − 0 1 2 + 0 −1 1 = 21 = 0. 0 −1 1 2 2 2 1 1 2 1 1 2 2 1 2 2 1 Thus, it follows from Corollary 4.5.17 that the vectors are linearly independent. 1 4 7 11. Since 2 5 8 = 0, the given set of vectors is linearly dependent in R3 . Further, since the vectors 3 6 9 are not collinear, it follows that span{v1 , v2 , v3 } is a plane 3-space. 2 1 −3 3 −9 = 0, the given set of vectors is linearly dependent. 12. (a). Since −1 5 −4 12 (b). By inspection, we see that v3 = −3v2 . Hence v2 and v3 are collinear and therefore span{v2 , v3 } is the line through the origin that has the direction of v2 . Further, since v1 is not proportional to either of these vectors, it does not lie along the same line, hence v1 is not in span{v2 , v3 }. 1 0 1 1 0 1 2 k − 1 = (3 − k)(k + 4). Now, the determinant is zero when 13. 1 2 k = 0 2 k − 1 = k 6−k k k 6 0 k 6−k k = 3 or k = −4. Consequently, by Corollary 4.5.17, the vectors are linearly dependent if and only if k = 3 or k = −4. 14. {(1, 0, 1, k), (−1, 0, k, 1), (2, 0, 1, 3)}. These vectors are elements in R4 . Let a, b, c ∈ R. a(1, 0, 1, k) + b(−1, 0, k, 1) + c(2, 0, 1, 3) = (0, 0, 0, 0) =⇒ (a, 0, a, ka) + (−b, 0, kb, b) + (2c, 0, c, 3c) = (0, 0, 0, 0) =⇒ (a − b + 2c, 0, a + kb + c, ka + b + 3c) ⎧ = (0, 0, 0, 0). ⎪ ⎨ a − b + 2c = 0 a + kb + c = 0 . Evaluating the determinant of the coefficient The last equality results in the system: ⎪ ⎩ ka + b + 3c = 0 (c)2017 Pearson Education. Inc. 288 matrix, we obtain 1 1 k −1 2 1 k 1 = 1 1 3 k 0 k+1 k+1 0 −1 = 2(k + 1)(2 − k). 3 − 2k Consequently, the system has only the trivial solution, hence the given set of vectors are linearly independent if and only if k = 2, −1. 15. {(1, 1, 0, −1), (1, k, 1, 1), (2, 1, k, 1), (−1, 1, 1, k)}. These vectors are elements in R4 . Let a, b, c, d ∈ R. a(1, 1, 0, −1) + b(1, k, 1, 1) + c(2, 1, k, 1) + d(−1, 1, 1, k) = (0, 0, 0, 0) =⇒ (a, a, 0, −a) + (b, kb, b, b) + (2c, c, kc, c) + (−d, d, d, kd) = (0, 0, 0, 0) =⇒ (a + b + 2c − d, a + kb + c + d, b + kc + d, −a + b + c + kd) = (0, 0, 0, 0). ⎧ a + b + 2c − d = 0 ⎪ ⎪ ⎪ ⎨ a + kb + c + d = 0 The last equality results in the system: . By Corollary 3.2.6, this system has the ⎪ b + kc + d = 0 ⎪ ⎪ ⎩ −a + b + c + kd = 0 trivial solution if and only if the determinant of the coefficient matrix is zero. Evaluating the determinant of we obtain: the coefficient matrix, 1 1 2 −1 1 1 2 −1 2 0 −k 2 + k − 1 3 − k k − 1 −1 0 k − 1 −1 1 k 1 2 1 = 1 = k 1 = 1 k 1 0 1 k 1 k 1 1 0 2 3 k−1 0 3 − 2k k−3 −1 1 1 2 3 k−1 k 0 2 2 k −k+1 1 k −k+1 k−3 = (k − 3)(k − 1)(k + 2). = (k − 3) = 3 − 2k k−3 3 − 2k 1 For the original set of vectors to be linearly independent, we need a = b = c = 0. We see that this condition will be true provided that k ∈ R such that k = 3, 1, −2. 1 1 2 −1 3 6 0 0 16. Let a, b, c ∈ R. a +b +c = 0 1 0 1 0 4 0 0 ⎧ ⎪ ⎨ a + 2b + 3c = 0 a + 2b + 3c a − b + 6c 0 0 a − b + 6c = 0 . =⇒ = . The last equation results in the system: 0 a + b + 4c 0 0 ⎪ ⎩ a + b + 4c = 0 ⎤ ⎡ 1 0 5 0 The REDUCED ROW ECHELON FORM of the augmented matrix of this system is ⎣ 0 1 −1 0 ⎦, 0 0 0 0 which implies that the system has an infinite number of solutions. Consequently, the given matrices are linearly dependent. 2 −1 −1 2 0 0 17. Let a, b ∈ R. a +b = 3 4 1 3 0 0 ⎧ 2a − b = 0 ⎪ ⎪ ⎪ ⎨ 3a + b = 0 2a − b −a + 2b 0 0 . This system =⇒ = . The last equation results in the system: 3a + b 4a + 3b 0 0 ⎪ −a + 2b = 0 ⎪ ⎪ ⎩ 4a + 3b = 0 has only the trivial solution a = b = 0, thus the given matrices are linearly independent in M2 (R). 1 0 −1 1 2 1 0 0 18. Let a, b, c ∈ R. a +b +c = 1 2 2 1 5 7 0 0 (c)2017 Pearson Education. Inc. 289 ⎧ a − b + 2c = 0 ⎪ ⎪ ⎪ ⎨ a + 2b + 5c = 0 0 0 a − b + 2c b+c . . The last equation results in the system: = =⇒ 0 0 a + 2b + 5c 2a + b + 7c ⎪ 2a + b + 7c = 0 ⎪ ⎪ ⎩ b+c=0 ⎡ ⎤ 1 0 3 0 ⎢ 0 1 1 0 ⎥ ⎥ The REDUCED ROW ECHELON FORM of the augmented matrix of this system is ⎢ ⎣ 0 0 0 0 ⎦, which 0 0 0 0 implies that the system has an infinite number of solutions. Consequently, the given matrices are linearly dependent. 19. Let a, b ∈ R. ap1 (x) + bp2 (x) = 0 =⇒ a(1 − x) + b(1 + x) = 0 =⇒ (a + b) + (−a + b)x = 0. a+b=0 Equating like coefficients, we obtain the system: . Since the only solution to this system is −a + b = 0 a = b = 0, it follows that the given vectors are linearly independent. 20. Let a, b ∈ R. ap1 (x) + bp2 (x) = 0 =⇒ a(2 + 3x) + b(4 + 6x) = 0 =⇒ (2a + 4b) + (3a + 6b)x = 0. 2a + 4b = 0 Equating like coefficients, we obtain the system: . The REDUCED ROW ECHELON FORM 3a + 6b = 0 1 2 0 , which implies that the system has an infinite number of the augmented matrix of this system is 0 0 0 of solutions. Thus, the given vectors are linearly dependent. 21. Let c1 , c2 , c3 ∈ R. c1 p1 (x) + c2 p2 (x) + c3 p3 (x) = 0 =⇒ c1 (1 − 3x2 ) + c2 (2x + x2 ) + c3 (5) = 0. Equating like coefficients, we obtain c1 + 5c3 = 0, 2c2 = 0, −3c1 + c2 = 0. It is easy to solve this system of equations to find that c1 = c2 = c3 = 0. Therefore, the given set of vectors is linearly independent. 22. In Section 4.6, we will see that four vectors in P2 (R) must be linearly dependent, with no calculations necessary. However, for now we proceed as follows. Let c1 , c2 , c3 , c4 ∈ R. c1 p1 (x) + c2 p2 (x) + c3 p3 (x) + c4 p4 (x) = 0 =⇒ c1 (3x + 5x2 ) + c2 (1 + x + x2 ) + c3 (2 − x) + c4 (1 + 2x2 ) = 0. Equating like coefficients, we obtain c2 + 2c3 + c4 = 0, 3c1 + c2 − c3 = 0, 5c1 + c2 + 2c4 = 0. ⎡ ⎤ 0 1 2 1 0 This linear system has augmented matrix A# = ⎣ 3 1 −3 0 0 ⎦. Because the matrix A# only has 5 1 0 2 0 three rows, its RREF can have at most three pivots, which implies that at least one of the columns A will be unpivoted. The resulting presence of free variables implies that this system has an infinite number of solutions for c1 , c2 , c3 , and c4 . Thus, the given set of vectors is linearly dependent. 23. Let c1 , c2 ∈ R. c1 p1 (x) + c2 p2 (x) = 0 =⇒ c1 (a + bx) + c2 (c + dx) = 0 =⇒ (ac1 + cc2 ) + (bc1 + dc2 )x = 0. ac1 + cc2 = 0 Equating like coefficients, we obtain the system: . The determinant of the matrix of coefbc1 + dc2 = 0 a c = ad − bc. Consequently, the system has just the trivial solution, and hence p1 (x) and ficients is b d p2 (x) are linearly independent if and only if ad − bc = 0. (c)2017 Pearson Education. Inc. 290 24. Since cos 2x = cos2 x−sin2 x, f1 (x) = f3 (x)−f2 (x) so it follows that f1 , f2 , and f3 are linearly dependent in C ∞ (−∞, ∞). 25. Let v1 = (1, 2, 3), v2 = (−3, 4, 5), v3 = (1, − 43 , − 53 ). By inspection, we see that v2 = −3v3 . Further, since v1 and v2 are not proportional they are linearly independent. Consequently, {v1 , v2 } is a linearly independent set of vectors and span{v1 , v2 } =span{v1 , v2 , v3 }. 26. Let v1 = (3, 1, 5), v2 = (0, 0, 0), v3 = (1, 2, −1), v4 = (−1, 2, 3). Since v2 = 0, it is certainly true that span{v1 , v3 v4 } =span{v1 , v2 , v3 v4 }. Further, since det[v1 , v3 , v4 ] = 42 = 0, {v1 , v3 , v4 } is a linearly independent set. 27. Since we have four vectors in R3 , the given set is linearly dependent. We could determine the specific linear dependency between the vectors to find a linearly independent subset, but in this case, if we just take any 1 1 3 three of the vectors, say (1, −1, 1), (1, −3, 1), (3, 1, 2), then −1 −3 1 = 2 = 0, so that these vectors are 1 1 2 linearly independent. Consequently, span{(1, −1, 1), (1, −3, 1), (3, 1, 2)} = span{(1, 1, 1), (1, −1, 1), (1, −3, 1), (3, 1, 2)}. 28. Let v1 = (1, 1, −1, 1),v2 = (2, −1, 3, 1), v3 = (1, 1, 2, 1), v4 = (2, −1, 2, 1). 1 2 1 2 1 −1 1 −1 = 0, the set {v1 , v2 , v3 , v4 } is linearly dependent. We now determine the linearly Since 3 2 2 −1 1 1 1 1 dependent relationship. The REDUCED ROW ECHELON FORM of the augmented matrix corresponding to the system c1 (1, 1, −1, 1) + c2 (2, −1, 3, 1) + c3 (1, 1, 2, 1) + c4 (2, −1, 2, 1) = (0, 0, 0, 0) ⎤ 1 1 0 0 0 3 ⎢ 0 1 0 1 0 ⎥ ⎥ is ⎢ ⎣ 0 0 1 − 1 0 ⎦, so that c1 = −r, c2 = −3r, c3 = r, c4 = 3r, where r is a free variable. It follows 3 0 0 0 0 0 that a linearly dependent relationship between the given set of vectors is ⎡ −v1 − 3v2 + v3 + 3v4 = 0 so that v1 = −3v2 + v3 + 3v4 . Consequently, span{v2 , v3 , v4 } = span{v1 , v2 , v3 , v4 }, and {v2 , v3 , v4 } is a linearly independent set. 1 2 −1 2 3 2 29. Let A1 = , A2 = , A3 = . Then 3 4 5 7 1 1 c1 A1 + c2 A2 + c3 A3 = 02 requires that c1 − c2 + 3c3 = 0, 2c1 + 2c2 + 2c3 = 0, 3c1 + 5c2 + c3 = 0, 4c1 + 7c2 + c3 = 0. ⎡ ⎤ 1 0 2 0 ⎢ 0 1 −1 0 ⎥ ⎥. The REDUCED ROW ECHELON FORM of the augmented matrix of this system is ⎢ ⎣ 0 0 0 0 ⎦ 0 0 0 0 Consequently, the matrices are linearly dependent. Solving the system gives c1 = −2r, c2 = c3 = r. Hence, a linearly dependent relationship is −2A1 + A2 + A3 = 02 . (c)2017 Pearson Education. Inc. 291 30. We first determine whether the given set of polynomials is linearly dependent. Let p1 (x) = 2 − 5x, p2 (x) = 3 + 7x, p3 (x) = 4 − x. Then c1 (2 − 5x) + c2 (3 + 7x) + c3 (4 − x) = 0 requires 2c1 + 3c2 + 4c3 = 0 and − 5c1 + 7c2 − c3 = 0. This system has solution (−31r, −18r, 29r), where r is a free variable. Consequently, the given set of polynomials is linearly dependent, and a linearly dependent relationship is −31p1 (x) − 18p2 (x) + 29p3 (x) = 0, or equivalently, 1 [31p1 (x) + 18p2 (x)]. 29 Hence, the linearly independent set of vectors {2 − 5x, 3 + 7x} spans the same subspace of P1 (R) as that spanned by {2 − 5x, 3 + 7x, 4 − x}. p3 (x) = 31. We first determine whether the given set of polynomials is linearly dependent. Let p1 (x) = 2 + x2 , p2 (x) = 4 − 2x + 3x2 , p3 (x) = 1 + x. Then c1 (2 + x2 ) + c2 (4 − 2x + 3x2 ) + c3 (1 + x) = 0 leads to the system 2c1 + 4c2 + c3 = 0, −2c2 + c3 = 0, c1 + 3c2 = 0. This system has solution (−3r, r, 2r) where r is a free variable. Consequently, the given set of vectors is linearly dependent, and a specific linear relationship is −3p1 (x) + p2 (x) + 2p3 (x) = 0, or equivalently, p2 (x) = 3p1 (x) − 2p3 (x). Hence, the linearly independent set of vectors {2 + x2 , 1 + x} spans the same subspace of P2 (R) as that spanned by the given set of vectors. 1 x x2 32. W [f1 , f2 , f3 ](x) = 0 1 2x = 2. Since W [f1 , f2 , f3 ](x) = 0 on I, it follows that the functions are 0 0 2 linearly independent on I. sin x W [f1 , f2 , f3 ](x) = cos x − sin x 33. cos x − sin x − cos x sin x tan x = cos x sec2 x 2 tan x sec2 x 0 cos x − sin x 0 tan x sec2 x tan x(2 sec2 x + 1) = − tan x(2 sec2 x + 1). W [f1 , f2 , f3 ](x) is not always zero over I, so the vectors are linearly independent by Theorem 4.5.23. (c)2017 Pearson Education. Inc. 292 1 3x x2 − 1 3 2x = 6 = 0 on I. Consequently, {f1 , f2 , f3 } is a linearly 2x = 34. W [f1 , f2 , f3 ](x) = 0 3 0 2 0 0 2 independent set on I by Theorem 4.5.23. 2x e 1 1 1 0 1 e3x e−x 0 2x 3e3x −e−x = e4x 2 3 −1 = e4x 2 1 −3 = 12e4x . Since the 35. W [f1 , f2 , f3 ](x) = 2e 4 9 4e2x 9e3x 4 5 −3 1 e−x Wronskian is never zero, the functions are linearly independent on (−∞, ∞). 36. On [0, ∞), f2 = 7f1 , so that the functions are linearly dependent on this interval. Therefore W [f1 , f2 ](x) = 3x3 7x2 = −21x4 = 0. Since the Wronskian is 0 for x ∈ [0, ∞). However, on (−∞, 0), W [f1 , f2 ](x) = 9x2 14x not zero for all x ∈ (−∞, ∞), the functions are linearly independent on that interval. 1 x 2x − 1 = 0. By inspection, we see that f3 = 2f2 − f1 , so that the functions 2 37. W [f1 , f2 , f3 ](x) = 0 1 0 0 0 are linearly dependent on (−∞, ∞). 38. We show that the Wronskian (the determinant can be computed by cofactor expansion along the first row) is identically zero: x −x e x e −x cosh x e x −e−x sinh x = −(cosh x + sinh x) − (cosh x − sinh x) + 2 cosh x = 0. e e cosh x Thus, the Wronskian is identically zero on (−∞, ∞). Furthermore, {f1 , f2 , f3 } is a linearly dependent set because 1 1 1 1 1 ex + e−x 1 = 0 for all x ∈ I. − f1 (x) − f2 (x) + f3 (x) = − ex − e−x + cosh x = − ex − e−x + 2 2 2 2 2 2 2 39. We show that the Wronskian is identically zero for f1 (x) = ax3 and f2 (x) = bx3 , which covers the functions in this problem as a special case: ax3 bx3 5 5 3ax2 3bx2 = 3abx − 3abx = 0. Next, let a, b ∈ R. If x ≥ 0, then af1 (x) + bf2 (x) = 0 =⇒ 2ax3 + 5bx3 = 0 =⇒ (2a + 5b)x3 = 0 =⇒ 2a + 5b = 0. If x ≤ 0, then af1 (x) + bf2 (x) = 0 =⇒ 2ax3 − 3bx3 = 0 =⇒ (2a − 3b)x3 = 0 =⇒ 2a − 3b = 0. Solving the resulting system, we obtain a = b = 0; therefore, {f1 , f2 } is a linearly independent set of vectors on (−∞, ∞). 40. (a). When x > 0, f2 (x) = 1 and when x < 0, f2 (x) = −1; thus f2 (0) does not exist, which implies that f2 ∈ / C 1 (−∞, ∞). (c)2017 Pearson Education. Inc. 293 (b). Let a, b ∈ R. On the interval (−∞, 0), ax + b(−x) = 0, which has more than the trivial solution for a and b. Thus, {f1 , f2 } is a linearly dependent set of vectors on (−∞, 0). On the interval [0, ∞), ax + bx = 0 =⇒ a + b = 0, which has more than the trivial solution for a and b. Therefore {f1 , f2 } is a linearly dependent set of vectors on [0, ∞). a−b=0 On the interval (−∞, ∞), a and b must satisfy: ax + b(−x) = 0 and ax + bx = 0, that is, . Since a+b=0 this system has only a = b = 0 as its solution, {f1 , f2 } is a linearly independent set of vectors on (−∞, ∞). y y = f2(x) = -x f1(x) = - f2(x) on (-', 0) y = f1(x) = x = f2(x) f1(x) = f2(x) on (0, ') x y = f1(x) = x Figure 0.0.63: Figure for Problem 40 41. Let a, b ∈ R. ax + bx = 0 if x = 0 (a + b)x = 0 =⇒ a(0) + b(1) = 0 if x = 0 b=0 =⇒ a = 0 and b = 0 so {f1 , f2 } is a linearly independent set on I. af1 (x) + bf2 (x) = 0 =⇒ 42. Let a, b, c ∈ R and x ∈ (−∞, ∞). af1 (x) + bf2 (x) + cf3 (x) = 0 =⇒ a(x − 1) + b(2x) + c(3) = 0 for x ≥ 1 2a(x − 1) + b(2x) + c(3) = 0 for x < 1 (a + 2b)x + (−a + 3c) = 0 a + 2b = 0 and − a + 3c = 0 =⇒ (2a + 2b)x + (−2a + 3c) = 0 2a + 2b = 0 and − 2a + 3c = 0. Since the only solution to this system of equations is a = b = c = 0, it follows that the given functions are linearly independent on (−∞, ∞). The domain space may be divided into three types of intervals: (1) interval subsets of (−∞, 1), (2) interval subset of [1, ∞), (3) intervals containing 1 where 1 is not an endpoint of the intervals. =⇒ For intervals of type (3): Intervals such as type (3) are treated as above [with domain space of (−∞, ∞)]: vectors are linearly independent. For intervals of type (1): a(2(x − 1)) + b(2x) + c(3) = 0 =⇒ (2a + 2b)x + (−2a + 3c) = 0 =⇒ 2a + 2b = 0, and −2a + 3c = 0. Since this system has three variables with only two equations, the solution to the system is not unique, hence intervals of type (1) result in linearly dependent vectors. For intervals of type (2): (c)2017 Pearson Education. Inc. 294 a(x−1)+b(2x)+c(3) = 0 =⇒ a+2b = 0 and −a+3c = 0. As in the last case, this system has three variables with only two equations, so it must be the case that intervals of type (2) result in linearly dependent vectors. 43. (a). Let f0 (x) = 1, f1 (x) = x2 , f3 (x) = x3 . = x, f2 (x) 1 x x2 x 3 0 1 2x 3x2 = 12 = 0. Hence, {f1 , f2 , f3 , f4 } is linearly independent on Then W [f1 , f2 , f3 , f4 ](x) = 6x 0 0 2 0 0 0 6 any interval. 1 x x2 · · · xn n−1 0 1 2x · · · nx 0 0 2 · · · n(n − 1)xn−2 (b). W [f0 , f1 , f2 , . . . , fn ](x) = . .. .. .. .. . . . . 0 0 0 ··· n! The matrix corresponding to this determinant is upper triangular, so the value of the determinant is given by the product of all of the diagonal entries. W [f0 , f1 , f2 , . . . , fn ](x) = 1 · 1 · 2 · 6 · 24 · · · n!, which is not zero regardless of the actual domain of x. Consequently, the functions are linearly independent on any interval. 44. (a). Let f1 (x) = er1 x , f2 (x) = er2 x and f 3 (x) = er3 x . Then e r1 x 1 e r2 x er3 x r1 x r2 x r3 x r1 x r2 x r3 x r2 e r3 e = e e e r1 W [f1 , f2 , f3 ](x) = r1 e r12 er1 x r22 er2 x r32 er3 x r12 1 r2 r22 1 r3 r32 = e(r1 +r2 +r3 )x (r3 − r1 )(r3 − r2 )(r2 − r1 ). If ri = rj for i = j, then W [f1 , f2 , f3 ](x) is never zero, and hence the functions are linearly independent on any interval. If, on the other hand, ri = rj with i = j, then fi − fj = 0, so that the functions are linearly dependent. Thus, r1 , r2 , r3 must all be different in order that f1 , f2 , f3 are linearly independent. (b). e r1 x r1 e r 1 x r12 er1 x W [f1 , f2 , f3 , . . . , fn ](x) = .. . n−1 r e r1 x e r2 x r2 e r 2 x r22 er2 x .. . ··· ··· ··· ern x rn ern x rn2 ern x .. . rn−1 ern x r2n−1 er2 x · · · n 1 1 ··· r1 r ··· 2 2 2 r1 x r 2 x r n x r1 r ··· 2 = e e ···e .. .. . n−1 . n−1 r r ··· 1 1 2 n−1 r 1 rn rn2 .. . n ∗ = er1 x er2 x · · · ern x V (r1 , r2 , . . . , rn ) = e r1 x e r2 x · · · e rn x (rm − ri ). 1≤i<m≤n Now from the last equality, we see that if ri = rj for i = j, where i, j ∈ {1, 2, 3, . . . , n}, then W [f1 , f2 , . . . , fn ](x) is never zero, and hence the functions are linearly independent on any interval. If, on the other hand, ri = rj with i = j, then f1 − fj = 0, so that the functions are linearly dependent. Thus {f1 , f2 , . . . , fn } is a linearly independent set if and only if all the ri are distinct for i ∈ {1, 2, 3, . . . , n}. (c)2017 Pearson Education. Inc. 295 (*Note: V (r1 , r2 , . . . , rn ) is the n × n Vandermonde determinant. See Problem 30 in Section 3.3). 45. Let a, b ∈ R. Assume that av + bw = 0. Then a(αv1 + v2 ) + b(v1 + αv2 ) = 0 which implies that (αa + b)v1 + (a + bα)v2 = 0. Now since it is given that v1 and v2 are linearly independent, αa + b = 0 and α 1 = 0. That is, if and a + bα = 0. This system has only the trivial solution for a and b if and only if 1 α only if α2 − 1 = 0, or α = ±1. Therefore, the vectors are linearly independent provided that α = ±1. 46. It is given that v1 and v2 are linearly independent. Let u1 = a1 v1 + b1 v2 , u2 = a2 v1 + b2 v2 , and u3 = a3 v1 + b3 v2 where a1 , a2 , a3 , b1 , b2 , b3 ∈ R. Let c1 , c2 , c3 ∈ R. Then c1 u 1 + c2 u 2 + c3 u 3 = 0 =⇒ c1 (a1 v1 + b1 v2 ) + c2 (a2 v1 + b2 v2 ) + c3 (a3 v1 + b3 v2 ) = 0 =⇒ (c1 a1 + c2 a2 + c3 a3 )v1 + (c1 b1 + c2 b2 + c3 b3 )v2 = 0. c1 a1 + c2 a2 + c3 a3 = 0 Now since v1 and v2 are linearly independent: There are an infinite number c1 b1 + c2 b2 + c3 b3 = 0. of solutions to this homogeneous system since there are three unknowns but only two equations. Hence, {u1 , u2 , u3 } is a linearly dependent set of vectors. 47. Given that {v1 , v2 , . . . , vm } is a linearly independent set of vectors in a vector space V , and uk = m , aik vi , k ∈ {1, 2, . . . , n}. (47.1) i=1 (a). Let ck ∈ R, k ∈ {1, 2, . . . , n}. Using system (47.1) and n , ck uk = 0, we obtain: k=1 n , k=1 ck m , aik vi i=1 = 0 ⇐⇒ n m , , i=1 aik ck vi = 0. k=1 Since the vi for each i ∈ {1, 2, . . . , m} are linearly independent, n , aik ck = 0, 1 ≤ i ≤ m. (47.2) k=1 But this is a system of m equations with n unknowns c1 , c2 , . . . , cn . Since n > m, the system has more unknowns than equations, and so has nontrivial solutions. Thus, {u1 , u2 , . . . , un } is a linearly dependent set. (b). If m = n, then the system (47.2) has a trivial solution ⇐⇒ the coefficient matrix of the system is invertible ⇐⇒ det[aij ] = 0. (c). If n < m, then the homogeneous system (47.2) has just the trivial solution if and only if rank(A) = n. Recall that for a homogeneous system, rank(A# ) = rank(A). (d). Corollary 4.5.17. 48. We assume that c1 (Av1 ) + c2 (Av2 ) + · · · + cn (Avn ) = 0. Our aim is to show that c1 = c2 = · · · = cn = 0. We manipulate the left side of the above equation as (c)2017 Pearson Education. Inc. 296 follows: c1 (Av1 ) + c2 (Av2 ) + · · · + cn (Avn ) = 0 A(c1 v1 ) + A(c2 v2 ) + · · · + A(cn vn ) = 0 A(c1 v1 + c2 v2 + · · · + cn vn ) = 0. Since A is invertible, we can left multiply the last equation by A−1 to obtain c1 v1 + c2 v2 + · · · + cn vn = 0. Since {v1 , v2 , . . . , vn } is linearly independent, we can now conclude directly that c1 = c2 = · · · = cn = 0, as required. 49. Assume that c1 v1 + c2 v2 + c3 v3 = 0. We must show that c1 = c2 = c3 = 0. Let us suppose for the moment that c3 = 0. In that case, we can solve the above equation for v3 : v3 = − c1 c2 v1 − v2 . c3 c3 However, this contradicts the assumption that v3 does not belong to span{v1 , v2 }. Therefore, we conclude that c3 = 0. Our starting equation therefore reduces to c1 v1 + c2 v2 = 0. Now the assumption that {v1 , v2 } is linearly independent shows that c1 = c2 = 0. Therefore, c1 = c2 = c3 = 0, as required. 50. Assume that c1 v1 + c2 v2 + · · · + ck vk + ck+1 vk+1 = 0. We must show that c1 = c2 = · · · = ck+1 = 0. Let us suppose for the moment that ck+1 = 0. In that case, we can solve the above equation for vk+1 : vk+1 = − c1 c2 ck v1 − v2 − · · · − vk . ck+1 ck+1 ck+1 However, this contradicts the assumption that vk+1 does not belong to span{v1 , v2 , . . . , vk }. Therefore, we conclude that ck+1 = 0. Our starting equation therefore reduces to c1 v1 + c2 v2 + · · · + ck vk = 0. Now the assumption that {v1 , v2 , . . . , vk } is linearly independent shows that c1 = c2 = · · · = ck = 0. Therefore, c1 = c2 = · · · = ck = ck+1 = 0, as required. 51. Let {v1 , v2 , . . . , vk } be a set of vectors with k ≥ 2. Suppose that vk can be expressed as a linear combination of {v1 , v2 , . . . , vk−1 }. We claim that span{v1 , v2 , . . . , vk } = span{v1 , v2 , . . . , vk−1 }. Since every vector belonging to span{v1 , v2 , . . . , vk−1 } evidently belongs to span{v1 , v2 , . . . , vk }, we focus on showing that every vector in span{v1 , v2 , . . . , vk } also belongs to span{v1 , v2 , . . . , vk−1 }: Let v ∈ span{v1 , v2 , . . . , vk }. We therefore may write v = c1 v1 + c2 v2 + · · · + ck vk . By assumption, we may write vk = d1 v1 + d2 v2 + · · · + dk−1 vk−1 . Therefore, we obtain v = c1 v1 + c2 v2 + · · · + ck vk = c1 v1 + c2 v2 + · · · + ck−1 vk−1 + ck (d1 v1 + d2 v2 + · · · + dk−1 vk−1 ) = (c1 + ck d1 )v1 + (c2 + ck d2 )v2 + · · · + (ck−1 + ck dk−1 )vk−1 ∈ span{v1 , v2 , . . . , vk−1 }. This shows that every vector belonging to span{v1 , v2 , . . . , vk } also belongs to span{v1 , v2 , . . . , vk−1 }, as needed. (c)2017 Pearson Education. Inc. 297 52. We first prove part 1 of Proposition 4.5.8. Suppose that we have a set {u, v} of two vectors in a vector space V . If {u, v} is linearly dependent, then we have cu + dv = 0, where c and d are not both zero. Without loss of generality, suppose that c = 0. Then we have d u = − v, c so that u and v are proportional. Conversely, if u and v are proportional, then v = cu for some constant c. Thus, cu − v = 0, which shows that {u, v} is linearly dependent. For part 2 of Proposition 4.5.8, suppose the zero vector 0 belongs to a set S of vectors in a vector space V . Then 1 · 0 is a linear dependency among the vectors in S, and therefore S is linearly dependent. 53. Suppose that {v1 , v2 , . . . , vk } spans V and let v be any vector in V . By assumption, we can write v = c 1 v 1 + c2 v 2 + · · · + ck vk , for some constants c1 , c2 , . . . , ck . Therefore, c1 v 1 + c2 v 2 + · · · + ck vk − v = 0 is a linear dependency among the vectors in {v, v1 , v2 , . . . , vk }. Thus, {v, v1 , v2 , . . . , vk } is linearly dependent. 54. Let S = {p1 , p2 , . . . , pk } and assume without loss of generality that the polynomials are listed in decreasing order by degree: deg(p1 ) > deg(p2 ) > · · · > deg(pk ). To show that S is linearly independent, assume that c1 p1 + c2 p2 + · · · + ck pk = 0. We wish to show that c1 = c2 = · · · = ck = 0. We require that each coefficient on the left side of the above equation is zero, since we have 0 on the right-hand side. Since p1 has the highest degree, none of the terms c2 p2 ,c3 p3 , . . . , ck pk can cancel the leading coefficient of p1 . Therefore, we conclude that c1 = 0. Thus, we now have c2 p2 + c3 p3 + · · · + ck pk = 0, and we can repeat this argument again now to show successively that c2 = c3 = · · · = ck = 0. Solutions to Section 4.6 (a): FALSE. It is not enough that S spans V . It must also be the case that S is linearly independent. (b): FALSE. For example, R2 is not a subspace of R3 , since R2 is not even a subset of R3 . (c): TRUE. Any set of two non-proportional vectors in R2 will form a basis for R2 . (d): FALSE. We have dim[Pn (R)] = n + 1 and dim[Rn ] = n. (e): FALSE. For example, if V = R2 , then the set S = {(1, 0), (2, 0), (3, 0)}, consisting of 3 > 2 vectors, fails to span V , a 2-dimensional vector space. (c)2017 Pearson Education. Inc. 298 (f ): TRUE. We have dim[P3 (R)] = 4, and so any set of more than four vectors in P3 (R) must be linearly dependent (a maximal linearly independent set in a 4-dimensional vector space consists of four vectors). (g): FALSE. For instance, the two vectors 1 + x and 2 + 2x in P3 (R) are linearly dependent. (h): FALSE. If the differential equation is not homogeneous, such as, for instance y + y = 1, then the zero function is not a solution, so the set of all solutions to the differential equation fails the zero vector check and does not form a vector space. (i): FALSE. Only linearly independent sets with fewer than n vectors can be extended to a basis for V . (j): TRUE. We can build such a subset by choosing vectors from the set as follows. Choose v1 to be any vector in the set. Now choose v2 in the set such that v2 ∈ span{v1 }. Next, choose v3 in the set such that v3 ∈ span{v1 , v2 }. Proceed in this manner until it is no longer possible to find a vector in the set that is not spanned by the collection of previously chosen vectors. This point will occur eventually, since V is finite-dimensional. Moreover, the chosen vectors will form a linearly independent set, since each vi is chosen from outside span{v1 , v2 , . . . , vi−1 }. Thus, the set we obtain in this way is a linearly independent set of vectors that spans V , hence forms a basis for V . (k): FALSE. The set of all 3 × 3 upper triangular matrices forms a 6-dimensional subspace of M3 (R), not a 3-dimensional subspace. One basis is given by {E11 , E12 , E13 , E22 , E23 , E33 }. Problems: 1. NO. Since dim[R2 ] = 2, any basis must contain exactly two vectors. Here, we are only given one vector, so this set cannot form a basis. 2 2 2. YES. dim[R ] = 2. There are two vectors, so if they are to form a basis for R , they need to be linearly 1 −1 = 2 = 0. This implies that the vectors are linearly independent, hence they form a independent: 1 1 basis for R2 . 3 3 3. YES. dim[R ] = 3. There are three vectors, so if they are to form a basis for R , they need to be linearly 1 3 1 1 = 13 = 0. This implies that the vectors are linearly independent, hence they independent: 2 −1 1 2 −1 3 form a basis for R . 3 4. NO. dim[R 3 ] = 3. There are three vectors, so if they are to form a basis for R , they need to be linearly 1 2 3 5 11 = 0. This implies that the vectors are linearly dependent, hence they do not independent: −1 1 −2 −5 form a basis for R3 . 5. NO. dim[R4 ] = 4. We need 4 linearly independent vectors in order to span R4 . However, there are only 3 vectors in this set. Thus, the vectors cannot be a basis for R4 . 4 6. YES. dim[R vectors, so if they are to form a basis for R3 , they need to be linearly ] = 4. There are four 1 2 −1 2 2 −1 2 −7 0 −5 1 −1 2 −3 1 1 1 −1 0 −1 2 −3 1 = −11. 1 = 3 1 independent: = = 3 1 3 1 1 3 1 1 −5 0 −2 0 −5 0 −2 0 2 −1 −2 0 −5 2 0 −2 Since this determinant is nonzero, the given vectors are linearly independent. Consequently, they form a basis for R4 . (c)2017 Pearson Education. Inc. 299 7. NO. Since dim[R3 ] = 3, any basis must contain exactly three vectors. Here, we are given four vectors, so this set cannot form a basis. 8. dim[R4 ] = 4. There are four vectors, so if they are to form a basis for R3 , they need to be linearly 0 1 0 k −1 1 0 0 0 k −1 0 1 0 = − 0 1 2 − −1 1 0 = −(−1 + 2k) − (−k 2 ) = k 2 − 2k + 1 = independent: 0 1 1 2 k 0 1 k 0 1 k 0 0 1 (k − 1)2 = 0 when k = 1. Thus, the vectors will form a basis for R4 provided k = 1. 9. NO. Since dim[P1 (R)] = 2, any basis must contain exactly two vectors. Here, we are given three vectors, so this set cannot form a basis. 10. YES. We will check whether or not S is linearly independent. Suppose c1 (1 − 3x2 ) + c2 (2x + 5x2 ) + c3 (1 − x + 3x2 ) = 0. Equating like powers of x on each side, we have c1 + c3 = 0, 2c2 − c3 = 0, −3c1 + 5c2 + 3c3 = 0. It is easy to verify that this linear system has only the trivial solution (the determinant of the coefficient 1 0 1 matrix, 0 2 −1 = 17 = 0, for instance, is nonzero). That is, c1 = c2 = c3 = 0. Thus, S is linearly −3 5 3 independent. By Corollary 4.6.13, we conclude that S is a basis. 11. YES. We will check whether or not S is linearly independent. Suppose c1 (5x2 )+c2 (1+6x)+c3 (−3−x2 ) = 0. Equating like powers of x on each side, we have c2 − 3c3 = 0, 6c2 = 0, 5c1 − c3 = 0. It is easy to verify that this linear system has only the trivial solution (the second equation implies c2 = 0, from which the first implies that c3 = 0, and then the third implies that c1 = 0). That is, c1 = c2 = c3 = 0. Thus, S is linearly independent. By Corollary 4.6.13, we conclude that S is a basis. 12. NO. Since dim[P2 (R)] = 3, any basis must contain exactly three vectors. Here, we are given four vectors, so this set cannot form a basis. 13. NO. Since dim[P3 (R)] = 4, any basis must contain exactly four vectors. Here, we are given three vectors, so this set cannot form a basis. 14. YES. We will check whether or not S is linearly independent. Suppose c1 (1 + x + 2x3 ) + c2 (2 + x + 3x2 − x3 ) + c3 (−1 + x + x2 − 2x3 ) + c4 (2 − x + x2 + 2x3 ) = 0. Equating like powers of x, we obtain th e system c1 + 2c2 − c3 + 2c4 = 0, c1 + c2 + c3 − c4 = 0, 3c2 + c3 + c4 = 0, 2c1 − c2 − 2c3 + 2c4 = 0. ⎡ ⎤ 1 2 −1 2 0 ⎢ 1 1 1 −1 0 ⎥ ⎥. One can check, by using Gaussian The augmented matrix for this linear system is ⎢ ⎣ 0 3 1 1 0 ⎦ 2 −1 −2 2 0 Elimination or determinants, that the coefficient matrix for this linear system is invertible. Therefore, we conclude that c1 = c2 = c3 = c4 = 0. Thus, S is linearly independent. By Corollary 4.6.13, we conclude that S is a basis. (c)2017 Pearson Education. Inc. 300 15. NO. Since dim[M2 (R)] = 4, any basis must contain exactly four vectors. Here, we are given five vectors, so this set cannot form a basis. 16. YES. We will check whether or not S is linearly independent. Suppose −2 −8 0 1 −5 0 3 −2 0 c1 + c2 + c3 + c4 = 1 4 −1 1 5 −4 4 −1 0 0 0 . The corresponding linear system of equations is −2c1 − 5c3 + 3c4 = 0, −8c1 + c2 − 2c4 = 0, c1 − c2 + 5c3 + 4c4 = 0, 4c1 + c2 − 4c3 − c4 = 0. ⎤ ⎡ −2 0 −5 3 ⎢ −8 1 0 −2 ⎥ ⎥. Using any method to compute the The matrix of coefficients for this linear system is ⎢ ⎣ 1 −1 5 4 ⎦ 4 1 −4 −1 determinant of this coefficient matrix shows that it is nonzero, and hence the coefficient matrix is invertible. Thus, c1 = c2 = c3 = c4 = 0. Thus, S is linearly independent. By Corollary 4.6.13, we conclude that S is a basis. 17. NO. Since dim[M32 (R)] = 6, any basis must contain exactly six vectors. Here, we are given five vectors, so this set cannot form a basis. 1 1 1 2 2 2 18. NO. By inspection, note that S contains two proportional vectors, and . 1 1 1 2 2 2 Therefore, S is linearly dependent, and hence, S cannot form a basis. 19. A row-echelon form of A is given by 1 −9/8 3/8 3/8 −5/8 , which has four unpivoted columns. Therefore, the linear system Ax = 0 has four free variables. We conclude that the dimension of the null space of A is 4. 1 3 x1 0 1 3 0 20. Ax = 0 =⇒ ∼ = . The augmented matrix for this system is −2 −6 −2 −6 0 x2 0 1 3 0 ; thus, x1 + 3x2 = 0, or x1 = −3x2 . Let x2 = r so that x1 = −3r where r ∈ R. Consequently, 0 0 0 S = {x ∈ R2 : x = r(−3, 1), r ∈ R} = span{(−3, 1)}. It follows that {(−3, 1)} is a basis for S and dim[S] = 1. ⎤ ⎤ ⎡ ⎤⎡ ⎡ 0 0 0 x1 0 21. Ax = 0 =⇒ ⎣ 0 0 0 ⎦ ⎣ x2 ⎦ = ⎣ 0 ⎦. The REDUCED ROW ECHELON FORM of the augx 0 1 0 0 ⎤ ⎡3 0 1 0 0 mented matrix for this system is ⎣ 0 0 0 0 ⎦. We see that x2 = 0, and x1 and x3 are free variables: 0 0 0 0 x1 = r and x2 = s. Hence, (x1 , x2 , x3 ) = (r, 0, s) = r(1, 0, 0) + s(0, 0, 1), so that the solution set of the system is S = {x ∈ R3 : x = r(1, 0, 0) + s(0, 0, 1), r, s ∈ R}. Therefore we see that {(1, 0, 0), (0, 0, 1)} is a basis for S and dim[S] = 2. ⎡ ⎤⎡ ⎤ ⎤ ⎡ 1 −1 4 x1 0 3 −2 ⎦ ⎣ x2 ⎦ = ⎣ 0 ⎦. The REDUCED ROW ECHELON FORM of the 22. Ax = 0 =⇒ ⎣ 2 1 2 −2 0 x ⎤ ⎡ 3 1 0 2 0 augmented matrix for this system is ⎣ 0 1 −2 0 ⎦. If we let x3 = r then (x1 , x2 , x3 ) = (−2r, 2r, r) = 0 0 0 0 (c)2017 Pearson Education. Inc. 301 r(−2, 2, 1), so that the solution set of the system is S = {x ∈ R3 : x = r(−2, 2, 1), r ∈ R}. Therefore we see that {(−2, 2, 1)} is a basis for S and dim[S] = 1. ⎤ ⎤⎡ ⎡ ⎤ ⎡ x1 1 −1 2 3 0 ⎥ ⎢ 2 −1 3 4 ⎥ ⎢ x2 ⎥ ⎢ ⎥⎢ ⎥ = ⎢ 0 ⎥. The REDUCED ROW ECHELON FORM of the 23. Ax = 0 =⇒ ⎢ ⎣ 1 ⎣ 0 ⎦ 0 1 1 ⎦ ⎣ x3 ⎦ 0 3 −1 4 5 ⎡ x4 ⎤ 1 0 1 1 0 ⎢ 0 1 −1 −2 0 ⎥ ⎥. If we let x3 = r, x4 = s then x2 = r + 2s and augmented matrix for this system is ⎢ ⎣ 0 0 0 0 0 ⎦ 0 0 0 0 0 x1 = −r − s. Hence, the solution set of the system is S = {x ∈ R4 : x = r(−1, 1, 1, 0) + s(−1, 2, 0, 1), r, s ∈ R} = span{(−1, 1, 1, 0), (−1, 2, 0, 1)}. Further, the vectors v1 = (−1, 1, 1, 0), v2 = (−1, 2, 0, 1) are linearly independent since c1 v1 + c2 v2 = 0 =⇒ c1 (−1, 1, 1, 0) + c2 (−1, 2, 0, 1) = (0, 0, 0, 0) =⇒ c1 = c2 = 0. Consequently, {(−1, 1, 1, 0), (−1, 2, 0, 1)} is a basis for S and dim[S] = 2. 24. If we let y = r and z = s where r, s ∈ R, then x = 3r − s. It follows that any ordered triple in S can be written in the form: (x, y, z) = (3r − s, r, s) = (3r, r, 0) + (−s, 0, s) = r(3, 1, 0) + s(−1, 0, 1), where r, s ∈ R. If we let v1 = (3, 1, 0) and v2 = (−1, 0, 1), then S = {v ∈ R3 : v = r(3, 1, 0) + s(−1, 0, 1), r, s ∈ R} = span{v1 , v2 }; moreover, v1 and v2 are linearly independent for if a, b ∈ R and av1 + bv2 = 0, then a(3, 1, 0) + b(−1, 0, 1) = (0, 0, 0), which implies that (3a, a, 0) + (−b, 0, b) = (0, 0, 0), or (3a − b, a, b) = (0, 0, 0). In other words, a = b = 0. Since {v1 , v2 } spans S and is linearly independent, it is a basis for S. Also, dim[S] = 2. 25. S = {x ∈ R3 : x = (r, r − 2s, 3s − 5r), r, s ∈ R} = {x ∈ R3 : x = (r, r, −5r) + (0, −2s, 3s), r, s ∈ R} = {x ∈ R3 : x = r(1, 1, −5) + s(0, −2, 3), r, s ∈ R}. Thus, S = span{(1, 1, −5), (0, −2, 3)}. The vectors v1 = (1, 1, −5) and v2 = (0, −2, 3) are linearly independent for if a, b ∈ R and av1 + bv2 = 0, then a(1, 1, −5)+b(0, −2, 3) = (0, 0, 0) =⇒ (a, a, −5a)+(0, −2b, 3b) = (0, 0, 0) =⇒ (a, a−2b, 3b−5a) = (0, 0, 0) =⇒ a = b = 0. It follows that {v1 , v2 } is a basis for S and dim[S] = 2. 26. We claim that S = {1, x, x2 , x3 } is a basis for P3 (R). We will verify that S is linearly independent and that S spans P3 (R). S is linearly independent: Assume c0 (1) + c1 (x) + c2 (x2 ) + c3 (x3 ) = 0. It follows immediately from this equation that c0 = c1 = c2 = c3 = 0. Therefore, S is linearly independent. S spans P3 (R): Consider an arbitrary polynomial p(x) = a + bx + cx2 + dx3 ∈ P3 (R). We can write p(x) as a linear combination of vectors in S in the obvious way: p(x) = a(1) + b(x) + c(x2 ) + d(x3 ). Therefore, S spans P3 (R). Since S is linearly independent and spans P3 (R), we conclude that S forms a basis for P3 (R). Since S contains four vectors, we conclude that dim[P3 (R)] = 4. 27. Many acceptable bases are possible here. One example is S = {x3 , x3 + 1, x3 + x, x3 + x2 }. (c)2017 Pearson Education. Inc. 302 All of the polynomials in this set have degree 3. We verify that S is a basis: Assume that c1 (x3 ) + c2 (x3 + 1) + c3 (x3 + x) + c4 (x3 + x2 ) = 0. Thus, (c1 + c2 )x3 + c4 x2 + c3 x + c2 = 0, from which we quickly see that c1 = c2 = c3 = c4 = 0. Thus, S is linearly independent. Since P3 (R) is 4-dimensional, we can now conclude from Corollary 4.6.13 that S is a basis for P3 (R). a b 28. S = {A ∈ M2 (R) : A = , a, b, c ∈ R}. Each vector in S can be written as 0 c 1 0 0 1 0 0 A=a +b +c , 0 0 0 0 0 1 1 0 0 1 0 0 so that S = span , , . Since the vectors in this set are clearly linearly inde0 0 0 0 0 1 1 0 0 1 0 0 pendent, it follows that a basis for S is , , , and therefore dim[S] = 3. 0 0 0 0 0 1 a b 29. S = {A ∈ M2 (R) : A = , a, b, c ∈ R}. Each vector in S can be written as c −a 1 0 0 1 0 0 A=a +b +c , 0 −1 0 0 1 0 1 0 0 1 0 0 so that S = span , , . Since this set of vectors is clearly linearly independent, 0 −1 0 0 1 0 1 0 0 1 0 0 it follows that a basis for S is , , , and therefore dim[S] = 3. 0 −1 0 0 1 0 30. We see directly that v3 = 2v1 . Let v be an arbitrary vector in S. Then v = c1 v1 + c2 v2 + c3 v3 = (c1 + 2c3 )v1 + c2 v2 = d1 v1 + d2 v2 , where d1 = (c1 + 2c3 ) and d2 = c2 . Hence S = span{v1 , v2 }. Further, v1 and v2 are linearly independent for if a, b ∈ R and av1 + bv2 = 0, then a(1, 0, 1) + b(0, 1, 1) = (0, 0, 0) =⇒ (a, 0, a) + (0, b, b) = (0, 0, 0) =⇒ (a, b, a + b) = (0, 0, 0) =⇒ a = b = 0. Consequently, {v1 , v2 } is a basis for S and dim[S] = 2. f1 (x) − f2 (x) ex − e−x 31. f3 depends on f1 and f2 since sinh x = . Thus, f3 (x) = . 2 2 x e e−x = −2 = 0 for all x ∈ R, so {f1 , f2 } is a linearly independent set. Thus, {f1 , f2 } W [f1 , f2 ](x) = x e −e−x is a basis for S and dim[S] = 2. 32. The given set of matrices is linearly dependent because it contains the zero vector. Consequently, the 1 3 −1 4 5 −6 , A2 = , A3 = span the same subspace of M2 (R) as that matrices A1 = −1 2 1 1 −5 1 spanned by the original set. We now determine whether {A1 , A2 , A3 } is linearly independent. The vector equation: c1 A1 + c2 A2 + c3 A3 = 02 (c)2017 Pearson Education. Inc. 303 leads to the linear system c1 − c2 + 5c3 = 0, 3c1 + 4c2 − 6c3 = 0, −c1 + c2 − 5c3 = 0, 2c1 + c2 + c3 = 0. This system has solution (−2r, 3r, r), where r is a free variable. Consequently, {A1 , A2 , A3 } is linearly dependent with linear dependency relation −2A1 + 3A2 + A3 = 0, or equivalently, A3 = 2A1 − 3A2 . It follows that the set of matrices {A1 , A2 } spans the same subspace of M2 (R) as that spanned by the original set of matrices. Further {A1 , A2 } is linearly independent by inspection, and therefore it is a basis for the subspace. 33. (a). We must show that every vector (x, y) ∈ R2 can be expressed as a linear combination of v1 and v2 . Mathematically, we express this as c1 (1, 1) + c2 (−1, 1) = (x, y) which implies that c1 − c2 = x and c1 + c2 = y. Adding the equations here, we obtain 2c1 = x + y or c1 = 1 (x + y). 2 Now we can solve for c2 : 1 (y − x). 2 Therefore, since we were able to solve for c1 and c2 in terms of x and y, we see that the system of equations is consistent for all x and y. Therefore, {v1 , v2 } spans R2 . 1 −1 = 2 = 0, so the vectors are linearly independent. (b). 1 1 c2 = y − c1 = (c). We can draw this conclusion from part (a) alone by using Theorem 4.6.12 or from part (b) alone by using Theorem 4.6.10. 34. (a). We must show that every vector (x, y) ∈ R2 can be expressed as a linear combination of v1 and v2 . Mathematically, we express this as c1 (2, 1) + c2 (3, 1) = (x, y) which implies that 2c1 + 3c2 = x and c1 + c2 = y. From this, we can solve for c1 and c2 in terms of x and y: c1 = 3y − x and c2 = x − 2y. Therefore, since we were able to solve for c1 and c2 in terms of x and y, we see that the system of equations is consistent for all x and y. Therefore, {v1 , v2 } spans R2 . (c)2017 Pearson Education. Inc. 304 2 (b). 1 3 = −1 = 0, so the vectors are linearly independent. 1 (c). We can draw this conclusion from part (a) alone by using Theorem 4.6.12 or from part (b) alone by using Theorem 4.6.10. 35. dim[R3 ] = 3. There so if they are linearly independent, then they are a basis of R3 by are 3 vectors, 0 3 6 Theorem 4.6.10. Since 6 0 −3 = 81 = 0, the given vectors are linearly independent, and so {v1 , v2 , v3 } 3 3 0 3 is a basis for R . 36. dim[P2 (R)] = 3. There are 3 vectors, so {p1 , p2 , p3 } may be a basis for P2 (R) depending on α. To be a basis, the set of vectors must be linearly independent. Let a, b, c ∈ R. Then ap1 (x) + bp2 (x) + cp3 (x) = 0 =⇒ a(1 + αx2 ) + b(1 + x + x2 ) + c(2 + x) = 0 =⇒ (a +⎧b + 2c) + (b + c)x + (aα + b)x2 = 0. Equating like coefficients in the last equality, we obtain the ⎪ ⎨ a + b + 2c = 0 b + c = 0 Reduce the augmented matrix of this system. system: ⎪ ⎩ aα + b = 0. ⎡ 1 ⎣ 0 α ⎤ ⎡ 1 2 0 1 1 1 0 ⎦∼⎣ 0 1 0 0 0 1 1 α−1 2 1 2α ⎤ ⎡ 0 1 1 2 1 0 ⎦∼⎣ 0 1 0 0 α+1 0 ⎤ 0 0 ⎦. 0 For this system to have only the trivial solution, the last row of the matrix must not be a zero row. This means that α = −1. Therefore, for the given set of vectors to be linearly independent (and thus a basis), α can be any value except −1. 37. dim[P2 (R)] = 3. There are 3 vectors, so in order to form a basis for P2 (R), they must be linearly independent. Since we are dealing with functions, we will use the Wronskian. 1+x W [p1 , p2 , p3 ](x) = 1 0 x(x − 1) 2x − 1 2 1 + 2x2 1 + x = 1 4x 0 4 x(x − 1) 2x − 1 2 1 + 2x = −2. 2 0 Since the Wronskian is nonzero, {p1 , p2 , p3 } is linearly independent, and hence forms a basis for P2 (R). 1 x 3x2 −1 2 38. W [p0 , p1 , p2 ](x) = 0 1 3x = 3 = 0, so {p0 , p1 , p2 } is a linearly independent set. Since 0 0 3 dim[P2 (R)] = 3, it follows that {p0 , p1 , p2 } is a basis for P2 (R). 39. (a). Suppose that a, b,c, d ∈ R. −1 1 1 3 1 0 0 −1 a +b +c +d = 02 0 1 −1 0 1 2 2 3 =⇒ −a + b + c −b + c + 2d ⎧ −a + b + c = 0 ⎪ ⎪ ⎪ ⎨ a + 3b − d = 0 a + 3b − c = 02 . The last matrix results in the system: The a + 2c + 3d ⎪ −b + c + 2d = 0 ⎪ ⎪ ⎩ a + 2c + 3d = 0. (c)2017 Pearson Education. Inc. 305 ⎡ ⎤ 1 −1 −1 0 0 ⎢ 0 4 1 −1 0 ⎥ ⎥, REDUCED ROW ECHELON FORM of the augmented matrix of this system is ⎢ ⎣ 0 0 5 7 0 ⎦ 0 0 0 1 0 which implies that a = b = c = d = 0. Hence, the given set of vectors is linearly independent. Since dim[M2 (R)] = 4, it follows that {A1 , A2 , A3 , A4 } is a basis for M2 (R). 5 6 (b). We wish to express the vector in the form 7 8 5 7 6 8 =a −1 0 1 1 +b 1 −1 3 0 +c 1 1 0 2 +d 0 −1 2 3 . Matching entries on each side of this equation (upper left, upper right, lower left, and lower right), we obtain a linear system with augmented matrix ⎡ −1 1 1 0 ⎢ 1 3 0 −1 ⎢ ⎣ 0 −1 1 2 1 0 2 3 ⎤ 5 6 ⎥ ⎥. 7 ⎦ 8 55 56 Solving this system by Gaussian elimination, we find that a = − 34 3 , b = 12, c = − 3 , and d = 3 . Thus, we have 34 −1 1 55 1 0 56 0 −1 5 6 1 3 − =− + 12 + . 7 8 0 1 1 2 2 3 −1 0 3 3 3 40. ⎡ ⎤ ⎤ x ⎤ ⎡ 1 1 −1 1 ⎢ 1 ⎥ 0 x2 ⎥ ⎣ ⎦ 5 −6 ⎦ ⎢ (a). Ax = 0 =⇒ ⎣ 2 −3 ⎣ x3 ⎦ = 0 . The augmented matrix for this linear system is 5 0 2 −3 0 x4 ⎡ ⎡ ⎤ ⎡ 1 1 −1 1 0 1 1 ⎣ 2 −3 5 −6 0 ⎦ ∼ ⎣ 0 5 0 2 −3 0 0 ⎤ ⎡ ⎤ ⎡ ⎤ 1 −1 1 0 1 1 −1 1 0 1 1 −1 1 0 2 3 −5 7 −8 0 ⎦ ∼ ⎣ 0 −5 7 −8 0 ⎦ ∼ ⎣ 0 1 −1.4 1.6 0 ⎦ . −5 7 −8 0 0 0 0 0 0 0 0 0 0 0 1. A12 (−2), A13 (−5) 2. A23 (−1) 3. M2 (− 15 ) We see that x3 = r and x4 = s are free variables, and therefore nullspace(A) is 2-dimensional. Now we can check directly that Av1 = 0 and Av2 = 0, and since v1 and v2 are linearly independent (they are non-proportional), they therefore form a basis for nullspace(A) by Corollary 4.6.13. (b). An arbitrary vector (x, y, z, w) in nullspace(A) can be expressed as a linear combination of the basis vectors: (x, y, z, w) = c1 (−2, 7, 5, 0) + c2 (3, −8, 0, 5), where c1 , c2 ∈ R. 41. (c)2017 Pearson Education. Inc. 306 ⎡ a c (a). An arbitrary matrix in S takes the form ⎣ −a − c ⎡ 1 = a⎣ 0 −1 ⎤ ⎡ 0 −1 0 0 0 ⎦ + b⎣ 0 0 1 0 ⎤ −a − b ⎦ −c − d a+b+c+d b d −b − d ⎤ ⎡ 1 −1 0 0 0 ⎦ + c⎣ 1 −1 1 −1 Therefore, we have the following basis for S: ⎧⎡ ⎤ ⎡ ⎤ ⎡ 0 1 −1 0 1 0 −1 ⎨ ⎣ 0 0 0 0 ⎦,⎣ 1 0 ⎦,⎣ 0 ⎩ 0 −1 1 −1 −1 0 1 ⎤ ⎡ ⎤ 0 0 0 0 0 0 −1 ⎦ + d ⎣ 0 1 −1 ⎦ . 0 1 0 −1 1 ⎤ ⎡ 0 0 0 0 −1 ⎦ , ⎣ 0 0 1 0 ⎤⎫ 0 0 ⎬ 1 −1 ⎦ . ⎭ −1 1 From this, we see that dim[S] = 4. (b). We see that each of the matrices ⎡ ⎤ ⎡ 1 0 0 0 1 ⎣ 0 0 0 ⎦,⎣ 0 0 0 0 0 0 0 ⎤ ⎡ 0 0 0 ⎦,⎣ 0 0 0 0 0 0 ⎤ ⎡ 1 0 0 ⎦,⎣ 1 0 0 0 0 0 ⎤ ⎡ 0 0 0 ⎦,⎣ 0 0 1 0 0 0 ⎤ 0 0 ⎦ 0 has a different row or column that does not sum to zero, and thus none of these matrices belong to S, and they are linearly independent from one another. Therefore, supplementing the basis for S in part (a) with the five matrices here extends the basis in (a) to a basis for M3 (R). 42. (a). A typical element in S has the form ⎡ a d A=⎣ −(a + d) ⎡ 1 = a⎣ 0 −1 0 0 0 ⎤ ⎡ ⎤ ⎡ 0 0 1 0 0 0 ⎦+b ⎣ 0 0 0 ⎦+c ⎣ 0 0 0 −1 0 0 Therefore, a basis for S is ⎧⎡ ⎤ ⎡ ⎤ ⎡ 1 0 0 0 1 0 0 ⎨ ⎣ 0 0 0 ⎦,⎣ 0 0 0 ⎦,⎣ 0 ⎩ −1 0 0 0 −1 0 0 b e −(b + e) ⎤ c ⎦ f −(c + f ) ⎤ ⎡ 0 1 0 0 0 ⎦+d ⎣ 1 0 −1 −1 ⎤ ⎡ 0 1 0 0 0 ⎦,⎣ 1 0 −1 −1 0 0 0 0 0 0 ⎤ ⎡ ⎤ ⎡ 0 0 0 0 0 0 ⎦+e ⎣ 0 1 0 ⎦+f ⎣ 0 0 0 −1 0 0 ⎤ ⎡ ⎤ ⎡ 0 0 0 0 0 0 ⎦,⎣ 0 1 0 ⎦,⎣ 0 0 0 −1 0 0 ⎤ 0 0 0 1 ⎦. 0 −1 ⎤⎫ 0 0 ⎬ 0 1 ⎦ . ⎭ 0 −1 Thus, dim[S] = 6. (b). To extend the basis for S in part (a) to a basis for V , note that dim[V ] = 9, so that three additional matrices must be added. We can add one vector for which the first column fails to sum to zero, one vector for which the second column fails to sum to zero, and ⎡ one vector for the third ⎤ which ⎡ ⎤ column ⎡ fails to ⎤sum to 1 0 0 0 1 0 0 0 1 zero. The easiest way to do this is to add the vectors ⎣ 0 0 0 ⎦, ⎣ 0 0 0 ⎦, and ⎣ 0 0 0 ⎦. It is 0 0 0 0 0 0 0 0 0 possible to express each of the matrices Eij (with 1 ≤ i, j ≤ 3) by using these three matrices together with (c)2017 Pearson Education. Inc. 307 the six members of the basis for S found in part (a). Therefore, we see that the nine matrices formed span V. a b ∈ Sym2 (R); a, b, c ∈ R. This vector can be represented as 43. Let A ∈ Sym2 (R) so A = AT . A = b c 1 0 0 1 0 0 a linear combination of , , : 1 0 0 1 0 0 1 0 0 1 0 0 1 0 0 1 0 0 A=a +b +c . Since , , is a linearly independent 0 0 1 0 0 1 0 0 1 0 0 1 a basis forSym2 (R). Thus, dim[Sym Let B ∈ Skew2 (R) so B = set that also spansSym2 (R), it is 2 (R)] = 3. 0 b 0 1 0 1 =b , where b ∈ R. The set is linearly independent −B T . Then B = −b 0 −1 0 −1 0 0 1 . Consequently, dim[Skew2 (R)] = 1. Now, and spans Skew2 (R) so that a basis for Skew2 (R) is −1 0 dim[Sym2 (R)] = 4, and hence dim[Sym2 (R)] + dim[Skew2 (R)] = 3 + 1 = 4 = dim[M2 (R)]. 44. We know that dim[Mn (R)] = n2 . Let S ∈ Symn (R) and let [Sij ] be the matrix with ones in the (i, j) and (j, i) positions and zeroes elsewhere. Then the general n × n symmetric matrix can be expressed as: S = a11 S11 + a12 S12 +a13 S13 + · · · + a1n S1n + a22 S22 +a23 S23 + · · · + a2n S2n + ··· + an−1 n−1 Sn−1 n−1 + an−1 n Sn−1 n + ann Snn . We see that S has been resolved into a linear combination of n(n + 1) linearly independent matrices, which therefore form a basis for n + (n − 1) + (n − 2) + · · · + 1 = 2 n(n + 1) . Symn (R); hence dim[Sym2 (R)] = 2 Now let T ∈ Skewn (R) and let [Tij ] be the matrix with one in the (i, j)-position, negative one in the (j, i)position, and zeroes elsewhere, including the main diagonal. Then the general n × n skew-symmetric matrix can be expressed as: T = a12 T12 + a13 T13 +a14 T14 + · · · + a1n T1n + a23 T23 +a24 T24 + · · · + a2n T2n + ··· + an−1 n Tn−1 n (n − 1)n We see that T has been resolved into a linear combination of (n−1)+(n−2)+(n−3)+· · ·+2+1 = 2 (n − 1)n . linearly independent vectors, which therefore form a basis for Skewn (R); hence dim[Skewn (R)] = 2 Consequently, using these results, we have n(n + 1) (n − 1)n + = n2 = dim[Mn (R)]. dim[Symn (R)] + dim[Skew2 (R)] = 2 2 45. (a). S is a two-dimensional subspace of R3 . Consequently, any two linearly independent vectors lying in this subspace determine a basis for S. By inspection we see, for example, that v1 = (4, −1, 0) and v2 = (3, 0, 1) both lie in the plane. Further, since they are not proportional, these vectors are linearly independent. Consequently, a basis for S is {(4, −1, 0), (3, 0, 1)}. (b). To extend the basis obtained in Part (a) to obtain a basis for R3 , we require one more vector that does not lie in S. For example, since v3 = (1, 0, 0) does not lie on the plane it is an appropriate vector. Consequently, a basis for R3 is {(4, −1, 0), (3, 0, 1), (1, 0, 0)}. (c)2017 Pearson Education. Inc. 308 46. Each vector in S can be written as a b 1 =a b a 0 0 1 +b 0 1 1 0 . 1 0 0 1 Consequently, a basis for S is given by the linearly independent set , . To extend this 0 1 1 0 1 0 0 1 basis to M2 (R), we can choose, for example, the two vectors and . Then the linearly 0 0 0 0 1 0 0 1 1 0 0 1 independent set , , , is a basis for M2 (R). 0 1 1 0 0 0 0 0 47. The vectors in S can be expressed as (2a1 + a2 )x2 + (a1 + a2 )x + (3a1 − a2 ) = a1 (2x2 + x + 3) + a2 (x2 + x − 1), and since {2x2 + x + 3, x2 + x − 1} is linearly independent (these polynomials are non-proportional) and certainly span S, they form a basis for S. To extend this basis to V = P2 (R), we must include one additional vector (since P2 (R) is 3-dimensional). Any polynomial that is not in S will suffice. For example, x ∈ S, since x cannot be expressed in the form (2a1 + a2 )x2 + (a1 + a2 )x + (3a1 − a2 ), since the equations 2a1 + a2 = 0, a1 + a2 = 1, 3a1 − a2 = 0 are inconsistent. Thus, the extension we use as a basis for V here is {2x2 + x + 3, x2 + x − 1, x}. Many other correct answers can also be given here. 48. Let y(x) = erx , so that y (x) = rerx and y (x) = r2 erx . Substituting into the given differential equation, we obtain y + 2y − 3y = r2 erx + 2rerx − 3erx = erx (r2 + 2r − 3). From erx (r2 + 2r − 3) = 0, we conclude that r2 + 2r − 3 = 0. Factoring, we obtain (r + 3)(r − 1) = 0, so that r = −3 and r = 1. Thus, y1 (x) = e−3x and y2 (x) = ex are two non-proportional solutions. Since the solution space is a two-dimensional vector space, we conclude that a basis is given by {e−3x , ex }. 49. Let y(x) = erx , so that y (x) = rerx and y (x) = r2 erx . Substituting into the given differential equation, we obtain y + 6y = r2 erx + 6rerx = erx (r2 + 6r). From erx (r2 + 6r) = 0, we conclude that r2 + 6r = 0. Factoring, we obtain r(r + 6) = 0, so that r = 0 and r = −6. Thus, y1 (x) = 1 and y2 (x) = e−6x are two non-proportional solutions. Since the solution space is a two-dimensional vector space, we conclude that a basis is given by {1, e−6x }. 50. Let y(x) = erx , so that y (x) = rerx and y (x) = r2 erx . Substituting into the given differential equation, we obtain y − 2y = r2 erx − 2erx = erx (r2 − 2). From erx (r2 − 2) = 0, (c)2017 Pearson Education. Inc. 309 √ √ √ √ we conclude that r2 −2 = 0. Factoring, we obtain (r+ 2)(r− 2) = 0, so that r = ± 2. Thus, y1 (x) = e 2x √ and y2 (x) = e− 2x are two non-proportional solutions. Since the solution space is a two-dimensional vector √ √ 2x − 2x space, we conclude that a basis is given by {e ,e }. 51. A general vector in S has the form y(x) = c(sin 4x + 5 cos 4x), where c is a real constant. Since the solution space is a two-dimensional vector space, we need to add one additional solution to the differential equation that is not proportional to y1 (x) = sin 4x + 5 cos 4x. There are many valid choices. One of the easiest is y2 (x) = sin 4x. Therefore, our extended basis is given by {sin 4x + 5 cos 4x, sin 4x}. 52. Since S is a basis for Pn−1 (R), S contains n vectors. Therefore, S ∪{xn } is a set of n+1 vectors, which is precisely the dimension of Pn (R). Moreover, xn does not lie in Pn−1 (R) = span(S), and therefore, S ∪ {xn } is linearly independent by Problem 50 in Section 4.5. By Corollary 4.6.13, we conclude that S ∪ {xn } is a basis for Pn (R). 53. Since S is a basis for Pn−1 (R), S contains n vectors. Therefore, S ∪ {p} is a set of n + 1 vectors, which is precisely the dimension of Pn (R). Moreover, p does not lie in Pn−1 (R) = span(S), and therefore, S ∪ {p} is linearly independent by Problem 50 in Section 4.5. By Corollary 4.6.13, we conclude that S ∪ {p} is a basis for Pn (R). 54. (a). Let ek denote the kth standard basis vector. Then a basis for Cn with scalars in R is given by {e1 , e2 , . . . , en , ie1 , ie2 , . . . , ien }, and the dimension is 2n. (b). Using the notation in part (a), a basis for Cn with scalars in C is given by {e1 , e2 , . . . , en }, and the dimension is n. Solutions to Section 4.7 True-False Review: (a): TRUE. This is the content of Theorem 4.7.1. The existence of such a linear combination comes from the fact that a basis for V must span V , and the uniqueness of such a linear combination follows from the linear independence of the vectors comprising a basis. (b): TRUE. This follows from the Equation [v]B = PB←C [v]C , which is just Equation (4.7.6) with the roles of B and C reversed. (c): TRUE. The number of columns in the change-of-basis matrix PC←B is the number of vectors in B, while the number of rows of PC←B is the number of vectors in C. Since all bases for the vector space V contain the same number of vectors, this implies that PC←B contains the same number of rows and columns. (d): TRUE. If V is an n-dimensional vector space, then PC←B PB←C = In = PB←C PC←B , which implies that PC←B is invertible. (e): TRUE. This follows from the linearity properties: [v − w]B = [v + (−1)w]B = [v]B + [(−1)w]B = [v]B + (−1)[w]B = [v]B − [w]B . (c)2017 Pearson Education. Inc. 310 (f ): FALSE. It depends on the order in which the vectors in the bases B and C are listed. For instance, if 2 we consider the bases B = {(1, 0), (0, 1)} and C = {(0, 1), (1, 0)} for R ,then although B and C contain the 1 0 while [v]C = . same vectors, if we let v = (1, 0), then [v]B = 0 1 2 (g): FALSE. For instance, if we consider the bases B = {(1, 0),(0, 1)} and C = {(0, 1), (1, 0)} for R , and 1 = [w]C . if we let v = (1, 0) and w = (0, 1), then v = w, but [v]B = 0 (h): TRUE. If B = {v1 , v2 , . . . , vn }, then the column vector [vi ]B is the ith standard basis vector (1 in the ith position and zeroes elsewhere). Thus, for each i, the ith column of PB←B consists of a 1 in the ith position and zeroes elsewhere. This describes precisely the identity matrix. Problems: 1. Write (27, 6) = c1 (7, −1) + c2 (−9, −2). Then 7c1 − 9c2 = 27 and − c1 − 2c2 = 6. Solving this system of equations gives c1 = 0 and c2 = −3. Thus, 0 [v]B = . −3 2. Write (5, −10) = c1 (2, −2) + c2 (1, 4). Then 2c1 + c2 = 5 and − 2c1 + 4c2 = −10. Solving this system of equations gives c1 = 3 and c2 = −1. Thus, 3 [v]B = . −1 3. Write (8, −2) = c1 (−1, 3) + c2 (3, 2). Then −c1 + 3c2 = 8 and 3c1 + 2c2 = −2. Solving this system of equations gives c1 = −2 and c2 = 2. Thus, −2 [v]B = . 2 4. Write (−9, 1, −8) = c1 (1, 0, 1) + c2 (1, 1, −1) + c3 (2, 0, 1). Then c1 + c2 + 2c3 = −9 and c2 = 1 and (c)2017 Pearson Education. Inc. c1 − c2 + c3 = −8. 311 Solving this system of equations gives c1 = −4, c2 = 1, and c3 = −3. Thus, ⎡ ⎤ −4 [v]B = ⎣ 1 ⎦ . −3 5. Write (1, 7, 7) = c1 (1, −6, 3) + c2 (0, 5, −1) + c3 (3, −1, −1). Then c1 + 3c3 = 1 and − 6c1 + 5c2 − c3 = 7 and 3c1 − c2 − c3 = 7. Using Gaussian elimination to solve this system of equations gives c1 = 4, c2 = 6, and c3 = −1. Thus, ⎡ ⎤ 4 [v]B = ⎣ 6 ⎦ . −1 6. Write (1, 7, 7) = c1 (3, −1, −1) + c2 (1, −6, 3) + c3 (0, 5, −1). Then 3c1 + c2 = 1 and − c1 − 6c2 + 5c3 = 7 and − c1 + 3c2 − c3 = 7. Using Gaussian elimination to solve this system of equations gives c1 = −1, c2 = 4, and c3 = 6. Thus, ⎡ ⎤ −1 [v]B = ⎣ 4 ⎦ . 6 7. Write (5, 5, 5) = c1 (−1, 0, 0) + c2 (0, 0, −3) + c3 (0, −2, 0). Then −c1 = 5 − 2c3 = 5 and Therefore, c1 = −5, c2 = − 53 , and c3 = − 52 . Thus, and − 3c2 = 5. ⎤ −5 [v]B = ⎣ −5/3 ⎦ . −5/2 8. Write ⎡ −4x2 + 2x + 6 = c1 (x2 + x) + c2 (2 + 2x) + c3 (1). Equating the powers of x on each side, we have c1 = −4 and c1 + 2c2 = 2 and 2c2 + c3 = 6. Solving this system of equations, we find that c1 = −4, c2 = 3, and c3 = 0. Hence, ⎤ ⎡ −4 [p(x)]B = ⎣ 3 ⎦ . 0 (c)2017 Pearson Education. Inc. 312 9. Write 15 − 18x − 30x2 = c1 (5 − 3x) + c2 (1) + c3 (1 + 2x2 ). Equating the powers of x on each side, we have 5c1 + c2 + c3 = 15 and − 3c1 = −18 2c3 = −30. and Solving this system of equations, we find that c1 = 6, c2 = 0, and c3 = −15. Hence, ⎡ ⎤ 6 0 ⎦. [p(x)]B = ⎣ −15 10. Write 4 − x + x2 − 2x3 = c1 (1) + c2 (1 + x) + c3 (1 + x + x2 ) + c4 (1 + x + x2 + x3 ). Equating the powers of x on each side, we have c1 + c2 + c3 + c4 = 4 c2 + c3 + c4 = −1 and and c3 + c4 = 1 c4 = −2. and Solving this system of equations, we find that c1 = 5, c2 = −2, c3 = 3, and c4 = −2. Hence, ⎤ ⎡ 5 ⎢ −2 ⎥ ⎥ [p(x)]B = ⎢ ⎣ 3 ⎦. −2 11. Write 8 + x + 6x2 + 9x3 = c1 (x3 + x2 ) + c2 (x3 − 1) + c3 (x3 + 1) + c4 (x3 + x). Equating the powers of x on each side, we have −c2 + c3 = 8 and c4 = 1 and c1 = 6 and c1 + c2 + c3 + c4 = 9. Solving this system of equations, we find that c1 = 6, c2 = −3, c3 = 5, and c4 = 1. Hence ⎤ ⎡ 6 ⎢ −3 ⎥ ⎥ [p(x)]B = ⎢ ⎣ 5 ⎦. 1 12. Write −3 −2 −1 2 = c1 1 1 1 1 + c2 1 1 1 0 + c3 1 0 1 0 + c4 1 0 0 0 . Equating the individual entries of the matrices on each side of this equation (upper left, upper right, lower left, and lower right, respectively) gives c1 + c2 + c3 + c4 = −3 and c1 + c2 + c3 = −2 and c1 + c2 = −1 and c1 = 2. Solving this system of equations, we find that c1 = 2, c2 = −3, c3 = −1, and c4 = −1. Thus, ⎤ ⎡ 2 ⎢ −3 ⎥ ⎥ [A]B = ⎢ ⎣ −1 ⎦ . −1 (c)2017 Pearson Education. Inc. 313 13. Write −10 16 −15 −14 = c1 2 −1 3 5 + c2 0 −1 4 1 + c3 1 1 1 1 + c4 3 −1 2 5 . Equating the individual entries of the matrices on each side of this equation (upper left, upper right, lower left, and lower right, respectively) gives 2c1 + c3 + 3c4 = −10, −c1 + 4c2 + c3 − c4 = 16, 3c1 − c2 + c3 + 2c4 = −15, 5c1 + c2 + c3 + 5c4 = −14. Solving this system of equations, we find that c1 = −2, c2 = 4, c3 = −3, and c4 = −1. Thus, ⎤ ⎡ −2 ⎢ 4 ⎥ ⎥ [A]B = ⎢ ⎣ −3 ⎦ . −1 14. Write 5 7 6 8 = c1 −1 0 1 1 + c2 1 3 −1 0 + c3 1 1 0 2 + c4 0 −1 2 3 . Equating the individual entries of the matrices on each side of this equation (upper left, upper right, lower left, and lower right, respectively) gives −c1 + c2 + c3 = 5 and c1 + 3c2 − c4 = 6 and − c2 + c3 + 2c4 = 7 and c1 + 2c3 + 3c4 = 8. Solving this system of equations, we find that c1 = −34/3, c2 = 12, c3 = −55/3, and c4 = 56/3. Thus, ⎤ ⎡ −34/3 ⎢ 12 ⎥ ⎥ [A]B = ⎢ ⎣ −55/3 ⎦ . 56/3 15. Write (x, y, z) = c1 (0, 6, 3) + c2 (3, 0, 3) + c3 (6, −3, 0). Then 6c1 − 3c3 = y and 3c1 + 3c2 = z. ⎤ ⎡ 3 3 0 z The augmented matrix for this linear system is ⎣ 6 0 −3 y ⎦. We can reduce this to row-echelon form 0 3 6 z ⎤ ⎡ z/3 1 1 0 ⎦. Solving this system by back-substitution gives x/3 as ⎣ 0 1 2 0 0 9 y + 2x − 2z 3c2 + 6c3 = x c1 = 1 2 1 x+ y− z 9 9 9 and and 1 2 4 c2 = − x − y + z 9 9 9 and Hence, denote the ordered basis {v1 , v2 , v3 } by B, we have ⎤ ⎡ 1 2 1 9x + 9y − 9z [v]B = ⎣ − 19 x − 29 y + 49 z ⎦ . 2 1 2 9x + 9y − 9z (c)2017 Pearson Education. Inc. c3 = 2 1 2 x + y − z. 9 9 9 314 16. Write a0 + a1 x + a2 x2 = c1 (1 + x) + c2 x(x − 1) + c3 (1 + 2x2 ). Equating powers of x on both sides of this equation, we have c 1 + c3 = a0 c 1 − c2 = a1 c2 + 2c3 = a2 . ⎤ ⎡ 1 0 1 a0 The augmented matrix corresponding to this system of equations is ⎣ 1 −1 0 a1 ⎦. We can reduce this 0 1 2 a2 ⎤ ⎡ a0 1 0 1 ⎦. Thus, solving by back-substitution, we have a2 to row-echelon form as ⎣ 0 1 2 0 0 1 −a0 + a1 + a2 c1 = 2a0 − a1 − a2 and and and c2 = 2a0 − 2a1 − a2 and c3 = −a0 + a1 + a2 . Hence, relative to the ordered basis B = {p1 , p2 , p3 }, we have ⎤ ⎡ 2a0 − a1 − a2 [p(x)]B = ⎣ 2a0 − 2a1 − a2 ⎦ . −a0 + a1 + a2 17. Let v1 = (9, 2) and v2 = (4, −3). Setting (9, 2) = c1 (2, 1) + c2 (−3, 1) 3 and solving, we find c1 = 3 and c2 = −1. Thus, [v1 ]C = . Next, setting −1 (4, −3) = c1 (2, 1) + c2 (−3, 1) −1 and solving, we find c1 = −1 and c2 = −2. Thus, [v2 ]C = . Therefore, −2 3 −1 PC←B = . −1 −2 18. Let v1 = (−5, −3) and v2 = (4, 28). Setting (−5, −3) = c1 (6, 2) + c2 (1, −1) −1 and solving, we find c1 = −1 and c2 = 1. Thus, [v1 ]C = . Next, setting 1 (4, 28) = c1 (6, 2) + c2 (1, −1) 4 and solving, we find c1 = 4 and c2 = −20. Thus, [v2 ]C = . Therefore, −20 −1 4 PC←B = . 1 −20 (c)2017 Pearson Education. Inc. 315 19. Let v1 = (2, −5, 0), v2 = (3, 0, 5), and v3 = (8, −2, −9). Setting (2, −5, 0) = c1 (1, −1, 1) + c2 (2, 0, 1) + c3 (0, 1, 3) ⎡ ⎤ 4 and solving, we find c1 = 4, c2 = −1, and c3 = −1. Thus, [v1 ]C = ⎣ −1 ⎦. Next, setting −1 (3, 0, 5) = c1 (1, −1, 1) + c2 (2, 0, 1) + c3 (0, 1, 3) ⎡ ⎤ 1 and solving, we find c1 = 1, c2 = 1, and c3 = 1. Thus, [v2 ]C = ⎣ 1 ⎦. Finally, setting 1 (8, −2, −9) = c1 (1, −1, 1) + c2 (2, 0, 1) + c3 (0, 1, 3) ⎡ ⎤ −2 and solving, we find c1 = −2, c2 = 5, and c3 = −4. Thus, [v3 ]C = ⎣ 5 ⎦. Therefore, −4 ⎡ ⎤ 4 1 −2 5 ⎦. PC←B = ⎣ −1 1 −1 1 −4 20. Let v1 = (−7, 4, 4), v2 = (4, 2, −1), and v3 = (−7, 5, 0). Setting (−7, 4, 4) = c1 (1, 1, 0) + c2 (0, 1, 1) + c3 (3, −1, −1) ⎡ ⎤ 0 and solving, we find c1 = 0, c2 = 5/3 and c3 = −7/3. Thus, [v1 ]C = ⎣ 5/3 ⎦. Next, setting −7/3 (4, 2, −1) = c1 (1, 1, 0) + c2 (0, 1, 1) + c3 (3, −1, −1) and solving, we find c1 = 0, c2 = 19/3 and c3 = −7/3. Setting (4, 2, −1) = c1 (1, 1, 0) + c2 (0, 1, 1) + c3 (3, −1, −1) ⎡ ⎤ 3 and solving, we find c1 = 3, c2 = −2/3, and c3 = 1/3. Thus, [v2 ]C = ⎣ −2/3 ⎦. Finally, setting 1/3 (−7, 5, 0) = c1 (1, 1, 0) + c2 (0, 1, 1) + c3 (3, −1, −1) ⎡ ⎤ 5 and solving, we find c1 = 5, c2 = −4, and c3 = −4. Hence, [v3 ]C = ⎣ −4 ⎦. Therefore, −4 ⎤ ⎡ 0 3 5 PC←B = ⎣ 53 − 23 −4 ⎦ . 1 −4 − 73 3 (c)2017 Pearson Education. Inc. 316 21. Let v1 = 7 − 4x and v2 = 5x. Setting 7 − 4x = c1 (1 − 2x) + c2 (2 + x) 3 . Next, setting and solving, we find c1 = 3 and c2 = 2. Thus, [v1 ]C = 2 5x = c1 (1 − 2x) + c2 (2 + x) −2 and solving, we find c1 = −2 and c2 = 1. Hence, [v2 ]C = . Therefore, 1 PC←B = 3 −2 2 1 . 22. Let v1 = −4 + x − 6x2 , v2 = 6 + 2x2 , and v3 = −6 − 2x + 4x2 . Setting −4 + x − 6x2 = c1 (1 − x + 3x2 ) + c2 (2) + c3 (3 + x2 ) ⎡ ⎤ −1 and solving, we find c1 = −1, c2 = 3, and c3 = −3. Thus, [v1 ]C = ⎣ 3 ⎦. Next, setting −3 6 + 2x2 = c1 (1 − x + 3x2 ) + c2 (2) + c3 (3 + x2 ) ⎡ ⎤ 0 and solving, we find c1 = 0, c2 = 0, and c3 = 2. Thus, [v2 ]C = ⎣ 0 ⎦. Finally, setting 2 −6 − 2x + 4x2 = c1 (1 − x + 3x2 ) + c2 (2) + c3 (3 + x2 ) ⎡ ⎤ 2 and solving, we find c1 = 2, c2 = −1, and c3 = −2. Thus, [v3 ]C = ⎣ −1 ⎦. Therefore, −2 ⎡ −1 PC←B = ⎣ 3 −3 ⎤ 0 2 0 −1 ⎦ . 2 −2 23. Let v1 = −2 + 3x + 4x2 − x3 , v2 = 3x + 5x2 + 2x3 , v3 = −5x2 − 5x3 , and v4 = 4 + 4x + 4x2 . Setting −2 + 3x + 4x2 − x3 = c1 (1 − x3 ) + c2 (1 + x) + c3 (x + x2 ) + c4 (x2 + x3 ) ⎤ ⎡ 0 ⎢ −2 ⎥ ⎥ and solving, we find c1 = 0, c2 = −2, c3 = 5, and c4 = −1. Thus, [v1 ]C = ⎢ ⎣ 5 ⎦. Next, setting −1 3x + 5x2 + 2x3 = c1 (1 − x3 ) + c2 (1 + x) + c3 (x + x2 ) + c4 (x2 + x3 ) (c)2017 Pearson Education. Inc. 317 ⎤ 0 ⎢ 0 ⎥ ⎥ and solving, we find c1 = 0, c2 = 0, c3 = 3, and c4 = 2. Thus, [v2 ]C = ⎢ ⎣ 3 ⎦. Next, solving 2 ⎡ −5x2 − 5x3 = c1 (1 − x3 ) + c2 (1 + x) + c3 (x + x2 ) + c4 (x2 + x3 ) ⎤ ⎡ 0 ⎢ 0 ⎥ ⎥ and solving, we find c1 = 0, c2 = 0, c3 = 0, and c4 = −5. Thus, [v3 ]C = ⎢ ⎣ 0 ⎦. Finally, setting −5 4 + 4x + 4x2 = c1 (1 − x3 ) + c2 (1 + x) + c3 (x + x2 ) + c4 (x2 + x3 ) ⎤ ⎡ 2 ⎢ 2 ⎥ ⎥ and solving, we find c1 = 2, c2 = 2, c3 = 2, and c4 = 2. Thus, [v4 ]C = ⎢ ⎣ 2 ⎦. Therefore, 2 ⎤ ⎡ 0 0 0 2 ⎢ −2 0 0 2 ⎥ ⎥. PC←B = ⎢ ⎣ 5 3 0 2 ⎦ −1 2 −5 2 24. Let v1 = 2 + x2 , v2 = −1 − 6x + 8x2 , and v3 = −7 − 3x − 9x2 . Setting 2 + x2 = c1 (1 + x) + c2 (−x + x2 ) + c3 (1 + 2x2 ) ⎡ ⎤ 3 and solving, we find c1 = 3, c2 = 3, and c3 = −1. Thus, [v1 ]C = ⎣ 3 ⎦. Next, solving −1 −1 − 6x + 8x2 = c1 (1 + x) + c2 (−x + x2 ) + c3 (1 + 2x2 ) ⎡ ⎤ −4 and solving, we find c1 = −4, c2 = 2, and c3 = 3. Thus, [v2 ]C = ⎣ 2 ⎦. Finally, solving 3 −7 − 3x − 9x2 = c1 (1 + x) + c2 (−x + x2 ) + c3 (1 + 2x2 ) ⎡ ⎤ −2 and solving, we find c1 = −2, c2 = 1, and c3 = −5. Thus, [v3 ]C = ⎣ 1 ⎦. Therefore, −5 ⎡ ⎤ 3 −4 −2 2 1 ⎦. PC←B = ⎣ 3 −1 3 −5 25. Let v1 = 1 0 0 −1 3 5 −2 −4 , v2 = , v3 = , and v4 = . Setting −1 −2 3 0 0 0 0 0 1 0 1 1 1 1 1 1 1 0 = c1 + c2 + c3 + c4 −1 −2 1 1 1 0 0 0 0 0 (c)2017 Pearson Education. Inc. 318 ⎤ −2 ⎢ 1 ⎥ ⎥ and solving, we find c1 = −2, c2 = c3 = c4 = 1. Thus, [v1 ]C = ⎢ ⎣ 1 ⎦. Next, setting 1 0 −1 1 1 1 1 1 1 1 0 c1 + c2 + c3 + c4 3 0 1 1 1 0 0 0 0 0 ⎤ ⎡ 0 ⎢ 3 ⎥ ⎥ and solving, we find c1 = 0, c2 = 3, c3 = −4, and c4 = 1. Thus, [v2 ]C = ⎢ ⎣ −4 ⎦. Next, setting 1 3 5 1 1 1 1 1 1 1 0 c1 + c2 + c3 + c4 0 0 1 1 1 0 0 0 0 0 ⎤ ⎡ 0 ⎢ 0 ⎥ ⎥ and solving, we find c1 = 0, c2 = 0, c3 = 5, and c4 = −2. Thus, [v3 ]C = ⎢ ⎣ 5 ⎦. Finally, setting −2 −2 −4 1 1 1 1 1 1 1 0 c1 + c2 + c3 + c4 0 0 1 1 1 0 0 0 0 0 ⎤ ⎡ 0 ⎢ 0 ⎥ ⎥ and solving, we find c1 = 0, c2 = 0, c3 = −4, and c4 = 2. Thus, [v4 ]C = ⎢ ⎣ −4 ⎦. Therefore, we have 2 ⎤ ⎡ −2 0 0 0 ⎥ ⎢ 1 3 0 0 ⎥. PC←B = ⎢ ⎣ 1 −4 5 −4 ⎦ 1 1 −2 2 ⎡ 26. Let v1 = E12 , v2 = E22 , v3 = E21 , and v4 = E11 . We see by inspection that ⎤ ⎤ ⎤ ⎤ ⎡ ⎡ ⎡ ⎡ 0 1 0 0 ⎢ 0 ⎥ ⎢ 0 ⎥ ⎢ 0 ⎥ ⎢ 1 ⎥ ⎥ ⎥ ⎥ ⎥ ⎢ ⎢ ⎢ [v1 ]C = ⎢ ⎣ 0 ⎦ , [v2 ]C = ⎣ 0 ⎦ , [v3 ]C = ⎣ 1 ⎦ , [v4 ]C = ⎣ 0 ⎦ . 1 0 0 0 Therefore, ⎡ 0 ⎢ 0 PC←B = ⎢ ⎣ 0 1 1 0 0 0 0 0 1 0 ⎤ 0 1 ⎥ ⎥. 0 ⎦ 0 27. We could simply compute the inverse of the matrix obtained in Problem 17. For instructive purposes, however, we proceed directly. Let w1 = (2, 1) and w2 = (−3, 1). Setting (2, 1) = c1 (9, 2) + c2 (4, −3) (c)2017 Pearson Education. Inc. 319 and solving, we obtain c1 = 2/7 and c2 = −1/7. Thus, [w1 ]B = 2/7 −1/7 . Next, setting (−3, 1) = c1 (9, 2) + c2 (4, −3) −1/7 . Therefore, and solving, we obtain c1 = −1/7 and c2 = −3/7. Thus, [w2 ]B = −3/7 ⎤ ⎡ 2 − 17 7 ⎦. PB←C = ⎣ 1 3 −7 −7 28. We could simply compute the inverse of the matrix obtained in Problem 18. For instructive purposes, however, we proceed directly. Let w1 = (6, 2) and w2 = (1, −1). Setting (6, 2) = c1 (−5, −3) + c2 (4, 28) −5/4 . Next, setting and solving, we obtain c1 = −5/4 and c2 = −1/16. Thus, [w1 ]B = −1/16 (1, −1) = c1 (−5, −3) + c2 (4, 28) −1/4 . Therefore, and solving, we obtain c1 = −1/4 and c2 = −1/16. Thus, [w2 ]B = −1/16 ⎤ ⎡ − 14 − 54 ⎦. PB←C = ⎣ 1 1 − 16 − 16 29. We could simply compute the inverse of the matrix obtained in Problem 19. For instructive purposes, however, we proceed directly. Let w1 = (1, −1, 1), w2 = (2, 0, 1), and w3 = (0, 1, 3). Setting (1, −1, 1) = c1 (2, −5, 0) + c2 (3, 0, 5) + c3 (8, −2, −9) ⎡ ⎤ 1/5 and solving, we find c1 = 1/5, c2 = 1/5, and c3 = 0. Thus, [w1 ]B = ⎣ 1/5 ⎦. Next, setting 0 (2, 0, 1) = c1 (2, −5, 0) + c2 (3, 0, 5) + c3 (8, −2, −9) ⎡ ⎤ −2/45 and solving, we find c1 = −2/45, c2 = 2/5, and c3 = 1/9. Thus, [w2 ]B = ⎣ 2/5 ⎦. Finally, setting 1/9 (0, 1, 3) = c1 (2, −5, 0) + c2 (3, 0, 5) + c3 (8, −2, −9) ⎡ ⎤ −7/45 and solving, we find c1 = −7/45, c2 = 2/5, and c3 = −1/9. Thus, [w3 ]B = ⎣ 2/5 ⎦. Therefore, −1/9 ⎡ 1 ⎤ 2 7 − 45 − 45 5 ⎢ ⎥ ⎢ 1 ⎥ 2 2 ⎥ ⎢ PB←C = ⎢ 5 5 5 ⎥. ⎣ ⎦ 1 1 0 −9 9 (c)2017 Pearson Education. Inc. 320 30. We could simply compute the inverse of the matrix obtained in Problem 21. For instructive purposes, however, we proceed directly. Let w1 = 1 − 2x and w2 = 2 + x. Setting 1 − 2x = c1 (7 − 4x) + c2 (5x) and solving, we find c1 = 1/7 and c2 = −2/7. Thus, [w1 ]B = 1/7 −2/7 . Setting 2 + x = c1 (7 − 4x) + c2 (5x) and solving, we find c1 = 2/7 and c2 = 3/7. Thus, [w2 ]B = ⎡ PB←C = ⎣ 1 7 2 7 − 27 3 7 2/7 3/7 . Therefore, ⎤ ⎦. 31. Referring to Problem 23, we have ⎡ 0 ⎢ −2 −1 PB←C = (PC←B ) = ⎢ ⎣ 5 −1 ⎤−1 ⎡ 0 0 2 1/2 ⎢ −7/6 0 0 2 ⎥ ⎥ =⎢ ⎣ −11/30 3 0 2 ⎦ 2 −5 2 1/2 −1/2 5/6 13/30 0 ⎤ 0 0 1/3 0 ⎥ ⎥. 2/15 −1/5 ⎦ 0 0 32. We could simply compute the inverse of the matrix obtained in Problem 26. For instructive purposes, however, we proceed directly. Let w1 = E22 , w2 = E11 , w3 = E21 , and w4 = E12 . We see by inspection that ⎤ ⎤ ⎤ ⎤ ⎡ ⎡ ⎡ ⎡ 0 0 0 1 ⎢ 1 ⎥ ⎢ 0 ⎥ ⎢ 0 ⎥ ⎢ 0 ⎥ ⎥ ⎥ ⎥ ⎥ ⎢ ⎢ ⎢ [w1 ]B = ⎢ ⎣ 0 ⎦ , [w2 ]B = ⎣ 0 ⎦ , [w3 ]B = ⎣ 1 ⎦ , [w4 ]B = ⎣ 0 ⎦ . 0 1 0 0 Therefore ⎡ 0 ⎢ 1 PB←C = ⎢ ⎣ 0 0 0 0 0 1 0 0 1 0 ⎤ 1 0 ⎥ ⎥. 0 ⎦ 0 33. We first compute [v]B and [v]C directly. Setting (−5, 3) = c1 (9, 2) + c2 (4, −3) ⎡ and solving, we obtain c1 = −3/35 and c2 = −37/35. Thus, [v]B = ⎣ (−5, 3) = c1 (2, 1) + c2 (−3, 1) (c)2017 Pearson Education. Inc. 3 − 35 − 37 35 ⎤ ⎦. Setting 321 ⎡ and solving, we obtain c1 = 4/5 and c2 = 11/5. Thus, [v]C = ⎣ PC←B = 3 −1 −1 −2 4 5 11 5 ⎤ ⎦. Now, according to Problem 17, , so PC←B [v]B = 3 −1 −1 −2 ⎡ ⎣ 3 − 35 − 37 35 ⎤ ⎡ ⎦=⎣ 4 5 11 5 ⎤ ⎦ = [v]C , which confirms Equation (4.7.6). 34. We first compute [v]B and [v]C directly. Setting (−1, 2, 0) = c1 (−7, 4, 4) + c2 (4, 2, −1) + c3 (−7, 5, 0) ⎡ 3 ⎤ 43 ⎢ ⎥ ⎢ 12 ⎥ ⎥ and solving, we obtain c1 = 3/43, c2 = 12/43, and c3 = 10/43. Thus, [v]B = ⎢ ⎢ 43 ⎥. Setting ⎣ ⎦ 10 43 (−1, 2, 0) = c1 (1, 1, 0) + c2 (0, 1, 1) + c3 (3, −1, −1) ⎡ ⎤ 2 and solving, we obtain c1 = 2, c2 = −1, and c3 = −1. Thus, [v]C = ⎣ −1 ⎦. Now, according to Problem −1 ⎤ ⎡ ⎡ ⎡ ⎤ ⎤⎡ 3 ⎤ 0 3 5 0 3 5 2 43 ⎥ ⎢ ⎢ ⎢ ⎥ ⎥ ⎥⎢ ⎥ ⎢ ⎢ 5 ⎢ ⎥ ⎥ ⎥⎢ 2 ⎥, so PC←B [v]B = ⎢ 5 − 2 −4 ⎥ ⎢ 12 ⎥ = ⎢ −1 ⎥ = [v]C , which − −4 20, PC←B = ⎢ 3 3 ⎥ ⎢ ⎢ 3 ⎢ 3 ⎥ ⎥ ⎢ 43 ⎥ ⎦ ⎣ ⎣ ⎣ ⎦ ⎦ ⎦⎣ 1 1 7 10 −1 −4 −4 − 73 − 3 3 3 43 confirms Equation (4.7.6). 35. We first compute [v]B and [v]C directly. Setting 6 − 4x = c1 (7 − 4x) + c2 (5x) ⎤ ⎡ 6 and solving, we obtain c1 = 6/7 and c2 = −4/35. Thus, [v]B = ⎣ 7 4 − 35 ⎦. Next, setting 6 − 4x = c1 (1 − 2x) + c2 (2 + x) ⎡ 14 ⎤ and solving, we obtain c1 = 14/5 and c2 = 8/5. Thus, [v]C = ⎣ PC←B = 3 −2 2 1 5 8 5 ⎦. Now, according to Problem 21, , so ⎡ PC←B [v]B = ⎣ 3 2 −2 1 ⎤⎡ ⎦⎣ 6 7 4 − 35 ⎤ ⎡ 14 ⎤ ⎦=⎣ (c)2017 Pearson Education. Inc. 5 8 5 ⎦ = [v]C , 322 which confirms Equation (4.7.6). 36. We first compute [v]B and [v]C directly. Setting 5 − x + 3x2 = c1 (−4 + x − 6x2 ) + c2 (6 + 2x2 ) + c3 (−6 − 2x + 4x2 ) ⎡ ⎤ 1 and solving, we obtain c1 = 1, c2 = 5/2, and c3 = 1. Thus, [v]B = ⎣ 52 ⎦. Next, setting 1 5 − x + 3x2 = c1 (1 − x + 3x2 ) + c2 (2) + c3 (3 + x2 ) ⎡ ⎤ 1 and solving, we obtain c1 = 1, c2 = 2, and c3 = 0. Thus, [v]C = ⎣ 2 ⎦. Now, according to Problem 22, 0 ⎡ ⎤ −1 0 2 PC←B = ⎣ 3 0 −1 ⎦, so −3 2 −2 ⎤ ⎡ ⎤⎡ ⎤ 1 −1 0 2 1 PC←B [v]B = ⎣ 3 0 −1 ⎦ ⎣ 52 ⎦ = ⎣ 2 ⎦ = [v]C , −3 2 −2 0 1 ⎡ which confirms Equation (4.7.6). 37. We first compute [v]B and [v]C directly. Setting 1 0 0 −1 3 −1 −1 + c2 + c3 = c1 −1 −2 3 0 0 −4 5 5 0 + c4 −2 −4 0 0 ⎤ −5/2 ⎢ −13/6 ⎥ ⎥ and solving, we obtain c1 = −5/2, c2 = −13/6, c3 = 37/6, and c4 = 17/2. Thus, [v]B = ⎢ ⎣ 37/6 ⎦. Next, 17/2 setting −1 −1 1 1 1 1 1 1 1 0 = c1 + c2 + c3 + c4 −4 5 1 1 1 0 0 0 0 0 ⎤ ⎡ 5 ⎢ −9 ⎥ ⎥ and solving, we find c1 = 5, c2 = −9, c3 = 3, and c4 = 0. Thus, [v]C = ⎢ ⎣ 3 ⎦. Now, according to Problem 0 ⎤ ⎡ −2 0 0 0 ⎢ 1 3 0 0 ⎥ ⎥, and so 25, PC←B = ⎢ ⎣ 1 −4 5 −4 ⎦ 1 1 −2 2 ⎡ ⎤ ⎤ ⎡ ⎤⎡ 5 −5/2 −2 0 0 0 ⎥ ⎥ ⎢ ⎢ ⎢ 1 3 0 0 ⎥ ⎥ ⎢ −13/6 ⎥ = ⎢ −9 ⎥ = [v]C , PC←B [v]B = ⎢ ⎦ ⎣ ⎦ ⎣ ⎣ 1 −4 3 ⎦ 37/6 5 −4 0 17/2 1 1 −2 2 ⎡ (c)2017 Pearson Education. Inc. 323 which confirms Equation (4.7.6). 38. Write x = a1 v1 + a2 v2 + · · · + an vn . We have cx = c(a1 v1 + a2 v2 + · · · + an vn ) = (ca1 )v1 + (ca2 )v2 + · · · + (can )vn . Hence, ⎡ ⎤ ⎤ a1 ca1 ⎢ a2 ⎥ ⎢ ca2 ⎥ ⎢ ⎢ ⎥ ⎥ [cx]B = ⎢ . ⎥ = c ⎢ . ⎥ = c[x]B . ⎣ .. ⎦ ⎣ .. ⎦ ⎡ can an 39. We must show that {v1 , v2 , . . . , vn } is linearly independent and spans V . Check linear independence: Assume that c1 v1 + c2 v2 + · · · + cn vn = 0. We wish to show that c1 = c2 = · · · = cn = 0. Now by assumption, the zero vector 0 can be uniquely written as a linear combination of the vectors in {v1 , v2 , . . . , vn }. Since 0 = 0 · v1 + 0 · v2 + · · · + 0 · vn , we therefore conclude that c1 = c2 = · · · = cn = 0, as needed. Check spanning property: Let v be an arbitrary vector in V . By assumption, it is possible to express v (uniquely) as a linear combination of the vectors in {v1 , v2 , . . . , vn }; say v = a1 v 1 + a2 v 2 + · · · + an v n . Therefore, v lies in span{v1 , v2 , . . . , vn }. Since v is an arbitrary member of V , we conclude that {v1 , v2 , . . . , vn } spans V . 40. Let B = {v1 , v2 , . . . , vn } and let C = {vσ(1) , vσ(2) , . . . , vσ(n) }, where σ is a permutation of the set {1, 2, . . . , n}. We will show that PC←B contains exactly one 1 in each row and column, and zeroes elsewhere (the argument for PB←C is essentially identical, or can be deduced from the fact that PB←C is the inverse of PC←B ). Let i be in {1, 2, . . . , n}. The ith column of PC←B is [vi ]C . Suppose that ki ∈ {1, 2, . . . , n} is such that σ(ki ) = i. Then [vi ]C is a column vector with a 1 in the ki th position and zeroes elsewhere. Since the values k1 , k2 , . . . , kn are distinct, we see that each column of PC←B contains a single 1 (with zeroes elsewhere) in a different position from any other column. Hence, when we consider all n columns as a whole, each position in the column vector must have a 1 occurring exactly once in one of the columns. Therefore, PC←B contains exactly one 1 in each row and column, and zeroes elsewhere. Solutions to Section 4.8 True-False Review: (a): TRUE. Note that rowspace(A) is a subspace of Rn and colspace(A) is a subspace of Rm , so certainly if rowspace(A) = colspace(A), then Rn and Rm must be the same. That is, m = n. (c)2017 Pearson Education. Inc. 324 (b): FALSE. A basis for the row space of A consists of the nonzero row vectors of any row-echelon form of A. (c): FALSE. The nonzero column vectors of the original matrix A that correspond to the nonzero column vectors of a row-echelon form of A form a basis for colspace(A). (d): TRUE. Both rowspace(A) and colspace(A) have dimension equal to rank(A), the number of nonzero rows in a row-echelon form of A. Equivalently, their dimensions are both equal to the number of pivots occurring in a row-echelon form of A. (e): TRUE. For an invertible n × n matrix, rank(A) = n. That means there are n nonzero rows in a rowechelon form of A, and so rowspace(A) is n-dimensional. Therefore, we conclude that rowspace(A) = Rn . (f ): TRUE. This follows immediately from the (true) statements in True-False Review Questions 4-5 above. Problems: 1 − 16 . Consequently, a basis for rowspace(A) is {(1, − 16 )}, whereas a 1. A row-echelon form for A is 0 0 basis for colspace(A) is {(6, 12)}. Rowspace(A) consists of points on the line in the xy-plane with equation y = − 16 x, while colspace(A) consists of points on the line in the xy-plane with equation y = 2x. 1 −2 2. A row-echelon form of A is . Consequently, a basis for rowspace(A) is {(1, −2)}, whereas a 0 0 basis for colspace(A) is {(1, −3)}. 3. (a). n = 5; a basis for rowspace(A) is given by {(1, 2, 3, 4, 5)}. (b). m = 1; a basis for colspace(A) is given by {1}. 4. (a). n = 1; a basis for rowspace(A) is given by {1}. (b). m = 3; a basis for colspace(A) is given by {(−3, 1, 7)}. 1 1 −3 2 . Consequently, a basis for rowspace(A) is 5. (a). n = 4; a row-echelon form of A is 0 1 −2 1 {(1, 1, −3, 2), (0, 1, −2, 1)}. (b). m = 2; using the row-echelon form found in part (a), we can immediately find a basis for colspace(A): {(1, 3), (1, 4)}. ⎡ ⎤ 1 2 3 6. (a). n = 3; a row-echelon form of A is ⎣ 0 1 2 ⎦. Consequently, a basis for rowspace(A) is 0 0 0 {(1, 2, 3), (0, 1, 2)}. (b). m = 3; using the row-echelon form found in part (a), we can immediately find a basis for colspace(A): {(1, 5, 9), (2, 6, 10)}. ⎤ ⎡ 0 1 13 7. (a). n = 3; a row-echelon form of A is ⎣ 0 0 0 ⎦. Consequently, a basis for rowspace(A) is {(0, 1, 13 )}. 0 0 0 (b). m = 3; using the row-echelon form found in part (a), we can immediately find a basis for colspace(A): {(3, −6, 12)}. (c)2017 Pearson Education. Inc. 325 ⎡ 1 ⎢ 0 ⎢ 8. (a). n = 4; a row-echelon form of A is ⎣ 0 0 {(1, 2, −1, 3), (0, 0, 0, 1)}. ⎤ 2 −1 3 0 0 1 ⎥ ⎥. Consequently, a basis for rowspace(A) is 0 0 0 ⎦ 0 0 0 (b). m = 4; using the row-echelon form found in part (a), we can immediately find a basis for colspace(A): {(1, 3, 1, 5), (3, 5, −1, 7)}. ⎡ ⎤ 1 −1 2 3 2 −4 3 ⎦ (Note: This is not row-echelon form, but it is not 9. (a). n = 4; we can reduce A to ⎣ 0 0 0 6 −13 necessary to bring the leading nonzero element in each row to 1.). Consequently, a basis for rowspace(A) is {(1, −1, 2, 3), (0, 2, −4, 3), (0, 0, 6, −13)}. (b). m = 3; using the row-echelon form found in part (a), we can immediately find a basis for colspace(A): {(1, 1, 3), (−1, 1, 1), (2, −2, 4)}. ⎡ ⎤ 1 −1 2 1 ⎦. A row-echelon form of 10. (a). We determine a basis for the row space of the matrix ⎣ 5 −4 7 −5 −4 ⎡ ⎤ 1 −1 2 1 −9 ⎦. Consequently, a basis for the subspace spanned by the given vectors is this matrix is ⎣ 0 0 0 0 {(1, −1, 2), (0, 1, −9)}. ⎤ ⎡ 1 5 7 (b). We determine a basis for the column space of the matrix A = ⎣ −1 −4 −5 ⎦. A row-echelon form 2 1 −4 ⎡ ⎤ 1 5 7 for this matrix is ⎣ 0 1 2 ⎦. Since the first two columns contain the pivots, we may take as a basis for 0 0 0 colspace(A), and hence as a basis of R3 spanned by the given set of vectors, the set {(1, −1, 2), (5, −4, 1)}. ⎤ ⎡ 1 3 3 ⎢ 1 5 −1 ⎥ ⎥. A row-echelon form of 11. (a). We determine a basis for the rowspace of the matrix ⎢ ⎣ 2 7 4 ⎦ 1 4 1 ⎤ ⎡ 1 3 3 ⎢ 0 1 −2 ⎥ ⎥. Consequently, a basis for the subspace spanned by the given vectors is this matrix is ⎢ ⎣ 0 0 0 ⎦ 0 0 0 {(1, 3, 3), (0, 1, −2)}. ⎡ ⎤ 1 1 2 1 5 7 4 ⎦. A row-echelon form (b). We determine a basis for the column space of the matrix A = ⎣ 3 3 −1 4 1 ⎡ ⎤ 1 1 2 1 for this matrix is ⎣ 0 1 1/2 1/2 ⎦. Since the first two columns contain the pivots, we may take as a basis 0 0 0 0 for colspace(A), and hence as a basis of R3 spanned by the given set of vectors, the set {(1, 3, 3), (1, 5, −1)}. (c)2017 Pearson Education. Inc. 326 ⎡ ⎤ 1 1 −1 2 3 −4 ⎦. A row-echelon form of 12. (a). We determine a basis for the rowspace of the matrix ⎣ 2 1 1 2 −6 10 ⎡ ⎤ 1 1 −1 2 this matrix is ⎣ 0 1 −5 8 ⎦. Consequently, a basis for the subspace spanned by the given vectors is 0 0 0 0 {(1, 1, −1, 2), (0, 1, −5, 8)}. ⎤ ⎡ 1 2 1 ⎢ 1 1 2 ⎥ ⎥. A row-echelon form (b). We determine a basis for the column space of the matrix A = ⎢ ⎣ −1 3 −6 ⎦ 2 −4 10 ⎤ ⎡ 1 2 1 ⎢ 0 1 −1 ⎥ ⎥. Since the first two columns contain the pivots, we may take as a basis for ⎢ for this matrix is ⎣ 0 0 0 ⎦ 0 0 0 colspace(A), and hence as a basis of R4 spanned by the given set of vectors, the set {(1, 1, −1, 2), (2, 1, 3, −4)}. ⎤ ⎡ 1 4 1 3 ⎢ 2 8 3 5 ⎥ ⎥ 13. (a). We determine a basis for the rowspace of the matrix ⎢ ⎣ 1 4 0 4 ⎦. A row-echelon form of 2 8 2 6 ⎤ ⎡ 1 4 1 3 ⎢ 0 0 1 −1 ⎥ ⎥. Consequently, a basis for the subspace spanned by the given vectors is this matrix is ⎢ ⎣ 0 0 0 0 ⎦ 0 0 0 0 {(1, 4, 1, 3), (0, 0, 1, −1)}. ⎤ ⎡ 1 2 1 2 ⎢ 4 8 4 8 ⎥ ⎥ (b). We determine a basis for the column space of the matrix A = ⎢ ⎣ 1 3 0 2 ⎦. A row-echelon form for 3 5 4 6 ⎡ ⎤ 1 2 1 2 ⎢ 0 1 −1 0 ⎥ ⎥. Since the first two columns contain the pivots, we may take as a basis for this matrix is ⎢ ⎣ 0 0 0 0 ⎦ 0 0 0 0 colspace(A), and hence as a basis of R4 spanned by the given set of vectors, the set {(1, 4, 1, 3), (2, 8, 3, 5)}. ⎡ ⎤ 1 2 4 14. (a). A row-echelon form of A is ⎣ 0 1 1 ⎦. Consequently, a basis for rowspace(A) is 0 0 0 {(1, 2, 4), (0, 1, 1)}, whereas a basis for colspace(A) is {(1, 5, 3), (2, 11, 7)}. (b). We see that both of these subspaces are two-dimensional, and therefore each corresponds geometrically to a plane. By inspection, we see that the two basis vectors for rowspace(A) satisfy the equations 2x+y −z = 0, and therefore rowspace(A) corresponds to the plane with this equation. Similarly, we see that the two basis vectors for colspace(A) satisfy the equations 2x − y + z = 0, and therefore colspace(A) corresponds to the plane with this equations. 0 1 15. Many examples are possible here. One possibility is A = . We see that rowspace(A) consists 0 0 (c)2017 Pearson Education. Inc. 327 of all vectors of the form (0, y), where y ∈ R. However, colspace(A) consists of all vectors of the form (x, 0), where x ∈ R. There are no nonzero vectors in common to both and column space of A. the row space 1 1 Here is a second example that the reader can verify: A = . We have rowspace(A) = {(r, r) : −1 −1 r ∈ R}, while colspace(A) = {(r, −r) : r ∈ R}. Thus, rowspace(A) and colspace(A) have no nonzero vectors in common. 1 1 16. If A = , colspace(A) is spanned by (1, 2), but if we permute the two rows of A, we obtain a 2 2 new matrix whose column space is spanned by (2, 1). On the other hand, if we multiply the first row by 2, then we obtain a new matrix whose column space is spanned by (2, 2). And if we add −2 times the first row to the second row, we obtain a new matrix whose column space is spanned by (1, 0). Therefore, in all cases, colspace(A) is altered by the row operations performed. 17. Since colspace(A) is a subspace of Rm and nullspace(A) is a subspace of Rn , if colspace(A) = nullspace(A), we must have m = n. 18. To the contrary, if A is invertible, then nullspace(A) = {0}. The assumption that rowspace(A) = nullspace(A) therefore implies that rowspace(A) = {0}. In other words, A = 0n . But the zero matrix is not invertible, so this is a contradiction. Hence, we conclude that A cannot be invertible. Solutions to Section 4.9 True-False Review: (a): FALSE. For example, consider the 7 × 3 zero matrix, 07×3 . We have rank(07×3 ) = 0, and therefore by the Rank-Nullity Theorem, nullity(07×3 ) = 3. But |m − n| = |7 − 3| = 4. Many other examples can be given. In particular, provided that m > 2n, the m × n zero matrix will show the falsity of the statement. (b): FALSE. In this case, rowspace(A) is a subspace of R9 , hence it cannot possibly be equal to R7 . The correct conclusion here should have been that rowspace(A) is a 7-dimensional subspace of R9 . (c): TRUE. By the Rank-Nullity Theorem, rank(A) = 7, and therefore, rowspace(A) is a 7-dimensional subspace of R7 . Hence, rowspace(A) = R7 . 0 1 (d): FALSE. For instance, the matrix A = is an upper triangular matrix with two zeros appearing 0 0 on the main diagonal. However, since rank(A) = 1, we also have nullity(A) = 1. (e): TRUE. An invertible matrix A must have nullspace(A) = {0}, but if colspace(A) is also {0}, then A would be the zero matrix, which is certainly not invertible. 1 0 0 1 (f ): FALSE. For instance, if we take A = and B = , then nullity(A)+ nullity(B) = 0 0 0 0 1 1 1 + 1 = 2, but A + B = , and nullity(A + B) = 1. 0 0 1 0 0 0 (g): FALSE. For instance, if we take A = and B = , then nullity(A)·nullity(B) = 1 · 1 = 0 0 0 1 1, but AB = 02 , and nullity(AB) = 2. (h): TRUE. If x belongs to the nullspace of B, then Bx = 0. Therefore, (AB)x = A(Bx) = A0 = 0, so that x also belongs to the nullspace of AB. Thus, nullspace(B) is a subspace of nullspace(AB). Hence, nullity(B) ≤ nullity(AB), as claimed. (c)2017 Pearson Education. Inc. 328 (i): TRUE. If y belongs to nullspace(A), then Ay = 0. Hence, if Axp = b, then A(y + xp ) = Ay + Axp = 0 + b = b, which demonstrates that y + xp is also a solution to the linear system Ax = b. Problems: 1. By inspection, we can see that nullspace(A) = {0}, the trivial subspace of R. Hence, nullity(A) = 0. We also see by inspection that rank(A) = 1. Thus, rank(A) + nullity(A) = 1 + 0 = 1 = n, which verifies the Rank-Nullity Theorem. 2. The matrix is already in row-echelon form. A vector (x, y, z, w) in nullspace(A) must satisfy x−6z−w = 0. We see that y,z, and w are free variables, and ⎫ ⎧⎡ ⎧⎡ ⎤ ⎡ ⎤ ⎤⎫ ⎤ ⎡ 0 6z + w 1 ⎪ 6 ⎪ ⎪ ⎪ ⎪ ⎪ ⎪ ⎪ ⎬ ⎨⎢ ⎨⎢ ⎢ 0 ⎥ ⎢ 0 ⎥⎬ ⎥ 1 ⎥ y ⎥ ⎢ ⎥ ⎢ ⎢ ⎥ ⎥ ⎢ nullspace(A) = ⎣ ⎦ : y, z, w ∈ R⎪ = span ⎪⎣ 0 ⎦ , ⎣ 1 ⎦ , ⎣ 0 ⎦⎪ . z ⎪ ⎪ ⎪ ⎪ ⎪ ⎭ ⎭ ⎩ ⎩ 0 w 1 0 Therefore, nullity(A) = 3. Moreover, since this row-echelon form contains one nonzero row, rank(A) = 1. Since the number of columns of A is 4 = 1 + 3, the Rank-Nullity Theorem is verified. 3. We bring A to row-echelon form: REF(A) = 1 − 12 0 0 . A vector (x, y) in nullspace(A) must satisfy x − 12 y = 0. Setting y = 2t, we get x = t. Therefore, t 1 nullspace(A) = : t ∈ R = span . 2t 2 Therefore, nullity(A) = 1. Since REF(A) contains one nonzero row, rank(A) = 1. Since the number of columns of A is 2 = 1 + 1, the Rank-Nullity Theorem is verified. 4. We bring A to row-echelon form: ⎡ 1 REF(A) = ⎣ 0 0 ⎤ 1 −1 1 7 ⎦. 0 1 Since there are no unpivoted columns, there are no free variables in the associated homogeneous linear system, and so nullspace(A) = {0}. Therefore, nullity(A) = 0. Since REF(A) contains three nonzero rows, rank(A) = 3. Since the number of columns of A is 3 = 3 + 0, the Rank-Nullity Theorem is verified. 5. We bring A to row-echelon form: ⎤ 1 4 −1 3 1 1 ⎦. REF(A) = ⎣ 0 1 0 0 0 0 ⎡ (c)2017 Pearson Education. Inc. 329 A vector (x, y, z, w) in nullspace(A) must satisfy x+4y−z+3w = 0 and y+z+w = 0. We see that z and w are free variables. Set z = t and w = s. Then y = −z−w = −t−s and x = z−4y−3w = t−4(−t−s)−3s = 5t+s. Therefore, ⎧⎡ ⎫ ⎧⎡ ⎤⎫ ⎤ ⎤ ⎡ 1 ⎪ 5t + s 5 ⎪ ⎪ ⎪ ⎪ ⎪ ⎪ ⎪ ⎨⎢ ⎬ ⎨⎢ ⎥⎬ ⎥ ⎢ −t − s ⎥ ⎥ : s, t ∈ R = span ⎢ −1 ⎥ , ⎢ −1 ⎥ . nullspace(A) = ⎢ ⎦ ⎣ ⎣ 1 ⎦ ⎣ 0 ⎦⎪ t ⎪ ⎪ ⎪ ⎪ ⎪ ⎪ ⎪ ⎭ ⎩ ⎭ ⎩ 1 s 0 Therefore, nullity(A) = 2. Moreover, since REF(A) contains two nonzero rows, rank(A) = 2. Since the number of columns of A is 4 = 2 + 2, the Rank-Nullity Theorem is verified. 6. Since all rows (or columns) of this matrix are proportional to the first one, rank(A) = 1. Since A has two columns, we conclude from the Rank-Nullity Theorem that nullity(A) = 2 − rank(A) = 2 − 1 = 1. 7. The first and last rows of A are not proportional, but the middle rows are proportional to the first row. Therefore, rank(A) = 2. Since A has five columns, we conclude from the Rank-Nullity Theorem that nullity(A) = 5 − rank(A) = 5 − 2 = 3. 8. Since the second and third columns are not proportional and the first column is all zeros, we have rank(A) = 2. Since A has three columns, we conclude from the Rank-Nullity Theorem that nullity(A) = 3 − rank(A) = 3 − 2 = 1. 9. This matrix (already in row-echelon form) has one nonzero row, so rank(A) = 1. Since it has four columns, we conclude from the Rank-Nullity Theorem that nullity(A) = 4 − rank(A) = 4 − 1 = 3. ⎡ 1 10. The augmented matrix for this linear system is ⎣ 2 1 matrix to row-echelon form: ⎡ 1 3 −1 ⎣ 0 1 11 0 0 0 ⎤ 3 −1 4 7 9 11 ⎦. We quickly reduce this augmented 5 21 10 ⎤ 4 3 ⎦. 0 A solution (x, y, z) to the system will have a free variable corresponding to the third column: z = t. Then y + 11t = 3, so y = 3 − 11t. Finally, x + 3y − z = 4, so x = 4 + t − 3(3 − 11t) = −5 + 34t. Thus, the solution set is ⎧⎡ ⎫ ⎧ ⎡ ⎫ ⎤ ⎤ ⎡ ⎤ 34 −5 ⎨ −5 + 34t ⎬ ⎨ ⎬ ⎣ 3 − 11t ⎦ : t ∈ R = t ⎣ −11 ⎦ + ⎣ 3 ⎦ : t ∈ R . ⎩ ⎭ ⎩ ⎭ t 1 0 ⎡ ⎤ ⎡ ⎤ −5 34 Observe that xp = ⎣ 3 ⎦ is a particular solution to Ax = b, and that ⎣ −11 ⎦ forms a basis for 0 1 nullspace(A). Therefore, the set of solution vectors obtained does indeed take the form (4.9.3). (c)2017 Pearson Education. Inc. 330 ⎤ 1 −1 2 3 6 11. The augmented matrix for this linear system is ⎣ 1 −2 5 5 13 ⎦. We quickly reduce this aug2 −1 1 4 5 mented matrix to row-echelon form: ⎤ ⎡ 6 1 −1 2 3 ⎣ 0 1 −3 −2 −7 ⎦ . 0 0 0 0 0 ⎡ A solution (x, y, z, w) to the system will have free variables corresponding to the third and fourth columns: z = t and w = s. Then y − 3z − 2w = −7 requires that y = −7 + 3t + 2s, and x − y + 2z + 3w = 6 requires that x = 6 + (−7 + 3t + 2s) − 2t − 3s = −1 + t − s. Thus, the solution set is ⎫ ⎧⎡ ⎫ ⎧ ⎡ ⎤ ⎤ ⎡ ⎤ ⎡ ⎤ −1 −1 1 −1 + t − s ⎪ ⎪ ⎪ ⎪ ⎪ ⎪ ⎪ ⎪ ⎬ ⎨⎢ ⎬ ⎨ ⎢ ⎥ ⎥ ⎢ ⎥ ⎢ ⎥ ⎢ −7 + 3t + 2s ⎥ : s, t ∈ R = t ⎢ 3 ⎥ + s ⎢ 2 ⎥ + ⎢ −7 ⎥ : s, t ∈ R . ⎣ 0 ⎦ ⎣ 0 ⎦ ⎦ ⎣ t ⎪ ⎪ ⎣ 1 ⎦ ⎪ ⎪ ⎪ ⎪ ⎪ ⎭ ⎩ ⎩ ⎭ ⎪ 0 1 0 s ⎤ −1 ⎢ −7 ⎥ ⎥ Observe that xp = ⎢ ⎣ 0 ⎦ is a particular solution to Ax = b, and that 0 ⎡ ⎧⎡ ⎤⎫ ⎤ ⎡ −1 ⎪ 1 ⎪ ⎪ ⎪ ⎨⎢ ⎥⎬ ⎥ ⎢ ⎢ 3 ⎥,⎢ 2 ⎥ ⎦ ⎣ 0 ⎦⎪ ⎣ ⎪ ⎪ ⎪ 1 ⎭ ⎩ 1 0 is a basis for nullspace(A). Therefore, the set of solution vectors obtained does indeed take the form (4.9.3). ⎡ ⎤ 1 1 −2 −3 ⎢ 3 −1 −7 2 ⎥ ⎥. We quickly reduce this augmented 12. The augmented matrix for this linear system is ⎢ ⎣ 1 1 1 0 ⎦ 2 2 −4 −6 matrix to row-echelon form: ⎤ ⎡ −3 1 1 −2 1 ⎦. ⎣ 0 1 − 11 4 4 0 0 1 1 There are no free variables in the solution ⎧⎡set (since ⎤⎫ none of the first three columns is unpivoted), and we find 2 ⎬ ⎨ the solution set by back-substitution: ⎣ −3 ⎦ . It is easy to see that this is indeed a particular solution: ⎩ ⎭ 1 ⎡ ⎤ 2 xp = ⎣ −3 ⎦. Since the row-echelon form of A has three nonzero rows, rank(A) = 3. Thus, nullity(A) = 0. 1 Hence, nullspace(A) = {0}. Thus, the only term in the expression (4.9.3) that appears in the solution is xp , and this is precisely the unique solution we obtained in the calculations above. 13. By inspection, we see that a particular solution to this (homogeneous) linear system is xp = 0. We (c)2017 Pearson Education. Inc. 331 quickly reduce this augmented matrix to row-echelon form: ⎤ ⎡ 1 1 −1 5 0 ⎣ 0 1 −1 7 0 ⎦ . 2 2 0 0 0 0 0 A solution (x, y, z, w) to the system will have free variables corresponding to the third and fourth columns: z = t and w = s. Then y − 12 z + 72 w = 0 requires that y = 12 t − 72 s, and x + y − z + 5w = 0 requires that x = t − ( 12 t − 72 s) − 5s = 12 t − 32 s. Thus, the solution set is ⎫ ⎧ ⎡ 1 ⎤ ⎫ ⎧⎡ 1 ⎤ ⎡ 3 ⎤ −2 t − 32 s ⎪ ⎪ ⎪ ⎪ 2 2 ⎪ ⎪ ⎪ ⎪ ⎬ ⎨ ⎢ 1 ⎥ ⎬ ⎨⎢ 1 7 ⎥ ⎢ −7 ⎥ t − s 2 ⎢ 2 ⎥ : s, t ∈ R = t ⎢ 2 ⎥ + s ⎢ 2 ⎥ : s, t ∈ R . ⎣ ⎦ ⎣ 0 ⎦ t ⎪ ⎣ 1 ⎦ ⎪ ⎪ ⎪ ⎪ ⎪ ⎪ ⎩ ⎭ ⎪ ⎭ ⎩ s 0 1 ⎡ 1 ⎤ ⎡ 3 ⎤ −2 2 ⎢ 1 ⎥ ⎢ −7 ⎥ 2 ⎥ ⎢ 2 ⎥ Since the vectors ⎢ ⎣ 1 ⎦ and ⎣ 0 ⎦ form a basis for nullspace(A), our solutions to take the proper form 0 1 given in (4.9.3). 14. By the Rank-Nullity Theorem, rank(A) = 7 − nullity(A) = 7 − 4 = 3, and hence, colspace(A) is 3-dimensional. But since A has three rows, colspace(A) is a subspace of R3 . Therefore, since the only 3-dimensional subspace of R3 is R3 itself, we conclude that colspace(A) = R3 . Now rowspace(A) is also 3-dimensional, but it is a subspace of R5 . Therefore, it is not accurate to say that rowspace(A) = R3 . 15. By the Rank-Nullity Theorem, rank(A) = 4 − nullity(A) = 4 − 0 = 4, so we conclude that rowspace(A) is 4-dimensional. Since rowspace(A) is a subspace of R4 (since A contains four columns), and it is 4-dimensional, we conclude that rowspace(A) = R4 . Although colspace(A) is 4-dimensional, colspace(A) is a subspace of R6 , and therefore it is not accurate to say that colspace(A) = R4 . 16. If rowspace(A) = nullspace(A), then we know that rank(A) = nullity(A). Therefore, rank(A)+ nullity(A) must be even. But rank(A)+ nullity(A) is the number of columns of A. Therefore, A contains an even number of columns. 17. We know that rank(A) + nullity(A) = 7. But since A only has five rows, rank(A) ≤ 5. Therefore, nullity(A) ≥ 2. However, since nullspace(A) is a subspace of R7 , nullity(A) ≤ 7. Therefore, 2 ≤ nullity(A) ≤ 7. There are ⎤ many examples of a 5 × 7 matrix A with nullity(A) = 2; one example is ⎡ 1 0 0 0 0 0 0 ⎢ 0 1 0 0 0 0 0 ⎥ ⎥ ⎢ ⎢ 0 0 1 0 0 0 0 ⎥. The only 5 × 7 matrix with nullity(A) = 7 is 05×7 , the 5 × 7 zero matrix. ⎥ ⎢ ⎣ 0 0 0 1 0 0 0 ⎦ 0 0 0 0 1 0 0 18. We know that rank(A) + nullity(A) = 8. But since A only has three rows, rank(A) ≤ 3. Therefore, nullity(A) ≥ 5. However, since nullspace(A) is a subspace of R8 , nullity(A) ≤ 8. Therefore, 5 ≤ nullspace(A) ≤ 8. There are ⎡ ⎤ many examples of a 3 × 8 matrix A with nullity(A) = 5; one example is 1 0 0 0 0 0 0 0 ⎣ 0 1 0 0 0 0 0 0 ⎦. The only 3 × 8 matrix with nullity(A) = 8 is 03×8 , the 3 × 8 zero matrix. 0 0 1 0 0 0 0 0 19. If Bx = 0, then ABx = A0 = 0. This observation shows that nullspace(B) is a subspace of nullspace(AB). On the other hand, if ABx = 0, then Bx = (A−1 A)Bx = A−1 (AB)x = A−1 0 = 0, so (c)2017 Pearson Education. Inc. 332 Bx = 0. Therefore, nullspace(AB) is a subspace of nullspace(B). As a result, since nullspace(B) and nullspace(AB) are subspaces of each other, they must be equal: nullspace(AB) = nullspace(B). Therefore, nullity(AB) = nullity(B). Solutions to Section 4.10 True-False Review: (a): TRUE. This is the equivalence of statements (a) and (m) in the Invertible Matrix Theorem. 1 2 3 (b): TRUE. Here is an example to show that this is possible: A = . The two rows of A are 1 2 4 non-proportional, so they form a linearly independent set. However, the three columns of A include two identical columns, so of course the columns for a linearly dependent set of vectors. (c): FALSE. If the matrix has n linearly independent rows, then by the equivalence of (a) and (j) in the Invertible Matrix Theorem, such a matrix would be invertible. But if that were so, then by part (m) of the Invertible Matrix Theorem, such a matrix would have to have n linearly independent columns. (d): FALSE. An n × n matrix A with det(A) = 0 is not invertible by part (g) of the Invertible Matrix Theorem. Therefore, by the equivalence of (a) and (l) in the Invertible Matrix Theorem, the columns of A do not form a basis for Rn . (e): TRUE. If rowspace(A) = Rn , then by the equivalence of (a) and (n) in the Invertible Matrix Theorem, A is not invertible. Therefore, A is not row-equivalent to the identity matrix. Since B is row-equivalent to A, then B is not row-equivalent to the identity matrix, and therefore, B is not invertible. Hence, by part (k) of the Invertible Matrix Theorem, we conclude that colspace(B) = Rn . (f ): FALSE. If nullspace(A) = {0}, then A is invertible by the equivalence of (a) and (c) in the Invertible Matrix Theorem. Since E is an elementary matrix, it is also invertible. Therefore, EA is invertible. By part (g) of the Invertible Matrix Theorem, det(EA) = 0, contrary to the statement given. (g): FALSE. The matrix [A|B] has 2n columns, but only n rows, and therefore rank([A|B]) ≤ n. Hence, by the Rank-Nullity Theorem, nullity([A|B]) ≥ n > 0. (h): TRUE. The first and third rows are proportional, and hence statement (m) in the Invertible Matrix Theorem is false. Therefore, by the equivalence with (a), the matrix is not invertible. ⎤ ⎡ 0 1 0 0 ⎢ 1 0 0 0 ⎥ ⎥ (i): FALSE. For instance, the matrix ⎢ ⎣ 0 0 0 1 ⎦ is of the form given, but satisfies any (and all) of 0 0 1 0 the statements of the Invertible Matrix Theorem. ⎡ ⎤ 1 2 1 (j): FALSE. For instance, the matrix ⎣ 3 6 2 ⎦ is of the form given, but has a nonzero determinant, 1 0 0 and so by part (g) of the Invertible Matrix Theorem, it is invertible. Solutions to Section 4.11 Problems: (c)2017 Pearson Education. Inc. 333 1. Write v = (a1 , a2 , a3 , a4 , a5 ) ∈ R5 . Then we have (r + s)v = (r + s)(a1 , a2 , a3 , a4 , a5 ) = ((r + s)a1 , (r + s)a2 , (r + s)a3 , (r + s)a4 , (r + s)a5 ) = (ra1 + sa1 , ra2 + sa2 , ra3 + sa3 , ra4 + sa4 , ra5 + sa5 ) = (ra1 , ra2 , ra3 , ra4 , ra5 ) + (sa1 , sa2 , sa3 , sa4 , sa5 ) = r(a1 , a2 , a3 , a4 , a5 ) + s(a1 , a2 , a3 , a4 , a5 ) = rv + sv. 2. Write v = (a1 , a2 , a3 , a4 , a5 ) and w = (b1 , b2 , b3 , b4 , b5 ) in R5 . Then we have r(v + w) = r((a1 , a2 , a3 , a4 , a5 ) + (b1 , b2 , b3 , b4 , b5 )) = r(a1 + b1 , a2 + b2 , a3 + b3 , a4 + b4 , a5 + b5 ) = (r(a1 + b1 ), r(a2 + b2 ), r(a3 + b3 ), r(a4 + b4 ), r(a5 + b5 )) = (ra1 + rb1 , ra2 + rb2 , ra3 + rb3 , ra4 + rb4 , ra5 + rb5 ) = (ra1 , ra2 , ra3 , ra4 , ra5 ) + (rb1 , rb2 , rb3 , rb4 , rb5 ) = r(a1 , a2 , a3 , a4 , a5 ) + r(b1 , b2 , b3 , b4 , b5 ) = rv + rw. 3. YES. To show that V forms a vector space, we will show that it is a subspace of P2 (R) by appealing to Theorem 4.3.2. Note that V is non-empty since the polynomial p(x) = 0 belongs to V . Next, we show that V is closed under addition and scalar multiplication: Closure under Addition: Suppose p(x) and q(x) belong to V . Then we have p(3) = q(3) = 0 and p (5) = q (5) = 0. Hence, (p + q)(3) = p(3) + q(3) = 0 + 0 = 0 and (p + q) (5) = p (5) + q (5) = 0 + 0 = 0. Therefore, p(x) + q(x) belongs to V . Closure under Scalar Multiplication: Let k be a scalar and let p(x) be as above. Then (kp)(3) = k · p(3) = k · 0 = 0 and (kp) (5) = k · p (5) = k · 0 = 0. Therefore, k · p(x) belongs to V . 4. NO. This set of polynomials is not closed under scalar multiplication. For example, the polynomial p(x) = 2x belongs to the set, but 13 p(x) = 23 x does not belong to the set (since 23 is not an even integer). 5. YES. This set of polynomials forms a subspace of the vector space P5 (R). To confirm this, we will check that this set is closed under addition and scalar multiplication: Closure under Addition: Let p(x) = a0 + a1 x + a4 x4 + a5 x5 and q(x) = b0 + b1 x + b4 x4 + b5 x5 be polynomials in the set under consideration (their x2 and x3 terms are zero). Then p(x) + q(x) = (a0 + b0 ) + (a1 + b1 )x + · · · + (a4 + b4 )x4 + (a5 + b5 )x5 is again in the set (since it still has no x2 or x3 terms). So closure under addition holds. Closure under Scalar Multiplication: Let p(x) = a0 + a1 x + a4 x4 + a5 x5 be in the set, and let k be a scalar. Then kp(x) = (ka0 ) + (ka1 )x + (ka4 )x4 + (ka5 )x5 , (c)2017 Pearson Education. Inc. 334 which is again in the set (since it still has no x2 or x3 terms). So closure under scalar multiplication holds. 6. NO. We can see immediately that the zero vector (0, 0, 0) is not a solution to this linear system (the first equation is not satisfied by the zero vector), and therefore, we know at once that this set cannot be a vector space. 7. YES. The set of solutions to this linear system forms a subspace of R3 . To confirm this, we will check that this set is closed under addition and scalar multiplication: Closure under Addition: Let (a1 , a2 , a3 ) and (b1 , b2 , b3 ) be solutions to the linear system. This means that 4a1 − 7a2 + 2a3 = 0, 5a1 − 2a2 + 9a3 = 0 4b1 − 7b2 + 2b3 = 0, 5b1 − 2b2 + 9b3 = 0. and Adding the equations on the left, we get 4(a1 + b1 ) − 7(a2 + b2 ) + 2(a3 + b3 ) = 0, so the vector (a1 + b1 , a2 + b2 , a3 + b3 ) satisfies the first equation in the linear system. Likewise, adding the equations on the right, we get 5(a1 + b1 ) − 2(a2 + b2 ) + 9(a3 + b3 ) = 0, so (a1 + b1 , a2 + b2 , a3 + b3 ) also satisfies the second equation in the linear system. Therefore, (a1 + b1 , a2 + b2 , a3 + b3 ) is in the solution set for the linear system, and closure under addition therefore holds. Closure under Scalar Multiplication: Let (a1 , a2 , a3 ) be a solution to the linear system, and let k be a scalar. We have 4a1 − 7a2 + 2a3 = 0, 5a1 − 2a2 + 9a3 = 0, and so, multiplying both equations by k, we have k(4a1 − 7a2 + 2a3 ) = 0, k(5a1 − 2a2 + 9a3 ) = 0, or 4(ka1 ) − 7(ka2 ) + 2(ka3 ) = 0, 5(ka1 ) − 2(ka2 ) + 9(ka3 ) = 0. Thus, the vector (ka1 , ka2 , ka3 ) is a solution to the linear system, and closure under scalar multiplication therefore holds. 1 1 −1 1 8. NO. This set is not closed under addition. For example, the vectors and both 1 1 1 1 belong to the set (their entries are all nonzero), but 1 1 −1 1 0 2 + = , 1 1 1 1 2 2 which does not belong to the set (some entries are zero, and some are nonzero). So closure under addition fails, and therefore, this set does not form a vector space. 1 2 9. YES. The set of 2 × 2 real matrices that commute with C = forms a subspace of M2 (R). To 2 2 confirm this, we will check that this set is closed under addition and scalar multiplication: (c)2017 Pearson Education. Inc. 335 Closure under Addition: Let A and B be 2 × 2 real matrices that commute with C. That is, AC = CA and BC = CB. Then (A + B)C = AC + BC = CA + CB = C(A + B), so A + B commutes with C, and therefore, closure under addition holds. Closure under Scalar Multiplication: Let A be a 2 × 2 real matrix that commutes with C, and let k be a scalar. Then since AC = CA, we have (kA)C = k(AC) = k(CA) = C(kA), so kA is still in the set. Thus, the set is also closed under scalar multiplication. 10. YES. The set of functions f : [0, 1] → [0, 1] such that f (0) = f 41 = f 12 = f 34 = f (1) = 0 is a subspace of the vector space of all functions [0, 1] → [0, 1]. We confirm this by checking that this set is closed under addition and scalar multiplication: Closure under Addition: Let g and h be functions such that 1 1 3 g(0) = g =g =g = g(1) = 0 4 2 4 and 1 1 3 h(0) = h =h =h = h(1) = 0. 4 2 4 Now (g + h)(0) = g(0) + h(0) = 0 + 0 = 0, 1 1 1 =g +h = 0 + 0 = 0, (g + h) 4 4 4 1 1 1 =g +h = 0 + 0 = 0, (g + h) 2 2 2 3 3 3 =g +h = 0 + 0 = 0, (g + h) 4 4 4 (g + h)(1) = g(1) + h(1) = 0 + 0 = 0. Thus, g + h belongs to the set, and so the set is closed under addition. Closure under Scalar Multiplication: Let g be a function with 1 1 3 g(0) = g =g =g = g(1) = 0, 4 2 4 and let k be a scalar. Then (kg)(0) = k · g(0) = k · 0 = 0, 1 1 (kg) =k·g = k · 0 = 0, 4 4 1 1 =k·g = k · 0 = 0, (kg) 2 2 3 3 =k·g = k · 0 = 0, (kg) 4 4 (kg)(1) = k · g(1) = k · 0 = 0. (c)2017 Pearson Education. Inc. 336 Thus, kg belongs to the set, and so the set is closed under scalar multiplication. 11. NO. This set is not closed under addition. For example, the function f defined by f (x) = x belongs to the set, and the function g defined by x, if x ≤ 1/2 g(x) = 0, if x > 1/2 belongs to the set. But (f + g)(x) = 2x, x, if x ≤ 1/2 if x > 1/2. Then f + g : [0, 1] → [0, 1], but for 0 < x ≤ 12 , |(f + g)(x)| = 2x > x, so f + g is not in the set. Therefore, the set is not closed under addition. 12. NO. This set is not closed under addition. For example, if we let 1 0 0 1 A= and B= , 0 1 0 0 then A2 = A is symmetric, and B 2 = 02 is symmetric, but 2 1 1 1 = (A + B)2 = 0 1 0 2 1 is not symmetric, so A + B is not in the set. Thus, the set in question is not closed under addition. 13. We must check each of the vector space axioms (A1)-(A10). Axiom (A1): Assume that (a1 , a2 ) and (b1 , b2 ) belong to V . Then a2 , b2 > 0. Hence, (a1 , a2 ) + (b1 , b2 ) = (a1 + b1 , a2 b2 ) ∈ V, since a2 b2 > 0. Thus, V is closed under addition. Axiom (A2): Assume that (a1 , a2 ) ∈ V , and let k be a scalar. Note that since a2 > 0, the expression ak2 > 0 for every k ∈ R. Hence, k(a1 , a2 ) = (ka1 , ak2 ) ∈ V , thereby showing that V is closed under scalar multiplication. Axiom (A3): Let (a1 , a2 ), (b1 , b2 ) ∈ V . We have (a1 , a2 ) + (b1 , b2 ) = (a1 + b1 , a2 b2 ) = (b1 + a1 , b2 a2 ) = (b1 , b2 ) + (a1 , a2 ), as required. Axiom (A4): Let (a1 , a2 ), (b1 , b2 ), (c1 , c2 ) ∈ V . We have ((a1 , a2 ) + (b1 , b2 )) + (c1 , c2 ) = (a1 + b1 , a2 b2 ) + (c1 , c2 ) = ((a1 + b1 ) + c1 , (a2 b2 )c2 ) = (a1 + (b1 + c1 ), a2 (b2 c2 )) = (a1 , a2 ) + (b1 + c1 , b2 c2 ) = (a1 , a2 ) + ((b1 , b2 ) + (c1 , c2 )), (c)2017 Pearson Education. Inc. 337 as required. Axiom (A5): We claim that (0, 1) is the zero vector in V . To see this, let (b1 , b2 ) ∈ V . Then (0, 1) + (b1 , b2 ) = (0 + b1 , 1 · b2 ) = (b1 , b2 ). Since this holds for every (b1 , b2 ) ∈ V , we conclude that (0, 1) is the zero vector. Axiom (A6): We claim that the additive inverse of a vector (a1 , a2 ) in V is the vector (−a1 , a−1 2 ) (Note that a−1 > 0 since a > 0.) To check this, we compute as follows: 2 2 −1 (a1 , a2 ) + (−a1 , a−1 2 ) = (a1 + (−a1 ), a2 a2 ) = (0, 1). Axiom (A7): We have 1 · (a1 , a2 ) = (a1 , a12 ) = (a1 , a2 ) for all (a1 , a2 ) ∈ V . Axiom (A8): Let (a1 , a2 ) ∈ V , and let r and s be scalars. Then we have (rs)(a1 , a2 ) = ((rs)a1 , ars 2 ) = (r(sa1 ), (as2 )r ) = r(sa1 , as2 ) = r(s(a1 , a2 )), as required. Axiom (A9): Let (a1 , a2 ) and (b1 , b2 ) be members of V , and let r be a scalar. We have r((a1 , a2 ) + (b1 , b2 )) = r(a1 + b1 , a2 b2 ) = (r(a1 + b1 ), (a2 b2 )r ) = (ra1 + rb1 , ar2 br2 ) = (ra1 , ar2 ) + (rb1 , br2 ) = r(a1 , a2 ) + r(b1 , b2 ), as required. Axiom (A10): Let (a1 , a2 ) ∈ V , and let r and s be scalars. We have (r + s)(a1 , a2 ) = ((r + s)a1 , ar+s 2 ) = (ra1 + sa1 , ar2 as2 ) = (ra1 , ar2 ) + (sa1 , as2 ) = r(a1 , a2 ) + s(a1 , a2 ), as required. 14. We must show that S is closed under addition and closed under scalar multiplication: Closure under Addition: Let (a, 2a ) and (b, 2b ) be elements of S. Now consider the sum of these elements: (a, 2a ) + (b, 2b ) = (a + b, 2a 2b ) = (a + b, 2a+b ) ∈ S, which shows that S is closed under addition. (c)2017 Pearson Education. Inc. 338 Closure under Scalar Multiplication: Let (a, 2a ) be an element of S, and let k be a scalar. Then k(a, 2a ) = (ka, (2a )k ) = (ka, 2ka ) ∈ S, which shows that S is closed under scalar multiplication. Thus, S is a subspace of V . 15. Note that 3(1, 2) = (3, 23 ) = (3, 8), so the second vector is a multiple of the first one under the vector space operations of V from Problem 13. Therefore, {(1, 2), (3, 8)} is linearly dependent. 16. We show that S = {(1, 4), (2, 1)} is linearly independent and spans V . S is linearly independent: Assume that c1 (1, 4) + c2 (2, 1) = (0, 1). This can be written (c1 , 4c1 ) + (2c2 , 1c2 ) = (0, 1) or (c1 + 2c2 , 4c1 ) = (0, 1). In order for 4c1 = 1, we must have c1 = 0. And then in order for c1 + 2c2 = 0, we must have c2 = 0. Therefore, S is linearly independent. S spans V : Consider an arbitrary vector (a1 , a2 ) ∈ V , where a2 > 0. We must find constants c1 and c2 such that c1 (1, 4) + c2 (2, 1) = (a1 , a2 ). Thus, (c1 , 4c1 ) + (2c2 , 1c2 ) = (a1 , a2 ) or (c1 + 2c2 , 4c1 ) = (a1 , a2 ). Hence, c1 + 2c2 = a1 and 4c 1 = a 2 . From the second equation, we conclude that c1 = log4 (a2 ). Thus, from the first equation, 1 (a1 − log4 (a2 )). 2 Hence, since we were able to find constants c1 and c2 in order that c2 = c1 (1, 4) + c2 (2, 1) = (a1 , a2 ), we conclude that {(1, 4), (2, 1)} spans V . 17. This really hinges on whether or not the given vectors are linearly dependent or linearly independent. If we assume that c1 (2 + x2 ) + c2 (4 − 2x + 3x2 ) + c3 (1 + x) = 0, (c)2017 Pearson Education. Inc. 339 then (2c1 + 4c2 + c3 ) + (−2c2 + c3 )x + (c1 + 3c2 )x2 = 0. Thus, we have − 2c2 + c3 = 0 2c1 + 4c2 + c3 = 0 Since the matrix of coefficients c1 + 3c2 = 0. ⎡ ⎤ 2 4 1 ⎣ 0 −2 1 ⎦ 1 3 0 fails to be invertible, we conclude that there will be non-trivial solutions for c1 , c2 , and c3 . Thus, the polynomials are linearly dependent. We can therefore remove a vector from the set without decreasing the span. We remove 1 + x, leaving us with {2 + x2 , 4 − 2x + 3x2 }. Since these polynomials are not proportional, they are now linearly independent, and hence, they are a basis for their span. Hence, dim{2 + x2 , 4 − 2x + 3x2 , 1 + x} = 2. 18. NO. This set is not closed under scalar multiplication. For example, (1, 1) belongs to S, but 2 · (1, 1) = (2, 2) does not belong to S. 19. NO. This set is not closed under scalar multiplication. For example, (1, 1) belongs to S, but 2 · (1, 1) = (2, 2) does not belong to S. 20. NO. The zero vector (zero matrix) is not an orthogonal matrix. Any subspace must contain a zero vector. 21. YES. We show that S is closed under addition and closed under scalar multiplication. Closure under Addition: Assume that f and g belong to S. Thus, f (a) = 2f (b) and g(a) = 2g(b). We must show that f + g belongs to S. We have (f + g)(a) = f (a) + g(a) = 2f (b) + 2g(b) = 2[f (b) + g(b)] = 2(f + g)(b), so f + g ∈ S. So S is closed under addition. Closure under Scalar Multiplication: Assume that f belongs to S and k is a scalar. Thus, f (a) = 2f (b). Moreover, (kf )(a) = kf (a) = k(2f (b)) = 2(kf (b)) = 2(kf )(b), so kf ∈ S. Thus, S is closed under scalar multiplication. 22. YES. We show that S is closed under addition and closed under scalar multiplication. Closure under Addition: Assume that f and g belong to S. Thus, b b f (x)dx = 0 and g(x)dx = 0. a Hence, a b b (f + g)(x)dx = a b f (x)dx + a g(x)dx = 0 + 0 = 0, a so f + g ∈ S. (c)2017 Pearson Education. Inc. 340 Closure under Scalar Multiplication: Assume that f belongs to S and k is a scalar. Thus, b f (x)dx = 0. a Therefore, b b (kf )(x)dx = a b f (x)dx = k · 0 = 0, kf (x)dx = k a a so kf ∈ S. 23. YES. We show that S is closed under addition and scalar multiplication. Closure under Addition: Assume that ⎡ a1 ⎣ c1 e1 ⎤ ⎡ b1 a2 d1 ⎦ , ⎣ c 2 f1 e2 ⎤ b2 d2 ⎦ ∈ S. f2 Thus, a 1 + b1 = c 1 + f 1 , a1 − c 1 = e 1 − f 1 − d 1 , a 2 + b2 = c 2 + f 2 , a2 − c 2 = e 2 − f 2 − d 2 . Hence, (a1 + a2 ) − (c1 + c2 ) = (e1 + e2 ) − (f1 + f2 ) − (d1 + d2 ). (a1 + a2 ) + (b1 + b2 ) = (c1 + c2 ) + (f1 + f2 ) and Thus, ⎤ ⎡ b2 a1 + a2 d2 ⎦ = ⎣ c 1 + c 2 f2 e1 + e2 ⎡ a1 ⎣ c1 e1 ⎤ ⎡ b1 a2 d1 ⎦ + ⎣ c 2 f1 e2 ⎤ b 1 + b2 d1 + d2 ⎦ ∈ S, f1 + f2 which means S is closed under addition. Closure under Scalar Multiplication: Assume that ⎡ a ⎣ c e ⎤ b d ⎦∈S f and k is a scalar. Then we have a + b = c + f and a − c = e − f − d. Thus, ka + kb = kc + kf and ka − kc = ke − kf − kd. Therefore, ⎤ ⎡ ⎡ ⎤ a b ka kb k ⎣ c d ⎦ = ⎣ kc kd ⎦ ∈ S, e f ke kf so S is closed under scalar multiplication. 24. (a) NO, (b) YES. Since 3 vectors are required to span R3 , S cannot span V . However, since the vectors are not proportional, they are linearly independent. ⎡ ⎤ 6 −3 2 1 1 ⎦, 25. (a) YES, (b) YES. If we place the three vectors into the columns of a 3 × 3 matrix ⎣ 1 1 −8 −1 we observe that the matrix is invertible. Hence, its columns are linearly independent. Since we have 3 linearly independent vectors in the 3-dimensional space R3 , we have a basis for R3 . (c)2017 Pearson Education. Inc. 341 26. (a) NO, (b) YES. Since we have only 3 vectors in a 4-dimensional vector space, they cannot possibly span R4 . To check linear independence, we place the vectors into the columns of a matrix: ⎤ ⎡ 6 1 1 ⎢ −3 1 −8 ⎥ ⎥ ⎢ ⎣ 2 1 −1 ⎦ . 0 0 0 The rows for the same invertible matrix as in the previous problem, so the reduced row-echelon form ⎡ first three ⎤ 1 0 0 ⎢ 0 1 0 ⎥ ⎥ is ⎢ ⎣ 0 0 1 ⎦, so there are no free variables, and hence, the column vectors form a linearly independent 0 0 0 set. 27. (a) YES, (b) NO. Since (0, 0, 0) is a member of S, S cannot be linearly independent. However, the set {(10, −6, 5), (3, −3, 2), (6, 4, −1)} is linearly independent (these vectors can be made to form a 3 × 3 matrix that is invertible), and thus, S must span at least a 3-dimensional space. Since dim[R3 ] = 3, we know that S spans R3 . 28. (a) YES, (b) YES. Consider the linear equation c1 (2x − x3 ) + c2 (1 + x + x2 ) + 3c3 + c4 x = 0. Then (c2 + 3c3 ) + (2c1 + c2 + c4 )x + c2 x2 − c1 x3 = 0. From the latter equation, we see that c1 = c2 = 0 (looking at the x2 and x3 coefficients) and thus, c3 = c4 = 0 (looking at the constant term and x coefficient). Thus, c1 = c2 = c3 = c4 = 0, and hence, S is linearly independent. Since we have four linearly independent vectors in the 4-dimensional vector space P3 (R), we conclude that these vectors also span P3 (R). 29. (a) NO, (b) NO. The set S only contains four vectors, although dim[P4 (R)] = 5, so it is impossible for S to span P4 (R). Alternatively, simply note that none of the polynomials in S contain an x3 term. To check linear independence, consider the equation c1 (x4 + x + 2 + 1) + c2 (x2 + x + 1) + c3 (x + 1) + c4 (x4 + 2x + 3) = 0. Rearranging this, we have (c1 + c4 )x4 + (c1 + c2 )x2 + (c2 + c3 + 2c4 )x + (c1 + c2 + 3c4 ) = 0, and so we look to solve the linear system with augmented matrix ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 1 1 1 3 0 1 1 1 3 0 1 1 1 3 0 ⎢ 0 1 1 2 0 ⎥ ⎢ 0 ⎢ ⎥ 1 1 2 0 ⎥ ⎢ ⎥∼⎢ ⎥ = ⎢ 0 1 1 2 0 ⎥, ⎣ 1 1 0 0 0 ⎦ ⎣ 0 0 −1 −3 0 ⎦ ⎣ 0 0 1 3 0 ⎦ 1 0 0 1 0 0 −1 −1 −2 0 0 0 0 0 0 where we have added −1 times the first row to each of the third and fourth rows in the first step, and zeroed at the last row in the second step. We see that c4 is a free variable, which means that a nontrivial solution to the linear system exists, and therefore, the original vectors are linearly dependent. (c)2017 Pearson Education. Inc. 342 30. (a) NO, (b) YES. The vector space M2×3 (R) is 6-dimensional, and since only four vectors belong to S, S cannot possibly span M2×3 (R). On the other hand, if we form the linear system with augmented matrix ⎡ ⎤ −1 3 −1 −11 0 ⎢ 0 2 −2 −6 0 ⎥ ⎢ ⎥ ⎢ 0 1 −3 −5 0 ⎥ ⎢ ⎥, ⎢ 0 1 3 1 0 ⎥ ⎢ ⎥ ⎣ 1 2 2 −2 0 ⎦ 1 3 1 −5 0 row reduction shows that each column of a row-echelon form contains a pivot, and therefore, the vectors are linearly independent. 31. (a) NO, (b) NO. Since M2 (R) is only 4-dimensional and S contains 5 vectors, S cannot possibly be linearly independent. Moreover, each matrix in S is symmetric, and therefore, only symmetric matrices can be found in the span(S). Thus, S fails to span M2 (R). 32. Assume that {v1 , v2 , v3 } is linearly independent, and that v4 does not lie in span{v1 , v2 , v3 }. We will show that {v1 , v2 , v3 , v4 } is linearly independent. To do this, assume that c1 v1 + c2 v2 + c3 v3 + c4 v4 = 0. We must prove that c1 = c2 = c3 = c4 = 0. If c4 = 0, then we rearrange the above equation to show that v4 = − c1 c2 c3 v1 − v2 − v3 , c4 c4 c4 which implies that v4 ∈ span{v1 , v2 , v3 }, contrary to our assumption. Therefore, we know that c4 = 0. Hence the equation above reduces to c1 v1 + c2 v2 + c3 v3 = 0, and the linear independence of {v1 , v2 , v3 } now implies that c1 = c2 = c3 = 0. Therefore, c1 = c2 = c3 = c4 = 0, as required. 33. Write A = [a1 , a2 , . . . , an ]. We first assume that AT A is invertible, and we show that {a1 , a2 , . . . , an } is linearly independent. Suppose that c1 a1 + c2 a2 + · · · + cn an = 0. ⎡ ⎤ c1 ⎢ c2 ⎥ ⎢ ⎥ We can write this as Ac = 0, where c = ⎢ . ⎥. Left multiplying both sides of Ac = 0 by AT , we obtain ⎣ .. ⎦ cn AT Ac = AT 0 = 0. From the assumption that AT A is invertible, it therefore follows at once that c = 0. That is, c1 = c2 = · · · = cn = 0, which shows that {a1 , a2 , . . . , an } is linearly independent. Conversely, assume that the columns of A are linearly independent. To show that AT A is invertible, we will show that nullspace(AT A) = {0}. To do this, suppose that AT Ax = 0. We will show that x = 0. To do this, left multiply both sides of AT Ax = 0 by xT to get xT AT Ax = xT 0 = 0, (c)2017 Pearson Education. Inc. 343 or (Ax)T (Ax) = 0. This means that the sum⎡ of the ⎤ squares of the entries of the column vector Ax is zero. This means that x1 ⎢ x2 ⎥ ⎢ ⎥ Ax = 0. If we write x = ⎢ . ⎥, then ⎣ .. ⎦ xn x1 a1 + x2 a2 + · · · + xn an = 0. Since the columns of A are linearly independent, we conclude that x1 = x2 = · · · = xn = 0. Thus, x = 0, as desired. 34. (a). Clearly, S is a nonempty subset of M4 (R), so we can apply Theorem 4.3.2 and verify that S is closed under addition and scalar multiplication: Closure under Addition: Suppose that A and B belong to S. This implies that AT = −A and T B = −B. Therefore, (A + B)T = AT + B T = (−A) + (−B) = −(A + B), which proves that A + B belongs to S. Closure under Scalar Multiplication: Let k be a real constant, and let A be as above. Then (k · A)T = k · AT = k · (−A) = −(k · A), which proves that k · A belongs to S. ⎤ 0 a b c ⎢ −a 0 d e ⎥ ⎥. Since we have six free variables, we know (b). A typical element in S has the form ⎢ ⎣ −b −d 0 f ⎦ −c −e −f 0 that dim[S] = 6. To obtain a basis, we will use the notation Eij to denote the matrix whose (i, j)-element is 1, and all other entries are 0. Then we obtain the following basis for S by extracting the free variables: ⎡ {E12 − E21 , E13 − E31 , E14 − E41 , E23 − E32 , E24 − E42 , E34 − E43 }. (c). Since dim[M4 (R)] = 16 and dim[S] = 6, we need 10(= 16 − 6) additional matrices in M4 (R) to extend the basis from part (b). Each matrix added must be outside of S and must not be expressible in terms of any of the other matrices in the basis. There are many acceptable ways to do this, but the most natural is to add the following 10 matrices: E12 , E13 , E14 , E23 , E24 , E34 , E11 , E22 , E33 , E44 . 35. (a). Clearly, S is a nonempty subset of M4 (R), so we can apply Theorem 4.3.2 and verify that S is closed under addition and scalar multiplication: Closure under Addition: Suppose that A and B belong to S. Write the elements A and B in general form as ⎡ ⎤ a1 b1 c1 −(a1 + b1 + c1 ) ⎢ ⎥ d1 e1 f1 −(d1 + e1 + f1 ) ⎥ A=⎢ ⎣ ⎦ g1 h1 i1 −(g1 + h1 + i1 ) −(a1 + d1 + g1 ) −(b1 + e1 + h1 ) −(c1 + f1 + i1 ) a1 + b1 + c1 + d1 + e1 + f1 + g1 + h1 + i1 (c)2017 Pearson Education. Inc. 344 and ⎡ a2 ⎢ d2 B=⎢ ⎣ g2 −(a2 + d2 + g2 ) b2 e2 h2 −(b2 + e2 + h2 ) c2 f2 i2 −(c2 + f2 + i2 ) ⎤ −(a2 + b2 + c2 ) ⎥ −(d2 + e2 + f2 ) ⎥. ⎦ −(g2 + h2 + i2 ) a 2 + b2 + c 2 + d 2 + e 2 + f 2 + g 2 + h 2 + i 2 Using these forms for A and B, it is easy to see that the entries in A + B will have all rows and all columns summing to zero. Hence, A + B belongs to S. Closure under Scalar Multiplication: Let k be a real constant, and let A be as above. Then it is easy to see by using the form of A given above that, upon scaling all entries through by the constant k, the new matrix k · A still has row sums and column sums of zero. Thus, k · A belongs to S. (b). Using the typical form of elements of S (see the matrix A in part (a)), we have nine free variables, so that dim[S] = 9. To obtain a basis, we will use the notation Eij to denote the matrix whose (i, j)-element is 1, and all other entries are 0. Then we obtain the following basis for S by extracting the free variables: {E11 − E14 − E41 + E44 , E12 − E14 − E42 + E44 , E13 − E14 − E43 + E44 , E21 − E24 − E41 + E44 , E22 − E24 − E42 + E44 , E23 − E24 − E43 + E44 , E31 − E34 − E41 + E44 , E32 − E34 − E42 + E44 , E33 − E34 − E43 + E44 }. (c). Since dim[M4 (R)] = 16 and dim[S] = 9, we need 7(= 16 − 9) additional matrices in M4 (R) to extend the basis from part (b). Each matrix added must be outside of S and must not be expressible in terms of any of the other matrices in the basis. There are many acceptable ways to do this, but one choice is E11 , E12 , E21 , E13 , E31 , E14 , E41 . 36. (a). We must verify the axioms (A1)-(A10) for a vector space: Axiom (A1): Assume that (v1 , w1 ) and (v2 , w2 ) belong to V ⊕ W . Since v1 +V v2 ∈ V and w1 +W w2 ∈ W , then the sum (v1 , w1 ) + (v2 , w2 ) = (v1 +V v2 , w1 +W w2 ) lies in V ⊕ W . Axiom (A2): Assume that (v, w) belongs to V ⊕ W , and let k be a scalar. Since k ·V v ∈ V and k ·W w ∈ W , the scalar multiplication k · (v, w) = (k ·V v, k ·W w) lies in V ⊕ W . For the remainder of the axioms, we will omit the ·V and ·W notations. They are to be understood. Axiom (A3): Assume that (v1 , w1 ), (v2 , w2 ) ∈ V ⊕ W . Then (v1 , w1 ) + (v2 , w2 ) = (v1 + v2 , w1 + w2 ) = (v2 + v1 , w2 + w1 ) = (v2 , w2 ) + (v1 , w1 ), as required. Axiom (A4): Assume that (v1 , w1 ), (v2 , w2 ), (v3 , w3 ) ∈ V ⊕ W . Then ((v1 , w1 ) + (v2 , w2 )) + (v3 , w3 ) = (v1 + v2 , w1 + w2 ) + (v3 , w3 ) = ((v1 + v2 ) + v3 , (w1 + w2 ) + w3 ) = (v1 + (v2 + v3 ), w1 + (w2 + w3 )) = (v1 , w1 ) + (v2 + v3 , w2 + w3 ) = (v1 , w1 ) + ((v2 , w2 ) + (v3 , w3 )), (c)2017 Pearson Education. Inc. 345 as required. Axiom (A5): We claim that the zero vector in V ⊕ W is (0V , 0W ), where 0V is the zero vector in the vector space V and 0W is the zero vector in the vector space W . To check this, let (v, w) ∈ V ⊕ W . Then (0V , 0W ) + (v, w) = (0V + v, 0W + w) = (v, w), which confirms that (0V , 0W ) is the zero vector for V ⊕ W . Axiom (A6): We claim that the additive inverse of the vector (v, w) ∈ V ⊕ W is the vector (−v, −w), where −v is the additive inverse of v in the vector space V and −w is the additive inverse of w in the vector space W . We check this: (v, w) + (−v, −w) = (v + (−v), w + (−w)) = (0V , 0W ), as required. Axiom (A7): For every vector (v, w) ∈ V ⊕ W , we have 1 · (v, w) = (1 · v, 1 · w) = (v, w), where in the last step we have used the fact that Axiom (A7) holds in each of the vector spaces V and W . Axiom (A8): Let (v, w) be a vector in V ⊕ W , and let r and s be scalars. Using the fact that Axiom (A8) holds in V and W , we have (rs)(v, w) = ((rs)v, (rs)w) = (r(sv), r(sw)) = r(sv, sw) = r(s(v, w)). Axiom (A9): Let (v1 , w1 ) and (v2 , w2 ) be vectors in V ⊕ W , and let r be a scalar. Then r((v1 , w1 ) + (v2 , w2 )) = r(v1 + v2 , w1 + w2 ) = (r(v1 + v2 ), r(w1 + w2 )) = (rv1 + rv2 , rw1 + rw2 ) = (rv1 , rw1 ) + (rv2 , rw2 ) = r(v1 , w1 ) + r(v2 , w2 )), as required. Axiom (A10): Let (v, w) ∈ V ⊕ W , and let r and s be scalars. Then (r + s)(v, w) = ((r + s)v, (r + s)w) = (rv + sv, rw + sw) = (rv, rw) + (sv, sw) = r(v, w) + s(v, w), as required. (b). We show that {(v, 0) : v ∈ V } is a subspace of V ⊕ W , by checking closure under addition and closure under scalar multiplication: Closure under Addition: Suppose (v1 , 0) and (v2 , 0) belong to {(v, 0) : v ∈ V }, where v1 , v2 ∈ V . Then (v1 , 0) + (v2 , 0) = (v1 + v2 , 0) ∈ {(v, 0) : v ∈ V }, (c)2017 Pearson Education. Inc. 346 which shows that the set is closed under addition. Closure under Scalar Multiplication: Suppose (v, 0) ∈ {(v, 0) : v ∈ V } and k is a scalar. Then k(v, 0) = (kv, 0) is again in the set. Thus, {(v, 0) : v ∈ V } is closed under scalar multiplication. Therefore, {(v, 0) : v ∈ V } is a subspace of V ⊕ W . (c). Let {v1 , v2 , . . . , vn } be a basis for V , and let {w1 , w2 , . . . , wm } be a basis for W . We claim that S = {(vi , 0) : 1 ≤ i ≤ n} ∪ {(0, wj ) : 1 ≤ j ≤ m} is a basis for V ⊕ W . To show this, we will verify that S is a linearly independent set that spans V ⊕ W : Check that S is linearly independent: Assume that c1 (v1 , 0) + c2 (v2 , 0) + · · · + cn (vn , 0) + d1 (0, w1 ) + d2 (0, w2 ) + · · · + dm (0, wm ) = (0, 0). We must show that c1 = c2 = · · · = cn = d1 = d2 = · · · = dm = 0. Adding the vectors on the left-hand side, we have (c1 v1 + c2 v2 + · · · + cn vn , d1 w1 + d2 w2 + · · · + dm wm ) = (0, 0), so that c1 v 1 + c 2 v2 + · · · + cn vn = 0 and d1 w1 + d2 w2 + · · · + dm wm = 0. Since {v1 , v2 , . . . , vn } is linearly independent, c1 = c2 = · · · = cn = 0, and since {w1 , w2 , . . . , wm } is linearly independent, d1 = d2 = · · · = dm = 0. Thus, S is linearly independent. Check that S spans V ⊕ W : Let (v, w) ∈ V ⊕ W . We must express (v, w) as a linear combination of the vectors in S. Since {v1 , v2 , . . . , vn } spans V , there exist scalars c1 , c2 , . . . , cn such that v = c1 v1 + c2 v2 + · · · + cn vn , and since {w1 , w2 , . . . , wm } spans W , there exist scalars d1 , d2 , . . . , dm such that w = d1 w 1 + d2 w 2 + · · · + dm w m . Then (v, w) = c1 (v1 , 0) + c2 (v2 , 0) + · · · + cn (vn , 0) + d1 (0, w1 ) + d2 (0, w2 ) + · · · + dm (0, wm ). Therefore, (v, w) is a linear combination of vectors in S, so S spans V ⊕ W . Therefore, S is a basis for V ⊕ W . Since S contains n + m vectors, dim[V ⊕ W ] = n + m. 37. There are many examples here. One such example is S = {x3 , x3 − x2 , x3 − x, x3 − 1}, a basis for P3 (R) whose vectors all have degree 3. To see that S is a basis, note that S is linearly independent, since if c1 x3 + c2 (x3 − x2 ) + c3 (x3 − x) + c4 (x3 − 1) = 0, (c)2017 Pearson Education. Inc. 347 then (c1 + c2 + c3 + c4 )x3 − c2 x2 − c3 x − c4 = 0, and so c1 = c2 = c3 = c4 = 0. Since S is a linearly independent set of 4 vectors and dim[P3 (R)] = 4, S is a basis for P3 (R). 38. Let A be an m × n matrix. By the Rank-Nullity Theorem, dim[colspace(A)] + dim[nullspace(A)] = n. Since, by assumption, colspace(A) = nullspace(A) = r, n = 2r must be even. 39. The ith row of the matrix A is bi (c1 c2 . . . cn ). Therefore, each row of A is a multiple of the first row, and so rank(A) = 1. Thus, by the Rank-Nullity Theorem, nullity(A) = n − 1. 1 2 40. A row-echelon form of A is given by . Thus, a basis for the rowspace of A is given by {(1, 2)}, 0 0 −3 , and a basis for the nullspace of A is given by a basis for the columnspace of A is given by −6 −2 . All three subspaces are one-dimensional. 1 ⎡ ⎤ 1 −6 −2 0 1 1/3 5/21 ⎦. Thus, a basis for the rowspace of A is 41. A row-echelon form of A is given by ⎣ 0 0 0 0 0 ⎧⎡ ⎤ ⎡ ⎤⎫ 6 ⎬ ⎨ −1 5 )}, and a basis for the column space of A is given by ⎣ 3 ⎦ , ⎣ 3 ⎦ . For {(1, −6, −2, 0), (0, 1, 13 , 21 ⎩ ⎭ 7 21 the nullspace, we observe that the equations corresponding to the row-echelon form of A can be written as x − 6y − 2z = 0 and 5 1 y + z + w = 0. 3 21 5 t and x = − 10 Set w = t and z = s. Then y = − 13 s − 21 7 t. Thus, nullspace(A) = − 10 1 5 t, − s − t, s, t 7 3 21 : s, t ∈ R = 10 5 1 t − , − , 0, 1 + s 0, − , 1, 0 . 7 21 3 5 1 Hence, a basis for the nullspace of A is given by {(− 10 7 , − 21 , 0, 1), (0, − 3 , 1, 0)}. ⎤ ⎡ 1 −2.5 −5 ⎢ 0 1 1.3 ⎥ ⎥. Thus, a basis for the rowspace of A is given by 42. A row-echelon form of A is given by ⎢ ⎣ 0 0 1 ⎦ 0 0 0 {(1, −2.5, −5), (0, 1, 1.3), (0, 0, 1)}, and a basis for the columnspace of A is given by ⎧⎡ ⎤⎫ ⎤ ⎡ ⎤ ⎡ 0 3 ⎪ −4 ⎪ ⎪ ⎪ ⎨⎢ ⎥ ⎢ ⎥ ⎢ ⎥⎬ ⎢ 0 ⎥ , ⎢ 10 ⎥ , ⎢ 13 ⎥ . ⎣ 6 ⎦ ⎣ 5 ⎦ ⎣ 2 ⎦⎪ ⎪ ⎪ ⎪ ⎩ ⎭ 10 −2 5 Moreover, we have that the nullspace of A is 0-dimensional, and so the basis is empty. (c)2017 Pearson Education. Inc. 348 ⎤ 1 0 2 2 1 ⎢ 0 1 −1 −4 −3 ⎥ ⎥. Thus, a basis for the rowspace of A is 43. A row-echelon form of A is given by ⎢ ⎣ 0 0 1 4 3 ⎦ 0 0 0 1 0 given by {(1, 0, 2, 2, 1), (0, 1, −1, −4, −3), (0, 0, 1, 4, 3), (0, 0, 0, 1, 0)}, ⎡ a basis for the columnspace of A is given by ⎧⎡ ⎤ ⎡ ⎤⎫ ⎤ ⎡ ⎤ ⎡ 5 2 ⎪ 3 5 ⎪ ⎪ ⎪ ⎨⎢ ⎥ ⎢ ⎥⎬ ⎥ ⎢ ⎥ ⎢ ⎢ 1 ⎥,⎢ 0 ⎥,⎢ 2 ⎥,⎢ 2 ⎥ . ⎣ 1 ⎦ ⎣ 1 ⎦ ⎣ 1 ⎦ ⎣ −2 ⎦⎪ ⎪ ⎪ ⎪ ⎭ ⎩ 0 −2 −2 −4 For the nullspace, if the variables corresponding the columns are (x, y, z, u, v), then the row-echelon form tells us that v = t is a free variable, u = 0, z = −3t, y = 0, and x = 5t. Thus, nullspace(A) = {(5t, 0, −3t, 0, t) : t ∈ R} = {t(5, 0, −3, 0, 1) : t ∈ R}, and so a basis for the nullspace of A is {(5, 0, −3, 0, 1)}. 44. Any of the conditions (a)-(p) appearing in the Invertible Matrix Theorem would be appropriate at this point in the text. Chapter 5 Solutions Solutions to Section 5.1 (a): FALSE. Many counterexamples exist. For instance, the vectors v = (1, 1) and w = (1, 0) in R2 are linearly independent, but they are not orthogonal. (b): FALSE. We have kv, kw = kv, kw = kkw, v = k 2 w, v = k 2 v, w, where we have used the axioms of an inner product to carry out these steps. Therefore, the result conflicts with the given statement, which must therefore be a false statement. (c): TRUE. We have c1 v1 + c2 v2 , w = c1 v1 , w + c2 v2 , w = c1 v1 , w + c2 v2 , w = c1 · 0 + c2 · 0 = 0. (d): TRUE. We have x + y, x − y = x, x − x, y + y, x − y, y = x, x − y, y = ||x||2 − ||y||2 . This will be negative if and only if ||x||2 < ||y||2 , and since ||x|| and ||y|| are nonnegative real numbers, ||x||2 < ||y||2 if and only if ||x|| < ||y||. (e): FALSE. For example, if V is the inner product space of integrable functions on (−∞, ∞), then the formula b f, g = f (t)g(t)dt a (c)2017 Pearson Education. Inc. 349 is a valid any product for any choice of real numbers a < b. See also Problems 18 and 27 in this section, in which a “non-standard” inner product on R2 is given. (f ): TRUE. The angle between the vectors −2v and −2w is cos θ = (−2v) · (−2w) (−2)2 (v · w) v·w = = , || − 2v|||| − 2w|| (−2)2 ||v||||w|| ||v||||w|| and this is the angle between the vectors v and w. (g): FALSE. This definition of p, q will not satisfy the requirements of an inner product. For instance, if we take p = x, then p, p = 0, but p = 0. Problems: π 1. v, w = 0, so that cos θ = 0. That is, θ = rad. 2 √ √ 2. v, w = 8, ||v|| = 3 3, ||w|| = 7. Hence, cos θ = π π 3. f, g = 0 v, w 8 = √ =⇒ θ ≈ 0.95 radians. ||v||||w|| 3 21 x sin xdx = π, ||f || = 2 sin xdx 0 3 π π π 2 . Hence, = x dx = , ||g|| = 2 3 0 √ π 6 cos θ = 1/2 3 1/2 = ≈ 0.68 radians. π π π 2 3 4. π/2 f, g = π/2 sin x(2 cos x+4)dx = 2 0 π/2 sin x cos xdx+4 0 0 ||f || = f, f = ||g|| = g, g = 0 (4 cos2 x + 16 cos x + 16)dx = 0 cos θ = √ b π/2 sin2 xdx = π/2 Thus, 5. π/2 sin xdx = (sin2 x)|0 −4 cos x|0 π π √ 4 · + 16 · 1 + 16 · = 9π + 16. 4 2 5·2 √ =⇒ θ ≈ 0.559 rad. π · 9π + 16 b xm · xn dx = = (1−0)−4(0−1) = 5, √ π π = , 4 2 1 (bm+n+1 − am+n+1 ), m+n+1 a a b 1 2m x dx = ||f || = f, f = (b2m+1 − a2m+1 ), 2m + 1 a f, g = π/2 xm+n dx = (c)2017 Pearson Education. Inc. 350 ||g|| = g, g = b x2n dx = a Thus, θ = cos −1 1 (b2n+1 − a2n+1 ). 2n + 1 √ √ 2m + 1 2n + 1(bm+n+1 − am+n+1 ) √ √ . (m + n + 1) b2m+1 − a2m+1 b2n+1 − a2n+1 6. v, w i)(−1 − i) + (3 − 2i)(1 + 3i) + (4 + i)(3 + i) = 19 + 11i. ||v|| = = (2 + √ ||w|| = w, w = 22. √ v, v = 35. 7. v, w = (6 − 3i)(−i) + (4)(−2i) + (−2 + 5i)(−3i) + (3i)(−4i) = (−3 − 6i) + (−8i) + (15 + 6i) + 12 = 24 − 8i, √ √ √ ||v|| = v, v = (6 − 3i)(6 + 3i) + (4)(4) + (−2 + 5i)(−2 − 5i) + (3i)(−3i) = 45 + 16 + 29 + 9 = 99 = 3 11, √ √ ||w|| = w, w = (i)(−i) + (2i)(−2i) + (3i)(−3i) + (4i)(−4i) = 1 + 4 + 9 + 16 = 30. 8. We can show that this mapping fails the first property of Definition 5.1.3 with an example. Let A = 0 1 . Then the given mapping dictates that A, A = 0 even though A = 0. On the other hand, the 0 0 remaining properties (2) - (4) in Definition 5.1.3 do in fact hold. With A and B as given in the problem, k c11 c12 , we can verify this as follows: a constant, and C = c21 c22 A, B = a11 b11 = b11 a11 = B, A, kA, B = (ka11 )b11 = k(a11 b11 ) = kA, B, A + B, C = (a11 + b11 )c11 = a11 c11 + b11 c11 = A, C + B, C. 9. We can show that this mapping fails the first property of Definition 5.1.3 with an example. Let A = 0 1 . Then the given mapping dictates that A, A = det(A2 ) = det(02 ) = 0 even though A = 0. 0 0 Property (2) in Definition 5.1.3 holds, since A, B = det(AB) = (det A)(det B) = (det B)(det A) = det(BA) = B, A. In general, property (3) fails. Here is an example to illustrate this: 2I2 , I2 = det(2I2 · I2 ) = 4 but 2I2 , I2 = 2det(I2 · I2 ) = 2. Finally, for property (4), consider the example A = B = C = I2 . We have A + B, C = 2I2 , I2 = 4, but A, C + B, C = 1 + 1 = 2. Thus, property (4) fails. Hence, we conclude that, in general, only property (2) holds for the mapping in this problem. 1 0 , then according to the given mapping, we have 10. Property (1) fails. For instance, if we set A = 0 −1 A, A = −2, violating the requirement that v, v ≥ 0 for all vectors v. Now we consider the remaining properties. First, we see that property (2) in Defintion 5.1.3 holds: A, B = a11 b22 + a12 b21 + a21 b12 + a22 b11 = b11 a22 + b12 a21 + b21 a12 + b22 a11 = B, A. Next, observe that property (3) in Definition 5.1.3 holds: kA, B = (ka11 )b22 + (ka12 )b21 + (ka21 )b12 + (ka22 )b11 = k(a11 b22 + a12 b21 + a21 b12 + a22 b11 ) = kA, B. (c)2017 Pearson Education. Inc. 351 Finally, observe that property (4) in Definition 5.1.3 holds: A + B, C = (a11 + b11 )c22 + (a12 + b12 )c21 + (a21 + b21 )c12 + (a22 + b22 )c11 = (a11 c22 + a12 c21 + a21 c12 + a22 c11 ) + (b11 c22 + b12 c21 + b21 c12 + b22 c11 ) = A, C + B, C. In conclusion, only property (1) fails in this example. 11. Let A, B, C ∈ M2 (R). (1): A, A = a211 + a212 + a221 + a222 ≥ 0, and A, A = 0 ⇐⇒ a11 = a12 = a21 = a22 = 0 ⇐⇒ A = 0. (2): A, B = a11 b11 + a12 b12 + a21 b21 + a22 b22 = b11 a11 + b12 a12 + b21 a21 + b22 a22 = B, A. (3): Let k ∈ R. kA, B = ka11 b11 + ka12 b12 + ka21 b21 + ka22 b22 = k(a11 b11 + a12 b12 + a21 b21 + a22 b22 ) = kA, B. (4): (A + B), C = (a11 + b11 )c11 + (a12 + b12 )c12 + (a21 + b21 )c21 + (a22 + b22 )c22 = (a11 c11 + b11 c11 ) + (a12 c12 + b12 c12 ) + (a21 c21 + b21 c21 ) + (a22 c22 + b22 c22 ) = (a11 c11 + a12 c12 + a21 c21 + a22 c22 ) + (b11 c11 + b12 c12 + b21 c21 + b22 c22 = A, C + B, C. √ √ 12. A, B = 13, ||A|| = 33, ||B|| = 7. 13. A, B = 2 · 3 +(−1)1 + 3(−1) + 5 · 2 = 12. √ ||A|| = A, A = 2 · 2 + (−1)(−1) + 3 · 3 + 5 · 5 = 39. √ ||B|| = B, B = 3 · 3 + 1 · 1 + (−1)(−1) + 2 · 2 = 15. 14. Let p1 , p2 , p3 ∈ P1 (R) where p1 (x) = a + bx, p2 (x) = c + dx, and p3 (x) = e + f x. Define p1 , p2 = ac + bd. (14.1) The properties 1 through 4 of Definition 5.1.3 must be verified. (1): p1 , p1 = a2 + b2 ≥ 0 and p1 , p1 = 0 ⇐⇒ a = b = 0 ⇐⇒ p1 (x) = 0. (2): p1 , p2 = ac + bd = ca + db = p2 , p1 . (3): Let k ∈ R. kp1 , p2 = kac + kbd = k(ac + bd) = kp1 , p2 . (4): p1 + p2 , p3 = (a + c)e + (b + d)f = ae + ce + bf + df = (ae + bf ) + (ce + df ) = p1 , p3 + p2 , p3 . Hence, the mapping defined by (14.1) is an inner product in P1 (R). 15. NO. Property (1) in Definition 5.1.3 fails. For instance, consider the function defined by 0, if 0 ≤ x ≤ 1/2 , f (x) = x − 12 , if 1/2 ≤ x ≤ 1 . Clearly f = 0. However, f, f = 1/2 0 [f (x)]2 dx = 0. 16. YES. We verify the four properties in Definition 5.1.3. For property (1), note that 1 x[f (x)]2 dx ≥ 0, f, f = 0 (c)2017 Pearson Education. Inc. 352 since on the interval 0 ≤ x ≤ 1, we have that x[f (x)]2 ≥ 0, so a positive function is being integrated. However, in order for f, f = 0, we must have f (x) = 0 for all x ∈ [0, 1]. For property (2), we observe that f, g = 1 1 xf (x)g(x)dx = 0 0 xg(x)f (x) = g, f . For property (3), we observe that kf, g = 1 1 x(kf )(x)g(x)dx = k 0 xf (x)g(x)dx = kf, g. 0 For property (4), we observe that f + g, h = 1 1 x(f (x) + g(x))h(x)dx = 0 1 xf (x)h(x)dx + 0 0 xg(x)h(x)dx = f, h + g, h. Since properties (1) - (4) have all been verified, we conclude that the given mapping does form a valid inner product on V . 17. NO. Property (1) in Definition 5.1.3 fails. To see an explicit example, observe that x, x = 0 x3 dx = −1 1 40 1 1 x |−1 = (0 − 1) = − < 0. 4 4 4 Many other illustrations could also suffice here. 18. Property 1: u, u = 2u1 u1 + u1 u2 + u2 u1 + 2u2 u2 = 2u21 + 2u1 u2 + 2u22 = (u1 + u2 )2 + u21 + u22 ≥ 0. u, u = 0 ⇐⇒ (u1 + u2 )2 + u21 + u22 = 0 ⇐⇒ u1 = 0 = u2 ⇐⇒ u = 0. Property 2: u, v = 2u1 v1 + u1 v2 + u2 v1 + 2u2 v2 = 2u2 v2 + u2 v1 + u1 v2 + 2u1 v1 = v, u. Property 3: ku, v = k(2u1 v1 + u1 v2 + u2 v1 + 2u2 v2 ) = 2ku1 v1 + ku1 v2 + ku2 v1 + 2ku2 v2 = 2(ku1 )v1 + (ku1 )v2 + (ku2 )v1 + 2(ku2 )v2 = ku, v = 2ku1 v1 + ku1 v2 + ku2 v1 + 2ku2 v2 = 2u1 (kv1 ) + u1 (kv2 ) + u2 (kv1 ) + 2u2 (kv2 ) = u, kv. Property 4: (u + v), w = (u1 + v1 , u2 + v2 ), (w1 , w2 ) = 2(u1 + v1 )w1 + (u1 + v1 )w2 + (u2 + v2 )w1 + 2(u2 + v2 )w2 = 2u1 w1 + 2v1 w1 + u1 w2 + v1 w2 + u2 w1 + v2 w1 + 2u2 w2 + 2v2 w2 = 2u1 w1 + u1 w2 + u2 w1 + 2u2 w2 + 2v1 w1 + v1 w2 + v2 w1 + 2v2 w2 = u, w + v, w. Therefore u, v = 2u1 v1 + u1 v2 + u2 v1 + 2u2 v2 defines an inner product on R2 . 19. (a). Using the defined inner product: v, w = (1, 0), (−1, 2) = 2 · 1(−1) + 1 · 2 + 0(−1) + 2 · 0 · 2 = 0. (b). Using the standard inner product: v, w = (1, 0), (−1, 2) = 1(−1) + 0 · 2 = −1 = 0. (c)2017 Pearson Education. Inc. 353 20. (a). Using the defined inner product: v, w = 2 · 2 · 3 + 2 · 6 + (−1)3 + 2(−1)6 = 9 = 0. (b). Using the standard inner product: v, w = 2 · 3 + (−1)6 = 0. 21. (a). Using the defined inner product: v, w = 2 · 1 · 2 + 1 · 1 + (−2) · 2 + 2 · (−2) · 1 = −3 = 0. (b). Using the standard inner product: v, w = 1 · 2 + (−2)(1) = 0. 22. (a). Show symmetry: v, w = w, v. v, w = (v1 , v2 ), (w1 , w2 ) = v1 w1 − v2 w2 = w1 v1 − w2 v2 = (w1 , w2 ), (v1 , v2 ) = w, v. (b). Show kv, w = kv, w = v, kw. Note that kv = k(v1 , v2 ) = (kv1 , kv2 ) and kw = k(w1 , w2 ) = (kw1 , kw2 ). kv, w = (kv1 , kv2 ), (w1 , w2 ) = (kv1 )w1 − (kv2 )w2 = k(v1 w1 − v2 w2 ) = k(v1 , v2 ), (w1 , w2 ) = kv, w. Also, v, kw = (v1 , v2 ), (kw1 , kw2 ) = v1 (kw1 ) − v2 (kw2 ) = k(v1 w1 − v2 w2 ) = k(v1 , v2 ), (w1 , w2 ) = kv, w. (c). Show (u + v), w = u, w + v, w. Let w = (w1 , w2 ) and note that u + v = (u1 + v1 , u2 + v2 ). (u + v), w = (u1 + v1 , u2 + v2 ), (w1 , w2 ) = (u1 + v1 )w1 − (u2 + v2 )w2 = u1 w 1 + v 1 w 1 − u2 w 2 − v 2 w 2 = u1 w 1 − u2 w 2 + v 1 w 1 − v 2 w 2 = (u1 , v2 ), (w1 , w2 ) + (v1 , v2 ), (w1 , w2 ) = u, w + v, w. Property 1 fails since, for example, u, u < 0 whenever |u2 | > |u1 |. 23. v, v = 0 =⇒ (v1 , v2 ), (v1 , v2 ) = 0 =⇒ v12 − v22 = 0 =⇒ v12 = v22 =⇒ |v1 | = |v2 |. Thus, in this space, null vectors are given by {(v1 , v2 ) ∈ R2 : |v1 | = |v2 |} or equivalently, v = r(1, 1) or v = s(1, −1) where r, s ∈ R. 24. v, v < 0 =⇒ (v1 , v2 ), (v1 , v2 ) < 0 =⇒ v12 − v22 < 0 =⇒ v12 < v22 . In this space, timelike vectors are given by {(v1 , v2 ) ∈ R2 : v12 < v22 }. 25. v, v > 0 =⇒ (v1 , v2 ), (v1 , v2 ) > 0 =⇒ v12 − v22 > 0 =⇒ v12 > v22 . In this space, spacelike vectors are given by {(v1 , v2 ) ∈ R2 : v12 > v22 }. 26. This is illustrated in the figure. 27. We verify the four properties in Definition 5.1.3. For property (1), note that v, v = v12 − v1 v2 − v2 v1 + 4v22 = v12 − 2v1 v2 + 4v22 = (v1 − v2 )2 + 3v22 ≥ 0, and in order for v, v to be zero, we must have v1 − v2 = 0 and v2 = 0. That is v1 = v2 = 0, which means v = 0. This verifies property (1). For property (2), we have v, w = v1 w1 − v1 w2 − v2 w1 + 4v2 w2 = w1 v1 − w1 v2 − w2 v1 + 4w2 v2 = w, v. For property (3), we have kv, w = (kv1 )w1 − (kv1 )w2 − (kv2 )w1 + 4(kv2 )w2 = k(v1 w1 − v1 w2 − v2 w1 + 4v2 w2 ) = kv, w. (c)2017 Pearson Education. Inc. 354 y Null Vectors Null Vectors Timelike Vectors Spacelike Vectors Spacelike Vectors x Spacelike Vectors Spacelike Vectors Timelike Vectors Null Vectors Null Vectors Figure 0.0.64: Figure for Problem 26 Finally, for property (4), if we write u = (u1 , u2 ), then u + v, w = (u1 + v1 )w1 − (u1 + v1 )w2 − (u2 + v2 )w1 + 4(u2 + v2 )w2 = (u1 w1 − u1 w2 − u2 w1 + 4u2 w2 ) + (v1 w1 − v1 w2 − v2 w1 + 4v2 w2 ) = u, w + v, w, as needed. 28. Suppose that some ki ≤ 0. If ei denotes the standard basis vector in Rn with 1 in the ith position and zeros elsewhere, then the given formula yields ei , ei = ki ≤ 0, violating the first axiom of an inner product. Therefore, if the given formula defines a valid inner product, then we must have that ki > 0 for all i. Now we prove the converse. Suppose that each ki > 0. We verify the axioms of an inner product for the given proposed inner product. We have v, v = k1 v12 + k2 v22 + · · · + kn vn2 ≥ 0, and we have equality if and only if v1 = v2 = · · · = vn = 0. For the second axiom, we have w, v = k1 w1 v1 + k2 w2 v2 + · · · + kn wn vn = k1 v1 w1 + k2 v2 w2 + · · · + kn vn wn = v, w. For the third axiom, we have kv, w = k1 (kv1 )w1 + k2 (kv2 )w2 + · · · + kn (kvn )wn = k[k1 v1 w1 + k2 v2 w2 + · · · + kn vn wn ] = kv, w. Finally, for the fourth axiom, we have u + v, w = k1 (u1 + v1 )w1 + k2 (u2 + v2 )w2 + · · · + kn (un + vn )wn = [k1 u1 w1 + k2 u2 w2 + · · · + kn un wn ] + [k1 v1 w1 + k2 v2 w2 + · · · + kn vn wn ] = u, w + v, w. 29. We have v, 0 = v, 0 + 0 = 0 + 0, v = 0, v + 0, v = v, 0 + v, 0 = 2v, 0, (c)2017 Pearson Education. Inc. 355 which implies that v, 0 = 0. 30. We have u, v + w = v + w, u = v, u + w, u = v, u + w, u = u, v + u, w, as required. 31. We have u + v, w + x = u + v, w + u + v, x = u, w + v, w + u, x + v, x. 32. (a). We have ||v + w|| = = = = √ v + w, v + w ||v||2 + v, w + w, v + ||w||2 32 + (−2) + (−2) + 16 29. (b). We have 3v + w, −2v + 3w = 3v, −2v + 3v, 3w + w, −2v + w, 3w = −6||v||2 + 9v, w − 2w, v + 3||w||2 = −6 · 9 + 9 · (−2) − 2 · (−2) + 3 · 16 = −54 − 18 + 4 + 48 = −20. (c). We have −w, 5v + 2w = −w, 5v + −w, 2w = −5w, v − 2||w||2 = −5 · (−2) − 2 · 16 = −22. 33. (a). We have w − 2v, w − 2v = w, w − 2w, v − 2v, w + 4v, v = ||w||2 − 4v, w + 4||v||2 = 62 − 4 · 3 + 4 · 22 √ √ = 40 = 2 10. ||w − 2v|| = (c)2017 Pearson Education. Inc. 356 (b). We have ||2v + 5w|| = 2v + 5w, 2v + 5w 4||v||2 + 10v, w + 10w, v + 25||w||2 = 4 · 22 + 10 · 3 + 10 · 3 + 25 · 62 √ = 16 + 30 + 30 + 900 √ = 976. = (c). We have 2v − w, 4w = 8v, w − 4||w||2 = 8 · 3 − 4 · 62 = 24 − 144 = −120. 34. (a). We have u + v + w, u + v + w = ||u||2 + ||v||2 + ||w||2 + 2(u, v + u, w + v, w) = 62 + 22 + 32 + 2(4 + 5 + 0) √ = 36 + 4 + 9 + 18 √ = 47. ||u + v + w|| = (b). We have 2u − 3v − w, 2u − 3v − w = 4||u||2 + 9||v||2 + ||w||2 − 12u, v − 4u, w + 6v, w = 4 · 62 + 9 · 22 + 32 − 12 · 4 − 4 · 5 + 6 · 0 √ = 144 + 36 + 9 − 48 − 20 + 0 √ = 121 = 11. ||2u − 3v − w|| = (c). We have u + 2w, 3u − 3v + w = 3||u||2 + 2||w||2 − 3u, v + u, w + 6w, u − 6w, v = 3 · 62 + 2 · 32 − 3 · 4 + 5 + 6 · 5 − 6 · 0 = 108 + 18 − 12 + 5 + 30 = 149. 35. (a). For all v, w ∈ V . (c)2017 Pearson Education. Inc. 357 ||v + w||2 = (v + w), (v + w) = v, v + w + w, v + w by Property 4 = v + w, v + v + w, w by Property 2 = v, v + w, v + v, w + w, w by Property 4 = v, v + v, w + w, v + w, w by Property 2 = ||v||2 + 2v, w + ||w||2 . (b). This follows immediately by substituting v, w = 0 in the formula given in part (a). 36.(a). From Problem 35(a), it follows that ||v + w||2 = ||v||2 + 2v, w + ||w||2 and ||v − w||2 = ||v + 2 (−w)||2 = ||v||2 + 2v, −w + ||− w||2 = ||v||2 − 2v, w + ||w|| . 2 2 2 2 Thus, ||v + w|| − ||v − w|| = ||v|| + 2v, w + ||w|| − ||v||2 − 2v, w + ||w||2 = 4v, w. (b). + w||2 + ||v − w||2 = ||v||2 + 2v, w + ||w||2 + ||v||2 − 2v, w + ||w||2 = 2||v||2 + 2||w||2 = ||v 2 ||v||2 + ||w||2 . 37. For all v, w ∈ V and vi , wi ∈ C. ||v + w||2 = v + w, v + w = v, v + w + w, v + w by Property 4 = v + w, v + v + w, w by Property 2 = v, v + w, v + v, w + w, w by Property 4 = v, v + w, v + v, w + w, w = v, v + w, w + v, w + v, w = ||v||2 + ||w||2 + 2Re{v, w} = ||v||2 + 2Re{v, w} + ||v||2 . Solutions to Section 5.2 (a): TRUE. An orthonormal basis is simply an orthogonal basis consisting of unit vectors. (b): FALSE. The converse of this statement is true, but for the given statement, many counterexamples exist. For instance, the set of vectors {(1, 1), (1, 0)} in R2 is linearly independent, but does not form an orthogonal set. (c): TRUE. We can verify easily that π cos t sin tdt = 0 sin2 t π | = 0, 2 0 which means that {cos x, sin x} is an orthogonal set. Moreover, since they are non-proportional functions, they are linearly independent. Therefore, they comprise an orthogonal basis for the 2-dimensional inner product space span{cos x, sin x}. (d): FALSE. The vector w − P(w, v) is orthogonal to v, not w. (e): TRUE. This is the content of Theorem 5.2.7. (f ): TRUE. The vector P(w, v) is a scalar multiple of the vector v, and since v is orthogonal to u, P(w, v) is also orthogonal to u, so that its projection onto u must be 0. (c)2017 Pearson Education. Inc. 358 (g): TRUE. We have P(w1 + w2 , v) = w1 , v + w2 , v w1 , v w2 , v w1 + w2 , v v= v= v+ v = P(w1 , v) + P(w2 , v). v2 v2 v2 v2 Problems: 1. YES. Observe that (1, that √ (1, 2) is √ orthogonal to = 0, thus showing √ 2), (−4, 2) √ = (1)(−4) + (2)(2) √ (−4, 2). Since ||(1, 2)|| = 12 + 22 = 5 and ||(−4, 2)|| = (−4)2 + 22 = 16 + 4 = 20 = 2 5, we obtain the orthonormal set 1 1 √ (1, 2), √ (−4, 2) . 5 2 5 2. YES. (2, −1, 1), (1, 1, −1) = 2 + (−1) + (−1) = 0; (2, −1, 1), (0, 1, 1) = 0 + (−1) + 1 = 0; (1, 1, −1), (0, 1, 1) = 0 + 1 + (−1) = 0. Since each vector in the set is orthogonal to every other vector in the set, the vectors form an orthogonal set. To generate √ set, we divide each vector by its norm: √ an orthonormal ||(2, −1, 1)|| = √4 + 1 + 1 = √6, ||(1, 1, −1)|| =√ 1 + 1 + 1 =√ 3, and ||(0, 1, 1)|| = 0 + 1 + 1 = 2. Thus, the corresponding orthonormal set is: √ √ 6 3 2 (2, −1, 1), (1, 1, −1), (0, 1, 1) . 6 3 2 √ 3. YES. (1, 3, −1, 1), (−1, 1, 1, −1) = −1 + 3 + (−1) + (−1) = 0; (1, 3, −1, 1), (1, 0, 2, 1) = 1 + 0 + (−2) + 1 = 0; (−1, 1, 1, −1), (1, 0, 2, 1) = −1 + 0 + 2 + (−1) = 0. Since all vectors in the set are orthogonal to each other, they form an orthogonal set. To generate an orthonormal set, √ we divide each vector √ by its √ norm: ||(1, 3, −1, 1)|| = √ 1+9+1+1= √ 12 = 2 3, ||(−1, 1, 1, −1)||√= 1 + 1 + 1 + 1√= 4 = 2, and ||(1, 0, 2, 1)|| = 1 + 0 + 4 + 1 = 6. Thus, an orthonormal set is: √ 3 6 1 (1, 3, −1, 1), (−1, 1, 1, −1), (1, 0, 2, 1) . 6 2 6 √ 4. NO. (1, 2, −1, 0), (1, 0, 1, 2) = 1 + 0 + (−1) + 0 = 0; (1, 2, −1, 0), (−1, 1, 1, 0) = −1 + 2 + (−1) + 0 = 0; (1, 2, −1, 0), (1, −1, −1, 0) = 1 + (−2) + 1 + 0 = 0; (1, 0, 1, 2), (−1, 1, 1, 0) = −1 + 0 + 1 + 0 = 0; (1, 0, 1, 2), (1, −1, −1, 0) = 1 + 0 + (−1) + 0 = 0; (−1, 1, 1, 0), (1, −1, −1, 0) = −1 + (−1) + (−1) = −3. Hence, this is not an orthogonal set of vectors. 5. YES. (1, 2, −1, 0, 3), (1, 1, 0, 2, −1) = 1 + 2 + 0 + 0 + (−3) = 0; (1, 2, −1, 0, 3), (4, 2, −4, −5, −4) = 4 + 4 + 4 + 0 + (−12) = 0; (1, 1, 0, 2, −1), (4, 2, −4, −5, −4) = 4 + 2 + 0 + (−10) + 4 = 0. Since all vectors in the set are orthogonal to each other, they form an orthogonal set. To generate an (c)2017 Pearson Education. Inc. 359 orthonormal set, we√divide each vector by√its norm: ||(1, 2, −1, 0, 3)|| = √1 + 4 + 1 + 0 + 9 = √15, ||(1, 1, 0, 2, −1)|| = 1 √ + 1 + 0 + 4 + 1 = 7, and√ ||(4, 2, −4, −5, −4)|| = 16 + 4 + 16 + 25 + 16 = 77. Thus, an orthonormal set is: √ √ √ 15 7 77 (1, 2, −1, 0, 3), (1, 1, 0, 2, −1), (4, 2, −4, −5, −4) . 15 7 77 6. Write the nonzero vector w as w = (a, b). In order for w to be orthogonal to v, it is necessary and sufficient that v, w = 0; that is, 7a − 2b = 0, or b = 72 a. That is, any vector of the form w = (a, 72 a), where a = 0, satisfies the given requirement. Therefore, any nonzero multiple w of the vector (1, 72 ) will create the orthogonal set {v, w}. 7. The vectors w = (a, b, c) in R3 that are orthogonal to v must satisfy (−3, 6, 1), (a, b, c) = 0; that is, −3a + 6b + c = 0. Therefore, all vectors lying on the plane −3a + 6b + c = 0 in R3 are orthogonal to v. To obtain a basis for R3 , we need a set of three vectors. Along with v, we will add two additional vectors that lie on the plane to form the basis. In order to ensure that we arrive at an orthogonal basis, the two vectors we choose on the plane must be orthogonal to one another. There are many choices we can make. For instance, we can take w1 = (1, 0, 3) and w2 = (−3, − 53 , 1). Thus, one orthogonal basis for R3 that includes v is {(−3, 6, 1), (1, 0, 3), (−3, − 53 , 1)}. 8. We require that v1 , v2 = v1 , w = v2 , w = 0. Let w = (a, b, c) where a, b, c ∈ R. v1 , v2 = (1, 2, 3), (1, 1, −1) = 0. v1 , w = (1, 2, 3), (a, b, c) =⇒ a + 2b + 3c = 0. v2 , w = (1, 1, −1), (a, b, c) =⇒ a + b − c = 0. Letting the free variable c = t ∈ R, the system has the solution a = 5t, b = −4t, and c = t. Consequently, {(1, 2, 3), (1, 1, −1), (5t, −4t, t)} will form an orthogonal set whenever t = 0. To determine the corresponding orthonormal √ √ √ divide each √ √ √ set, we must √ vector by its√norm: ||v1 || = 1 + 4 + 9 = 14, ||v2 || = 1 + 1 + 1 = 3, ||w|| = 25t2 + 16t2 + t2 = 42t2 = |t| 42 = t 42 if t ≥ 0. Setting t = 1, an orthonormal set is: √ √ √ 14 3 42 (1, 2, 3), (1, 1, −1), (5, −4, 1) . 14 3 42 9. The vectors w = (a, b, c, d) in R4 that are orthogonal to v1 and v2 must satisfy (−4, 0, 0, 1), (a, b, c, d) = 0 and (1, 2, 3, 4), (a, b, c, d) = 0; that is −4a + d = 0 and a + 2b + 3c + 4d = 0. Solving this system of two equations by the usual process of Gaussian elimination, we find that d = t and c = s 1 are free variables. Then b = − 32 s− 17 8 t and a = 4 t. Thus, the set of vectors that are orthogonal to both v1 and 1 v2 are precisely those vectors of the form ( 4 t, − 32 s− 17 8 t, s, t), where s and t are real numbers. We must obtain two vectors of this form that are orthogonal to one another as well. By choosing t = 0, we obtain one vector as (0, − 32 , 1, 0). Taking the inner product of this vector with a general vector of the type we must choose, we 51 3 obtain 13 4 s + 16 t, so for the next vector to be orthogonal to (0, − 2 , 1, 0), we must choose s and t to satisfy 13 51 4 s+ 16 t = 0. One solution is s = −51 and t = 52. Hence, we take our second vector to be (13, −34, −51, 52). The orthogonal basis we have therefore obtained is {v1 , v2 , (0, − 32 , 1, 0), (13, −34, −51, 52)}. 10. We have (2 + i, −5i), (−2 − i, −i) = (2 + i)(−2 + i) + (−5i)(i) = −4 − 1 + 5 = 0. (c)2017 Pearson Education. Inc. 360 Thus, the two given vectors form an orthogonal set in C2 . To obtain an orthonormal set, we must normalize the given vectors. Note that √ √ ||(2 + i, −5i)|| = (2 + i, −5i), (2 + i, −5i) = (2 + i)(2 − i) + (−5i)(5i) = 5 + 25 = 30 and ||(−2 − i, −i)|| = √ √ (−2 − i, −i), (−2 − i, −i) = (−2 − i)(−2 + i) + (−i)(i) = 5 + 1 = 6. Thus, we obtain the orthonormal set 1 1 √ (2 + i, −5i), √ (−2 − i, −i) . 30 6 11. (1 − i, 3 + 2i), (2 + 3i, 1 − i) = (1 − i)(2 + 3i) + (3 + 2i)(1 − i) = (1 − i)(2 − 3i) + (3 + 2i)(1 + i) = (2 − 3i − 2i − 3) + (3 + 3i + 2i − 2) = (−1 − 5i) + (1 + √ 5i) = 0. The vectors √ are orthogonal. ||(1 − i, 3 + 2i)|| = (1 − i)(1 + i) + (3 + 2i)(3 − 2i) = 1 + 1 + 9 + 4 = 15. √ √ ||(2 + 3i, 1 − i)|| = (2 + 3i)(2 − 3i) + (1 − i)(1 + i) = 4 + 9 + 1 + 1 = 15. Thus, the corresponding orthonormal set is: √ √ 15 15 (1 − i, 3 + 2i), (2 + 3i, 1 − i) . 15 15 12. (1 − i, 1 + i, i), (0, i, 1 − i) = (1 − i) · 0 + (1 + i)(−i) + i(1 + i) = 0. (1 − i, 1 + i, i), (−3 + 3i, 2 + 2i, 2i) = (1 − i)(−3 − 3i) + (1 + i)(2 − 2i) + i(−2i) = 0. (0, i, 1 − i), (−3 + 3i, 2 + 2i, 2i) = 0 + i(2 − 2i) + (1 − i)(−2i) = (2i + 2) + (−2i − 2) = 0. Hence, the vectors are orthogonal. To obtain a corresponding orthonormal set, we divide each vector by its norm. √ √ ||(1 − i, 1 + i, i)||= (1 − i)(1 + i) + (1 + i)(1 − i) + i(−i) = 1 + 1 + 1 + 1 + 1 = 5. √ √ ||(0, i, 1 − i)|| = 0 + i(−i)+ (1 − i)(1 + i) = 1 + 1 + 1 = 3. √ √ ||(−3 + 3i, 2 + 2i, 2i)|| = (−3 + 3i)(−3 − 3i) + (2 + 2i)(2 − 2i) + 2i(−2i) = 9 + 9 + 4 + 4 + 4 = 30. Consequently, an orthonormal set is: √ √ √ 5 3 30 (1 − i, 1 + i, i), (0, i, 1 − i), (−3 + 3i, 2 + 2i, 2i) . 5 3 30 13. Let z = a + bi where a, b ∈ R. We require that v, w = 0. (1 − i, 1 + 2i), (2 + i, a + bi) = 0 =⇒ (1 − i)(2 − i) + (1 + 2i)(a − bi) = 0 =⇒ 1 − 3i + a + 2b + (2a − b)i = 0. a + 2b = −1 Equating real parts and imaginary parts from the last equality results in the system: 2a − b = 3. This system has the solution a = 1 and b = −1; hence z = 1 − i. Our desired orthogonal set is given by {(1 − i, 1 + 2i), (2 + i, 1 − i)}. √ √ ||(1 − i, 1 + 2i)|| = (1 − i)(1 + i) + (1 + 2i)(1 − 2i) = 1 + 1 + 1 + 4 = 7. √ √ ||(2 + i, 1 − i)|| = (2 + i)(2 − i) + (1 − i)(1 + i) = 4 + 1 + 1 + 1 = 7. The corresponding orthonormal set is given by: √ √ 7 7 (1 − i, 1 + 2i), (2 + i, 1 − i) . 7 7 (c)2017 Pearson Education. Inc. 361 1 − cos πx 14. f1 , f2 = 1, sin πx = sin πxdx = = 0. π −1 −1 1 1 sin πx cos πxdx = = 0. f1 , f3 = 1, cos πx = π −1 −1 1 1 1 1 −1 f2 , f3 = sin πx, cos πx = sin πx cos πxdx = sin 2πxdx = = 0. Thus, the cos 2πx 2 −1 4π −1 −1 vectors are orthogonal. 1 √ 1dx = [x]1−1 = 2. ||f1 || = −1 1 1 1 x 1 1 − cos 2πx 1 2 ||f2 || = sin πxdx = − = 1. dx = sin 2πx 2 2 −1 4π −1 −1 −1 1 1 1 x 1 1 1 + cos 2πx ||f3 || = cos2 πxdx = + = 1. dx = sin 2πx 2 2 −1 4π −1 −1 −1 √ 2 Consequently, , sin πx, cos πx is an orthonormal set of functions on [−1, 1]. 2 1 2 1 x = 0. 2 −1 −1 1 3 1 3x2 − 1 3x2 − 1 x −x f1 , f3 = 1, = 0. = dx = 2 2 2 −1 −1 1 1 3x2 − 1 3x2 − 1 1 3x4 x2 x· = 0. Thus, the vectors are orthogonal. = dx = − f2 , f3 = x, 2 2 2 4 2 −1 −1 1 √ ||f1 || = dx = [x]1−1 = 2. −1 √ 1 1 x3 6 2 x dx = = ||f2 || = . 3 −1 3 −1 √ 1 2 1 3x2 − 1 1 1 1 9x5 10 4 2 3 dx = (9x − 6x + 1)dx = = − 2x + x . ||f3 || = 2 4 −1 4 5 5 −1 −1 To obtain a set of orthonormal vectors, we divide each vector by its norm: 15. f1 , f2 = 1, x = 1 1 · xdx = √ √ √ 2 6 10 f2 f3 f1 = f1 , = f2 , = f3 . ||f1 || 2 ||f2 || 2 ||f3 || 2 √ √ 2 6 10 , x, (3x2 − 1) 2 2 4 √ Thus, 16. We have f1 , f2 = f1 , f3 = 1 1 is an orthonormal set of vectors. (2x)(1 + 2x2 )dx = 1 −1 −1 3 (2x)(x3 − x)dx = 5 −1 (2x + 4x3 )dx = (x2 + x4 )|1−1 = 0, 1 6 2 2 (2x4 − x2 )dx = ( x5 − x3 )|1−1 = 0, 5 5 5 −1 (c)2017 Pearson Education. Inc. 362 and 1 3 (1+2x )(x − x)dx = f2 , f3 = 5 −1 2 3 1 3 6 (x − x+2x5 − x3 )dx = 5 5 −1 3 1 1 3 1 3 1 (− x3 − x+2x5 )dx = (− x4 − x2 + x6 )|1−1 = 0. 5 5 20 10 3 −1 Thus, {f1 , f2 , f3 } forms an orthogonal set of vectors. To create an orthonormal set, we must normalize each of these vectors. Note that 1 4 31 8 2 ||f1 || = f1 , f1 = (2x) dx = x |−1 = , 3 3 −1 ||f2 || = f2 , f2 = 1 (1 + 2x2 )2 dx = 1 −1 (1 + 4x2 + 4x4 )dx = −1 4 4 (x + x3 + x5 )|1−1 = 3 5 94 , 15 and ||f3 || = f3 , f3 = 1 3 (x3 − x)2 dx = 5 −1 1 −1 (x6 − 6 4 9 x + x2 )dx = 5 25 1 6 3 ( x7 − x5 + x3 )|1−1 = 7 25 25 Thus, we obtain the orthonormal set 3 15 175 3 3 2 (2x), (1 + 2x ), (x − x) . 8 94 8 5 1 1 17. f1 , f2 = cos πx, cos 2πx = cos πx cos 2πxdx = (cos 3πx + cos πx)dx = 0. 2 1 −1 1 −1 1 f1 , f3 = cos πx, cos 3πx = cos πx cos 3πxdx = (cos 4πx + cos 2πx)dx = 0. 2 −1 −1 1 1 1 f2 , f3 = cos 2πx, cos 3πx = cos 2πx cos 3πxdx = (cos 5πx + cos πx)dx = 0. 2 −1 −1 Therefore, set. {f1 , f2 , f3 } is anorthogonal 1 1 1 sin 2πx 1 1 2 x+ ||f1 || = cos πxdx = (1 + cos 2πx)dx = = 1. 2 2 2π −1 −1 −1 1 1 1 1 1 sin 4πx 2 ||f2 || = cos 2πxdx = (1 + cos 4πx)dx = = 1. x+ 2 2 4π −1 −1 −1 1 1 sin 6πx 1 1 1 2 x+ ||f3 || = cos 3πxdx = (1 + cos 6πx)dx = = 1. 2 −1 2 6π −1 −1 Thus, it follows that{f1 , f2 , f3 } is an orthonormal set of vectors on [−1, 1]. 1 1 1 18. f1 , f2 = sin πx sin 2πxdx = (cos 3πx − cos πx)dx = 0. 2 1 −1 1 −1 1 f1 , f3 = sin πx sin 3πxdx = (cos 4πx − cos 2πx)dx = 0. 2 −1 −1 1 1 1 f2 , f3 = sin 2πx sin 3πxdx = (cos 5πx − cos πx)dx = 0. 2 −1 −1 Therefore, {f1 , f2 , f3 } is an orthogonal set. 1 (c)2017 Pearson Education. Inc. 8 . 175 363 1 1 sin 2πx 1 1 x− ||f1 || = sin πxdx = = 1. (1 − cos 2πx)dx = 2 2 2π −1 −1 −1 1 1 1 sin 4πx 1 1 2 x− ||f2 || = sin 2πxdx = = 1. (1 − cos 4πx)dx = 2 2 4π −1 −1 −1 1 1 1 sin 6πx 1 1 2 x− ||f3 || = sin 3πxdx = = 1. (1 − cos 6πx)dx = 2 6π −1 −1 2 −1 Thus, it follows that {f1 , f2 , f3 } is an orthonormal set of vectors on [−1, 1]. 2 1 19. We must have 2a − b − c = 0 and a + b + c = 0. Sovling this system via Gaussian elimination, we take c = t to be a free variable, from which we find that b = −t and a = 0. Thus, we must have r(x) = −t + tx2 = t(−1 + x2 ). Hence, {p(x), q(x), r(x)} is an orthogonal set precisely when r(x) is a multiple of −1 + x2 . 20. It is easily verified that A1 , A2 = 0, A1 , A3 = 0, A2 , A3 = 0. Thus we require a, b, c, d such that A1 , A4 = 0 =⇒ a + b − c + 2d = 0, A2 , A4 = 0 =⇒ −a + b + 2c + d = 0, A3 , A4 = 0 =⇒ a − 3b + 2d = 0. 3 1 Solving this system for a, b, c, and d, we obtain: a = c, b = − c, d = 0. Thus, 2 2 3 1 3 −1 3 −1 c −2c 2 = 2c =k , A4 = 2 0 2 0 c 0 where k is any nonzero real number. 21. (a). Let us determine a vector that is parallel to L. We can create a vector joining two points on L. By choose x = 0 and x = 1, respectively, we will obtain two points on L. These are the points (0, b) and (1, m + b). Hence, a parallel vector to L can be obtained by subtraction: (1, m + b) − (0, b) = (1, m) = v, as required. (b). We consider a point Q(x1 , y1 ) on the line L (so that y1 = mx1 + b) and form the vector w from Q to P via vector subtraction: w = (x0 − x1 , y0 − y1 ). Now P(w, v) = w, v (x0 − x1 ) + m(y0 − y1 ) v= (1, m). 2 ||v|| 1 + m2 The distance from P (x0 , y0 ) to the line L is given by m 1 ||w − P(w, v)|| = || (mx0 − y0 + b), (y0 − mx0 − b) || 1 + m2 1 + m2 1 2 m (mx0 − y0 + b)2 + (y0 − mx0 − b)2 = 1 + m2 1 2 = (m + 1)(y0 − mx0 − b)2 1 + m2 |y0 − mx0 − b| √ = . 1 + m2 22. Using the formula derived in Problem 21(b), here we take x0 = −8, y0 = 0, m = 3, and b = −4 to obtain | − 3 · (−8) − (−4)| 28 √ =√ . 10 1 + 32 (c)2017 Pearson Education. Inc. 364 23. The equation of the given line can be written in the form y = − 45 x + 15 . Using the formula derived in Problem 21(b), here we take x0 = 1, y0 = −1, m = − 45 , and b = 15 to obtain | − 1 + 45 − 15 | 2 5 2 = √ − = √ . 5 41 41 1 + 16 25 24. The equation of the given line can be written in the form y = x − 3. Using the formula derived in Problem 21(b), here we take x0 = −6, y0 = 4, m = 1, and b = −3 to obtain 13 |4 − (−6) − (−3)| √ =√ . 2 2 1+1 25. Setting t = 0, we obtain a point on the line: Q(0, −4, 0). Therefore, we can form the vector pointing from Q to P : w = (4, 1, −1) − (0, −4, 0) = (4, 5, −1). A parallel vector to L is v = (2, −1, 3). Thus, we have P(w, v) = (4, 5, −1), (2, −1, 3) (2, −1, 3) = 0. ||(2, −1, 3)||2 Thus, the distance from P to L is given by ||w − P(w, v)|| = ||w|| = 42 + 52 + (−1)2 = √ 42. 26. Setting t = 0, we obtain a point on the line: Q(4, 6, 0). Therefore, we can form the vector pointing from Q to P : w = (9, 0, 0) − (4, 6, 0) = (5, −6, 0). A parallel vector to L is v = (3, 0, −1). Thus, we have P(w, v) = (5, −6, 0), (3, 0, −1) 15 9 3 (3, 0, −1) = (3, 0, −1) = ( , 0, − ). ||(3, 0, −1)||2 10 2 2 Thus, the distance from P to L is given by 9 3 1 3 ||w−P(w, v)|| = ||(5, −6, 0)−( , 0, − )|| = ||( , −6, )|| = 2 2 2 2 2 2 1 1 77 3 9 + (−6)2 + = + 36 + = . 2 2 4 4 2 27. Setting t = 0, we obtain a point on the line: Q(−3, 0, 1, 0, 2). Therefore, we can form the vector pointing from Q to P : w = (1, 2, 3, 2, 1)−(−3, 0, 1, 0, 2) = (4, 2, 2, 2, −1). A parallel vector to L is v = (0, −2, 2, −1, 5). Thus, we have (4, 2, 2, 2, −1), (0, −2, 2, −1, 5) 7 7 7 7 35 . P(w, v) = (0, −2, 2, −1, 5) = − (0, −2, 2, −1, 5) = 0, , − , , − ||(0, −2, 2, −1, 5)||2 34 17 17 34 34 Thus, the distance from P to L is given by 7 7 7 35 27 41 61 1 ||w − P(w, v)|| = ||(4, 2, 2, 2, −1) − 0, , − , , − || = ||(4, , , , )|| 17 17 34 34 17 17 34 34 √ √ 1362 + 542 + 822 + 612 + 12 31858 1 ||(136, 54, 82, 61, 1)|| = = ≈ 5.250. = 34 34 34 28. (c). Note that w = (x0 − x1 , y0 − y1 , z0 − z1 ) and n = (a, b, c). Now, we have P(w, n) = w, n n, ||n||2 (c)2017 Pearson Education. Inc. 365 so that ||P(w, n)|| = |w, n| |ax0 + by0 + cz0 + d| |a(x0 − x1 ) + b(y0 − y1 ) + c(z0 − z1 )| √ √ = . = ||n|| a 2 + b2 + c 2 a 2 + b2 + c 2 29. We have a = 1, b = 2, c = −4, d = −2, x0 = −4, y0 = 7, and z0 = −2. Therefore, from the formula in Problem 28(c), we obtain the requested distance: |(1)(−4) + (2)(7) + (−4)(−2) + (−2)| 16 =√ . 2 2 2 21 1 + 2 + (−4) 30. We have a = 3, b = −1, c = −1,d = −5, x0 = 0, y0 = −1, and z0 = 3. Therefore, from the formula in Problem 28(c), we obtain the requested distance: |(3)(0) + (−1)(−1) + (−1)(3) + (−5)| 7 =√ . 2 2 2 11 3 + (−1) + (−1) 31. We have a = 2, b = 0, c = −1, d = 0, x0 = −1, y0 = 1, and z0 = −1. Therefore, from the formula in Problem 28(c), we obtain the requested distance: |(2)(−1) + (0)(1) + (−1)(−1) + 0| 1 =√ . 2 2 2 5 2 + 0 + (−1) 32. We have a = 0, b = 1, c = −4, d = −2, x0 = 8, y0 = 8, and z0 = −1. Therefore, from the formula in Problem 28(c), we obtain the requested distance: 10 |(0)(8) + (1)(8) + (−4)(−1) + (−2)| =√ . 2 2 2 17 0 + 1 + (−4) 33. {u1 , u2 , v} is a linearly independent set of vectors, and u1 , u2 = 0. If we let u3 = v + λu1 + μu2 , then it must be the case that u3 , u1 = 0 and u3 , u2 = 0. u3 , u1 = 0 =⇒ v + λu1 + μu2 , u1 = 0 =⇒ v, u1 + λu1 , u1 + μu2 , u1 = 0 =⇒ v, u1 + λu1 , u1 + μ · 0 = 0 v, u1 . =⇒ λ = − ||u1 ||2 u3 , u2 = 0 =⇒ v + λu1 + μu2 , u2 = 0 =⇒ v, u2 + λu1 , u2 + μu2 , u2 = 0 =⇒ v, u2 + λ · 0 + μu2 , u2 = 0 v, u2 . =⇒ μ = − ||u2 ||2 v, u1 v, u2 Hence, if u3 = v − u1 − u2 , then {u1 , u2 , u3 } is an orthogonal basis for the subspace spanned 2 ||u1 || ||u2 ||2 by {u1 , u2 , v}. 34. It was shown in Remark 3 following Definition 5.2.1 that each vector ui is a unit vector. Moreover, for i = j, 1 1 1 ui , uj = vi , vj = vi , vj = 0, vi vj vi vj since vi , vj = 0. Therefore {u1 , u2 , . . . , uk } is an orthonormal set of vectors. (c)2017 Pearson Education. Inc. 366 35. (a). Using technology we find that π π π sin nx dx = 0, cos nx dx = 0, sin nx cos mx dx = 0. −π Further, for m = n, −π −π π π sin nx sin mx dx = 0 and −π cos nx cos mx dx = 0. −π Consequently the given set of vectors is orthogonal on [−π, π]. (b). Multiplying (5.2.5) by cos mx and integrating over [−π, π] yields π 1 f (x) cos mx dx = a0 2 −π π cos mx dx + −π π , ∞ −π n=1 (an cos nx + bn sin nx) cos mx dx. Assuming that interchange of the integral and infinite summation is permissible, this can be written π π ∞ π , 1 f (x) cos mx dx = a0 cos mx dx + (an cos nx + bn sin nx) cos mx dx. 2 −π −π n=1 −π which reduces to π f (x) cos mx dx = −π 1 a0 2 π −π π cos mx dx + am cos2 mx dx −π where we have used the results from part (a). When m = 0, this gives π π 1 1 π f (x) dx = a0 dx = πa0 =⇒ a0 = f (x) dx, 2 π −π −π −π whereas for m = 0, π π 1 π f (x) cos mx dx = am cos2 mx dx = πam =⇒ am = f (x) cos mx dx. π −π −π −π (c). Multiplying (5.2.5) by sin(mx), integrating over [−π, π], and interchanging the integration and summation yields π π ∞ π , 1 f (x) sin mx dx = a0 sin mx dx + (an cos nx + bn sin nx) sin mx dx. 2 −π −π n=1 −π Using the results from (a), this reduces to π π 1 π f (x) sin mx dx = bn sin2 mx dx = πbn =⇒ bn = f (x) sin mx dx. π −π −π −π (d). The Fourier coefficients for f are 1 π 1 π a0 = xdx = 0, an = x cos nx dx = 0, π −π π −π 1 π 2 2 x sin nx dx = − cos nπ = (−1)n+1 . bn = π −π n n (c)2017 Pearson Education. Inc. 367 The Fourier series for f is ∞ , 2 n n=1 (−1)n+1 sin nx. (e). The approximations using the first term, the first three terms, the first five terms, and the first ten terms in the Fourier series for f are shown in the accompanying figures. S3(x) 3 2 1 x 3 2 1 1 2 3 1 2 3 Figure 0.0.65: Figure for Problem 35(e) - 3 terms included These figures suggest that the Fourier series is converging to the function f (x) at all points in the interval (−π, π). Solutions to Section 5.3 (a): TRUE. In this case, all of the inner products appearing in the formulas in Theorem 5.3.2 are 0, so that we obtain vi = xi for each i. That is, the new set obtained via the Gram-Schmidt Process is identical to the original set we are given. (b): TRUE. We have (x1 + x2 ), x1 x1 ||x1 ||2 ||x1 ||2 + x2 , x1 = (x1 + x2 ) − x1 ||x1 ||2 = (x1 + x2 ) − x1 = x2 . v2 = (x1 + x2 ) − (c): FALSE. For instance, in R3 we can take the vectors x1 = (1, 0, 0), x2 = (1, 1, 0), and x3 = (0, 0, 1). Applying the Gram-Schmidt process to the ordered set {x1 , x2 , x3 } yields the standard basis {e1 , e2 , e3 }. However, applying the Gram-Schmidt process to the ordered set {x3 , x2 , x1 } yields the basis {x3 , x2 , ( 12 , − 12 , 0)} instead. (c)2017 Pearson Education. Inc. 368 S5(x) 3 2 1 3 2 1 1 2 3 x 1 2 3 Figure 0.0.66: Figure for Problem 35(e) - 5 terms included S10(x) 3 2 1 3 2 1 1 2 3 x 1 2 3 Figure 0.0.67: Figure for Problem 35(e) - 10 terms included (d): TRUE. This is assumed in Theorem 5.3.2. If the given vectors were not linearly independent, then we could not expect to obtain a basis after carrying out the Gram-Schmidt Process. (e): FALSE. Here is a counterexample (there are many such). In R2 , take x1 = y1 = v1 = (1, 0), and take x2 = (1, 1) and y2 = (−1, 1). Then B1 = B2 , but applying the Gram-Schmidt Process to B1 and B2 , we obtain the same orthogonal basis: {(1, 0), (0, 1)}. (f ): TRUE. We have v1 = x1 , and thus, 2v1 = 2x1 . Now, to find the second vector when the Gram-Schmidt (c)2017 Pearson Education. Inc. 369 Process is applied to B2 , we must compute 2x2 − P(2x2 , 2v1 ). We are given already that v2 = x2 − P(x2 , v1 ) = x2 − Now, 2x2 − P(2x2 , 2v1 ) = 2x2 − x2 , v1 v1 . ||v1 ||2 2x2 , 2v1 x2 , v1 (2v1 ) = 2x2 − (2v1 ) = 2v2 . ||2v1 ||2 ||v1 ||2 Thus, we have obtained the orthogonal set {2v1 , 2v2 }, as requested. Problems: 1. The given set of vectors is already an orthogonal set, so we √ need not apply the √ Gram-Schmidt pro2 + 22 + 32 = 1 14 and ||(6, −3, 0)|| = cess. Instead, we proceed directly to normalization: ||(1, 2, 3)|| = √ √ 62 + (−3)2 + 02 = 45 = 3 5. Thus, we obtain the orthonormal basis 1 1 √ (1, 2, 3), √ (6, −3, 0) . 14 3 5 = (2, 1, −1). 2. Let x1 = (1, −1, −1) and let x2 √ v1 = x1 = (1, −1, −1), ||v1 || = 12 + (−1)2 + (−1)2 = 3. x2 , v1 = (2, 1, −1), (1, −1, −1) = (2)(1) + (1)(−1) + (−1)(−1) = 2. x2 , v1 2 1 v1 = (2, 1, −1) − (1, −1, −1) = (4, 5, −1). v2 = x2 − ||v1 ||2 3 3 √ ||v2 || = 13 42 + 52 + (−1)2 = 342 . Hence, an orthonormal basis is 1 1 √ (1, −1, −1), √ (4, 5, −1) . 3 42 3. Let x1 = (2, 1, −2) and let x2 = (1, 3, −1). v1 = x1 = (2, 1, −2), ||v1 || = 22 + 12 + (−2)2 = 3. x2 , v1 = (1, 3, −1), (2, 1, −2) = (1)(2) + (3)(1) + (−1)(−2) = 7. x2 , v1 7 5 v1 = (1, 3, −1) − (2, 1, −2) = (−1, 4, 1). v2 = x2 − 2 ||v || 9 9 √ 1 ||v2 || = 59 (−1)2 + 42 + 12 = 5 3 2 . Hence, an orthonormal basis is 1 1 (2, 1, −2), √ (−1, 4, 1) . 3 3 2 4. Note that this particular set of vectors can be shown to be linearly independent, and since we have three vectors in R3 , they therefore span all of R3 . As such, we could merely replace this basis with the orthonormal basis {(1, 0, 0), (0, 1, 0), (0, 0, 1)} for the same space. However, our aim here is to demonstrate the Gram-Schmidt Process. Let x1 = (1, −5, −3), let x2 = (0, −1, 3), and let x3 = (−6, √ 0, −2). v1 = x1 = (1, −5, −3), ||v1 || = 12 + (−5)2 + (−3)2 = 35. x2 , v1 = (0, −1, 3), (1, −5, −3) = (0)(1) + (−1)(−5) + (3)(−3) = −4. x2 , v1 4 1 v1 = (0, −1, 3) + (1, −5, −3) = v2 = x2 − (4, −55, 93). ||v1 ||2 35 35 √ 1 42 + (−55)2 + 932 = 11690 ||v2 || = 35 35 . (c)2017 Pearson Education. Inc. 370 x3 , v1 = (−6, 0, −2), (1, −5, −3) = (−6)(1) + (0)(−5) + (−2)(−3) = 0. 1 1 x3 , v2 = 35 (−6, 0, −2), (4, −55, 93) = 35 (−210) = −6. x3 , v1 x3 , v2 105 1 3 v1 − v2 = (−6, 0, −2)−(0, 0, 0)+ v3 = x3 − · (4, −55, 93) = (−6, 0, −2)+ (4, −55, 93) = ||v1 ||2 ||v2 ||2 167 35 167 55 (−18, −3, −1). 167 √ √ 55 55 √ 2 . Hence, an orthonormal basis is (−18)2 + (−3)2 + (−1)2 = 167 · 334 = 55 ||v3 || = 167 167 1 1 1 √ (1, −5, −3), √ (4, −55, 93), √ (−18, −3, −1) . 35 11690 334 5. Note that this particular set of vectors can be shown to be linearly independent, and since we have three vectors in R3 , they therefore span all of R3 . As such, we could merely replace this basis with the orthonormal basis {(1, 0, 0), (0, 1, 0), (0, 0, 1)} for the same space. However, our aim here is to demonstrate the Gram-Schmidt Process. Let x1 = (2, 0, 1), let x2 = (−3, x3 = (1, −3, 8). √ 1, 1), and let √ v1 = x1 = (2, 0, 1), ||v1 || = 22 + 02 + 12 = 5. x2 , v1 = (−3, 1, 1), (2, 0, 1) = (−3)(2) + (1)(0) + (1)(1) = −5. x2 , v1 v 2 = x2 − v1 = (−3, 1, 1) + (2, 0, 1) = (−1, 1, 2). 2 ||v1 || √ ||v2 || = (−1)2 + 12 + 22 = 6. x3 , v1 = (1, −3, 8), (2, 0, 1) = (1)(2) + (−3)(0) + (8)(1) = 10. x3 , v2 = (1, −3, 8), (−1, 1, 2) = (1)(−1) + (−3)(1) + (8)(2) = 12. x3 , v1 x3 , v2 v 3 = x3 − v1 − v2 = (1, −3, 8) − 2(2, 0, 1) − 2(−1, 1, 2) = (−1, −5, 2). 2 ||v2 ||2 √ ||v1 || ||v3 || = (−1)2 + (−5)2 + 22 = 30. Hence, an orthonormal basis is 1 1 1 √ (2, 0, 1), √ (−1, 1, 2), √ (−1, −5, 2) . 5 6 30 = (1, 2, 1, 2). 6. Let x1 = (−1, 1, 1, 1) and let x2 v1 = x1 = (−1, 1, 1, 1), ||v1 || = (−1)2 + 12 + 12 + 12 = 2. x2 , v1 = (1, 2, 1, 2), (−1, 1, 1, 1) = (1)(−1) + (2)(1) + (1)(1) + (2)(1) = 4. x2 , v1 v2 = x2 − v1 = (1, 2, 1, 2) − (−1, 1, 1, 1) = (2, 1, 0, 1). 2 √ ||v1 || √ ||v2 || = 22 + 12 + 02 + 12 = 6. Hence, an orthonormal basis is 1 1 (−1, 1, 1, 1), √ (2, 1, 0, 1) . 2 6 7. Let x1 = (1, 0, −1, 0), let x2 = (1, 1, −1, 0), and let x3 = (−1, √ 1, 0, 1). v1 = x1 = (1, 0, −1, 0), ||v1 || = 12 + 02 + (−1)2 + 02 = 2. x2 , v1 = (1, 1, −1, 0), (1, 0, −1, 0) = (1)(1) + (1)(0) + (−1)(−1) + (0)(0) = 2. √ x2 , v1 v2 = x2 − v1 = (1, 1, −1, 0) − (1, 0, −1, 0) = (0, 1, 0, 0). ||v2 || = 02 + 12 + 02 + 02 = 1. 2 ||v1 || x3 , v1 = (−1, 1, 0, 1), (1, 0, −1, 0) = (−1)(1) + (1)(0) + (0)(−1) + (1)(0) = −1. x3 , v2 = (−1, 1, 0, 1), (0, 1, 0, 0) = (−1)(0) + (1)(1) + (0)(0) + (1)(0) = 1. (c)2017 Pearson Education. Inc. 371 x3 , v1 x3 , v2 1 1 v1 − v2 = (−1, 1, 0, 1) + (1, 0, −1, 0) − (0, 1, 0, 0) = (−1, 0, −1, 2). ||v3 || = v3 = x3 − 2 2 ||v1 || ||v√ 2 2 2 || 6 1 2 + 02 + (−1)2 + 22 = (−1) . Hence, an orthonormal basis is 2 2 1 1 √ (1, 0, −1, 0), (0, 1, 0, 0), √ (−1, 0, −1, 2) . 2 6 8. Let x1 = (1, 2, 0, 1), let x2 = (2, √ 1, 1, 0), and let x3 =√(1, 0, 2, 1). v1 = x1 = (1, 2, 0, 1), ||v1 || = 12 + 22 + 02 + 12 = 6. x2 , v1 = (2, 1, 1, 0), (1, 2, 0, 1) = (2)(1) + (1)(2) + (1)(0) + (0)(1) = 4. x2 , v1 4 1 v1 = (2, 1, 1, 0) − (1, 2, 0, 1) = (4, −1, 3, −2). v2 = x2 − 2 ||v || 6 3 1 √ ||v2 || = 13 42 + (−1)2 + 32 + (−2)2 = 13 30. x3 , v1 = (1, 0, 2, 1), (1, 2, 0, 1) = (1)(1) + (0)(2) + (2)(0) + (1)(1) = 2. x3 , v2 = (1, 0, 2, 1), 13 (4, −1, 3, −2) = (1)( 43 ) + (0)(− 13 ) + (2)(1) + (1)(− 23 ) = 83 . x3 , v1 x3 , v2 2 4 2 v3 = x3 − v1 − v2 = (1, 0, 2, 1) − (1, 2, 0, 1) − (4, −1, 3, −2) = (−1, −1, 3, 3). 2 2 ||v || ||v || 6 15 5 2 1 ||v3 || = 25 (−1)2 + (−1)2 + 32 + 32 = √45 . Hence, an orthonormal basis is 1 1 1 √ (1, 2, 0, 1), √ (4, −1, 3, −2), √ (−1, −1, 3, 3) . 6 30 2 5 9. Let x1 = (1, 1, −1, 0), let x2 = (−1, 0, 1, 1), and let x3 = (2, √−1, 2, 1). v1 = x1 = (1, 1, −1, 0), ||v1 || = 12 + 12 + (−1)2 + 02 = 3. x2 , v1 = (−1, 0, 1, 1), (1, 1, −1, 0) = (−1)(1) + (0)(1) + (1)(−1) + (1)(0) = −2. x2 , v1 2 1 v1 = (−1, 0, 1, 1) + (1, 1, −1, 0) = (−1, 2, 1, 3). v 2 = x2 − 2 ||v || 3 3 1 √ ||v2 || = 13 (−1)2 + 22 + 12 + 32 = 13 15. x3 , v1 = (2, −1, 2, 1), (1, 1, −1, 0) = (2)(1) + (−1)(1) + (2)(−1) + (1)(0) = −1. x3 , v2 = (2, −1, 2, 1), 13 (−1, 2, 1, 3) = (2)(− 13 ) + (−1)( 23 ) + (2)( 13 ) + (1)(1) = 13 . x3 , v1 x3 , v2 1 1 4 v3 = x3 − v1 − v2 = (2, −1, 2, 1) + (1, 1, −1, 0) − (−1, 2, 1, 3) = (3, −1, 2, 1). 2 2 ||v || ||v || 3 15 5 1 2 √ ||v3 || = 45 32 + (−1)2 + 22 + 12 = 45 15. Hence, an orthonormal basis is 1 1 1 √ (1, 1, −1, 0), √ (−1, 2, 1, 3), √ (3, −1, 2, 1) . 3 15 15 10. Let x1 = (1, 2, 3, 4, 5) and let x√ 2 = (−7, 0, 1, −2, 0). √ v1 = x1 = (1, 2, 3, 4, 5), ||v1 || = 12 + 22 + 32 + 42 + 52 = 55. x2 , v1 = (1, 2, 3, 4, 5), (−7, 0, 1, −2, 0) = (1)(−7) + (2)(0) + (3)(1) + (4)(−2) + (5)(0) = −12. x2 , v1 12 373 24 91 62 12 v1 = (−7, 0, 1, −2, 0) + (1, 2, 3, 4, 5) = (− v 2 = x2 − , , , − , ). ||v1 ||2 55 55 55 55 55 11 √ 1 (−373)2 + 242 + 912 + (−62)2 + (60)2 = 3 17,270 . Hence, an orthonormal basis is ||v2 || = 55 55 1 1 √ (1, 2, 3, 4, 5), √ (−373, 24, 91, −62, 60) . 55 3 17270 (c)2017 Pearson Education. Inc. 372 1 −3 2 0 −1 . Hence, a basis for rowspace(A) is {(1, −3, 2, 0, −1), (0, 1, −3, 13 , 2)}. 2 0 1 −3 13 The vectors in this basis are not orthogonal. We therefore use the Gram-Schmidt process to determine an orthogonal basis. Let x1 = (1, −3, 2, 0, −1) and x2 = (0, 1, −3, 13 , 2). √ Then v1 = x1 = (1, −3, 2, 0, −1). ||v1 || = 12 + (−3)2 + 22 + 02 + (−1)2 = 15. x2 , v1 = (0, 1, −3, 13 , 2), (1, −3, 2, 0, −1) = (0)(1) + (1)(−3) + (−3)(2) + ( 13 )(0) + (2)(−1) = −11. x2 , v1 1 11 1 v1 = (0, 1, −3, , 2)+ (1, −3, 2, 0, −1) = (11, −18, −23, 5, 19). Thus, an orthogonal v2 = x2 − ||v1 ||2 3 15 15 basis for rowspace(A) is given by {(1, −3, 2, 0, −1), (11, −18, −23, 5, 19)}. Since colspace(A) = R2 , we can simply take {(1, 0), (0, 1)} as an orthogonal basis for colspace(A). ⎤ ⎡ 1 5 ⎢ 0 1 ⎥ ⎥ ⎢ 2 ⎥ 12. A row-echelon form of A is ⎢ ⎢ 0 0 ⎥. Obviously, rowspace(A) = R , so we may simply take ⎣ 0 0 ⎦ 0 0 {(1, 0), (0, 1)} as orthogonal basis for rowspace(A). A basis for colspace(A) is simply composed of the two columns of A: ⎧⎡ ⎤⎫ ⎤ ⎡ 5 ⎪ 1 ⎪ ⎪ ⎪ ⎪ ⎪ ⎪⎢ 2 ⎥ ⎢ 4 ⎥⎪ ⎨ ⎥⎬ ⎥ ⎢ ⎢ ⎢ 3 ⎥,⎢ 3 ⎥ . ⎥ ⎥ ⎢ ⎢ ⎪ ⎪ ⎪⎣ 4 ⎦ ⎣ 2 ⎦⎪ ⎪ ⎪ ⎪ ⎪ ⎭ ⎩ 1 5 11. A row-echelon form of A is The vectors in this basis are not orthogonal. We therefore use the Gram-Schmidt process to determine an orthogonal basis. (We will render the columns as rows while we perform the steps to save space.) Let x1 = (1, 2, 3, 4, 5) and let x2 = (5, 4, 3, 2, 1). √ √ Then v1 = x1 = (1, 2, 3, 4, 5). ||v1 || = 12 + 22 + 32 + 42 + 52 = 55. x2 , v1 = (5, 4, 3, 2, 1), (1, 2, 3, 4, 5) = (5)(1) + (4)(2) + (3)(3) + (2)(4) + (1)(5) = 35. x2 , v1 35 1 v 2 = x2 − v1 = (5, 4, 3, 2, 1) − (1, 2, 3, 4, 5) = (48, 30, 12, −6, −24). Thus, an orthogonal ||v1 ||2 55 11 basis for colspace(A) is given by ⎧⎡ ⎤⎫ ⎤ ⎡ 48 ⎪ 1 ⎪ ⎪ ⎪ ⎪ ⎪ ⎥⎪ ⎥ ⎢ ⎪ ⎨⎢ ⎢ 2 ⎥ ⎢ 30 ⎥⎬ ⎢ 3 ⎥ , ⎢ 12 ⎥ . ⎥⎪ ⎥ ⎢ ⎢ ⎪ ⎪ ⎣ 4 ⎦ ⎣ −6 ⎦⎪ ⎪ ⎪ ⎪ ⎪ ⎭ ⎩ −24 5 ⎡ ⎤ 1 −2 1 1 17 ⎦. Hence, a basis for rowspace(A) is {(1, −2, 1), (0, 7, 1)}. 13. A row-echelon form of A is ⎣ 0 0 0 0 The vectors in this basis are not orthogonal. We therefore use the Gram-Schmidt process to determine an orthogonal basis. Let x1 = (1, −2, 1) and x2 = (0, 7, 1). We have v1 = x1 = (1, −2, 1). Now ||v1 || = √ x2 , v1 12 + (−2)2 + 12 = 6 and x2 , v1 = (1)(0) + (−2)(7) + (1)(1) = −13. Thus, v2 = x2 − v1 = ||v1 ||2 13 1 (0, 7, 1) + (1, −2, 1) = (13, 16, 19). 6 6 Then an orthogonal basis for rowspace(A) is {(1, −2, 1), (13, 16, 19)}. (c)2017 Pearson Education. Inc. 373 Next we consider colspace(A), and we will render the column vectors that arise as rows in order to save space. From the row-echelon form of A obtained above, a basis for colspace(A) is given by {(3, 1, 1), (1, −2, 5)}. The vectors in this basis are not orthogonal. We therefore use the Gram-Schmidt process to determine an orthogonal basis. √ Let x1 = (3, 1, 1) and x2 = (1, −2, 5). We have v1 = x1 = (3, 1, 1). Now ||v1 || = √ 32 + 12 + 12 = 11 and x2 , v1 = (1, −2, 5), (3, 1, 1) = (1)(3) + (−2)(1) + (5)(1) = 6. Thus, , x2 , v1 6 7 v1 = (1, −2, 5) − (3, 1, 1) = v2 = x2 − (−1, −4, 7). Then, an orthogonal basis for colspace(A) 2 ||v || 11 11 1 ⎧⎡ ⎤ ⎡ ⎤⎫ −1 ⎬ ⎨ 3 is ⎣ 1 ⎦ , ⎣ −4 ⎦ . ⎩ ⎭ 1 7 ⎡ ⎤ 1 −4 7 1 −3 ⎦. Hence, a basis for rowspace(A) is {(1, −4, 7), (0, 1, −3)}. 14. A row-echelon form of A is ⎣ 0 0 0 0 The vectors in this basis are not orthogonal. We therefore use the Gram-Schmidt process to determine an orthogonal basis. Let x1√ = (1, −4, 7) and x2 = (0, 1, −3). We have v1 = x1 = (1, −4, 7). Now ||v1 || = 12 + (−4)2 + 72 = 66 and x2 , v1 = (0, 1, −3), (1, −4, 7) = (0)(1) + (1)(−4) + (−3)(7) = −25. x2 , v1 25 1 v1 = (0, 1, −3) + (1, −4, 7) = Thus, v2 = x2 − (25, −34, −23). ||v1 ||2 66 66 Then an orthogonal basis for rowspace(A) is {(1, −4, 7), (25, −34, −23)}. Next we consider colspace(A), and we will render the column vectors that arise as rows in order to save space. From the row-echelon form of A obtained above, a basis for colspace(A) is given by {(1, −2, −1), (−4, 6, 0)}. The vectors in this basis are not orthogonal. We therefore use the Gram-Schmidt process to determine an orthogonal basis. Let x1 √ = (1, −2, −1) and x2 = (−4, 6, 0). We have v1 = x1 = (1, −2, −1). Now ||v1 || = 12 + (−2)2 + (−1)2 = 6 and x2 , v1 = (−4, 6, 0), (1, −2, −1) = (−4)(1) + (6)(−2) + (0)(−1) = −16. x2 , v1 16 1 v1 = (−4, 6, 0) + Thus, v2 = x2 − (1, −2, −1) = (−4, 2, −8). Then an orthogonal basis for 2 ||v || 6 3 1⎤ ⎡ ⎧⎡ ⎤⎫ 1 −4 ⎨ ⎬ colspace(A) is ⎣ −2 ⎦ , ⎣ 2 ⎦ . ⎩ ⎭ −1 −8 15. Let v1 = (1 + i, i, 2 − i) and v 2 = (1 + 2i, 1 − i, i). √ u1 = v1 = (1 + i, i, 2 − i), ||u1 || = (1 + i)(1 − i) + i(−i) + (2 − i)(2 + i) = 2 2. v2 , u1 = (1 + 2i)(1 − i) + (1 − i)(−i) + i(2 + i) = 1 + 2i. v2 , u1 1 1 u1 = (1 + 2i, 1 − i, i) − (1 + 2i)(1 + i, i, 2 − i) = (9 + 13i, 10 − 9i, −4 + 5i). u2 = v 2 − ||u1 ||2 8 8 √ 118 1 (9 + 13i)(9 − 13i) + (10 − 9i)(10 + 9i) + (−4 + 5i)(−4 − 5i) = . Hence, an orthonormal ||u2 || = 8 4 basis is: √ √ 2 118 (1 + i, i, 2 − i), (9 + 13i, 10 − 9i, −4 + 5i) . 4 236 16. Let v1 = (1 − i, 0, i) and v 2 = (1, 1 + i, 0). √ u1 = v1 = (1 − i, 0, i), ||u1 || = (1 − i)(1 + i) + 0 + i(−i) = 3. v2 , u1 = (1, 1 + i, 0), (1 − i, 0, i) = 1(1 + i) + (1 + i)0 + 0(−i) = 1 + i. v2 , u1 1+i 1 u1 = (1, 1 + i, 0) − u2 = v2 − (1 − i, 0, i) = (1, 3 + 3i, 1 − i). ||u1 ||2 3 3 (c)2017 Pearson Education. Inc. 374 ||u2 || = √ 1 21 1 + (3 + 3i)(3 − 3i) + (1 − i)(1 + i) = . Hence, an orthonormal basis is: 3 3 √ √ 3 21 (1 − i, 0, i), (1, 3 + 3i, 1 − i) . 3 21 17. Let f1 = 1 + 2x and f2 = −2 − x + x2 . We will define the orthogonal basis as {g1 , g2 } as follows. Set g1 = f1 = 1 + 2x. Then 1 1 13 2 2 ||g1 || = g1 , g1 = (1 + 2x) dx = (1 + 4x + 4x )dx = 3 0 0 and g1 , f2 = 1 (1 + 2x)(−2 + x + x2 )dx = 0 Thus, g 2 = f2 − 1 0 (−2 − 3x + 3x2 + 2x3 )dx = −2. g1 , f2 6 20 1 g1 = (−2 − x + x2 ) + (1 + 2x) = − − x + x2 . 2 ||g1 || 13 13 13 Thus, we have the orthogonal basis 1 + 2x, − 20 1 − x + x2 . 13 13 18. Let f1 = 1, f2 = x and f3 = x2 . g1 = f1 = 1; 2 1 1 1 1 x 2 dx = 1; f2 , g1 = xdx = = . ||g1 || = 2 2 0 0 0 f2 , g1 1 1 g 2 = f2 − g1 = x − = (2x − 1). ||g1 ||2 2 2 2 3 1 1 1 x 1 1 1 x2 x 2 2 dx = x− x −x+ dx = = ||g2 || = − + . 2 4 3 2 4 12 0 0 3 1 0 1 1 x f3 , g1 = x2 dx = = . 3 3 0 0 4 1 1 1 1 x x3 2 x x− = f3 , g2 = dx = − . 2 4 6 12 0 0 1 f3 , g1 f3 , g2 1 1 2 = (6x2 − 6x + 1). Thus, an orthogonal basis is given g1 − g2 = x − − x − g 3 = f3 − 2 2 ||g1 || ||g2 || 3 2 6 by: 1 1 2 1, (2x − 1), (6x − 6x + 1) . 2 6 19. Let f1 = 1, f2 = x2 and f3 = x4 for all x in [−1, 1]. 1 1 1 2 2 2 2 4 x dx = , f1 , f3 = x dx = , and f2 , f3 = x6 dx = . f1 , f2 = 3 5 7 −1 −1 1−1 1 2 ||f1 ||2 = dx = 2, ||f2 ||2 = x4 dx = . 5 −1 −1 (c)2017 Pearson Education. Inc. 375 Let g1 = f1 = 1. f2 , g1 x2 , 1 1 1 g 2 = f2 − g 1 = x2 − · 1 = x2 − = (3x − 1). 2 ||g1 || ||1||2 3 3 x4 , x2 − 13 f3 , g1 f3 , g2 x4 , 1 1 4 2 x − g1 − g2 = x − ·1− g 3 = f3 − ||g1 ||2 ||g2 ||2 ||1||2 3 ||x2 − 13 ||2 = x4 − 6x2 3 1 + = (35x4 − 30x2 + 3). 7 35 35 Thus, an orthogonal basis is given by: 1 1 2 4 2 1, (3x − 1), (35x − 30x + 3) . 3 35 π π 20. Let f1 = 1, f2 = sin x and f3 = cos x for all x in − , . 2 2 π/2 π/2 sin xdx = [− cos 2x]−π/2 = 0. Therefore, f1 and f2 are orthogonal. f1 , f2 = −π/2 π/2 1 sin x cos xdx = f2 , f3 = 2 −π/2 nal. π/2 f1 , f3 = π/2 1 sin 2xdx = − cos 2x = 0. Therefore, f2 and f3 are orthogo4 −π/2 −π/2 π/2 π/2 cos xdx = [sin x]−π/2 = 2. π/2 Let g1 = f1 = 1 so that ||g1 ||2 = dx = π. −π/2 π/2 π/2 1 − cos 2x π sin2 xdx = g2 = f2 = sin x, and ||g2 ||2 = dx = . 2 2 −π/2 −π/2 f3 , g1 f3 , g2 2 1 g1 − g2 = cos x − · 1 − 0 · sin x = (π cos x − 2). Thus, an orthogonal basis for g 3 = f3 − ||g1 ||2 ||g2 ||2 π π the subspace of C 0 [−π/2, π/2] spanned by {1, sin x, cos x} is: 1 1, sin x, (π cos x − 2) . π −π/2 1 −1 2 −3 21. Given A1 = and A2 = . Using the Gram-Schmidt procedure: 1 4 1 2 1 −1 , A2 , B1 = 10 + 6 + 24 + 5 = 45, and ||B1 ||2 = 5 + 2 + 12 + 5 = 24. B1 = 2 1 1 A2 , B1 15 1 −1 − 98 2 −3 8 = . Thus, an orthogonal basis for the B = − B2 = A2 − 1 1 2 1 4 1 − 78 ||B1 ||2 8 4 subspace of M2 (R) spanned by A1 and A2 is: 1 − 98 1 −1 8 , 1 . 2 1 − 78 4 0 1 0 1 1 , A2 = and A3 = 22. Given A1 = 1 0 1 1 1 0 1 2 , A2 , B1 = 5, and ||B1 || = 5. B1 = 1 0 1 0 . Using the Gram-Schmidt procedure: (c)2017 Pearson Education. Inc. 376 A2 , B1 0 1 0 1 0 0 B = − = . 1 1 1 1 0 0 1 ||B1 ||2 2 Also, A3 , B1 = 5, A3 , B2 = 0, and ||B 2 || =5, so that A3 , B2 A3 , B1 1 1 0 1 0 0 1 B − B = − − 0 = B3 = A3 − 1 2 1 0 1 0 0 1 0 ||B1 ||2 ||B2 ||2 basis for the subspace of M2 (R) spanned by A1 , A2 , and A3 is: 0 1 1 0 0 0 , , , 1 0 0 0 0 1 B2 = A2 − 0 0 . Thus, an orthogonal which is the subspace of all symmetric matrices in M2 (R). 23. Given p1 (x) = 1 − 2x + 2x2 and p2 (x) = 2 − x − x2 . Using the Gram-Schmidt procedure: q1 = 1 − 2x + 2x2 , p2 , q1 = 2 · 1 + (−1)(−2) + (−1)2 = 2, ||q1 ||2 = 12 + (−2)2 + 22 = 9. So, p2 , q1 2 1 q1 = 2 − x − x2 − (1 − 2x + 2x2 ) = (16 − 5x − 13x2 ). Thus, an orthogonal basis for the q 2 = p2 − 2 ||q1 || 9 9 subspace spanned by p1 and p2 is {1 − 2x + 2x2 , 16 − 5x − 13x2 }. 24. Given p1 (x) = 1 + x2 , p2 (x) = 2 − x + x3 , and p3 (x) = −x + 2x2 . Using the Gram-Schmidt procedure: q1 = 1 + x2 , p2 , q1 = 2 · 1 + (−1)(0) + 0 · 1 + 1 · 2 = 2, and q1 2 = 12 + 12 = 2. So, p2 , q1 q1 = 2 − x + x3 − (1 + x2 ) = 1 − x − x2 + x3 . Also, q 2 = p2 − ||q1 ||2 p3 , q1 = 0 · 1 + (−1)0 + 2 · 1 + 02 = 2 p3 , q2 = 0 · 1 + (−1)2 + 2(−1) + 0 · 1 = −1, and ||q2 ||2 = 12 + (−1)2 + (−1)2 + 12 = 4 so that p3 , q1 p3 , q2 1 1 q1 − q2 = −x + 2x2 − (1 + x2 ) + (1 − x − x2 + x3 ) = (−3 − 5x + 3x2 + x3 ). Thus, an q 3 = p3 − ||q1 ||2 ||q2 ||2 4 4 orthogonal basis for the subspace spanned by p1 , p2 , and p3 is {1 + x2 , 1 − x − x2 + x3 , −3 − 5x + 3x2 + x3 }. 25. Set z = x − P(x, v1 ) − P(x, v2 ) − · · · − P(x, vk ). To verify that z is orthogonal to vi , we compute the inner product of z with vi : z, vi = x − P(x, v1 ) − P(x, v2 ) − · · · − P(x, vk ), vi = x, vi − P(x, v1 ), vi − P(x, v2 ), vi − · · · − P(x, vk ), vi . Since P(x, vj ) is a multiple of vj and the set {v1 , v2 , . . . , vk } is orthogonal, all of the subtracted terms on the right-hand side of this expression are zero except P(x, vi ), vi = x, vi x, vi vi , vi = vi , vi = x, vi . vi 2 vi 2 Therefore, z, vi = x, vi − x, vi = 0, which implies that z is orthogonal to vi . Solutions to Section 5.4 (a): TRUE. The given conclusion is essentially the definition of a least squares solution. (b): FALSE. The correct formula is x0 = (AT A)−1 AT b. (c)2017 Pearson Education. Inc. 377 (c): TRUE. Any least squares solution x0 to Ax = b must satisfy the normal equations in the system AT Ax0 = AT b. Since A is invertible, so is AT , and thus we can solve this system for x0 which yields precisely x0 = A−1 b. (d): FALSE. Example 5.4.3 provides a counterexample to this statement. (e): TRUE. In the text, we observed the projection matrix P projects b onto the column space of A as follows: P b = A(AT A)−1 AT b = Ax0 ∈ colspace(A). (f ): TRUE. We have P = A(AT A)−1 AT = AA−1 (AT )−1 AT = In . Problems: 1. We construct A= 6 1 −2 1 , A quick computation shows that 40 4 AT A = , and thus, 4 2 x= a b , (AT A)−1 = and 1 64 3 0 b= . 2 −4 −4 40 = 1 32 1 −2 −2 20 Continuing with the computation, we now find 1 1 −2 6 −2 3 3/8 x0 = (AT A)−1 AT b = = . 1 1 0 3/4 32 −2 20 Therefore, the least squares line is 3 3 x+ . 8 4 In this case, this is precisely the line that joins the two given points in the xy-plane. y(x) = 2. We construct 1 2 1 1 A quick computation shows that 5 T A A= 3 3 2 A= , x= a b , and , T and thus, (A A) Continuing with the computation, we now find 2 −3 1 T −1 T x0 = (A A) A b = −3 5 1 10 20 b= −1 2 1 . 2 −3 −3 5 = 10 20 = 10 0 . Therefore, the least squares line is y(x) = 10x. In this case, this is precisely the line that joins the two given points in the xy-plane. 3. We construct ⎡ 1 A=⎣ 2 3 ⎤ 1 1 ⎦, 1 x= a b ⎤ 10 b = ⎣ 20 ⎦ . 10 ⎡ , and (c)2017 Pearson Education. Inc. . . 378 A quick computation shows that 14 6 , AT A = 6 3 1 6 (AT A)−1 = and thus, 3 −6 −6 14 . Continuing with the computation, we now find 1 x0 = (AT A)−1 AT b = 6 3 −6 −6 14 1 1 2 1 3 1 ⎡ ⎤ 10 0 ⎣ 20 ⎦ = . 40/3 10 Therefore, the least squares line is y(x) = 4. We construct ⎡ ⎤ 1 1 ⎦, 1 2 A=⎣ 1 0 A quick computation shows that 5 AT A = 3 x= a b 40 . 3 ⎡ , 3 3 , ⎤ −2 b = ⎣ −3 ⎦ . 0 and (AT A)−1 = and thus, 1 6 3 −3 −3 5 . Continuing with the computation, we now find 1 x0 = (AT A)−1 AT b = 6 3 −3 −3 5 2 1 1 1 0 1 ⎡ ⎤ −2 −1 ⎣ −3 ⎦ = . −2/3 0 Therefore, the least squares line is 2 y(x) = −x − . 3 5. We construct ⎡ 2 ⎢ 0 A=⎢ ⎣ 5 1 ⎤ 1 1 ⎥ ⎥, 1 ⎦ 1 x= A quick computation shows that 30 8 T A A= , and thus, 8 4 a b T (A A) ⎤ 5 ⎢ −1 ⎥ ⎥ b=⎢ ⎣ 3 ⎦. −3 ⎡ , −1 and 1 = 56 4 −8 −8 30 1 = 28 2 −4 −4 15 Continuing with the computation, we now find x0 = (AT A)−1 AT b = 1 28 2 −4 −4 15 2 1 0 1 5 1 Therefore, the least squares line is y(x) = x − 1. (c)2017 Pearson Education. Inc. 1 1 ⎤ 5 ⎢ −1 ⎥ 1 ⎥= ⎢ . ⎣ 3 ⎦ −1 −3 ⎡ . 379 ⎤ 1 1 ⎥ ⎥, 1 ⎦ 1 ⎡ 6. We construct 0 ⎢ 1 A=⎢ ⎣ 2 4 A quick computation shows that 21 T A A= 7 7 4 a b x= , and , ⎤ 3 ⎢ −1 ⎥ ⎥ b=⎢ ⎣ 6 ⎦. 6 ⎡ T and thus, (A A) −1 1 = 35 Continuing with the computation, we now find x0 = (AT A)−1 AT b = 1 35 4 −7 −7 21 0 1 1 1 2 1 4 1 4 −7 −7 21 . ⎤ 3 ⎢ −1 ⎥ ⎥ = 6/5 . ⎢ ⎣ 6 ⎦ 7/5 6 ⎡ Therefore, the least squares line is 6 7 x+ . 5 5 y(x) = −7 ⎢ −4 ⎢ A=⎢ ⎢ 2 ⎣ 3 6 A quick computation shows that 114 T A A= 0 0 5 ⎤ 3 ⎢ 0 ⎥ ⎥ ⎢ ⎥ b=⎢ ⎢ −1 ⎥ . ⎣ 6 ⎦ −1 ⎡ ⎤ 1 1 ⎥ ⎥ 1 ⎥ ⎥, 1 ⎦ 1 ⎡ 7. We construct x= a b , and , and thus, T (A A) −1 Continuing with the computation, we now find T x0 = (A A) −1 T A b= 1 114 0 = 1 114 0 0 1 5 . ⎤ 3 ⎢ 0 ⎥ ⎥ −11/114 0 −7 −4 2 3 6 ⎢ ⎥ ⎢ −1 ⎥ = . 1 7/5 1 1 1 1 1 ⎢ 5 ⎣ 6 ⎦ −1 ⎡ Therefore, the least squares line is y(x) = − 8. We construct ⎡ 1 A=⎣ 4 9 1 2 3 ⎤ 1 1 ⎦, 1 A quick computation shows that ⎤ ⎡ 98 36 14 AT A = ⎣ 36 14 6 ⎦ , 14 6 3 11 7 x+ . 114 5 ⎡ ⎤ a x = ⎣ b ⎦, c and thus, ⎡ and ⎤ 10 b = ⎣ 20 ⎦ . 10 ⎡ ⎤ 3 −12 10 1 49 −42 ⎦ . (AT A)−1 = ⎣ −12 2 10 −42 38 (c)2017 Pearson Education. Inc. 380 Continuing with the computation, we now find ⎡ ⎤⎡ 3 −12 10 1 1 49 −42 ⎦ ⎣ 1 x0 = (AT A)−1 AT b = ⎣ −12 2 10 −42 38 1 4 2 1 ⎤⎡ ⎤ ⎡ ⎤ 9 10 −10 3 ⎦ ⎣ 20 ⎦ = ⎣ 40 ⎦ . 1 10 −20 Therefore, the least squares parabola is y(x) = −10x2 + 40x − 20. In this case, this is precisely the unique parabola that joins the three given points in the xy-plane. 9. We construct ⎡ 4 A=⎣ 1 0 2 1 0 ⎤ 1 1 ⎦, 1 A quick computation shows that ⎡ ⎤ 17 9 5 AT A = ⎣ 9 5 3 ⎦ , 5 3 3 ⎡ ⎤ a x = ⎣ b ⎦, c ⎡ and ⎡ ⎤ 3 −6 1 1 (AT A)−1 = ⎣ −6 13 −3 ⎦ . 2 1 −3 2 and thus, Continuing with the computation, we now find ⎡ ⎤⎡ 3 −6 1 4 1 x0 = (AT A)−1 AT b = ⎣ −6 13 −3 ⎦ ⎣ 2 2 1 −3 2 1 Therefore, the least squares parabola is ⎤ −2 b = ⎣ −3 ⎦ . 0 1 1 1 ⎤⎡ ⎤ ⎡ ⎤ −2 0 2 0 ⎦ ⎣ −3 ⎦ = ⎣ −5 ⎦ . 1 0 0 y(x) = 2x2 − 5x. In this case, this is precisely the unique parabola that joins the three given points in the xy-plane. 13. We construct ⎤ 0 ⎢ 3 ⎥ ⎥ A=⎢ ⎣ 5 ⎦, 6 ⎤ 0 ⎢ −10 ⎥ ⎥ b=⎢ ⎣ −14 ⎦ . −19 ⎡ ⎡ x= a , and A quick computation shows that AT A = 70 , and thus, (AT A)−1 = 1 70 . Continuing with the computation, we now find ⎤ 0 ⎢ −10 ⎥ ⎥ ⎢ ⎣ −14 ⎦ = −19 ⎡ x0 = (AT A)−1 AT b = 1 70 0 3 5 6 − 107 35 . Thus, we can estimate the spring constant k for the spring with k = −a = 107 35 pounds per inch. 14. Since the population growth of a bacteria culture is known to follow an exponential model, we will seek to write P (t) in the form P (t) = Cekt , for some constants C and k. Using the notation y(t) = ln P (t), we compute the following table from the given data (all values are approximated to three decimal places): (c)2017 Pearson Education. Inc. 381 t y(t) 0 1.253 4 2.833 5 3.239 8 4.449 10 5.244 We can now formulate this least squares problem via the vector equation Ax = y, where ⎤ ⎡ ⎤ ⎡ 1.253 0 1 ⎢ 2.833 ⎥ ⎢ 4 1 ⎥ ⎥ ⎢ ⎥ ⎢ ⎢ 3.239 ⎥ . ⎥, x = k , 5 1 and y = A=⎢ ⎥ ⎢ ⎥ ⎢ b ⎣ 4.449 ⎦ ⎣ 8 1 ⎦ 5.244 10 1 We compute 205 27 AT A = 27 5 , (AT A)−1 = and thus, 1 296 5 −27 −27 205 . Continuing the computation, we now find ⎤ 1.253 ⎢ 2.833 ⎥ ⎥ .400 10 ⎢ ⎢ 3.239 ⎥ = . ⎥ 1.275 1 ⎢ ⎣ 4.449 ⎦ 5.244 ⎡ x0 = (AT A)−1 AT y = 1 296 5 −27 −27 205 0 1 4 1 5 1 8 1 Thus, we approximate k0 = 0.400 and b0 = 1.275. Hence, C0 = eb0 ≈ 3.579. Thus, our best fit exponential function describing the growth of this bacteria culture is P (t) = 3.579e0.400t thousands of bacteria after t hours. 15. Since the population growth of a bacteria culture is known to follow an exponential model, we will seek to write P (t) in the form P (t) = Cekt , for some constants C and k. Using the notation y(t) = ln P (t), we compute the following table from the given data (all values are approximated to three decimal places): t y(t) 0 .405 12 1.609 20 2.398 24 2.773 We can now formulate this least squares problem via the vector equation Ax = y, where ⎤ ⎤ ⎡ ⎡ .405 0 1 ⎢ 1.609 ⎥ ⎢ 12 1 ⎥ k ⎥ ⎥ and y = ⎢ A=⎢ ⎣ 2.398 ⎦ . ⎣ 20 1 ⎦ , x = b , 2.773 24 1 We compute T A A= 1120 56 56 4 , and thus, T (A A) −1 1 = 1344 4 −56 −56 1120 1 = 336 1 −14 −14 280 Continuing the computation, we now find x0 = (AT A)−1 AT y = 1 336 1 −14 −14 280 0 1 12 1 20 1 (c)2017 Pearson Education. Inc. 24 1 ⎤ .405 ⎢ 1.609 ⎥ ⎥ = .099 . ⎢ ⎣ 2.398 ⎦ .412 2.773 ⎡ . 382 Thus, we approximate k0 = 0.099 and b0 = .412. Hence, C0 = eb0 ≈ 1.510. Thus, our best fit exponential function describing the growth of this bacteria culture is P (t) = 1.510e0.099t thousands of bacteria after t minutes. 16. We will use the notation y(t) = ln A(t). We compute the following table from the given data (all values are approximated to three decimal places): t y(t) 0 4.605 1 3.199 3 .405 4 -.916 6 -2.303 We can now formulate this least squares problem via the vector equation Ax = y, where ⎤ ⎡ ⎡ ⎤ 4.605 0 1 ⎢ 3.199 ⎥ ⎢ 1 1 ⎥ ⎥ ⎢ ⎢ ⎥ k ⎢ ⎥ .405 ⎥ , and y = ⎢ A = ⎢ 3 1 ⎥, x = ⎥. ⎢ b ⎣ −.916 ⎦ ⎣ 4 1 ⎦ −2.303 6 1 We compute T A A= 62 14 14 5 , T and thus, (A A) −1 Continuing the computation, we now find T x0 = (A A) −1 1 A y= 114 T 5 −14 −14 62 1 = 114 5 −14 −14 62 . ⎤ 4.605 ⎢ 3.199 ⎥ ⎥ −1.186 6 ⎢ ⎥ ⎢ .405 ⎥ = . 4.319 1 ⎢ ⎣ −.916 ⎦ −2.303 ⎡ 0 1 1 1 3 1 4 1 Thus, we approximate k0 = −1.186 and b0 = 4.319. Hence, C0 = eb0 ≈ 75.113. Thus, our best fit exponential function describing the decay of this radioactive sample is A(t) = 75.113e−1.186t grams after t hours. 17. (a). Since A is m × n, AT A is n × n and AT is n × m, we conclude from matrix multiplication that P is m × m. (b). We have P A = [A(AT A)−1 AT ]A = A(AT A)−1 (AT A) = AI = A and P 2 = [A(AT A)−1 AT ][A(AT A)−1 AT ] = A(AT A)−1 (AT A)(AT A)−1 AT = A(AT A)−1 AT = P, as needed. (c). We will show that P T = P : P T = [A(AT A)−1 AT ]T = (AT )T [(AT A)−1 ]T AT = A[(AT A)T ]−1 AT = A[AT (AT )T ]−1 AT = A[AT A]−1 AT = P, as needed. 18. According to Theorem 5.4.2, the least squares solution to Ax = b must satisfy AT Ax0 = AT b. In this case, A is invertible, so we can solve explicitly for x0 : x0 = (AT A)−1 AT b = A−1 (AT )−1 AT b = A−1 b. This is the only least squares solution for this system. (c)2017 Pearson Education. Inc. 383 Solutions to Section 5.5 Problems: 1. We have u, v = (2, 3), (4, −1) = 2 · 4 + 3 · (−1) = 5, u = and so θ = cos −1 u, v uv = cos −1 5 √ √ 13 17 √ 13, v = √ 17, ≈ 1.23 radians. 2. We have u, v = (−2, −1, 2, 4), (−3, 5, 1, 1) = (−2) · (−3) + (−1) · 5 + 2 · 1 + 4 · 1 = 7, u = 5, v = 6, and so θ = cos −1 u, v uv = cos −1 7 5·6 = cos−1 (7/30) ≈ 1.34 radians. 3. For Problem 1, we have u, v = (2, 3), (4, −1) = 2 · 2 · 4 + 3 · (−1) = 13, u = and so θ = cos−1 u, v uv = cos−1 √ 13 √ 17 33 √ 17, v = √ 33, ≈ 0.99 radians. For Problem 2, we have u, v = (−2, −1, 2, 4), (−3, 5, 1, 1) = 2 · (−2) · (−3) + (−1) · 5 + 2 · 1 + 4 · 1 = 13, u = and so θ = cos−1 u, v uv = cos−1 √ 13 √ 29 · 45 √ 29, v = √ 45, ≈ 1.20 radians. 4. We will obtain bases for the ⎡ rowspace,⎤ columnspace, and nullspace and orthonormalize them. A row1 2 6 ⎢ 0 1 2 ⎥ ⎥ echelon form of A is given by ⎢ ⎣ 0 0 0 ⎦. We see that a basis for the rowspace of A is given by 0 0 0 {(1, 2, 6), (0, 1, 2)}. We apply Gram-Schmidt to this set, and thus we need to replace (0, 1, 2) by 14 13 2 14 . (0, 1, 2) − (1, 2, 6) = − , , − 41 41 41 41 So an orthogonal basis for the rowspace of A is given by 14 13 2 (1, 2, 6) , − , , − . 41 41 41 (c)2017 Pearson Education. Inc. 384 To replace this with an orthonormal basis, we must normalize each vector. The first one has norm the second one has norm √341 . Hence, an orthonormal basis for the rowspace of A is √ 41 and √ √ 14 13 41 41 2 − , ,− . (1, 2, 6), 41 3 41 41 41 Returning to the row-echelon form of A obtained above, we see that a basis for the columnspace of A is ⎧⎡ ⎤⎫ ⎤ ⎡ 2 ⎪ 1 ⎪ ⎪ ⎪ ⎨⎢ ⎥⎬ ⎥ ⎢ ⎢ 2 ⎥,⎢ 1 ⎥ . ⎣ 0 ⎦ ⎣ 1 ⎦⎪ ⎪ ⎪ ⎪ ⎭ ⎩ 0 1 ⎤ ⎡ 2 ⎢ 1 ⎥ ⎥ We apply Gram-Schmidt to this set, and thus we need to replace ⎢ ⎣ 1 ⎦ by 0 ⎤ ⎡ ⎤ ⎤ ⎡ ⎡ 2 1 4 ⎢ 1 ⎥ 4 ⎢ 2 ⎥ 1 ⎢ −1 ⎥ ⎥ ⎢ ⎥ ⎥ ⎢ ⎢ ⎣ 1 ⎦ − 6 ⎣ 0 ⎦ = 3 ⎣ 3 ⎦. 0 1 −2 So an orthogonal basis for the columnspace of A is given by ⎧⎡ ⎤ ⎤⎫ ⎡ 1 4 ⎪ ⎪ ⎪ ⎪ ⎨⎢ ⎥ ⎥⎬ ⎢ ⎢ 2 ⎥ , 1 ⎢ −1 ⎥ . ⎣ 0 ⎦ 3 ⎣ 3 ⎦⎪ ⎪ ⎪ ⎪ ⎩ ⎭ 1 −2 √ √ The norms of these vectors are, respectively, 6 and 30/3. Hence, we normalize the above orthogonal basis to obtain the orthonormal basis for the columnspace: ⎧ ⎤⎫ ⎡ ⎡ ⎤ 1 4 ⎪ ⎪ √ √ ⎪ ⎪ ⎨ 6⎢ ⎥⎬ ⎢ ⎥ ⎢ 2 ⎥ , 30 ⎢ −1 ⎥ . ⎪ 6 ⎣ 0 ⎦ 30 ⎣ 3 ⎦⎪ ⎪ ⎪ ⎭ ⎩ −2 1 Returning once more to the row-echelon form of A obtained above, we see that, in order to find the nullspace of A, we must solve the equations x + 2y + 6z = 0 and y + 2z = 0. Setting z = t as a free variable, we find that y = −2t and x = −2t. Thus, a basis for the nullspace of A is {(−2, −2, 1)}, which can be normalized to 2 2 1 − ,− , . 3 3 3 ⎤ ⎡ 1 3 5 ⎢ 0 1 3/2 ⎥ ⎥ ⎢ ⎥ 5. A row-echelon form for A is ⎢ ⎢ 0 0 1 ⎥. Note that since rank(A) = 3, nullity(A) = 0, and so ⎣ 0 0 0 ⎦ 0 0 0 there is no basis for nullspace(A). Moreover, rowspace(A) is a 3-dimensional subspace of R3 , and therefore, rowspace(A) = R3 . An orthonormal basis for this is {(1, 0, 0), (0, 1, 0), (0, 0, 1)}. (c)2017 Pearson Education. Inc. 385 Finally, consider the columnspace of A. We must apply the Gram-Schmidt process to the three columns of A. Thus, we replace the second column vector by ⎤ ⎤ ⎡ ⎤ ⎡ ⎡ −1 1 3 ⎢ −1 ⎥ ⎢ 1 ⎥ ⎢ −3 ⎥ ⎥ ⎥ ⎢ ⎥ ⎢ ⎢ ⎥ − 4⎢ 0 ⎥ = ⎢ 2 ⎥. 2 v2 = ⎢ ⎥ ⎥ ⎢ ⎥ ⎢ ⎢ ⎣ 1 ⎦ ⎣ 1 ⎦ ⎣ 5 ⎦ 1 1 5 Next, we replace the third column vector by ⎤ ⎤ ⎡ ⎡ ⎡ ⎤ ⎤ ⎡ 5 1 −1 3 ⎥ ⎥ ⎢ 1 ⎥ ⎢ ⎢ ⎥ ⎢ ⎢ ⎥ 7 ⎢ −1 ⎥ 3 ⎢ 1 ⎥ ⎢ 3 ⎥ ⎥ ⎢ ⎢ ⎢ ⎢ ⎥ ⎥ v3 = ⎢ 3 ⎥ − ⎢ 0 ⎥ − ⎢ 2 ⎥ = ⎢ 0 ⎥ ⎥. ⎣ 2 ⎦ 2 ⎣ 1 ⎦ 2 ⎣ 1 ⎦ ⎣ −3 ⎦ 3 1 8 1 Hence, an orthogonal basis for the columnspace of A is ⎧⎡ ⎤⎫ ⎤ ⎡ ⎤ ⎡ 3 ⎪ −1 1 ⎪ ⎪ ⎪ ⎪ ⎪ ⎥⎪ ⎥ ⎢ ⎥ ⎢ ⎪ ⎨⎢ ⎢ −1 ⎥ ⎢ 1 ⎥ ⎢ 3 ⎥⎬ ⎢ 0 ⎥,⎢ 2 ⎥,⎢ 0 ⎥ . ⎥ ⎥ ⎢ ⎥ ⎢ ⎢ ⎪ ⎪ ⎪ ⎣ 1 ⎦ ⎣ 1 ⎦ ⎣ −3 ⎦⎪ ⎪ ⎪ ⎪ ⎪ ⎭ ⎩ 3 1 1 Normalizing each vector yields the orthonormal basis ⎧ ⎡ ⎤ ⎤⎫ ⎤ ⎡ ⎡ 1 3 ⎪ −1 ⎪ ⎪ ⎪ ⎪ ⎪ ⎢ −1 ⎥ ⎪ ⎢ 3 ⎥⎪ ⎢ 1 ⎥ ⎨ ⎬ ⎥ ⎥ ⎥ ⎢ ⎢ ⎢ 1⎢ ⎥ , √1 ⎢ 2 ⎥ , 1 ⎢ 0 ⎥ . 0 ⎥ ⎥ ⎥ ⎢ ⎪ 2⎢ 8⎢ ⎪ ⎪ ⎣ 1 ⎦ ⎣ 1 ⎦ 6 ⎣ −3 ⎦⎪ ⎪ ⎪ ⎪ ⎪ ⎩ ⎭ 1 3 1 6. Let x1 = (5, −1, 2) and let x2 = (7, 1, 1). Using the Gram-Schmidt process, we have v1 = x1 = (5, −1, 2) and v2 = x2 − x2 , v1 36 6 12 11 7 v1 = (7, 1, 1) − (5, −1, 2) = (7, 1, 1) − (6, − , ) = (1, , − ). 2 |v1 | 30 5 5 5 5 Hence, an orthogonal basis is given by {(5, −1, 2), (1, 11 7 , − )}. 5 5 7. We can readily verify that S spans R3 , so therefore an obvious orthogonal basis for span(S) is {(1, 0, 0), (0, 1, 0), (0, 0, 1)}. Alternatively, for practice with Gram-Schmidt, we would proceed as follows: Let x1 = (6, −3, 2), x2 = (1, 1, 1), and x3 = (1, −8, −1). Using the Gram-Schmidt process, we have v1 = x1 = (6, −3, 2), (c)2017 Pearson Education. Inc. 386 v2 = x2 − x2 , v1 5 v1 = (1, 1, 1) − (6, −3, 2) = 2 v1 49 19 64 39 , , 49 49 49 and x3 , v1 x3 , v2 4 38 v3 = x3 − v1 − v2 = (1, −8, −1) − (6, −3, 2) + (19, 64, 39) = 2 2 v1 v2 7 427 , 45 36 81 − ,− , 61 61 61 . Hence, an orthogonal basis is given by 19 64 39 45 36 81 (6, −3, 2), , − ,− , . , , 49 49 49 61 61 61 8. We can check that S spans P3 (R), so therefore we can apply Gram-Schmidt to the basis {1, x, x2 , x3 } for P3 (R), instead of the given set of polynomials. Let x1 = 1, x2 = x, x3 = x2 , and x4 = x3 . Using the Gram-Schmidt process, we have v1 = x1 = 1, v2 = x2 − v3 = x3 − x2 , v1 1 v1 = x − , 2 v1 2 x3 , v1 x3 , v2 1 1 1 v1 − v2 = x2 − (x − ) − = x2 − x + , v1 2 v2 2 2 3 6 and v4 = x4 − x4 , v1 x4 , v2 x4 , v3 1 1 3 3 9 1 3 1 v1 − v2 − v3 = x3 − − (x − ) − (x2 − x + ) = x3 − x2 + x − . v1 2 v2 2 v3 2 4 10 2 2 6 2 5 20 Hence, an orthogonal basis is given by 1 1 3 3 1 . 1, x − , x2 − x + , x3 − x2 + x − 2 6 2 5 20 9. It is easy to see that the span of the set of vectors in this problem is the set of all 2 × 2 symmetric matrices. Therefore, we can simply give the orthogonal basis 1 0 0 1 0 0 , , 0 0 1 0 0 1 for the set of all 2 × 2 symmetric matrices. 10. (a). We must verify the four axioms for an inner product given in Definition 5.1.3. Axiom 1: We have p · p = p(t0 )p(t0 ) + p(t1 )p(t1 ) + · · · + p(tn )p(tn ) = p(t0 )2 + p(t1 )2 + · · · + p(tn )2 ≥ 0. Moreover, p(t0 )2 + p(t1 )2 + · · · + p(tn )2 = 0 ⇐⇒ p(t0 ) = p(t1 ) = · · · = p(tn ) = 0. But the only polynomial of degree ≤ n which has more than n roots is the zero polynomial. Thus, p · p = 0 ⇐⇒ p = 0. (c)2017 Pearson Education. Inc. 387 Axiom 2: We have p · q = p(t0 )q(t0 ) + p(t1 )q(t1 ) + · · · + p(tn )q(tn ) = q(t0 )p(t0 ) + q(t1 )p(t1 ) + · · · + q(tn )p(tn ) = q · p for all p, q ∈ Pn (R). Axiom 3: Let k be a scalar, and let p, q ∈ Pn (R). Then (kp) · q = (kp)(t0 )q(t0 ) + (kp)(t1 )q(t1 ) + · · · + (kp)(tn )q(tn ) = kp(t0 )q(t0 ) + kp(t1 )q(t1 ) + · · · + kp(tn )q(tn ) = k[p(t0 )q(t0 ) + p(t1 )q(t1 ) + · · · + p(tn )q(tn )] = k[p · q], as required. Axiom 4: Let p1 , p2 , q ∈ Pn (R). Then we have (p1 + p2 ) · q = (p1 + p2 )(t0 )q(t0 ) + (p1 + p2 )(t1 )q(t1 ) + · · · + (p1 + p2 )(tn )q(tn ) = [p1 (t0 ) + p2 (t0 )]q(t0 ) + [p1 (t1 ) + p2 (t1 )]q(t1 ) + · · · + [p1 (tn ) + p2 (tn )]q(tn ) = [p1 (t0 )q(t0 ) + p1 (t1 )q(t1 ) + · · · + p1 (tn )q(tn )] + [p2 (t0 )q(t0 ) + p2 (t1 )q(t1 ) + · · · + p2 (tn )q(tn )] = (p1 · q) + (p2 · q), as required. (b). We have p0 , p1 = p0 (−3)p1 (−3) + p0 (−1)p1 (−1) + p0 (1)p1 (1) + p0 (3)p1 (3) = (1)(−3) + (1)(−1) + (1)(1) + (1)(3) = 0, which proves that p0 and p1 are orthogonal. (c). The projection of p2 onto span{p0 , p1 } is given by p2 , p0 p2 , p1 20 0 p0 + p1 = p0 + p1 = 5. p0 2 p1 2 4 11 Therefore, we can take q = t2 − 5. 11. We are given the point P (2, 3, 4), and we can write the equation of the given line L parametrically via x(t) = 6t, −t, −4t. A parallel vector to L is given by v = (6, −1, −4). Since the point Q(0, 0, 0) lies on L, we form the vector pointing from Q to P : w = (2, 3, 4) − (0, 0, 0) = (2, 3, 4). Thus, we have P(w, v) = (2, 3, 4), (6, −1, −4) 7 (2, −1, 3) = − (6, −1, −4). 2 ||(6, −1, −4)|| 53 Thus, the distance from P to L is given by ||w − P(w, v)|| = ||(2, 3, 4) − (− 1 42 7 28 , , )|| = ||(148, 152, 184)|| = 53 53 53 53 √ 78864 . 53 12. We will use the formula obtained in Problem 28(c) in Section 5.2 with a = 2, b = −1, c = 3, d = −6, and x0 = y0 = z0 = 0. The requested distance is: |ax0 + by0 + cz0 + d| 6 | − 6| √ =√ . = 2 2 2 2 2 2 14 a +b +c 2 + (−1) + 3 (c)2017 Pearson Education. Inc. 388 13. We will use the formula obtained in Problem 28(c) in Section 5.2 with a = −1, b = 3, c = 3, d = −8, and x0 = −1, y0 = 3, and z0 = 5. The requested distance is: |(−1)(−1) + (3)(3) + (3)(5) + (−8)| 17 |ax0 + by0 + cz0 + d| √ = =√ . 19 a 2 + b2 + c 2 (−1)2 + 32 + 32 14. We must start by determining the equation of the given plane. We can obtain a normal vector to this plane by computing the cross product of v and w: v × w = −15, 14, −3. Thus, we can write the equation of the plane as −15(x − x0 ) + 14(y − y0 ) − 3(z − z0 ) = 0, or using the point (−1, 0, 5), −15(x + 1) + 14y − 3(z − 5) = −15x + 14y − 3z = 0. Next, we will use the formula obtained in Problem 28(c) in Section 5.2 with a = −15, b = 14, c = −3, d = 0, and x0 = −2, y0 = 8, and z0 = 0. The requested distance is: |(−15)(−2) + (14)(8) + (−3)(0) + 0| 142 |ax0 + by0 + cz0 + d| √ = =√ . 2 2 2 2 2 2 150 a +b +c (−15) + 14 + (−3) ⎡ 15. We construct 0 ⎢ 1 ⎢ A=⎢ ⎢ 2 ⎣ 3 4 ⎤ 1 1 ⎥ ⎥ 1 ⎥ ⎥, 1 ⎦ 1 A quick computation shows that 30 10 AT A = , and thus, 10 5 ⎤ −2 ⎢ −1 ⎥ ⎥ ⎢ ⎥ b=⎢ ⎢ 1 ⎥. ⎣ 2 ⎦ 2 ⎡ x= a b , and (AT A)−1 = 1 50 5 −10 −10 30 = 1 10 1 −2 −2 6 Continuing with the computation, we now find ⎤ −2 ⎢ −1 ⎥ ⎥ 4 ⎢ ⎢ 1 ⎥ = 11/10 . ⎥ −9/5 1 ⎢ ⎣ 2 ⎦ 2 ⎡ T x0 = (A A) −1 1 A b= 10 T 1 −2 −2 6 0 1 1 1 2 1 3 1 Therefore, the least squares line is y(x) = 16. We construct ⎡ −1 ⎢ 1 A=⎢ ⎣ 2 3 ⎤ 1 1 ⎥ ⎥, 1 ⎦ 1 x= 11 9 x− . 10 5 a b ⎤ 5 ⎢ 1 ⎥ ⎥ b=⎢ ⎣ 1 ⎦. −3 ⎡ , and (c)2017 Pearson Education. Inc. . 389 A quick computation shows that 15 T A A= 5 5 4 , T and thus, (A A) −1 1 = 35 4 −5 −5 15 . Continuing with the computation, we now find 1 1 4 −5 −64 −64/35 T −1 T x0 = (A A) A b = = = . 135 27/7 35 −5 15 35 Therefore, the least squares line is y(x) = − ⎡ 17. We construct −4 ⎢ −3 A=⎢ ⎣ −2 0 ⎤ 1 1 ⎥ ⎥, 1 ⎦ 1 A quick computation shows that 29 −9 T A A= , −9 4 x= 64 23 x+ . 35 7 a b ⎤ −1 ⎢ 1 ⎥ ⎥ b=⎢ ⎣ 3 ⎦. 7 ⎡ , and T and thus, (A A) −1 1 = 35 4 9 9 29 . Continuing with the computation, we now find x0 = (AT A)−1 AT b = 1 35 4 9 9 29 −4 −3 −2 0 1 1 1 1 ⎤ −1 ⎢ 1 ⎥ ⎥= 2 . ⎢ ⎣ 3 ⎦ 7 7 ⎡ Therefore, the least squares line is y(x) = 2x + 7. ⎡ 18. We construct −3 ⎢ −2 ⎢ A=⎢ ⎢ −1 ⎣ 0 2 ⎤ 1 1 ⎥ ⎥ 1 ⎥ ⎥, 1 ⎦ 1 A quick computation shows that 18 −4 AT A = , −4 5 ⎤ 1 ⎢ 0 ⎥ ⎥ ⎢ ⎥ b=⎢ ⎢ 1 ⎥. ⎣ −1 ⎦ −1 ⎡ x= a b and thus, , and (AT A)−1 = 1 74 5 4 4 18 . Continuing with the computation, we now find ⎤ 1 ⎢ 0 ⎥ ⎥ 4 −3 −2 −1 0 2 ⎢ ⎢ 1 ⎥ = −15/37 . ⎥ −12/37 18 1 1 1 1 1 ⎢ ⎣ −1 ⎦ −1 ⎡ T x0 = (A A) −1 1 A b= 74 T 5 4 (c)2017 Pearson Education. Inc. 390 Therefore, the least squares line is y(x) = − 15 12 x− . 37 37 20. Note that x − y, vi = x, vi − y, vi = 0, by assumption. Let v be an arbitrary vector in V , and write v = a1 v 1 + a2 v 2 + · · · + a n v n , for some scalars a1 , a2 , . . . , an . Observe that x − y, v = x − y, a1 v1 + a2 v2 + · · · + an vn = a1 x − y, v1 + a2 x − y, v2 + · · · + an x − y, vn = a1 · 0 + a2 · 0 + · · · + an · 0 = 0. Thus, x − y is orthogonal to every vector in V . In particular x − y, x − y = 0, and hence, x − y = 0. Therefore x = y. Chapter 6 Solutions Solutions to Section 6.1 True-False Review: (a): FALSE. The conditions T (u + v) = T (u) + T (v) and T (c · v) = c · T (v) must hold for all vectors u, v in V and for all scalars c, not just “for some”. (b): FALSE. The dimensions of the matrix A should be m × n, not n × m, as stated in the question. (c): FALSE. This will only necessarily hold for a linear transformation, not for more general mappings. (d): TRUE. This is precisely the definition of the matrix associated with a linear transformation, as given in the text. (e): TRUE. Since 0 = T (0) = T (v + (−v)) = T (v) + T (−v), we conclude that T (−v) = −T (v). (f ): TRUE. Using the properties of a linear transformation, we have T ((c + d)v) = T (cv + dv) = T (cv) + T (dv) = cT (v) + dT (v), as stated. Problems: 1. Let (x1 , x2 , x3 ), (y1 , y2 , y3 ) ∈ R3 and c ∈ R. (c)2017 Pearson Education. Inc. 391 T ((x1 , x2 , x3 ) + (y1 , y2 , y3 )) = T (x1 + y1 , x2 + y2 , x3 + y3 ) = (x1 + y1 + 3x2 + 3y2 + x3 + y3 , x1 + y1 − x2 − y2 ) = (x1 + 3x2 + x3 , x1 − x2 ) + (y1 + 3y2 + y3 , y1 − y2 ) = T (x1 , x2 , x3 ) + T (y1 , y2 , y3 ). T (c(x1 , x2 , x3 )) = T (cx1 , cx2 , cx3 ) = (cx1 +3cx2 +cx3 , cx1 −cx2 ) = c(x1 +3x2 +x3 , x1 −x2 ) = cT (x1 , x2 , x3 ). Thus, T is a linear transformation. 2. Let (x1 , x2 ), (y1 , y2 ) ∈ R2 and c ∈ R. T ((x1 , x2 ) + (y1 , y2 )) = T (x1 + y1 , x2 + y2 ) = (x1 + y1 + 2x2 + 2y2 , 2x1 + 2y1 − x2 − y2 ) = (x1 + 2x2 , 2x1 − x2 ) + (y1 + 2y2 , 2y1 − y2 ) = T (x1 , x2 ) + T (y1 , y2 ). T (c(x1 , x2 )) = T (cx1 , cx2 ) = (cx1 + 2cx2 , 2cx1 − cx2 ) = c(x1 + 2x2 , 2x1 − x2 ) = cT (x1 , x2 ). Thus, T is a linear transformation. 3. Let y1 , y2 ∈ C 2 (I) and c ∈ R. Then, T (y1 + y2 ) = (y1 + y2 ) + a1 (y1 + y2 ) + a2 (y1 + y2 ) = (y1 + a1 y1 + a2 y1 ) + (y2 + a1 y2 + a2 y2 ) = T (y1 ) + T (y2 ). T (cy1 ) = (cy1 ) + a1 (cy1 ) + a2 (cy1 ) = c(y1 + a1 y1 + a2 y1 ) = cT (y1 ). Consequently, T is a linear transformation. 4. Let y1 , y2 ∈ C 2 (I) and c ∈ R. Then, T (y1 + y2 ) = (y1 + y2 ) − 16(y1 + y2 ) = (y1 − 16y1 ) + (y2 − 16y2 ) = T (y1 ) + T (y2 ). T (cy1 ) = (cy1 ) − 16(cy1 ) = c(y1 − 16y1 ) = cT (y1 ). Consequently, T is a linear transformation. 5. Let A1 , A2 , B ∈ Mn (R) and c ∈ R. T (A1 + A2 ) = (A1 + A2 )B − B(A1 + A2 ) = A1 B + A2 B − BA1 − BA2 = (A1 B − BA1 ) + (A2 B − BA2 ) = T (A1 ) + T (A2 ). T (cA1 ) = (cA1 B) − B(cA1 ) = c(A1 B − BA1 ) = cT (A1 ). Consequently, T is a linear transformation. b b b b 6. Let f, g ∈ V and c ∈ R. Then T (f +g) = (f +g)(x)dx = [f (x)+g(x)]dx = f (x)dx+ g(x)dx = a a a T (f ) + T (g). b b T (cf ) = [cf (x)]dx = c f (x)dx = cT (f ). Therefore, T is a linear transformation. a a 7. Let A, B ∈ Mn (R) and c ∈ R. Then, n n n , , , T (A + B) = tr(A + B) = (akk + bkk ) = akk + bkk = tr(A) + tr(B) = T (A) + T (B). k=1 k=1 k=1 (c)2017 Pearson Education. Inc. a 392 T (cA) = tr(cA) = n , k=1 cakk = c n , akk = ctr(A) = cT (A). k=1 Consequently, T is a linear transformation. 8. Let A, B ∈ Mn (R) and c ∈ R. Then, S(A + B) = (A + B) + (A + B)T = A + AT + B + B T = S(A) + S(B). S(cA) = (cA) + (cA)T = c(A + AT ) = cS(A). Consequently, S is a linear transformation. 9. According to the given mapping, we have T (0) = 1, which violates part (1) of Theorem 6.1.11. The conclusion is that T must not be linear. 10. We show with a specific example addition (the first property in the that T fails to respect definition of 0 1 0 0 0 1 a linear transformation). Let A = and B = . We have A + B = , so that 0 0 1 0 1 0 2 0 1 T (A + B) = (A + B)2 = = I2 , 1 0 whereas T (A) + T (B) = A2 + B 2 = 02 + 02 = 02 . Therefore, in this example, we have T (A + B) = T (A) + T (B). (Many other examples can be given as well.) 11. According to the given mapping, we have T (0) = x, which violates part (1) of Theorem 6.1.11. The conclusion is that T must not be linear. 12. Let x = (x1 , x2 ), y = (y1 , y2 ) be in R2 . Then T (x + y) = T (x1 + y1 , x2 + y2 ) = (x1 + x2 + y1 + y2 , 2), whereas T (x) + T (y) = (x1 + x2 , 2) + (y1 + y2 , 2) = (x1 + x2 + y1 + y2 , 4). We see that T (x + y) = T (x) + T (y), hence T is not a linear transformation. 13. Let A ∈ M2 (R) and c ∈ R. Then T (cA) = det(cA) = c2 det(A) = c2 T (A). Since T (cA) = cT (A) in general, it follows that T is not a linear transformation. 3 −2 14. If T (x1 , x2 ) = (3x1 − 2x2 , x1 + 5x2 ), then A = [T (e1 ), T (e2 )] = . 1 5 ⎡ ⎤ 1 3 15. If T (x1 , x2 ) = (x1 + 3x2 , 2x1 − 7x2 , x1 ), then A = [T (e1 ), T (e2 )] = ⎣ 2 −7 ⎦. 1 0 1 −1 1 . 16. If T (x1 , x2 , x3 ) = (x1 − x2 + x3 , x3 − x1 ), then A = [T (e1 ), T (e2 ), T (e3 )] = −1 0 1 17. If T (x1 , x2 , x3 ) = x1 + 5x2 − 3x3 , then A = [T (e1 ), T (e2 ), T (e3 )] = 1 5 −3 . ⎡ −1 ⎢ −1 18. If T (x1 , x2 , x3 ) = (x3 − x1 , −x1 , 3x1 + 2x3 , 0), then A = [T (e1 ), T (e2 ), T (e3 )] = ⎢ ⎣ 3 0 (c)2017 Pearson Education. Inc. 0 0 0 0 ⎤ 1 0 ⎥ ⎥. 2 ⎦ 0 393 19. T (x) = Ax = 1 3 −4 7 x1 x2 = x1 + 3x2 −4x1 + 7x2 , which we write as T (x1 , x2 ) = (x1 + 3x2 , −4x1 + 7x2 ). 20. T (x) = Ax = 2 −1 5 3 1 −2 ⎤ x1 ⎣ x2 ⎦ = 2x1 − x2 + 5x3 , which we write as 3x1 + x2 − 2x3 x3 ⎡ T (x1 , x2 , x3 ) = (2x1 − x2 + 5x3 , 3x1 + x2 − 2x3 ). ⎤ ⎡ ⎤ ⎤⎡ 2 2 −3 x1 2x1 + 2x2 − 3x3 2 ⎦ ⎣ x2 ⎦ = ⎣ 4x1 − x2 + 2x3 ⎦ , which we write as 21. T (x) = Ax = ⎣ 4 −1 5 7 −8 x3 5x1 + 7x2 − 8x3 ⎡ T (x1 , x2 , x3 ) = (2x1 + 2x2 − 3x3 , 4x1 − x2 + 2x3 , 5x1 + 7x2 − 8x3 ). ⎤ ⎤ ⎡ −3x −3 ⎢ −2x ⎥ ⎢ −2 ⎥ ⎥ , which we write as ⎥ ⎢ 22. T (x) = Ax = ⎢ ⎣ 0 ⎦ [x] = ⎣ 0 ⎦ x 1 ⎡ T (x) = (−3x, −2x, 0, x). ⎡ 23. T (x) = Ax = 1 −4 −6 0 2 ⎤ x1 ⎢ x2 ⎥ ⎢ ⎥ ⎢ x3 ⎥ = [x1 − 4x2 − 6x3 + 2x5 ], which we write as ⎢ ⎥ ⎣ x4 ⎦ x5 T (x1 , x2 , x3 , x4 , x5 ) = x1 − 4x2 − 6x3 + 2x5 . 24. Let u be a fixed vector in V , v1 , v2 ∈ V , and c ∈ R. Then T (v1 + v2 ) = u, v1 + v2 = u, v1 + u, v2 = T (v1 ) + T (v2 ). T (cv1 ) = u, cv1 = cu, v1 = cT (v1 ). Thus, T is a linear transformation. 25. We must show that the linear transformation T respects addition and scalar multiplication: T respects addition: Let v1 and v2 be vectors in V . Then we have T (v1 + v2 ) = (u1 , v1 + v2 , u2 , v1 + v2 ) = (v1 + v2 , u1 , v1 + v2 , u2 ) = (v1 , u1 + v2 , u1 , v1 , u2 + v2 , u2 ) = (v1 , u1 , v1 , u2 ) + (v2 , u1 , v2 , u2 ) = (u1 , v1 , u2 , v1 ) + (u1 , v2 , u2 , v2 ) = T (v1 ) + T (v2 ). (c)2017 Pearson Education. Inc. 394 T respects scalar multiplication: Let v be a vector in V and let c be a scalar. Then we have T (cv) = (u1 , cv, u2 , cv) = (cv, u1 , cv, u2 ) = (cv, u1 , cv, u2 ) = c(v, u1 , v, u2 ) = c(u1 , v, u2 , v) = cT (v). 1 1 then det(D) = −2 = 0, so by Corollary 4.5.17 the vectors v1 = (1, 1) 26. (a). If D = [v1 , v2 ] = 1 −1 and v2 = (1, −1) are linearly independent. Since dim[R2 ] = 2, it follows from Theorem 4.6.10 that {v1 , v2 } is a basis for R2 . (b). Let x = (x1 , x2 ) be an arbitrary vector in R2 . Since {v1 , v2 } forms a basis for R2 , there exist c1 and c2 such that (x1 , x2 ) = c1 (1, 1) + c2 (1, −1), that is, such that c 1 + c 2 = x1 , c 1 − c 2 = x2 . Solving this system yields c1 = 1 1 (x1 + x2 ), c2 = (x1 − x2 ). Thus, 2 2 (x1 , x2 ) = 1 1 (x1 + x2 )v1 + (x1 − x2 )v2 , 2 2 1 1 T [(x1 , x2 )] = T (x1 + x2 )v1 + (x1 − x2 )v2 2 2 1 1 = (x1 + x2 )T (v1 ) + (x1 − x2 )T (v2 ) 2 2 1 1 = (x1 + x2 )(2, 3) + (x1 − x2 )(−1, 1) 2 2 x1 3x2 = + , 2x1 + x2 . 2 2 In particular, when (4, −2) is substituted for (x1 , x2 ), it follows that T (4, −2) = (−1, 6). so that 27. The matrix of T is the 4 × 2 matrix [T (e1 ), T (e2 )]. Therefore, we must determine T (1, 0) and T (0, 1), which we can determine from the given information by using the linear transformation properties. A quick calculation shows that (1, 0) = − 23 (−1, 1) + 13 (1, 2), so 2 1 2 1 2 1 5 1 5 T (1, 0) = T (− (−1, 1)+ (1, 2)) = − T (−1, 1)+ T (1, 2) = − (1, 0, −2, 2)+ (−3, 1, 1, 1) = (− , , , −1). 3 3 3 3 3 3 3 3 3 Similarly, we have (0, 1) = 13 (−1, 1) + 13 (1, 2), so 1 1 1 1 1 1 2 1 1 T (0, 1) = T ( (−1, 1) + (1, 2)) = T (−1, 1) + T (1, 2) = (1, 0, −2, 2) + (−3, 1, 1, 1) = (− , , − , 1). 3 3 3 3 3 3 3 3 3 (c)2017 Pearson Education. Inc. 395 Therefore, we have the matrix of T : ⎤ −5/3 −2/3 ⎢ 1/3 1/3 ⎥ ⎥ ⎢ ⎣ 5/3 −1/3 ⎦ . −1 1 ⎡ 28. The matrix of T is the 2 × 4 matrix [T (e1 ), T (e2 ), T (e3 ), T (e4 )]. Therefore, we must determine T (1, 0, 0, 0), T (0, 1, 0, 0), T (0, 0, 1, 0), and T (0, 0, 0, 1), which we can determine from the given information by using the linear transformation properties. We are given that T (1, 0, 0, 0) = (3, −2). Next, T (0, 1, 0, 0) = T (1, 1, 0, 0) − T (1, 0, 0, 0) = (5, 1) − (3, −2) = (2, 3), T (0, 0, 1, 0) = T (1, 1, 1, 0) − T (1, 1, 0, 0) = (−1, 0) − (5, 1) = (−6, −1), and T (0, 0, 0, 1) = T (1, 1, 1, 1) − T (1, 1, 1, 0) = (2, 2) − (−1, 0) = (3, 2). Therefore, we have the matrix of T : 3 −2 2 −6 3 −1 3 2 . 29. The matrix of T is the 3 × 3 matrix [T (e1 ), T (e2 ), T (e3 )]. Therefore, we must determine T (1, 0, 0), T (0, 1, 0), and T (0, 0, 1), which we can determine from the given information by using the linear transformation properties. A quick calculation shows that (1, 0, 0) = (1, 2, 0) − 6(0, 1, 1) + 2(0, 2, 3), so T (1, 0, 0) = T (1, 2, 0) − 6T (0, 1, 1) + 2T (0, 2, 3) = (2, −1, 1) − 6(3, −1, −1) + 2(6, −5, 4) = (32, −5, 4). Similarly, (0, 1, 0) = 3(0, 1, 1) − (0, 2, 3), so T (0, 1, 0) = 3T (0, 1, 1) − T (0, 2, 3) = 3(3, −1, −1) − (6, −5, 4) = (3, 2, −7). Finally, (0, 0, 1) = −2(0, 1, 1) + (0, 2, 3), so T (0, 0, 1) = −2T (0, 1, 1) + T (0, 2, 3) = −2(3, −1, −1) + (6, −5, 4) = (0, −3, 6). Therefore, we have the matrix of T : ⎡ ⎤ 32 3 0 ⎣ −5 2 −3 ⎦ . 4 −7 6 30. The matrix of T is the 4 × 3 matrix [T (e1 ), T (e2 ), T (e3 )]. Therefore, we must determine T (1, 0, 0), T (0, 1, 0), and T (0, 0, 1), which we can determine from the given information by using the linear transformation properties. A quick calculation shows that (1, 0, 0) = 41 (0, −1, 4) − 14 (0, 3, 3) + 14 (4, 4, −1), so T (1, 0, 0) = 1 1 1 1 1 1 3 1 1 T (0, −1, 4)− T (0, 3, 3)+ T (4, 4, −1) = (2, 5, −2, 1)− (−1, 0, 0, 5)+ (−3, 1, 1, 3) = (0, , − , − ). 4 4 4 4 4 4 2 4 4 4 (0, 3, 3), so Similarly, (0, 1, 0) = − 15 (0, −1, 4) + 15 1 4 2 2 17 T (0, 1, 0) = − (2, 5, −2, 1) + (−1, 0, 0, 5) = (− , −1, , ). 5 15 3 5 15 (c)2017 Pearson Education. Inc. 396 1 (0, 3, 3), so Finally, (0, 0, 1) = 15 (0, −1, 4) + 15 T (0, 0, 1) = 1 1 1 1 1 2 8 T (0, −1, 4) + T (0, 3, 3) = (2, 5, −2, 1) + (−1, 0, 0, 5) = ( , 1, − , ). 5 15 5 15 3 5 15 Therefore, we have the matrix of T : ⎤ 0 −2/3 1/3 ⎢ 3/2 −1 1 ⎥ ⎥. ⎢ ⎣ −1/4 2/5 −2/5 ⎦ −1/4 17/15 8/15 ⎡ 31. T (ax2 +bx+c) = aT (x2 )+bT (x)+cT (1) = a(3x+2)+b(x2 −1)+c(x+1) = bx2 +(3a+c)x+(2a−b+c). 32. Using the linearity of T , we have T (2v1 + 3v2 ) = v1 + v2 and T (v1 + v2 ) = 3v1 − v2 . That is, 2T (v1 ) + 3T (v2 ) = v1 + v2 and T (v1 ) + T (v2 ) = 3v1 − v2 . Solving this system for the unknowns T (v1 ) and T (v2 ), we obtain T (v2 ) = 3v2 − 5v1 and T (v1 ) = 8v1 − 4v2 . 33. Since T is a linear transformation we obtain: T (x2 ) − T (1) = x2 + x − 3, 2T (x) = 4x, (33.1) (33.2) 3T (x) + 2T (1) = 2(x + 3) = 2x + 6. (33.3) From Equation (33.2) it follows that T (x) = 2x, so upon substitution into Equation (33.3) we have 3(2x) + 2T (1) = 2(x + 3) or T (1) = −2x + 3. Substituting this last result into Equation (33.1) yields T (x2 ) − (−2x + 3) = x2 + x − 3 so T (x2 ) = x2 − x. Now if a, b and c are arbitrary real numbers, then T (ax2 + bx + c) = aT (x2 ) + bT (x) + cT (1) = a(x2 − x) + b(2x) + c(−2x + 3) = ax2 − ax + 2bx − 2cx + 3c = ax2 + (−a + 2b − 2c)x + 3c. 34. Let v ∈ V . Since {v1 , v2 } is a basis for V , there exists a, b ∈ R such that v = av1 + bv2 . Hence T (v) = T (av1 + bv2 ) = aT (v1 ) + bT (v2 ) = a(3v1 − v2 ) + b(v1 + 2v2 ) = 3av1 − av2 + bv1 + 2bv2 = (3a + b)v1 + (2b − a)v2 . 35. Let v be any vector in V . Since {v1 , v2 , . . . , vk } spans V , we can write v = c1 v1 + c2 v2 + · · · + ck vk for suitable scalars c1 , c2 , . . . , ck . Then T (v) = T (c1 v1 + c2 v2 + · · · + ck vk ) = c1 T (v1 ) + c2 T (v2 ) + · · · + ck T (vk ) = c1 S(v1 ) + c2 S(v2 ) + · · · + ck S(vk ) = S(c1 v1 + c2 v2 + · · · + ck vk ) = S(v), as required. 36. Let v be any vector in V . Since {v1 , v2 , . . . , vk } is a basis for V , we can write v = c1 v1 +c2 v2 +· · ·+ck vk for suitable scalars c1 , c2 , . . . , ck . Then T (v) = T (c1 v1 + c2 v2 + · · · + ck vk ) = c1 T (v1 ) + c2 T (v2 ) + · · · + ck T (vk ) = c1 0 + c2 0 + . . . ck 0 = 0, as required. (c)2017 Pearson Education. Inc. 397 37. Let v1 and v2 be arbitrary vectors in V . Then, (T1 + T2 )(v1 + v2 ) = T1 (v1 + v2 ) + T2 (v1 + v2 ) = T1 (v1 ) + T1 (v2 ) + T2 (v1 ) + T2 (v2 ) = T1 (v1 ) + T2 (v1 ) + T1 (v2 ) + T2 (v2 ) = (T1 + T2 )(v1 ) + (T1 + T2 )(v2 ). Further, if k is any scalar, then (T1 + T2 )(kv) = T1 (kv) + T2 (kv) = kT1 (v) + kT2 (v) = k[T1 (v) + T2 (v)] = k(T1 + T2 )(v). It follows that T1 + T2 is a linear transformation. Now consider the transformation cT , where c is an arbitrary scalar. (cT )(v1 + v2 ) = cT (v1 + v2 ) = c[T (v1 ) + T (v2 )] = cT (v1 ) + cT (v2 ) = (cT )(v1 ) + (cT )(v2 ). (cT )(kv1 ) = cT (kv1 ) = c[kT (v1 )] = (ck)T (v1 ) = (kc)T (v1 ) = k[cT (v1 )]. Thus, cT is a linear transformation. 38. (T1 + T2 )(x) = T1 (x) + T2 (x) = Ax + Bx = (A + B)x = Hence, (T1 + T2 )(x1 , x2 ) = (5x1 + 6x2 , 2x1 − 2x2 ). (cT1 )(x) = cT1 (x) = c(Ax) = (cA)x = 3c −c c 2c 5 6 2 −2 x1 x2 x1 x2 = = 5x1 + 6x2 2x1 − 2x2 3cx1 + cx2 −cx1 + 2cx2 . . Hence, (cT1 )(x1 , x2 ) = (3cx1 + cx2 , −cx1 + 2cx2 ). 39. (T1 + T2 )(x) = T1 (x) + T2 (x) = Ax + Bx = (A + B)x. (cT1 )(x) = cT1 (x) = c(Ax) = (cA)x. 40. Problem 37 establishes that if T1 and T2 are in L(V, W ) and c is any scalar, then T1 + T2 and cT1 are in L(V, W ). Consequently, Axioms (A1) and (A2) are satisfied. A3: Let v be any vector in L(V, W ). Then (T1 + T2 )(v) = T1 (v) + T2 (v) = T2 (v) + T1 (v) = (T2 + T1 )(v). Hence T1 + T2 = T2 + T1 , therefore the addition operation is commutative. A4: Let T3 ∈ L(V, W ). Then [(T1 + T2 ) + T3 ] (v) = (T1 + T2 )(v) + T3 (v) = [T1 (v) + T2 (v)] + T3 (v) = T1 (v) + [T2 (v) + T3 (v)] = T1 (v) + (T2 + T3 )(v) = [T1 + (T2 + T3 )](v). Hence (T1 + T2 ) + T3 = T1 + (T2 + T3 ), therefore the addition operation is associative. A5: The zero vector in L(V, W ) is the zero transformation, O : V → W , defined by O(v) = 0, for all v in V, where 0 denotes the zero vector in V . To show that O is indeed the zero vector in L(V, W ), let T be any transformation in L(V, W ). Then (T + O)(v) = T (v) + O(v) = T (v) + 0 = T (v) for all v ∈ V, so that T + O = T . A6: The additive inverse of the transformation T ∈ L(V, W ) is the linear transformation −T defined by −T = (−1)T , since [T + (−T )](v) = T (v) + (−T )(v) = T (v) + (−1)T (v) = T (v) − T (v) = 0, (c)2017 Pearson Education. Inc. 398 for all v ∈ V , so that T + (−T ) = O. A7-A10 are all straightforward verifications. Solutions to Section 6.2 True-False Review: (a): FALSE. For example, T (x1 , x2 ) = (0, 0) is a linear transformation that maps every line to the origin. (b): TRUE. All of the matrices Rx , Ry , Rxy , LSx , LSy , Sx , and Sy discussed in this section are elementary matrices. (c): FALSE. A shear parallel to the x-axis composed with a shear parallel to the y-axis is given by matrix 1 k 1 0 1 + kl k = , which is not a shear. 0 1 l 1 l 1 (d): TRUE. This is explained prior to Example 6.2.1. (e): FALSE. For example, Rxy · Rx = 0 1 1 0 1 0 0 −1 = 0 −1 1 0 , and this matrix is not in the form of a stretch. (f ): FALSE. For example, k 0 0 1 1 0 0 l = k 0 0 l is not a stretch. Problems: 1. T (1, 1) = (0, 3), T (2, 1) = (1, 4), T (2, 2) = (0, 6), T (1, 2) = (−1, 5). 2. T (1, 1) = (1, −1), T (2, 1) = (1, −2), T (2, 2) = (2, −2), T (1, 2) = (2, −1). 3. T (1, 1) = (−4, −2), T (2, 1) = (−6, −4), T (2, 2) = (−8, −4), T (1, 2) = (−6, −2). 4. T (1, 1) = (2, 0), T (2, 1) = (3, −1), T (2, 2) = (4, 0), T (1, 2) = (3, 1). 0 2 1 2 0 2 1 0 3 1 0 5. ∼ ∼ ∼ 2 0 0 0 0 2 0 1 1. P12 2. M1 (1/2) 3. M2 (1/2). So, T (x) = Ax = P12 M1 (2)M2 (2)x which corresponds to a stretch in the y-direction, followed by a stretch in the x-direction, followed by a reflection in y = x. 1 2 6. A = =⇒ T (x) = Ax corresponds to a shear parallel to the x-axis. 0 1 −1 0 1 1 0 2 1 0 1. M1 (−1) 2. M2 (−1) So, 7. ∼ ∼ 0 −1 0 −1 0 1 T (x) = Ax = M1 (−1)M2 (−1)x which corresponds to a reflection in the x-axis, followed by a reflection in the y-axis. (c)2017 Pearson Education. Inc. 399 y 6 5 4 3 2 1 x -2 -1 1 2 Figure 0.0.68: Figure for Problem 1 y 2 1 1 2 x -1 -2 Figure 0.0.69: Figure for Problem 2 8. A = 9. 1 3 1 0 =⇒ T (x) = Ax corresponds to a shear parallel to the y-axis. 3 1 2 1 1 2 2 1 0 3 1 0 1. A12 (−3) 2. A21 (1) ∼ ∼ ∼ 4 0 −2 0 −2 0 1 T (x) = Ax = A12 (3)A21 (−1)M2 (−2)x (c)2017 Pearson Education. Inc. 3. M2 (−1/2). So, 400 y 2 1 x -8 -6 -4 -2 1 2 -2 -4 Figure 0.0.70: Figure for Problem 3 y 2 1 1 2 3 4 x -1 Figure 0.0.71: Figure for Problem 4 which corresponds to a reflection in the x-axis followed by a stretch in they y-direction, followed by a shear parallel to the x-axis, followed by a shear parallel to the y-axis. 1 −3 1 1 −3 2 1 −3 3 1 0 1. A12 (2) 2. M2 (1/2) 3. A21 (3). So, 10. ∼ ∼ ∼ −2 8 0 2 0 1 0 1 T (x) = Ax = A12 (−2)M2 (2)A21 (−3)x which corresponds to a shear parallel to the x-axis, followed by a stretch in the y-direction, followed by a shear parallel to the y-axis. −1 −1 −1 −1 1 1 1 0 1 2 3 1. A12 (−1) 2. M1 (−1) 3. A21 (−1). 11. ∼ ∼ ∼ −1 0 0 1 0 1 0 1 So, T (x) = Ax = A12 (1)M1 (−1)A21 (1)x which corresponds to a shear parallel to the x-axis, followed by a reflection in the y-axis, followed by a shear parallel to the y-axis. (c)2017 Pearson Education. Inc. 401 1 0 1 0 1 0 = . So T (x) = Ax corresponds to a reflection in the x-axis followed by 0 −2 0 2 0 −1 a stretch in the y-direction. π 0 −1 13. The matrix for a counter-clockwise rotation through an angle θ = is . Now, 1 0 2 12. 0 −1 1 0 1 ∼ 1 0 0 −1 2 ∼ 1 0 0 1 1. P12 2. M2 (−1) So, T (x) = Ax = P12 M2 (−1)x which corresponds to a reflection in the x-axis followed by a reflection in y = x. 14. R(θ) = cos θ 0 0 1 cos θ 0 0 1 1 − tan θ 0 1 0 1 − tan θ cos θ 0 1 = 1 0 sec θ 0 1 sin θ 0 1 1 0 0 sec θ 1 − tan θ = sin θ cos θ cos θ − sin θ = sin θ cos θ which coincides with the matrix of the transformation of R2 corresponding to a rotation through an angle θ in the counter-clockwise direction. 1 sin θ Solutions to Section 6.3 True-False Review: (a): FALSE. The statement should read dim[Ker(T )] + dim[Rng(T )] = dim[V ], not dim[W ] on the right-hand side. (b): FALSE. As a specific illustration, we could take T : P4 (R) → R7 defined by T (a0 + a1 x + a2 x2 + a3 x3 + a4 x4 ) = (a0 , a1 , a2 , a3 , a4 , 0, 0), and it is easy to see that T is a linear transformation with Ker(T ) = {0}. Therefore, Ker(T ) is 0-dimensional. (c): FALSE. The solution set to the homogeneous linear system Ax = 0 is Ker(T ), not Rng(T ). (d): FALSE. Rng(T ) is a subspace of W , not V , since it consists of vectors of the form T (v), and these belong to W . (e): TRUE. From the given information, we see that Ker(T ) is at least 2-dimensional, and therefore, since M23 is 6-dimensional, the Rank-Nullity Theorem requires that Rng(T ) have dimension at most 6 − 2 = 4. (f ): TRUE. Any vector of the form T (v) where v belongs to Rn can be written as Av, and this in turn can be expressed as a linear combination of the columns of A. Therefore, T (v) belongs to colspace(A). Problems: (c)2017 Pearson Education. Inc. 402 ⎤ ⎤ ⎡ 0 1 2 ⎢ 0 ⎥ ⎢ 2 4 ⎥ −10 ⎥ ⎥ =⎢ 1. (a). T (−10, 5) = ⎢ ⎣ 0 ⎦. Thus, x does belong to Ker(T ). ⎣ 4 8 ⎦ 5 0 8 16 ⎤ ⎡ ⎤ ⎡ −1 1 2 ⎢ −2 ⎥ ⎢ 2 4 ⎥ 1 ⎥ ⎢ ⎥ (b). T (1, −1) = ⎢ ⎣ 4 8 ⎦ −1 = ⎣ −4 ⎦. Thus, x does not belong to Ker(T ). −8 8 16 ⎤ ⎡ ⎤ ⎡ 1 2 0 ⎢ 2 4 ⎥ ⎢ 0 ⎥ 2 ⎥ ⎥ ⎢ (c). T (2, −1) = ⎢ ⎣ 4 8 ⎦ −1 = ⎣ 0 ⎦. Thus, x does belong to Ker(T ). 8 16 0 ⎤ ⎡ 7 1 −1 2 ⎣ 0 ⎦ 5 = 2. (a). T (7, 5, −1) = =⇒ (7, 5, −1) ∈ Ker(T ). 1 −2 −3 0 −1 ⎡ ⎤ −21 1 −1 2 ⎣ −2 −15 ⎦ = (b). T (−21, −15, 2) = =⇒ (−21, −15, 2) ∈ / Ker(T ). 1 −2 −3 3 2 ⎡ ⎤ 35 1 −1 2 ⎣ 0 ⎦ 25 = (c). T (35, 25, −5) = =⇒ (35, 25, −5) ∈ Ker(T ). 1 −2 −3 0 −5 ⎡ 2 2 3. {x ∈ R : T (x) = 0} = {x ∈ R : Ax = 0}. The augmented matrix of the system Ax = 0 is: Ker(T ) = 3 6 0 1 2 0 , with reduced row-echelon form of . It follows that 1 2 0 0 0 0 Ker(T ) = {x ∈ R2 : x = (−2t, t), t ∈ R} = {x ∈ R2 : x = t(−2, 1), t ∈ R}. Geometrically, this is a line in R2 . It is the subspace of R2 spanned by the vector (−2, 1). dim[Ker(T )] = 1. For the given transformation, Rng(T ) = colspace(A). From the preceding reduced row-echelon form of A, we see that colspace(A) is generated by the first column vector of A. Consequently, Rng(T ) = {y ∈ R2 : y = r(3, 1), r ∈ R}. Geometrically, this is a line in R2 . It is the subspace of R2 spanned by the vector (3, 1). dim[Rng(T )] = 1. Since dim[Ker(T )]+ dim[Rng(T )] = 2 = dim[R2 ], Theorem 6.3.8 is satisfied. matrix of the system Ax = 0 4. Ker(T ) = {x ∈ R⎤3 : T (x) = 0} = {x ∈ R3 : Ax = 0}. ⎤ ⎡ ⎡ The augmented 1 −1 0 0 1 0 0 0 1 2 0 ⎦, with reduced row-echelon form ⎣ 0 1 0 0 ⎦. Thus x1 = x2 = x3 = 0, so is: ⎣ 0 2 −1 1 0 0 0 1 0 Ker(T ) = {0}. Geometrically, this describes a point (the origin) in R3 . dim[Ker(T )] = 0. For the given transformation, Rng(T ) = colspace(A). From the preceding reduced row-echelon form of A, we see that colspace(A) is generated by the first three column vectors of A. Consequently, Rng(T ) = R3 , dim[Rng(T )] = dim[R3 ] = 3, and Theorem 6.3.8 is satisfied since dim[Ker(T )]+ dim[Rng(T )] = 0 + 3 = dim[R3 ]. (c)2017 Pearson Education. Inc. 403 5. R3 : T (x) = 0} = {x ∈ R3 : Ax = 0}. ⎡ Ker(T ) = {x ∈ ⎤ ⎤ matrix of the system Ax = 0 is: ⎡ The augmented 1 −2 1 0 1 0 −5 0 ⎣ 2 −3 −1 0 ⎦, with reduced row-echelon form of ⎣ 0 1 −3 0 ⎦. Thus 5 −8 −1 0 0 0 0 0 Ker(T ) = {x ∈ R3 : x = t(5, 3, 1), t ∈ R}. Geometrically, this describes the line in R3 through the origin, spanned by (5, 3, 1). dim[Ker(T )] = 1. For the given transformation, Rng(T ) = colspace(A). From the preceding reduced row-echelon form of A, we see that a basis for colspace(A) is given by the first two column vectors of A. Consequently, Rng(T ) = {y ∈ R3 : y = r(1, 2, 5) + s(−2, −3, −8), r, s ∈ R}. Geometrically, this is a plane through the origin in R3 . dim[Rng(T )] = 2 and Theorem 6.3.8 is satisfied since dim[Ker(T )]+ dim[Rng(T )] = 1 + 2 = 3 = dim[R3 ]. 3 3 6. Ker(T ) = {x ∈ R : T (x) = 0} = {x ∈ R : Ax = 0}. The augmentedmatrix of the system Ax = 0 is: 1 −1 2 0 1 −1 2 0 , with reduced row-echelon form of . Thus −3 3 −6 0 0 0 0 0 Ker(T ) = {x ∈ R3 : x = r(1, 1, 0) + s(−2, 0, 1), r, s ∈ R}. Geometrically, this describes the plane through the origin in R3 , which is spanned by the linearly independent set {(1, 1, 0), (−2, 0, 1)}. dim[Ker(T )] = 2. For the given transformation, Rng(T ) = colspace(A). From the preceding reduced row-echelon form of A, we see that a basis for colspace(A) is given by the first column vector of A. Consequently, Rng(T ) = {y ∈ R2 : y = t(1, −3), t ∈ R}. Geometrically, this is the line through the origin in R2 spanned by (1, −3). dim[Rng(T )] = 1 and Theorem 6.3.8 is satisfied since dim[Ker(T )]+ dim[Rng(T )] = 2 + 1 = 3 = dim[R3 ]. 3 3 = 0}. The augmented matrix of the system Ax = 0 is: 7. Ker(T ) = {x ∈ R : T (x) = 0} = {x ∈ R : Ax 1 3 2 0 1 3 0 0 , with reduced row-echelon form of . Thus 2 6 5 0 0 0 1 0 Ker(T ) = {x ∈ R3 : x = r(−3, 1, 0), r ∈ R}. Geometrically, this describes the line through the origin in R3 , which is spanned by (−3, 1, 0). dim[Ker(T )] = 1. For the given transformation, Rng(T ) = colspace(A). From the preceding reduced row-echelon form of A, we see that a basis for colspace(A) is given by the first and third column vectors of A. Consequently, Rng(T ) = span{(1, 2), (2, 5)} = R2 , so that dim[Rng(T )] = 2. Geometrically, Rng(T ) is the xy-plane, and Theorem 6.3.8 is satisfied since dim[Ker(T )]+ dim[Rng(T )] = 1 + 2 = 3 = dim[R3 ]. ⎤ ⎡ −5/3 −2/3 ⎢ 1/3 1/3 ⎥ ⎥ 8. The matrix of T in Problem 27 of Section 6.1 is A = ⎢ ⎣ 5/3 −1/3 ⎦. Thus, −1 1 Ker(T ) = nullspace(A) = {0} (c)2017 Pearson Education. Inc. 404 ⎧⎡ ⎤⎫ ⎤ ⎡ −2/3 ⎪ −5/3 ⎪ ⎪ ⎪ ⎨⎢ ⎢ 1/3 ⎥⎬ 1/3 ⎥ ⎥ ⎢ ⎥ ⎢ . , Rng(T ) = colspace(A) = span ⎣ 5/3 ⎦ ⎣ −1/3 ⎦⎪ ⎪ ⎪ ⎪ ⎭ ⎩ 1 −1 and 9. The matrix of T in Problem 28 of Section 6.1 is A = 3 2 −6 −2 3 −1 3 2 . Thus, Ker(T ) = nullspace(A) = span{(16/13, 15/13, 1, 0), (−5/13, −12/13, 0, 1)} and Rng(T ) = colspace(A) = R2 . ⎡ ⎤ 32 3 0 2 −3 ⎦. Thus, 10. The matrix of T in Problem 29 of Section 6.1 is A = ⎣ −5 4 −7 6 Ker(T ) = nullspace(A) = {0} and Rng(T ) = colspace(A) = R3 . ⎡ 0 ⎢ 3/2 11. The matrix of T in Problem 30 of Section 6.1 is A = ⎢ ⎣ 0 −1/4 ⎤ −2/3 1/3 −1 1 ⎥ ⎥. Thus, 2/5 −2/5 ⎦ 17/15 8/15 Ker(T ) = nullspace(A) = {0} and ⎧⎡ ⎪ ⎪ ⎨⎢ Rng(T ) = colspace(A) = span ⎢ ⎣ ⎪ ⎪ ⎩ ⎤ ⎡ ⎤ ⎡ ⎤⎫ −2/3 0 1/3 ⎪ ⎪ ⎬ ⎢ ⎢ −1 ⎥ 3/2 ⎥ 1 ⎥ ⎥,⎢ ⎥,⎢ ⎥ . 2/5 ⎦ ⎣ −2/5 ⎦⎪ 0 ⎦ ⎣ ⎪ ⎭ 17/15 −1/4 8/15 12. (a). Ker(T ) = {v ∈ R3 : u, v = 0}. For v to be in the kernel of T , u and v must be orthogonal. Since u is any fixed vector in R3 , then v must lie in the plane orthogonal to u. Hence dim[Ker(T )] = 2. (b). Rng(T ) = {y ∈ R : y = u, v, v ∈ R3 }, and dim[Rng(T )] = 1. 13. (a). Ker(S) = {A ∈ Mn (R) : A − AT = 0} = {A ∈ Mn (R) : A = AT }. Hence, any matrix in Ker(S) is symmetric by definition. (b). Since any matrix in Ker(S) is symmetric, it has been shown that {A1 , A2 , A3 } is a spanning set for the set of all symmetric matrices in M2 (R), where 1 0 0 1 0 0 , A2 = , A3 = . A1 = 0 0 1 0 0 1 Thus, dim[Ker(S)] = 3. (c)2017 Pearson Education. Inc. 405 14. Ker(T ) = {A ∈ Mn (R) : AB − BA = 0} = {A ∈ M2 (R) : AB = BA}. This is the set of matrices that commute with B. 15. (a). Ker(T ) = {p ∈ P2 (R) : T (p) = 0} = {ax2 + bx + c ∈ P2 (R) : ax2 + (a + 2b + c)x + (3a − 2b − c) = 0, for all x}. Thus, for p(x) = ax2 + bx + c to be in Ker(T ), a, b, and c must satisfy the system: a a + 2b + c =0 =0 3a − 2b − c =0 Solving this system, we obtain that a = 0 and c = −2b. Consequently, all polynomials of the form 0x2 + bx + (−2b) are in Ker(T ), so Ker(T ) = {b(x − 2) : b ∈ R}. Since Ker(T ) is spanned by the nonzero vector x − 2, it follows that dim[Rng(T )] = 1. (b). In this case, Rng(T ) = {T (ax2 + bx + c) : a, b, c ∈ R} = {ax2 + (a + 2b + c)x + (3a − 2b − c) : a, b, c ∈ R} = {a(x2 + x + 3) + b(2x − 2) + c(x − 1) : a, b, c ∈ R} = {a(x2 + x + 3) + (2b + c)(x − 1) : a, b, c ∈ R} = span{x2 + x + 3, x − 1}. Since the vectors in this spanning set are linearly independent on any interval, it follows that the spanning set is a basis for Rng(T ). Hence, dim[Rng(T )] = 2. 16. Ker(T ) = {p ∈ P2 (R) : T (p) = 0} = {ax2 + bx + c ∈ P2 (R) : (a + b) + (b − c)x = 0, for all x}. Thus, a, b, and c must satisfy: a + b = 0 and b − c = 0 =⇒ a = −b and b = c. Letting c = r ∈ R, we have ax2 + bx + c = r(−x2 + x + 1). Thus, Ker(T ) = {r(−x2 + x + 1) : r ∈ R} and dim[Ker(T )] = 1. Rng(T ) = {T (ax2 + bx + c) : a, b, c ∈ R} = {(a + b) + (b − c)x : a, b, c ∈ R} = {c1 + c2 x : c1 , c2 ∈ R}. Consequently, a basis for Rng(T ) is {1, x}, so that Rng(T ) = P1 (R), and dim[Rng(T )] = 2. 17. Ker(T ) = {p ∈ P1 (R) : T (p) = 0} = {ax + b ∈ P1 (R) : (b − a) + (2b − 3a)x + bx2 = 0, for all x}. Thus, a and b must satisfy: b−a = 0, 2b−3a = 0, and b = 0 =⇒ a = b = 0. Thus, Ker(T ) = {0} and dim[Ker(T )] = 0. Rng(T ) = {T (ax + b) : a, b ∈ R} = {(b − a) + (2b − 3a)x + bx2 : a, b ∈ R} = {−a(1 + 3x) + b(1 + 2x + x2 ) : a, b ∈ R} = span{1 + 3x, 1 + 2x + x2 }. Since the vectors in this spanning set are linearly independent on any interval, it follows that the spanning set is a basis for Rng(T ), and dim[Rng(T )] = 2. 18. To determine Ker(T ), we set T a c b d = 0. Thus, (a − b + d) + (−a + b − d)x2 = 0, (c)2017 Pearson Education. Inc. 406 thus leading to the system of linear equations a−b+d=0 − a + b − d = 0. and We quickly find that b = r, c = s, and d = t are free variables, and thus, a = r − t. Hence, r−t r Ker(T ) = : r, s, t ∈ R . s t By extracting the free variables, a basis for Ker(T ) can be given as 1 1 0 0 −1 0 , , . 0 0 1 0 0 1 Hence, dim[Ker(T )] = 3. Next, Rng(T ) consists of all polynomials of the form (a − b + d) + (−a + b − d)x2 . The constant coefficient and x2 are additive inverses of one another, so we can describe Rng(T ) more simply as Rng(T ) = span{1 − x2 }, and thus, dim[Rng(T )] = 1. 19. To determine Ker(T ) we set T (x, y) = 0 0 0 0 0 0 . Thus, −x − y = 0, 2x + 2y = 0, 3x + 3y = 0, −9x − 9y = 0. All of these equations are equivalent to the single equation x + y = 0. That is, y = −x. Hence, Ker(T ) = {(x, y) : x, y ∈ R and y = −x} = {(x, −x) : x ∈ R}. A basis for Ker(T ) can be obtained by eliminating the free variable x to get the basis {(1, −1)}. Hence, dim[Ker(T ) = 1. −x − y 0 2x + 2y Next, Rng(T ) consists of all 2 × 3 matrices of the form . That is, 0 3x + 3y −9x − 9y −z 0 2z Rng(T ) consists of all 2 × 3 matrices of the form , a subspace of M23 (R) with basis 0 3z −9z −1 0 2 . 0 3 −9 20. To determine Ker(T ), we set T (A) = 0. That is, AT = 0, which implies that A = 0. Hence, Ker(T ) = {02×4 }, the trivial subspace. Hence, its dimension is zero. Next, Rng(T ) consists of all 4 × 2 matrices that arise as outputs from the transformation T . However, given any 4 × 2 matrix B, note that T (B T ) = (B T )T = B. Thus, Rng(T ) = M42 (R), which is an eightdimensional space. 21. T (v) = 0 ⇐⇒ T (av1 + bv2 + cv3 ) = 0 ⇐⇒ aT (v1 ) + bT (v2 ) + cT (v3 ) = 0 ⇐⇒ a(2w1 − w2 ) + b(w1 − w2 ) + c(w1 + 2w2 ) = 0 ⇐⇒ (2a + b + c)w1 + (−a − b + 2c)w2 = 0 ⇐⇒ 2a + b + c = 0 and a − b + 2c = 0. (c)2017 Pearson Education. Inc. 407 Reducing the augmented matrix of the system yields: 2 1 1 0 1 −1 2 0 1 0 1 0 ∼ ∼ . 1 −1 2 0 2 1 1 0 0 1 −1 0 Setting c = r =⇒ b = r, a = −r. Thus, Ker(T ) = {v ∈ V : v = r(−v1 + v2 + v3 ), r ∈ R} and dim[Ker(T )] = 1. Rng(T ) = {T (v) : v ∈ V } = {(2a + b + c)w1 + (−a − b + 2c)w2 : a, b, c ∈ V } = span{w1 , w2 } = W . Consequently, dim[Rng(T )] = 2. 22. (a). If w ∈ Rng(T ), then T (v) = w for some v ∈ V , and since {v1 , v2 , . . . , vn } is a basis for V , there exist c1 , c2 , . . . , cn ∈ R for which v = c1 v1 + c2 v2 + · · · + cn vn . Accordingly, w = T (v) = a1 T (v1 ) + a2 T (v2 ) + · · · + an T (vn ). Thus, Rng(T ) = span{T (v1 ), T (v2 ), . . . , T (vn )}. We must show that {T (v1 ), T (v2 ), . . . , T (vn )} is also linearly independent. Suppose that b1 T (v1 ) + b2 T (v2 ) + · · · + bn T (vn ) = 0. Then T (b1 v1 + b2 v2 + · · · + bn vn ) = 0, so since Ker(T ) = {0}, b1 v1 + b2 v2 + · · · + bn vn = 0. Since {v1 , v2 , . . . , vn } is a linearly independent set, the preceding equation implies that b1 = b2 = · · · = bn = 0. Consequently, {T (v1 ), T (v2 ), . . . , T (vn )} is a linearly independent set in W . Therefore, since we have already shown that it is a spanning set for Rng(T ), {T (v1 ), T (v2 ), . . . , T (vn )} is a basis for Rng(T ). (b). As an example, let T : R3 → R2 be defined by T ((a, b, c)) = (a, b) for all (a, b, c) in R3 . Then for the basis {e1 , e2 , e3 } of R3 , we have {T (e1 ), T (e2 ), T (e3 )} = {(1, 0), (0, 1), (0, 0)}, which is clearly not a basis for R2 . 23. Let v ∈ V . Since {v1 , v2 , . . . , vk } spans V , we can write v = c1 v 1 + c2 v2 + · · · + ck v k for some constants c1 , c2 , . . . , ck . Thus, T (v) = T (c1 v1 + c2 v2 + · · · + ck vk ) = c1 T (v1 ) + c2 T (v2 ) + · · · + ck T (vk ) = c1 S(v1 ) + c2 S(v2 ) + · · · + ck S(vk ) = S(c1 v1 + c2 v2 + · · · + ck vk ) = S(v), as required. 24. Let v ∈ V . Since {v1 , v2 , . . . , vk } spans V , we can write v = c1 v 1 + c2 v2 + · · · + ck v k for some constants c1 , c2 , . . . , ck . Thus, T (v) = T (c1 v1 + c2 v2 + · · · + ck vk ) = c1 T (v1 ) + c2 T (v2 ) + · · · + ck T (vk ) = c1 0 + c2 0 + · · · + ck 0 = 0, as required. (c)2017 Pearson Education. Inc. 408 Solutions to Section 6.4 True-False Review: (a): FALSE. Many one-to-one linear transformations T : P3 (R) → M32 can be constructed. One possible example would be to define ⎡ ⎤ a0 a1 T (a0 + a1 x + a2 x2 + a3 x3 ) = ⎣ a2 a3 ⎦ . 0 0 It is easy to check that with T so defined, T is a one-to-one linear transformation. (b): TRUE. We can define an isomorphism T : V → M32 via ⎛⎡ ⎤⎞ ⎡ a b c a T ⎝⎣ 0 d e ⎦⎠ = ⎣ c 0 0 f e ⎤ b d ⎦. f With T so defined, it is easy to check that T is an isomorphism. (c): TRUE. Both Ker(T1 ) and Ker(T2 T1 ) are subspaces of V1 , and since if T1 (v1 ) = 0, then (T2 T1 )v1 = T2 (T1 (v1 )) = T2 (0) = 0, we see that every vector in Ker(T1 ) belongs to Ker(T2 T1 ). Therefore, Ker(T1 ) is a subspace of Ker(T2 T1 ). (d): TRUE. Observe that T is not one-to-one since Ker(T ) = {0} (because Ker(T ) is 1-dimensional). Moreover, since M22 is 4-dimensional, then by the Rank-Nullity Theorem, Rng(T ) is 3-dimensional. Since P2 (R) is 3-dimensional, we conclude that Rng(T ) = P2 (R); that is, T is onto. (e): TRUE. Since M2 (R) is 4-dimensional, Rng(T ) can be at most 4-dimensional. However, P4 (R) is 5-dimensional. Therefore, any such linear transformation T cannot be onto. (f ): TRUE. If we assume that (T2 T1 )v = (T2 T1 )w, then T2 (T1 (v)) = T2 (T1 (w)). Since T2 is one-to-one, then T1 (v) = T1 (w). Next, since T1 is one-to-one, we conclude that v = w. Therefore, T2 T1 is one-to-one. (g): FALSE. This linear transformation is onto one-to-one. The reason is essentially because the derivative of any constant is zero. Therefore, Ker(T ) consists of all constant functions, and therefore, Ker(T ) = {0}. (h): TRUE. Since M23 is 6-dimensional and Rng(T ) is only 4-dimensional, T is not onto. Moreover, since P3 (R) is 4-dimensional, the Rank-Nullity Theorem implies that Ker(T ) is 0-dimensional. Therefore, Ker(T ) = {0}, and this means that T is one-to-one. (i): TRUE. Recall that dim[Rn ] = n and dim[Rm ] = m. In order for such an isomorphism to exist, Rn and Rm must have the same dimension; that is, m = n. (j): FALSE. For example, the vector space of all polynomials with real coefficients is an infinite-dimensional real vector space, and since Rn is finite-dimensional for all positive integers n, this statement is false. (k): FALSE. In order for this to be true, it would also have to be assumed that T1 is onto. For example, suppose V1 = V2 = V3 = R2 . If we define T2 (x, y) = (x, y) for all (x, y) in R2 , then T2 is onto. However, if we define T1 (x, y) = (0, 0) for all (x, y) in R2 , then (T2 T1 )(x, y) = T2 (T1 (x, y)) = T2 (0, 0) = (0, 0) for all (x, y) in R2 . Therefore T2 T1 is not onto, since Rng(T2 T1 ) = {(0, 0)}, even though T2 itself is onto. (l): TRUE. This is a direct application of the Rank-Nullity Theorem. Since T is assumed to be onto, Rng(T ) = R3 , which is three-dimensional. Therefore the dimension of Ker(T ) is 8 − 3 = 5. Problems: (c)2017 Pearson Education. Inc. 409 1. We have T2 T1 (x) = T2 (T1 (x)) = T2 (Ax) = (BA)x. Now, BA = −1 1 1 −1 3 2 Thus, T2 T1 (x) = (BA)x = 2 3 x1 x2 = 2 3 . = 2x1 + 3x2 . In order for T1 T2 to exist, the outputs from T2 must be received as inputs for T1 for the composition of functions T1 T2 to work. However, the outputs from T2 belong to R, whereas the domain of T1 is R2 . Therefore, T1 T2 does not exist. T2 T1 = BA. 2. T1 T2 (x) = T1 (T 2 (x)) = T1 (Bx) = (AB)x, so T1 T2 = AB. Similarly, −1 2 1 5 −5 −5 1 5 −1 2 14 7 AB = = . BA = = . 3 1 −2 0 −2 0 3 1 2 −4 1 15 −5x1 − 5x2 −5 −5 x1 = = (−5(x1 + x2 ), x1 + 15x2 ) and T1 T2 (x) = (AB)x = x2 x1 + 15x2 1 15 14x1 + 7x2 14 7 x1 = = (7(2x1 + x2 ), 2(x1 − 2x2 )). T2 T1 (x) = (BA)x = x2 2x1 − 4x2 2 −4 Clearly, T1 T2 = T2 T1 . 3. We have T1 T2 (x) = T1 (T2 (x)) = T1 (Bx) = (AB)x. Now, ⎤ ⎡ ⎤ 2 −1 −1 −8 5 0 −4 3 1 ⎦ 0 1 ⎦. AB = ⎣ 0 =⎣ 1 1 0 1 1 1 1 −4 4 Thus, ⎤ ⎡ ⎤ ⎤⎡ −1 −8 5 x1 −x1 − 8x2 + 5x3 ⎦. x1 + x3 0 1 ⎦ ⎣ x2 ⎦ = ⎣ T1 T2 (x) = (AB)x = ⎣ 1 1 −4 4 x3 x1 − 4x2 + 4x3 ⎡ ⎡ Now, Ker(T1 T2 ) = nullspace(AB), which we ⎡compute by first ⎤ finding a row-echelon form for the matrix AB. 1 0 1 One row-echelon form for AB is the matrix ⎣ 0 1 − 34 ⎦. Let z = t be a free variable. Then y = 34 t, and 0 0 0 then x = −t. Thus, 3 3 Ker(T1 T2 ) = nullspace(AB) = {(−t, t, t) : t ∈ R} = span{(−1, , 1)}. 4 4 Next, Rng(T1 T2 ) = colspace(AB). Using our calculations above, we see that the column space of AB is spanned by the first two columns of AB, so that ⎧⎡ ⎤ ⎡ ⎤⎫ −8 ⎬ ⎨ −1 Rng(T1 T2 ) = colspace(AB) = span ⎣ 1 ⎦ , ⎣ 0 ⎦ . ⎩ ⎭ 1 −4 (c)2017 Pearson Education. Inc. 410 Next, we turn our attention to T2 T1 . We have T2 T1 (x) = T2 (Ax) = (BA)x. Now BA = 0 −4 3 1 0 1 Thus, T2 T1 (x) = (BA)x = ⎡ ⎤ 2 −1 3 −1 ⎣ 0 ⎦ 1 = . 3 0 1 1 3 −1 3 0 x1 x2 = 3x1 − x2 3x1 . Now, Ker(T1 T2 ) = nullspace(BA), and since BA is an invertible matrix, its reduced row echelon form is I2 . Thus, Ker(T2 T1 ) = nullspace(AB) = {(0, 0)}. Also, Rng(T2 T1 ) = colspace(AB) = R2 . 0 1 −1 x1 = . This matrix equation results in 4. Ker(T1 ) = {x ∈ R : Ax = 0}. Ax = 0 =⇒ 0 x2 2 −2 the system: x1 − x2 = 0 and 2x1 − 2x2 = 0, or equivalently, x1 = x2 . Thus, 2 Ker(T1 ) = {x ∈ R2 : x = r(1, 1) where r ∈ R}. Geometrically, this is the line through the origin in R2 spanned by (1,1). 0 2 1 x 1 = . This matrix equation results in the Ker(T2 ) = {x ∈ R2 : Bx = 0}. Bx = 0 =⇒ x2 0 3 −1 system: 2x1 + x2 = 0 and 3x1 − x2 = 0, or equivalently, x1 = x2 = 0. Thus, Ker(T2 ) = {0}. Geometrically, this is a point (the origin). 0 1 −1 2 1 x1 2 = Ker(T1 T2 ) = {x ∈ R : (AB)x = 0}. (AB)x = 0 =⇒ x2 0 2 −2 3 −1 0 −1 2 x1 = . This matrix equation results in the system: =⇒ 0 x2 −2 4 −x1 + 2x2 = 0 and − 2x1 + 4x2 = 0, or equivalently, x1 = 2x2 . Thus, Ker(T1 T2 ) = {x ∈ R2 : x = s(2, 1) where s ∈ R}. Geometrically, this is the line through the origin in R2 spanned by the vector (2, 1). 0 2 1 1 −1 x1 2 = Ker(T2 T1 ) = {x ∈ R : (BA)x = 0}. (BA)x = 0 =⇒ 0 x2 3 −1 2 −2 0 4 −4 x1 = . This matrix equation results in the system: =⇒ x2 0 1 −1 4x1 − 4x2 = 0 and x1 − x2 = 0, or equivalently, x1 = x2 . Thus, Ker(T2 T1 ) = {x ∈ R2 : x = t(1, 1) where t ∈ R}. Geometrically, this is the line through the origin in R2 spanned by the vector (1, 1). (c)2017 Pearson Education. Inc. 411 5. (T2 T1 )(A) = T2 (T1 (A)) = T2 (A − AT ) = (A − AT ) + (A − AT )T = (A − AT ) + (AT − A) = 0n . d 6. (a). (T1 (f ))(x) = [sin(x − a)] = cos(x − a). dx x x (T2 (f ))(x) = sin(t − a)dt = [− cos(t − a)]a = 1 − cos(x − a). a (T1 T2 )(f )(x) = d [1 − cos(x − a)] = sin(x − a) = f (x). dxx x (T2 T1 )(f )(x) = a cos(t − a)dt = [sin(t − a)]a = sin(x − a) = f (x). Consequently, (T1 T2 )(f ) = (T2 T1 )(f ) = f . x d f (x)dx = f (x)dx = f (x). (b). (T1 T2 )(f ) = T1 (T2 (f )) = T1 dx a a x dg(x) dg(x) = dx = g(x) − g(a). (T2 T1 )(g) = T2 (T1 (g)) = T2 dx dx a x 7. Let v ∈ V . There exists a, b ∈ R such that v = av1 + bv2 . Then T2 T1 (v) = T2 [aT1 (v1 ) + bT1 (v2 )] = T2 [a(v1 − v2 ) + b(2v1 + v2 )] = T2 [(a + 2b)v1 + (b − a)v2 ] = (a + 2b)T2 (v1 ) + (b − a)T (v2 ) = (a + 2b)(v1 + 2v2 ) + (b − a)(3v1 − v2 ) = (5b − 2a)v1 + 3(a + b)v2 . 8. Let v ∈ V . There exists a, b ∈ R such that v = av1 + bv2 . Then T2 T1 (v) = T2 [aT1 (v1 ) + bT1 (v2 )] = T2 [a(3v1 + v2 )] = 3a(−5v2 ) + a(−v1 + 6v2 ) = −av1 − 9av2 . 9. ONTO only. The transformation T2 T1 : Mn (R) → R is given by (T2 T1 )(A) = 2 tr(A). Given any real number r, the matrix 12 r ·E11 (which has (1, 1)-entry 12 r and all other entries zero) satisfies (T2 T1 )( 12 r ·E11 ) = r. Thus, Rng(T2 T1 ) = R, which implies that T2 T1 is onto. However, for n > 1, T2 T1 cannot be one-to-one, since dim[Mn (R)] > 1 = dim(R). 10. ONE-TO-ONE only. By inspection, we see that Ker(T ) = {0}. Moreover, Rng(T ) = {Ax : x ∈ ⎧⎡ ⎫ ⎤ x ⎨ ⎬ R} = ⎣ 2x ⎦ : x ∈ R . Since Ker(T ) is trivial, T is one-to-one. However, Rng(T ) is only one-dimensional ⎩ ⎭ 3x ⎡ ⎤ 1 (spanned by ⎣ 2 ⎦), so that T is not onto. Since T fails to be onto, we know that T −1 does not exist. 3 11. BOTH. Ker(T ) = {x ∈ R2 : T (x) = 0} = {x ∈ R2 : Ax= 0}. Theaugmented matrix of the system 1 0 0 4 2 0 , with reduced row-echelon form of . It follows that Ker(T ) = {0}. Ax = 0 is: 1 3 0 0 1 0 Hence T is one-to-one by Theorem 6.4.8. For the given transformation, Rng(T ) = {y ∈ R2 : Ax = y is consistent}. The augmented matrix of the 1 0 (3y1 − y2 )/5 4 2 y1 with reduced row-echelon form of . The system is, system Ax = y is 0 1 (2y2 − y1 )/5 1 3 y2 (c)2017 Pearson Education. Inc. 412 therefore, consistent for all (y1 , y2 ), so that Rng(T ) = R2 . Consequently, T is onto. T −1 exists since T has been shown to be both one-to-one and onto. Using the Gauss-Jordan method for computing A−1 , we have 1 3 0 − 15 1 12 1 0 4 2 1 0 4 10 . ∼ ∼ 1 2 1 3 0 1 0 1 − 10 0 52 − 14 1 5 3 − 15 −1 10 Thus, A = , so T −1 (y) = A−1 y. 2 1 − 10 5 matrix of the 12. NEITHER. Ker(T ) = {x ∈ R2 : T (x) = 0} = {x ∈ R2 : Ax = 0}. The augmented 1 2 0 1 2 0 , with reduced row-echelon form of . It follows that system Ax = 0 is: −2 −4 0 0 0 0 Ker(T ) = {x ∈ R2 : x = t(−2, 1), t ∈ R}. By Theorem 6.4.8, since Ker(T ) = {0}, T is not one-to-one. This also implies that T −1 does not exist. 2 For the given transformation, Rng(T The augmented matrix of the ) = {y ∈ R : Ax = y is consistent}. y1 1 2 1 2 y1 with reduced row-echelon form of . The last row of system Ax = y is −2 −4 y2 0 0 2y1 + y2 this matrix implies that 2y1 + y2 = 0 is required for consistency. Therefore, it follows that Rng(T ) = {(y1 , y2 ) ∈ R2 : 2y1 + y2 = 0} = {y ∈ R2 : y = s(1, −2), s ∈ R}. T is not onto because dim[Rng(T )] = 1 = 2 = dim[R2 ]. matrix of the 13. ONTO only. Ker(T ) = {x ∈R3 : T (x) = 0} = {x ∈ R3 : Ax = 0}. The augmented 1 0 −7 0 1 2 −1 0 , with reduced row-echelon form . It follows that system Ax = 0 is: 2 5 1 0 0 1 3 0 Ker(T ) = {(x1 , x2 , x3 ) ∈ R3 : x1 = 7t, x2 = −3t, x3 = t, t ∈ R} = {x ∈ R3 : x = t(7, −3, 1), t ∈ R}. By Theorem 6.4.8, since Ker(T ) = {0}, T is not one-to-one. This also implies that T −1 does not exist. For the given transformation, Rng(T ) = {y ∈ R2 : Ax = y is consistent}. The augmented matrix of the system Ax = y is 1 2 −1 y1 1 0 −7 5y1 − 2y2 ∼ . 2 5 1 y2 0 1 3 y2 − 2y1 We clearly have a consistent system for all y = (y1 , y2 ) ∈ R2 , thus Rng(T ) = R2 . Therefore, T is onto. 14. NEITHER. Reducing A to row-echelon form, we obtain ⎤ ⎡ 1 3 5 ⎢ 0 1 2 ⎥ ⎥ REF(A) = ⎢ ⎣ 0 0 0 ⎦. 0 0 0 We quickly find that Ker(T ) = nullspace(A) = span{(1, −2, 1)}. Moreover, ⎧⎡ ⎤⎫ ⎤ ⎡ 1 ⎪ 0 ⎪ ⎪ ⎪ ⎨⎢ ⎥⎬ ⎢ 3 ⎥ ⎥,⎢ 4 ⎥ . Rng(T ) = colspace(A) = span ⎢ ⎦ ⎣ 4 ⎦⎪ ⎣ ⎪ ⎪ ⎪ 5 ⎭ ⎩ 1 2 (c)2017 Pearson Education. Inc. 413 Based on these calculations, we see that T is neither one-to-one nor onto. 15. We have and so 1 2 1 (0, 0, 1) = − (2, 1, −3) + (1, 0, 0) + (0, 1, 0), 3 3 3 2 1 1 T (0, 0, 1) = − T (2, 1, −3) + T (1, 0, 0) + T (0, 1, 0) 3 3 3 1 2 1 = − (7, −1) + (4, 5) + (−1, 1) 3 3 3 = (0, 4). (a). From the given information and the above calculation, we find that the matrix of T is A = 4 −1 0 5 1 4 . (b). Because A has more columns than rows, REF(A) must have an unpivoted column, which implies that nullspace(A) = {0}. Hence, T is not one-to-one. On the other hand, colspace(A) = R2 since the first two columns, for example, are linearly independent vectors in R2 . Thus, T is onto. 16. Show T is a linear transformation: Let x, y ∈ V and c, λ ∈ R where λ = 0. T (x + y) = λ(x + y) = λx + λy = T (x) + T (y), and T (cx) = λ(cx) = c(λx) = cT (x). Thus, T is a linear transformation. Show T is one-to-one: x ∈ Ker(T ) ⇐⇒ T (x) = 0 ⇐⇒ λx = 0 ⇐⇒ x = 0, since λ = 0. Thus, Ker(T ) = {0}. By Theorem 6.4.8, T is one-to-one. Show T is onto: dim[Ker(T )]+ dim[Rng(T )] = dim[V ] =⇒ dim[{0}]+ dim[Rng(T )] = dim[V ] =⇒ 0 + dim[Rng(T )] = dim[V ] =⇒ dim[Rng(T )] = dim[V ] =⇒ Rng(T ) = V . Thus, T is onto. Find T −1 : 1 1 1 T −1 (x) = x since T (T −1 (x)) = T x =λ x = x. λ λ λ 17. T : P1 (R) → P1 (R) where T (ax + b) = (2b − a)x + (b + a). Show T is one-to-one: Ker(T ) = {p ∈ P1 (R) : T (p) = 0} = {ax + b ∈ P1 (R) : (2b − a)x + (b + a) = 0}. Thus, a and b must satisfy the system 2b − a = 0, b + a = 0. The only solution to this system is a = b = 0. Consequently, Ker(T ) = {0}, so T is one-to-one. Show T is onto: Since Ker(T ) = {0}, dim[Ker(T )] = 0. Thus, dim[Ker(T )]+ dim[Rng(T )] = dim[P1 (R)] =⇒ dim[Rng(T )] = dim[P1 (R)], and since Rng(T ) is a subspace of P1 (R), it follows that Rng(T ) = P1 (R). Consequently, T is onto by Theorem 6.4.8. Determine T −1 : Since T is one-to-one and onto, T −1 exists. T (ax + b) = (2b − a)x + (b + a) A+B 2B − A and a = =⇒ T −1 [(2b − a)x + (b + a)] = ax + b. If we let A = 2b − a and B = a + b, then b = 3 3 1 1 −1 so that T [Ax + B] = (2B − A)x + (A + B). 3 3 18. T is not one-to-one: Ker(T ) = {p ∈ P2 (R) : T (p) = 0} = {ax2 + bx + c ∈ P2 (R) : c + (a − b)x = 0}. Thus, a, b, and c must satisfy the system c = 0 and a − b = 0. Consequently, Ker(T ) = {r(x2 + x) : r ∈ R} = {0}. Thus, by Theorem (c)2017 Pearson Education. Inc. 414 6.4.8, T is not one-to-one. T −1 does not exist because T is not one-to-one. T is onto: Since Ker(T ) = {r(x2 + x) : r ∈ R}, we see that dim[Ker(T )] = 1. Thus, dim[Ker(T )]+ dim[Rng(T )] = dim[P2 (R)] =⇒ 1+ dim[Rng(T )] = 3 =⇒ dim[Rng(T )] = 2. Since Rng(T ) is a subspace of P1 (R), 2 = dim[Rng(T )] ≤ dim[P1 (R)] = 2, and so equality holds: Rng(T ) = P1 (R). Thus, T is onto by Theorem 6.4.8. 19: ONTO only. Since dim[P2 (R)] = 3 > 2 = dim[R2 ], we know by inspection that T cannot be one-to-one. To see that T is onto, note that T (x2 ) = (1, 0) and T (x) = (−3, 1). Therefore, any linear combination of (1, 0) and (−3, 1) belongs to Rng(T ). Since {(1, 0), (−3, 1)} spans R2 , we conclude that Rng(T ) = R2 , which implies that T is onto. Since T fails to be one-to-one, T −1 does not exist. a b 20. ONE-TO-ONE and ONTO. Note that belongs to Ker(T ) if and only if ax2 + bx + c = 0; b c 0 0 that is, if and only if a = b = c = 0. Thus, Ker(T ) = . Therefore, T is one-to-one. Since 0 0 −1 that dim[V ] = 3 = dim[P2 (R)], Proposition 6.4.14 immediately implies T is also onto. Therefore, T exists, a b . and we have T −1 : P2 (R) → V given by T (ax2 + bx + c) = b c 21. ONE-TO-ONE only. We can see by inspection that Rng(T ) = M2 (R), since the outputs from T all have a zero in the lower right corner. Therefore, T is not onto. (Alternatively, this conclusion can bereached 0 0 by comparing dimensions for R3 and M2 (R).) To check if T is one-to-one, set T (a, b, c) = . Then 0 0 we have −a + 3c = 0, a − b − c = 0, 2a + b = 0. ⎡ ⎤ −1 0 3 0 This leads to the linear system of equations with augmented matrix ⎣ 1 −1 −1 0 ⎦. Since the coeffi2 1 0 0 cient matrix for this system is invertible (check this, for example, by using the determinant), we know that the system has only the trivial solution a = b = c = 0. Therefore, T is one-to-one. Since T fails to be onto, T −1 does not exist. a b 22. NEITHER. To check if T is one-to-one, set T = 0. This implies that c d (a − b + d) + (2a + b)x + cx2 + (4a − b + 2d)x3 = 0. This leads to the linear system of equations a − b + d = 0, 2a + b = 0, c = 0, The augmented matrix for this linear system is ⎡ 1 −1 0 1 ⎢ 2 1 0 0 ⎢ ⎣ 0 0 1 0 4 −1 0 2 4a − b + 2d = 0. ⎤ 0 0 ⎥ ⎥, 0 ⎦ 0 (c)2017 Pearson Education. Inc. 415 ⎤ 1 −1 0 1 ⎢ 2 1 0 0 ⎥ ⎥, we can use cofactor expansion along and considering the 4 × 4 coefficient matrix A = ⎢ ⎣ 0 0 1 0 ⎦ 4 −1 0 2 the third row, followed by the arrow method for 3 × 3 determinants, to verify that the coefficient matrix is not invertible. Therefore, the linear system has nontrivial solutions, so that Ker(T ) is nontrivial. Hence, T is not one-to-one. Moreover, since dim[M2 (R)] = 4 = dim[P3 (R)], we conclude that T is also not onto by Proposition 6.4.14. Since T is not one-to-one and not onto, we know that T −1 does not exist. ⎡ 23. {v1 , v2 } is a basis for V and T : V → V is a linear transformation. Show T is one-to-one: Any v ∈ V can be expressed as v = av1 + bv2 where a, b ∈ R. T (v) = 0 ⇐⇒ T (av1 + bv2 ) = 0 ⇐⇒ aT (v1 ) + bT (v2 ) = 0 ⇐⇒ a(v1 + 2v2 ) + b(2v1 − 3v2 ) = 0 ⇐⇒ (a + 2b)v1 + (2a − 3b)v2 = 0 ⇐⇒ a + 2b = 0 and 2a − 3b = 0 ⇐⇒ a = b = 0 ⇐⇒ v = 0. Hence Ker(T ) = {0}. Therefore T is one-to-one. Show T is onto: dim[Ker(T )] = dim[{0}] = 0 and dim[Ker(T )]+ dim[Rng(T )] = dim[V ] implies that 0+ dim[Rng(T )] = 2 or dim[Rng(T )] = 2. Since Rng(T ) is a subspace of V , it follows that Rng(T ) = V . Thus, T is onto by Theorem 6.4.8. Determine T −1 : Since T is one-to-one and onto, T −1 exists. T (av1 + bv2 ) = (a + 2b)v1 + (2a − 3b)v2 1 =⇒ T −1 [(a + 2b)v1 + (2a − 3b)v2 ] = (av1 + bv2 ). If we let A = a + 2b and B = 2a − 3b, then a = (3A + 2B) 7 1 1 1 and b = (2A − B). Hence, T −1 (Av1 + Bv2 ) = (3A + 2B)v1 + (2A − B)v2 . 7 7 7 24. Let v ∈ V . Then there exists a, b ∈ R such that v = av1 + bv2 . a b (T1 T2 )v = T1 [aT2 (v1 ) + bT2 (v2 )] = T1 (v1 + v2 ) + (v1 − v2 ) 2 2 a−b a+b a−b a+b = T1 (v1 ) + T1 (v2 ) = (v1 + v2 ) + (v1 − v2 ) 2 2 2 2 a+b a−b a+b a−b v1 + v2 = av1 + bv2 = v. = + − 2 2 2 2 (T2 T1 )v = T2 [aT1 (v1 ) + bT1 (v2 )] = T2 [a(v1 + v2 ) + b(v1 − v2 )] = T2 [(a + b)v1 + (a − b)v2 ] = (a + b)T2 (v1 ) + (a − b)T2 (v2 ) a−b a+b = (v1 + v2 ) + (v1 − v2 ) = av1 + bv2 = v. 2 2 Since (T1 T2 )v = v and (T2 T1 )v = v for all v ∈ V , it follows that T2 is the inverse of T1 , thus T2 = T1−1 . 25. An arbitrary vector in P1 (R) can be written as p(x) = ap0 (x) + bp1 (x), where p0 (x) = 1, and p1 (x) = x denote the standard basis vectors in P1 (R). Hence, we can define an isomorphism T : R2 → P1 (R) by T (a, b) = a + bx. 26. Let S denote the subspace of M2 (R) consisting of all upper triangular matrices. An arbitrary vector in (c)2017 Pearson Education. Inc. 416 S can be written as a 0 b c =a 1 0 0 0 0 0 1 0 a 0 b c +b +c 0 0 0 1 . Therefore, we define an isomorphism T : R3 → S by T (a, b, c) = . 27. Let S denote the subspace of M2 (R) consisting of all skew-symmetric matrices. An arbitrary vector in S can be written as 0 a 0 1 =a . −a 0 −1 0 Therefore, we can define an isomorphism T : R → S by 0 T (a) = −a a 0 . 28. Let S denote the subspace of M2 (R) consisting of all symmetric matrices. An arbitrary vector in S can be written as a b 1 0 0 1 0 0 =a +b +c . b c 0 0 1 0 0 1 Therefore, we can define an isomorphism T : R3 → S by a T (a, b, c) = b ⎡ a ⎢ 0 29. A typical vector in V takes the form A = ⎢ ⎣ 0 0 b e 0 0 c f h 0 b c . ⎤ d g ⎥ ⎥. Therefore, we can define T : V → R10 via i ⎦ j T (A) = (a, b, c, d, e, f, g, h, i, j). It is routine to verify that T is an invertible linear transformation. Therefore, we have n = 10. 30. A typical vector in V takes the form p = a0 + a2 x2 + a4 x4 + a6 x6 + a8 x8 . Therefore, we can define T : V → R5 via T (p) = (a0 , a2 , a4 , a6 , a8 ). It is routine to verify that T is an invertible linear transformation. Therefore, we have n = 5. 31. We can see that V is isomorphic to R3 via the isomorphism T : V → R3 given by ⎛⎡ ⎤⎞ 0 a b 0 c ⎦⎠ = (a, b, c). T ⎝⎣ −a −b −c 0 It is easy to verify that T is one-to-one and onto, and thus, T is an isomorphism. Therefore, we conclude that n = 3. (c)2017 Pearson Education. Inc. 417 32. We have T1−1 (x) = A−1 x = 33. We have T2−1 (x) = A−1 x = −1/6 2/3 3 −1 −2 1 x. x. ⎡ ⎤ 8 −29 3 19 −2 ⎦ x. T3−1 (x) = A−1 x = ⎣ −5 2 −8 1 34. We have ⎡ ⎤ 11/14 −8/7 1/14 5/7 1/7 ⎦ x. T4−1 (x) = A−1 x = ⎣ −3/7 3/14 1/7 −1/14 35. We have 36. The matrix of T2 T1 is −2 1 . −2 4 −1/3 1/3 −4 −1 2 2 1 2 1 3 = −6 −7 6 8 . The matrix of T1 T2 is ⎤ ⎤⎡ ⎡ 1 1 3 6 3 5 1 ⎦ ⎦ ⎣ ⎣ ⎣ 0 1 2 4 1 2 1 = 37. The matrix of T4 T3 is 3 5 −1 2 6 7 23 ⎤ ⎡ ⎤⎡ ⎤ ⎡ 3 5 1 10 25 23 1 1 3 ⎣ 0 1 2 ⎦ ⎣ 1 2 1 ⎦ = ⎣ 5 14 15 ⎦. 2 6 7 3 5 −1 12 19 1 ⎡ ⎤ 13 18 8 6 ⎦. 43 11 1 2 1 3 −4 −1 2 2 The matrix of T3 T4 is 38. (a). To show that T2 T1 : V1 → V3 is one-to-one, we show that Ker(T2 T1 ) = {0}. Suppose that v1 ∈ Ker(T2 T1 ). This means that (T2 T1 )(v1 ) = 0. Hence, T2 (T1 (v1 )) = 0. However, since T2 is one-to-one, we conclude that T1 (v1 ) = 0. Next, since T1 is one-to-one, we conclude that v1 = 0, which shows that the only vector in Ker(T2 T1 ) is 0, as expected. (b). To show that T2 T1 : V1 → V3 is onto, we begin with an arbitrary vector v3 in V3 . Since T2 : V2 → V3 is onto, there exists v2 in V2 such that T2 (v2 ) = v3 . Moreover, since T1 : V1 → V2 is onto, there exists v1 in V1 such that T1 (v1 ) = v2 . Therefore, (T2 T1 )(v1 ) = T2 (T1 (v1 )) = T2 (v2 ) = v3 , and therefore, we have found a vector, namely v1 , in V1 that is mapped under T2 T1 to v3 . Hence, T2 T1 is onto. (c). This follows immediately from parts (a) and (b). 39. We have (T2 T1 )(cu1 ) = T2 (T1 (cu1 )) = T2 (cT1 (u1 )) = c(T2 (T1 (u1 ))) = c(T2 T1 )(u1 ), (c)2017 Pearson Education. Inc. = 418 as needed. 40. We first prove that if T is one-to-one, then T is onto. The assumption that T is one-to-one implies that dim[Ker(T )] = 0. Hence, by Theorem 6.3.8, dim[W ] = dim[V ] = dim[Rng(T )], which implies that Rng(T ) = W . That is, T is onto. Next we show that if T is onto, then T is one-to-one. We have Rng(T ) = W , so that dim[Rng(T )] = dim[W ] = dim[V ]. Hence, by Theorem 6.3.8, we have dim[Ker(T )] = 0. Hence, Ker(T ) = {0}, which implies that T is one-to-one. 41. If T : V → W is linear, then show that T −1 : W → V is also linear. Let y, z ∈ W, c ∈ R, and T (u) = y and T (v) = z where u, v ∈ V . Then T −1 (y) = u and T −1 (z) = v. Thus, T −1 (y + z) = T −1 (T (u) + T (v)) = T −1 (T (u + v)) = u + v = T −1 (y) + T −1 (z), and T −1 (cy) = T −1 (cT (u)) = T −1 (T (cu)) = cu = cT −1 (y). Hence, T −1 is a linear transformation. 42. Let T : V → V be a one-to-one linear transformation. Since T is one-to-one, it follows from Theorem 6.4.8 that Ker(T ) = {0}, so dim[Ker(T )] = 0. By Theorem 6.3.8 and substitution, dim[Ker(T )]+ dim[Rng(T )] = dim[V ] =⇒ 0 + dim[Rng(T )] = dim[V ] =⇒ dim[Rng(T )] = dim[V ], and since Rng(T ) is a subspace of V , it follows that Rng(T ) = V , thus T is onto. T −1 exists because T is both one-to-one and onto. 43. To show that {T (v1 ), T (v2 ), . . . , T (vk )} is linearly independent, assume that c1 T (v1 ) + c2 T (v2 ) + · · · + ck T (vk ) = 0. We must show that c1 = c2 = · · · = ck = 0. Using the linearity properties of T , we can write T (c1 v1 + c2 v2 + · · · + ck vk ) = 0. Now, since T is one-to-one, we can conclude that c1 v1 + c2 v2 + · · · + ck vk = 0, and since {v1 , v2 , . . . , vk } is linearly independent, we conclude that c1 = c2 = · · · = ck = 0 as desired. 44. To prove that T is onto, let w be an arbitrary vector in W . We must find a vector v in V such that T (v) = w. Since {w1 , w2 , . . . , wm } spans W , we can write w = c 1 w1 + c 2 w2 + · · · + c m wm for some scalars c1 , c2 , . . . , cm . Therefore T (c1 v1 + c2 v2 + · · · + cm vm ) = c1 T (v1 ) + c2 T (v2 ) + · · · + cm T (vm ) = c1 w1 + c2 w2 + · · · + cm wm = w, which shows that v = c1 v1 + c2 v2 + · · · + cm vm maps under T to w, as desired. 45. Since T is a linear transformation, Rng(T ) is a subspace of W , but dim[W ] = dim[Rng(T )] = n, so W = Rng(T ), which means, by definition, that T is onto. 46. Let v ∈ V be arbitrary. Since T2 is onto, there exists w ∈ V such that T2 (w) = v. Applying T1 to both sides of this, we get T1 (T2 (w)) = T1 (v). (c)2017 Pearson Education. Inc. 419 However, by assumption, (T1 T2 )(w) = w. We conclude that T1 (v) = w. Applying T2 to both sides of this, we get (T2 T1 )(v) = T2 (w) = v, as required. 47. Suppose that x belongs to Rng(T ). This means that there exists a vector v in V such that T (v) = x. Applying T to both sides of this equation, we have T (T (v)) = T (x), or T 2 (v) = T (x). However, since T 2 = 0, we conclude that T 2 (v) = 0, and hence, T (x) = 0. By definition, this means that x belongs to Ker(T ). Thus, since every vector in Rng(T ) belongs to Ker(T ), Rng(T ) is a subset of Ker(T ). In addition, Rng(T ) is closed under addition and scalar multiplication, by Theorem 6.3.5, and therefore, Rng(T ) forms a subspace of Ker(T ). Solutions to Section 6.5 True-False Review: (a): FALSE. The matrix representation is an m × n matrix, not an n × m matrix. (b): FALSE. The matrix [T ]B C would only make sense if C was a basis for V and B was a basis for W , and this would only be true if V and W were the same vector space. Of course, in general V and W can be different, and so this statement does not hold. (c): FALSE. The correct equation is given in (6.5.2). −1 (d): FALSE. The correct statement is [T ]C = [T −1 ]B B C. (e): TRUE. Many examples are possible. A fairly simple one is the following. Let T1 : R2 → R2 be given by T1 (x, y) = (x, y), and let T2 : R2 → R2 be given by T2 (x, y) = (y,x). Clearly, T1 and T2 are different 1 0 1 . If B2 = {(1, 0), (0, 1)} and linear transformations. Now if B1 = {(1, 0), (0, 1)} = C1 , then [T1 ]C B1 = 0 1 1 0 2 . Thus, although T1 = T2 , we found suitable bases B1 , C1 , B2 , C2 C2 = {(0, 1), (1, 0)}, then [T2 ]C B2 = 0 1 C2 1 such that [T1 ]C B1 = [T2 ]B2 . (f ): TRUE. This is the content of part (b) of Corollary 6.5.10. Problems: 1. (a). We must determine T (E11 ), T (E12 ), T (E21 ), and T (E22 ), and find the components of the resulting vectors in P3 (R) relative to the basis C. We have T (E11 ) = 1 − x3 , T (E12 ) = 3x2 , T (E21 ) = x3 , T (E22 ) = −1. Therefore, relative to the standard basis C on P3 (R), we have ⎤ 1 ⎢ 0 ⎥ ⎥ [T (E11 )]C = ⎢ ⎣ 0 ⎦, −1 ⎡ ⎤ 0 ⎢ 0 ⎥ ⎥ [T (E12 )]C = ⎢ ⎣ 3 ⎦, 0 ⎡ ⎤ 0 ⎢ 0 ⎥ ⎥ [T (E21 )]C = ⎢ ⎣ 0 ⎦, 1 (c)2017 Pearson Education. Inc. ⎡ ⎤ −1 ⎢ 0 ⎥ ⎥ [T (E22 )]C = ⎢ ⎣ 0 ⎦. 0 ⎡ 420 Putting these results into the columns of [T ]C B , we obtain ⎤ ⎡ 1 0 0 −1 ⎢ 0 0 0 0 ⎥ ⎥. ⎢ [T ]C B =⎣ 0 3 0 0 ⎦ −1 0 1 0 (b). The values of T (E21 ), T (E11 ), T (E22 ), and T (E12 ) were determined in part (a). We must express those results in terms of the ordered basis C given in (b). We have ⎤ ⎤ ⎤ ⎤ ⎡ ⎡ ⎡ ⎡ 0 0 0 0 ⎥ ⎥ ⎥ ⎢ 0 ⎥ ⎢ ⎢ ⎢ ⎥ , [T (E11 )]C = ⎢ 1 ⎥ , [T (E22 )]C = ⎢ −1 ⎥ , [T (E12 )]C = ⎢ 0 ⎥ . [T (E21 )]C = ⎢ ⎣ 1 ⎦ ⎣ −1 ⎦ ⎣ 0 ⎦ ⎣ 0 ⎦ 0 0 0 3 Putting these results into the columns of [T ]C B , we obtain ⎤ ⎡ 0 0 0 0 ⎢ 0 1 −1 0 ⎥ ⎥. ⎢ [T ]C B = ⎣ 1 −1 0 0 ⎦ 0 0 0 3 2. (a). We must determine T (1), T (x), and T (x2 ), and find the components of the resulting vectors in R2 relative to the basis C. We have T (1) = (1, 2), T (x) = (0, 1), T (x2 ) = (−3, −2). Therefore, relative to the standard basis C on R2 , we have 1 0 [T (1)]C = , [T (x)]C = , 2 1 Therefore, [T ]C B = 1 2 0 −3 1 −2 2 [T (x )]C = −3 −2 . . (b). We must determine T (1), T (1 + x), and T (1 + x + x2 ), and find the components of the resulting vectors in R2 relative to the basis C. We have T (1) = (1, 2), T (1 + x) = (1, 3), T (1 + x + x2 ) = (−2, 1). Setting T (1) = (1, 2) = c1 (1, −1) + c2 (2, 1) −1 . Next, setting and solving, we find c1 = −1 and c2 = 1. Thus, [T (1)]C = 1 T (1 + x) = (1, 3) = c1 (1, −1) + c2 (2, 1) (c)2017 Pearson Education. Inc. 421 and solving, we find c1 = −5/3 and c2 = 4/3. Thus, [T (1 + x)]C = −5/3 4/3 . Finally, setting T (1 + x + x2 ) = (−2, 1) = c1 (1, −1) + c2 (2, 1) −4/3 and solving, we find c1 = −4/3 and c2 = −1/3. Thus, [T (1 + x + x2 )]C = . Putting the results for −1/3 [T (1)]C , [T (1 + x)]C , and [T (1 + x + x2 )]C into the columns of [T ]C B , we obtain −1 − 53 − 43 [T ]C . 4 B = 1 − 13 3 3. (a). We must determine T (1), T (x), and T (x2 ), and find the components of the resulting vectors relative to the standard basis C on P3 (R). We have T (1) = 1 + x, Therefore, ⎤ 1 ⎢ 1 ⎥ ⎥ [T (1)]C = ⎢ ⎣ 0 ⎦, 0 ⎡ T (x) = x + x2 , T (x2 ) = x2 + x3 . ⎤ 0 ⎢ 1 ⎥ ⎥ [T (x)]C = ⎢ ⎣ 1 ⎦, 0 ⎡ Putting these results into the columns of [T ]C B , we obtain ⎡ 1 0 ⎢ ⎢ 1 1 [T ]C B =⎣ 0 1 0 0 ⎤ 0 ⎢ 0 ⎥ ⎥ [T (x2 )]C = ⎢ ⎣ 1 ⎦. 1 ⎡ ⎤ 0 0 ⎥ ⎥. 1 ⎦ 1 (b). We must determine T (1), T (x − 1), and T ((x − 1)2 ), and find the components of the resulting vectors relative to the given basis C on P3 (R). We have T (1) = 1 + x, Setting T (x − 1) = −1 + x2 , T ((x − 1)2 ) = 1 − 2x − x2 + x3 . 1 + x = c1 (1) + c2 (x − 1) + c3 (x − 1)2 + c4 (x − 1)3 ⎡ ⎤ 2 ⎢ 1 ⎥ ⎥ and solving, we find c1 = 2, c2 = 1, c3 = 0, and c4 = 0. Therefore [T (1)]C = ⎢ ⎣ 0 ⎦. Next, setting 0 −1 + x2 = c1 (1) + c2 (x − 1) + c3 (x − 1)2 + c4 (x − 1)3 ⎤ 0 ⎢ 2 ⎥ ⎥ and solving, we find c1 = 0, c2 = 2, c3 = 1, and c4 = 0. Therefore, [T (x − 1)]C = ⎢ ⎣ 1 ⎦. Finally, setting 0 ⎡ 1 − 2x − x2 + x3 = c1 (1) + c2 (x − 1) + c3 (x − 1)2 + c4 (x − 1)3 (c)2017 Pearson Education. Inc. 422 ⎤ −1 ⎢ −1 ⎥ ⎥ and solving, we find c1 = −1, c2 = −1, c3 = 2, and c4 = 1. Therefore, [T ((x − 1)2 )]C = ⎢ ⎣ 2 ⎦. Putting 1 these results into the columns of [T ]C B , we obtain ⎤ ⎡ 2 0 −1 ⎢ 1 2 −1 ⎥ ⎥. ⎢ [T ]C B =⎣ 0 1 2 ⎦ 0 0 1 ⎡ 4. (a). We must determine T (1, 0, 0), T (0, 1, 0), and T (0, 0, 1), and find the components of the resulting vectors relative to the basis C. We have T (1, 0, 0) = cos x, T (0, 1, 0) = 3 sin x, T (0, 0, 1) = −2 cos x + sin x. Therefore, relative to the basis C, we have 1 0 [T (1, 0, 0)]C = , [T (0, 1, 0)]C = , 0 3 [T (0, 0, 1)]C = −2 1 . Putting these results into the columns of [T ]C B , we obtain 1 0 −2 = . [T ]C B 0 3 1 (b). We must determine T (2, −1, −1), T (1, 3, 5), and T (0, 4, −1), and find the components of the resulting vectors relative to the basis C. We have T (2, −1, −1) = 4 cos x − 4 sin x, T (1, 3, 5) = −9 cos x + 14 sin x, T (0, 4, −1) = 2 cos x + 11 sin x. Setting 4 cos x − 4 sin x = c1 (cos x − sin x) + c2 (cos x + sin x) 4 . Next, setting and solving, we find c1 = 4 and c2 = 0. Therefore [T (2, −1, −1)]C = 0 −9 cos x + 14 sin x = c1 (cos x − sin x) + c2 (cos x + sin x) −23/2 . Finally, setting and solving, we find c1 = −23/2 and c2 = 5/2. Therefore, [T (1, 3, 5)]C = 5/2 2 cos x + 11 sin x = c1 (cos x − sin x) + c2 (cos x + sin x) −9/2 . Putting these results and solving, we find c1 = −9/2 and c2 = 13/2. Therefore [T (0, 4, −1)]C = 13/2 into the columns of [T ]C B , we obtain − 92 4 − 23 2 . = [T ]C 13 5 B 0 2 2 (c)2017 Pearson Education. Inc. 423 5. (a). We must determine T (E11 ), T (E12 ), T (E21 ), and T (E22 ), and find the components of the resulting vectors relative to the standard basis C on R2 . We have T (E11 ) = (1, 1), T (E12 ) = (0, 0), T (E21 ) = (0, 0), T (E22 ) = (1, 1). Therefore, since C is the standard basis on R2 , we have 1 0 0 , [T (E12 )]C = , [T (E21 )]C = , [T (E11 )]C = 1 0 0 Putting these results into the columns of [T ]C B , we obtain 1 0 0 C [T ]B = 1 0 0 1 1 [T (E22 )]C = 1 1 . . (b). Let us denote the four matrices in the ordered basis for B as A1 , A2 , A3 , and A4 , respectively. We must determine T (A1 ), T (A2 ), T (A3 ), and T (A4 ), and find the components of the resulting vectors relative to the standard basis C on R2 . We have T (A1 ) = (−4, −4), T (A2 ) = (3, 3), T (A3 ) = (−2, −2), Therefore, since C is the standard basis on R2 , we have −4 3 −2 , [T (A2 )]C = , [T (A3 )]C = , [T (A1 )]C = −4 3 −2 Putting these results into the columns of [T ]C B , we obtain −4 3 −2 = [T ]C B −4 3 −2 0 0 T (A4 ) = (0, 0). [T (A4 )]C = 0 0 . . 6. (a). We must determine T (1), T (x), T (x2 ), and T (x3 ), and find the components of the resulting vectors relative to the standard basis C on P2 (R). We have T (1) = 0, T (x) = 1, T (x2 ) = 2x, T (x3 ) = 3x2 . Therefore, if C is the standard basis on P2 (R), then we have ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 0 0 1 [T (1)]C = ⎣ 0 ⎦ , [T (x)]C = ⎣ 0 ⎦ , [T (x2 )]C = ⎣ 2 ⎦ , 0 0 0 Putting these results into the columns of [T ]C B , we obtain ⎡ 0 1 0 ⎣ 0 0 2 [T ]C B = 0 0 0 ⎤ 0 0 ⎦. 3 (c)2017 Pearson Education. Inc. ⎡ ⎤ 0 [T (x3 )]C = ⎣ 0 ⎦ . 3 424 (b). We must determine T (x3 ), T (x3 + 1), T (x3 + x), and T (x3 + x2 ), and find the components of the resulting vectors relative to the given basis C on P2 (R). We have T (x3 ) = 3x2 , Setting T (x3 + 1) = 3x2 , T (x3 + x) = 3x2 + 1, T (x3 + x2 ) = 3x2 + 2x. 3x2 = c1 (1) + c2 (1 + x) + c3 (1 + x + x2 ) ⎡ ⎤ 0 and solving, we find c1 = 0, c2 = −3, and c3 = 3. Therefore [T (x3 )]C = ⎣ −3 ⎦. Likewise, [T (x3 + 1)]C = 3 ⎡ ⎤ 0 ⎣ −3 ⎦. Next, setting 3 3x2 + 1 = c1 (1) + c2 (1 + x) + c3 (1 + x + x2 ) ⎡ ⎤ 1 and solving, we find c1 = 1, c2 = −3, and c3 = 3. Therefore, [T (x3 + x)]C = ⎣ −3 ⎦. Finally, setting 3 3x2 + 2x = c1 (1) + c2 (1 + x) + c3 (1 + x + x2 ) ⎡ ⎤ −2 and solving, we find c1 = −2, c2 = −1, and c3 = 3. Therefore, [T (x3 + 2x)]C = ⎣ −1 ⎦. Putting these 3 , we obtain results into the columns of [T ]C B ⎡ ⎤ 0 0 1 −2 ⎣ −3 −3 −3 −1 ⎦ . [T ]C B = 3 3 3 3 7. (a). We must determine T (e2x ) and T (e−3x ), and find the components of the resulting vectors relative to the given basis C. We have T (e2x ) = 2e2x and T (e−3x ) = −3e−3x . Therefore, we have [T ]C B = 2 0 0 −3 . (b). We must determine T (e2x − 3e−3x ) and T (2e−3x ), and find the components of the resulting vectors relative to the given basis C. We have T (e2x − 3e−3x ) = 2e2x + 9e−3x and T (2e−3x ) = −6e−3x . Now, setting 2e2x + 9e−3x = c1 (e2x + e−3x ) + c2 (−e2x ) 9 2x −3x )]C = . Finally, setting and solving, we find c1 = 9 and c2 = 7. Therefore, [T (e − 3e 7 −6e−3x = c1 (e2x + e−3x ) + c2 (−e2x ) (c)2017 Pearson Education. Inc. 425 and solving, we find c1 = c2 = −6. Therefore, [T (2e−3x )]C = of [T ]C B , we obtain [T ]C B = 9 −6 7 −6 −6 −6 . Putting these results into the columns . 8. (a). We must determine T (E11 ), T (E12 ), T (E21 ), and T (E22 ), and find the components of the resulting vectors relative to the standard basis C on M2 (R). We have T (E11 ) = E11 , T (E12 ) = 2E12 − E21 , T (E21 ) = 2E21 − E12 , T (E22 ) = E22 . ⎤ 0 ⎢ 2 ⎥ ⎥ [T (E12 )]C = ⎢ ⎣ −1 ⎦ , 0 ⎤ 0 ⎢ −1 ⎥ ⎥ [T (E21 )]C = ⎢ ⎣ 2 ⎦, 0 ⎤ 0 ⎢ 0 ⎥ ⎥ [T (E22 )]C = ⎢ ⎣ 0 ⎦. 1 Therefore, we have ⎤ 1 ⎢ 0 ⎥ ⎥ [T (E11 )]C = ⎢ ⎣ 0 ⎦, 0 ⎡ ⎡ ⎡ ⎡ Putting these results into the columns of [T ]C B , we obtain ⎤ 1 0 0 0 ⎢ 0 2 −1 0 ⎥ ⎥. ⎢ [T ]C B = ⎣ 0 −1 2 0 ⎦ 0 0 0 1 ⎡ (b). Let us denote the four matrices in the ordered basis for B as A1 , A2 , A3 , and A4 , respectively. We must determine T (A1 ), T (A2 ), T (A3 ), and T (A4 ), and find the components of the resulting vectors relative to the standard basis C on M2 (R). We have −1 −2 1 0 0 −8 0 7 T (A1 ) = , T (A2 ) = , T (A3 ) = , T (A4 ) = . −2 −3 3 2 7 −2 −2 0 Therefore, we have ⎤ −1 ⎢ −2 ⎥ ⎥ [T (A1 )]C = ⎢ ⎣ −2 ⎦ , −3 ⎡ ⎤ 1 ⎢ 0 ⎥ ⎥ [T (A2 )]C = ⎢ ⎣ 3 ⎦, 2 ⎡ ⎤ 0 ⎢ −8 ⎥ ⎥ [T (A3 )]C = ⎢ ⎣ 7 ⎦, −2 ⎡ Putting these results into the columns of [T ]C B , we obtain ⎡ −1 ⎢ −2 C [T ]B = ⎢ ⎣ −2 −3 ⎤ 1 0 0 0 −8 7 ⎥ ⎥. 3 7 −2 ⎦ 2 −2 0 9. (c)2017 Pearson Education. Inc. ⎤ 0 ⎢ 7 ⎥ ⎥ [T (A4 )]C = ⎢ ⎣ −2 ⎦ . 0 ⎡ 426 (a). Let us first compute [T ]C B . We must determine T (1, 0, 0), T (0, 1, 0), and T (0, 0, 1), and find the components of the resulting vectors relative to the standard basis C = {1, x, x2 , x3 } on P3 (R). We have T (1, 0, 0) = 2 − x − x3 , Therefore, ⎤ 2 ⎢ −1 ⎥ ⎥ [T (1, 0, 0)]C = ⎢ ⎣ 0 ⎦, −1 T (0, 1, 0) = −x, T (0, 0, 1) = x + 2x3 . ⎤ 0 ⎢ −1 ⎥ ⎥ [T (0, 1, 0)]C = ⎢ ⎣ 0 ⎦, 0 ⎡ ⎤ 0 ⎢ 1 ⎥ ⎥ [T (0, 0, 1)]C = ⎢ ⎣ 0 ⎦. 2 ⎡ ⎡ Putting these results into the columns of [T ]C B , we obtain ⎤ ⎡ 2 0 0 ⎢ −1 −1 1 ⎥ ⎥. ⎢ [T ]C B =⎣ 0 0 0 ⎦ −1 0 2 Now, and therefore, ⎡ ⎤ 2 [v]B = [(2, −1, 5)]B = ⎣ −1 ⎦ , 5 ⎤ ⎤ ⎡ ⎤ 4 2 0 0 ⎡ 2 ⎥ ⎢ ⎢ −1 −1 1 ⎥ ⎥ ⎣ −1 ⎦ = ⎢ 4 ⎥ . ⎢ [T (v)]C = [T ]C B [v]B = ⎣ ⎣ 0 ⎦ 0 0 0 ⎦ 5 8 −1 0 2 ⎡ Therefore, (b). We have T (v) = 4 + 4x + 8x3 . T (v) = T (2, −1, 5) = 4 + 4x + 8x3 . 10. (a). Let us first compute [T ]C B . We must determine T (1) and T (x), and find the components of the resulting vectors relative to the standard basis C = {E11 , E12 , E21 , E22 } on M2 (R). We have 1 0 −1 0 T (1) = and T (x) = . 0 −1 −2 1 Therefore Hence, ⎤ 1 ⎢ 0 ⎥ ⎥ [T (1)]C = ⎢ ⎣ 0 ⎦ −1 ⎤ −1 ⎢ 0 ⎥ ⎥ and [T (x)]C = ⎢ ⎣ −2 ⎦ . 1 ⎡ ⎡ ⎤ 1 −1 ⎢ 0 0 ⎥ ⎥. ⎢ [T ]C B =⎣ 0 −2 ⎦ −1 1 ⎡ (c)2017 Pearson Education. Inc. 427 Now, [p(x)]B = [−2 + 3x]B = −2 3 . Therefore, we have ⎤ ⎡ ⎤ −5 1 −1 ⎥ ⎢ ⎢ 0 0 ⎥ ⎥ −2 = ⎢ 0 ⎥ . ⎢ [T (p(x))]C = [T ]C B [p(x)]B = ⎣ ⎣ ⎦ −6 ⎦ 3 0 −2 5 −1 1 ⎡ Thus, T (p(x)) = −5 −6 (b). We have 0 5 T (p(x)) = T (−2 + 3x) = . −5 −6 0 5 . 11. 2 (a). Let us first compute [T ]C B . We must determine T (1), T (x), and T (x ), and find the components of the 2 3 4 resulting vectors relative to the standard basis C = {1, x, x , x , x }. We have T (1) = x2 , Therefore, ⎤ 0 ⎢ 0 ⎥ ⎥ ⎢ ⎥ [T (1)]C = ⎢ ⎢ 1 ⎥, ⎣ 0 ⎦ 0 ⎡ T (x) = x3 , ⎤ 0 ⎢ 0 ⎥ ⎥ ⎢ ⎥ [T (x)]C = ⎢ ⎢ 0 ⎥, ⎣ 1 ⎦ 0 and therefore, ⎤ 0 ⎢ 0 ⎥ ⎥ ⎢ ⎥ [T (x2 )]C = ⎢ ⎢ 0 ⎥. ⎣ 0 ⎦ 1 ⎡ Putting these results into the columns of [T ]C B , we obtain ⎡ 0 0 ⎢ 0 0 ⎢ ⎢ [T ]C B =⎢ 1 0 ⎣ 0 1 0 0 Now, T (x2 ) = x4 . ⎡ ⎤ 0 0 ⎥ ⎥ 0 ⎥ ⎥. 0 ⎦ 1 ⎡ ⎤ −1 [p(x)]B = [−1 + 5x − 6x2 ]B = ⎣ 5 ⎦ , −6 ⎡ 0 ⎢ 0 ⎢ ⎢ [T (p(x))]C = [T ]C B [p(x)]B = ⎢ 1 ⎣ 0 0 0 0 0 1 0 ⎤ ⎡ ⎤ 0 0 ⎡ ⎤ ⎢ 0 ⎥ 0 ⎥ ⎥ ⎢ ⎥ −1 ⎥ ⎥ ⎣ ⎦ 5 =⎢ 0 ⎥ ⎢ −1 ⎥ . ⎣ ⎦ 5 ⎦ −6 0 −6 1 (c)2017 Pearson Education. Inc. 428 Hence, T (p(x)) = −x2 + 5x3 − 6x4 . (b). We have T (p(x)) = −x2 + 5x3 − 6x4 . 12. (a). Let us first compute [T ]C B . We must determine T (E11 ), T (E12 ), T (E21 ), and T (E22 ), and find the components of the resulting vectors relative to the standard basis C = {E11 , E12 , E21 , E22 }. We have 2 −1 −1 0 0 0 1 3 , T (E12 ) = , T (E21 ) = , T (E22 ) = . T (E11 ) = 0 −1 0 −1 0 3 0 0 Therefore, ⎤ 2 ⎢ −1 ⎥ ⎥ [T (E11 )]C = ⎢ ⎣ 0 ⎦, −1 ⎡ ⎤ −1 ⎢ 0 ⎥ ⎥ [T (E12 )]C = ⎢ ⎣ 0 ⎦, −1 ⎡ ⎤ 0 ⎢ 0 ⎥ ⎥ [T (E21 )]C = ⎢ ⎣ 0 ⎦, 3 ⎡ ⎤ 1 ⎢ 3 ⎥ ⎥ [T (E22 )]C = ⎢ ⎣ 0 ⎦. 0 ⎡ Putting these results into the columns of [T ]C B , we obtain ⎤ ⎡ 2 −1 0 1 ⎢ −1 0 0 3 ⎥ ⎥. ⎢ [T ]C B =⎣ 0 0 0 0 ⎦ −1 −1 3 0 Now, and therefore, ⎤ −7 ⎢ 2 ⎥ ⎥ [A]B = ⎢ ⎣ 1 ⎦, −3 ⎡ ⎤ ⎡ ⎤ ⎤⎡ −19 −7 2 −1 0 1 ⎥ ⎢ ⎥ ⎢ ⎢ −1 0 0 3 ⎥ ⎥ ⎢ 2 ⎥ = ⎢ −2 ⎥ . ⎢ [T (A)]C = [T ]C B [A]B = ⎣ ⎦ ⎣ ⎦ ⎣ 0 ⎦ 1 0 0 0 0 −3 8 −1 −1 3 0 ⎡ Hence, T (A) = (b). We have T (A) = −19 −2 0 8 −19 −2 0 8 . . 13. 2 3 4 (a). Let us first compute [T ]C B . We must determine T (1), T (x), T (x ), T (x ), and T (x ), and find the components of the resulting vectors relative to the standard basis C = {1, x, x2 , x3 }. We have T (1) = 0, T (x) = 1, T (x2 ) = 2x, T (x3 ) = 3x2 , (c)2017 Pearson Education. Inc. T (x4 ) = 4x3 . 429 Therefore, ⎤ 0 ⎢ 0 ⎥ ⎥ [T (1)]C = ⎢ ⎣ 0 ⎦, 0 ⎡ ⎤ 1 ⎢ 0 ⎥ ⎥ [T (x)]C = ⎢ ⎣ 0 ⎦, 0 ⎡ ⎤ 0 ⎢ 2 ⎥ ⎥ [T (x2 )]C = ⎢ ⎣ 0 ⎦, 0 ⎤ 0 ⎢ 0 ⎥ ⎥ [T (x3 )]C = ⎢ ⎣ 3 ⎦, 0 ⎡ ⎡ ⎤ 0 ⎢ 0 ⎥ ⎥ [T (x4 )]C = ⎢ ⎣ 0 ⎦. 4 ⎡ Putting these results into the columns of [T ]C B , we obtain ⎡ 0 ⎢ 0 C [T ]B = ⎢ ⎣ 0 0 Now, 1 0 0 0 0 2 0 0 0 0 3 0 ⎤ 0 0 ⎥ ⎥. 0 ⎦ 4 ⎤ 3 ⎢ −4 ⎥ ⎥ ⎢ ⎥ [p(x)]B = [3 − 4x + 6x2 + 6x3 − 2x4 ]B = ⎢ ⎢ 6 ⎥, ⎣ 6 ⎦ −2 ⎡ and therefore, ⎡ 0 ⎢ 0 0 0 ⎢ [T (p(x))]C = [T ]C B [p(x)]B = ⎣ 1 0 0 0 0 2 0 0 0 0 3 0 ⎤ ⎤ ⎡ 3 −4 0 ⎢ ⎥ −4 ⎥ ⎥ ⎢ ⎢ 0 ⎥ ⎥ ⎢ 6 ⎥ = ⎢ 12 ⎥ . ⎥ ⎢ ⎣ ⎦ 18 ⎦ 0 ⎣ 6 ⎦ −8 4 −2 ⎤ ⎡ Therefore, T (p(x)) = T (3 − 4x + 6x2 + 6x3 − 2x4 ) = −4 + 12x + 18x2 − 8x3 . (b). We have T (p(x)) = p (x) = −4 + 12x + 18x2 − 8x3 . 14. (a). Let us first compute [T ]C B . We must determine T (Eij ) for 1 ≤ i, j ≤ 3, and find components of the resulting vectors relative to the standard basis C = {1}. We have T (E11 ) = T (E22 ) = T (E33 ) = 1 and T (E12 ) = T (E13 ) = T (E23 ) = T (E21 ) = T (E31 ) = T (E32 ) = 0. Therefore [T (E11 )]C = [T (E22 ]C = [T (E33 ]C = [1] and all other component vectors are [0]. Putting these results into the columns of [T ]C B , we obtain [T ]C B = 1 0 0 0 1 0 0 0 (c)2017 Pearson Education. Inc. 1 . 430 Now, ⎤ 2 ⎢ −6 ⎥ ⎥ ⎢ ⎢ 0 ⎥ ⎥ ⎢ ⎢ 1 ⎥ ⎥ ⎢ ⎥ [A]B = ⎢ ⎢ 4 ⎥. ⎢ −4 ⎥ ⎥ ⎢ ⎢ 0 ⎥ ⎥ ⎢ ⎣ 0 ⎦ −3 ⎡ and therefore, ⎤ 2 ⎢ −6 ⎥ ⎥ ⎢ ⎢ 0 ⎥ ⎥ ⎢ ⎢ 1 ⎥ ⎥ ⎢ ⎢ 4 ⎥ = [3]. ⎥ ⎢ ⎢ −4 ⎥ ⎥ ⎢ ⎢ 0 ⎥ ⎥ ⎢ ⎣ 0 ⎦ −3 ⎡ [T (A)]C = [T ]C B [A]B = 1 0 0 0 1 0 0 0 1 Hence, T (A) = 3. (b). We have T (A) = 3. 15. 2 3 (a). Let us first compute [T ]C B . We must determine T (1), T (x), T (x ), and T (x ), and find the components of the resulting vectors relative to the standard basis C = {1}. We have T (1) = 1, T (x) = 2, T (x2 ) = 4, Therefore, [T (1)]C = [1], [T (x)]C = [2], T (x3 ) = 8. [T (x2 )]C = [4], [T (x3 )]C = [8]. Putting these results into the columns of [T ]C B , we obtain [T ]C B = Now, 1 2 4 8 . ⎤ 0 ⎢ 2 ⎥ ⎥ [p(x)]B = [2x − 3x2 ]B = ⎢ ⎣ −3 ⎦ , 0 ⎡ and therefore, ⎤ 0 ⎢ 2 ⎥ ⎥ ⎢ ⎣ −3 ⎦ = [−8]. 0 ⎡ [T (p(x))]C = [T ]C B [p(x)]B = 1 2 4 8 Therefore, T (p(x)) = −8. (c)2017 Pearson Education. Inc. 431 (b). We have p(2) = 2 · 2 − 3 · 22 = −8. 16. The linear transformation T2 T1 : P4 (R) → R is given by (T2 T1 )(p(x)) = p (2). Let A denote the standard basis on P4 (R), let B denote the standard basis on P3 (R), and let C denote the standard basis on R. (a). To determine [T2 T1 ]C A , we compute (T2 T1 )(1) = 0, (T2 T1 )(x2 ) = 4, (T2 T1 )(x) = 1, (T2 T1 )(x3 ) = 12, (T2 T1 )(x4 ) = 32. Therefore [(T2 T1 )(1)]C = [0], [(T2 T1 )(x2 )]C = [4], [(T2 T1 )(x)]C = [1], [(T2 T1 )(x3 )]C = [12], [(T2 T1 )(x4 )]C = [32]. Putting these results into the columns of [T2 T1 ]C A , we obtain [T2 T1 ]C A = 0 1 4 12 0 2 0 0 0 0 3 0 ⎤ 0 0 ⎥ ⎥= 0 ⎦ 4 32 . 0 1 (b). We have ⎡ B [T2 ]C B [T1 ]A = 1 2 4 8 0 ⎢ 0 ⎢ ⎣ 0 0 1 0 0 0 4 12 32 = [T2 T1 ]C A. 2 4 (c). Let p(x) ⎡ = 2⎤ + 5x − x + 3x . Then the component vector of p(x) relative to the standard basis A is 2 ⎢ 5 ⎥ ⎥ ⎢ ⎥ [p(x)]A = ⎢ ⎢ −1 ⎥. Thus, ⎣ 0 ⎦ 3 ⎤ 2 ⎢ 5 ⎥ ⎥ ⎢ ⎢ −1 ⎥ = [97]. ⎥ ⎢ ⎣ 0 ⎦ 3 ⎡ [(T2 T1 )(p(x))]C = [T2 T1 ]C A [p(x)]A = Therefore, Of course, 0 1 4 12 (T2 T1 )(2 + 5x − x2 + 3x4 ) = 97. p (2) = 5 − 2 · 2 + 12 · 23 = 97 by direct calculation as well. 17. The linear transformation T2 T1 : P1 (R) → R2 is given by (T2 T1 )(a + bx) = (0, 0). (c)2017 Pearson Education. Inc. 32 432 Let A denote the standard basis on P1 (R), let B denote the standard basis on M2 (R), and let C denote the standard basis on R2 . (a). To determine [T2 T1 ]C A , we compute (T2 T1 )(1) = (0, 0) and (T2 T1 )(x) = (0, 0). Therefore, we obtain [T2 T1 ]C A = 02 . (b). We have B [T2 ]C B [T1 ]A = 1 1 0 0 0 0 1 1 ⎤ 1 −1 ⎢ 0 0 ⎥ C ⎥ ⎢ ⎣ 0 −2 ⎦ = 02 = [T2 T1 ]A . −1 1 ⎡ (c). The component vector of p(x) = −3 + 8x relative to the standard basis A is [p(x)]A = [(T2 T1 )(p(x))]C = [T2 T1 ]C A [p(x)]A = 02 [p(x)]A = 0 0 −3 8 . Thus, . Therefore, (T2 T1 )(−3 + 8x) = (0, 0). Of course (T2 T1 )(−3 + 8x) = T1 −11 −16 0 11 = (0, 0) by direct calculation as well. 18. The linear transformation T2 T1 : P2 (R) → P2 (R) is given by (T2 T1 )(p(x)) = [(x + 1)p(x)] . Let A denote the standard basis on P2 (R), let B denote the standard basis on P3 (R), and let C denote the standard basis on P2 (R). (a). To determine [T2 T1 ]C A , we compute as follows: (T2 T1 )(1) = 1, Therefore, ⎡ ⎤ 1 [(T2 T1 )(1)]C = ⎣ 0 ⎦ , 0 (T2 T1 )(x) = 1 + 2x, (T2 T1 )(x2 ) = 2x + 3x2 . ⎡ ⎡ ⎤ 1 [(T2 T1 )(x)]C = ⎣ 2 ⎦ , 0 ⎤ 0 [(T2 T1 )(x2 )]C = ⎣ 2 ⎦ . 3 Putting these results into the columns of [T2 T1 ]C A , we obtain ⎡ ⎤ 1 1 0 C [T2 T1 ]A = ⎣ 0 2 2 ⎦ . 0 0 3 (b). We have ⎡ 0 B ⎣ 0 [T ] = [T2 ]C B 1 A 0 1 0 0 0 2 0 ⎡ ⎤ 1 0 ⎢ 1 0 ⎦⎢ ⎣ 0 3 0 0 1 1 0 ⎤ ⎡ 0 1 ⎥ 0 ⎥ ⎣ 0 = 1 ⎦ 0 1 (c)2017 Pearson Education. Inc. 1 2 0 ⎤ 0 2 ⎦ = [T2 T1 ]C A. 3 433 ⎡ ⎤ 7 (c). The component vector of p(x) = 7 − x + 2x2 relative to the standard basis A is [p(x)]A = ⎣ −1 ⎦. 2 Thus, ⎡ ⎤⎡ ⎤ ⎡ ⎤ 1 1 0 7 6 ⎣ ⎦ ⎣ ⎦ ⎣ 0 2 2 −1 2 ⎦. [p(x)] = [(T2 T1 )(p(x))]C = [T2 T1 ]C = A A 0 0 3 2 6 Therefore, (T2 T1 )(7 − x + 2x2 ) = 6 + 2x + 6x2 . Of course (T2 T1 )(7 − x + 2x2 ) = T2 ((x + 1)(7 − x + 2x2 )) = T2 (7 + 6x + x2 + 2x3 ) = 6 + 2x + 6x2 by direct calculation as well. (d). YES. Since the matrix [T2 T1 ]C A computed in part (a) is invertible, T2 T1 is invertible. 19. NO. The matrices [T ]C B obtained in Problem 1 are not invertible (they contain rows of zeros), and therefore, the corresponding linear transformation T is not invertible. 20. YES. We can explain this answer by using a matrix representation of T . Let B = {1, x, x2 } and let C = {(1, 0, 0), (0, 1, 0), (0, 0, 1)}. Then T (1) = (1, 1, 1), T (x) = (0, 1, 2), and T (x2 ) = (0, 1, 4), and so ⎡ ⎤ 1 0 0 ⎣ 1 1 1 ⎦. [T ]C B = 1 2 4 Since this matrix is invertible, the corresponding linear transformation T is invertible. 21. Note that w belongs to Rng(T ) ⇐⇒ w = T (v) for some v in V ⇐⇒ [w]C = [T (v)]C for some v in V ⇐⇒ [w]C = [T ]C B [v]B for some v in V . The right-hand side of this last expression can be expressed as a linear combination of the columns of [T ]C B, and therefore, w belongs to Rng(T ) if and only if [w]C can be expressed as a linear combination of the C columns of [T ]C B . That is, if and only if [w]C belongs to colspace([T ]B ). Solutions to Section 6.6 Additional Problems: 1. NO. Note that T (1, 1) = (2, 0, 0, 1) and T (2, 2) = (4, 0, 0, 4) = 2T (1, 1). Thus, T is not a linear transformation. 2. YES. The function T can be represented by the matrix function T (x) = Ax, where A= 2 −3 0 −1 0 0 (c)2017 Pearson Education. Inc. 434 and x is a vector in R3 . Every matrix transformation of the form T (x) = Ax is linear. Since the domain of T has larger dimension than the codomain of T , T cannot be one-to-one. However, since T (0, −1/3, 0) = (1, 0) and T (−1, −2/3, 0) = (0, 1), we see that T is onto. Thus, Rng(T ) = R2 (2-dimensional), and so a basis for Rng(T ) is {(1, 0), (0, 1)}. The kernel of T consists of vectors of the form (0, 0, z), and hence, a basis for Ker(T ) is {(0, 0, 1)} and Ker(T ) is 1-dimensional. 3. YES. The function T can be represented by the matrix function T (x) = Ax, where A= 0 −3 −1 5 0 2 and x is a vector in R3 . Every matrix transformation of the form T (x) = Ax is linear. Since the domain of T has larger dimension than the codomain of T , T cannot be one-to-one. However, since T (1, 1/3, −1/3) = (1, 0) and T (1, 1, 0) = (0, 1), we see that T is onto. Thus, Rng(T ) = R2 , and so a basis for Rng(T ) is {(1, 0), (0, 1)}, and Rng(T ) is 2-dimensional. The kernel of T consists of vectors of the form (t, 2t, 0), where t ∈ R, and hence, a basis for Ker(T ) is {(1, 2, 0)}. We have that Ker(T ) is 1-dimensional. 4. YES. The function T is a linear transformation, because if g, h ∈ C[0, 1], then T (g + h) = ((g + h)(0), (g + h)(1)) = (g(0) + h(0), g(1) + h(1)) = (g(0), g(1)) + (h(0), h(1)) = T (g) + T (h), and if c is a scalar, T (cg) = ((cg)(0), (cg)(1)) = (cg(0), cg(1)) = c(g(0), g(1)) = cT (g). Note that any function g ∈ C[0, 1] for which g(0) = g(1) = 0 (such as g(x) = x2 − x) belongs to Ker(T ), and hence, T is not one-to-one. However, given (a, b) ∈ R2 , note that g defined by g(x) = a + (b − a)x satisfies T (g) = (g(0), g(1)) = (a, b), so T is onto. Thus, Rng(T ) = R2 , with basis {(1, 0), (0, 1)} (2-dimensional). Now, Ker(T ) is infinitedimensional. We cannot list a basis for this subspace of C[0, 1]. 5. YES. The function T can be represented by the matrix function T (x) = Ax, where A= 1/5 1/5 and x is a vector in R2 . Every such matrix transformation is linear. Since the domain of T has larger dimension than the codomain of T , T cannot be one-to-one. However, T (5, 0) = 1, so we see that T is onto. (c)2017 Pearson Education. Inc. 435 Thus, Rng(T ) = R, a 1-dimensional space with basis {1}. The kernel of T consists of vectors of the form t(1, −1), and hence, a basis for Ker(T ) is {(1, −1)}. We have that Ker(T ) is 1-dimensional. 6. NO. For instance, note that 1 T 1 and so T However, 1 1 T 0 1 0 1 1 1 = (1, 0) +T 0 1 + and 1 0 1 1 1 0 1 1 T 1 0 1 1 = (0, 1), = (1, 0) + (0, 1) = (1, 1). =T 2 1 1 2 = (2, 2). Thus, T does not respect addition, and hence, T is not a linear transformation. Similar work could be given to show that T also fails to respect scalar multiplication. 7. YES. We can verify that T respects addition and scalar multiplication as follows: T respects addition: Let a1 + b1 x + c1 x2 and a2 + b2 x + c2 x2 belong to P2 (R). Then T ((a1 + b1 x + c1 x2 ) + (a2 + b2 x + c2 x2 )) = T ((a1 + a2 ) + (b1 + b2 )x + (c1 + c2 )x2 ) 0 −(a1 + a2 ) − (b1 + b2 ) = 3(c1 + c2 ) − (a1 + a2 ) −2(b1 + b2 ) 0 0 −a2 − b2 −a1 − b1 + = 3c1 − a1 −2b1 3c2 − a2 −2b2 = T (a1 + b1 x + c1 x2 ) + T (a2 + b2 x + c2 x2 ). T respects scalar multiplication: Let a + bx + cx2 belong to P2 (R) and let k be a scalar. Then we have −ka − kb 0 T (k(a + bx + cx2 )) = T ((ka) + (kb)x + (kc)x2 ) = 3(kc) − (ka) −2(kb) k(−a − b) 0 = k(3c − a) k(−2b) −a − b 0 =k 3c − a −2b = kT (a + bx + cx2 ). Next, observe that a + bx + cx2 belongs to Ker(T ) if and only if −a − b = 0, 3c − a = 0, and −2b = 0. These equations require that b = 0, a = 0, and c = 0. Thus, Ker(T ) = {0}, which implies that Ker(T ) is 0-dimensional (with basis ∅), and that T is one-to-one. However, since M2 (R) is 4-dimensional and P2 (R) is only 3-dimensional, we see immediately that T cannot be onto. By the Rank-Nullity Theorem, in fact, Rng(T ) must be 3-dimensional, and a basis is given by −1 0 −1 0 0 0 Basis for Rng(T ) = {T (1), T (x), T (x2 )} = , , . −1 0 0 −2 3 0 8. YES. We can verify that T respects addition and scalar multiplication as follows: (c)2017 Pearson Education. Inc. 436 T respects addition: Let A, B belong to M2 (R). Then T (A + B) = (A + B) + (A + B)T = A + B + AT + B T = (A + AT ) + (B + B T ) = T (A) + T (B). T respects scalar multiplication: Let A belong to M2 (R), and let k be a scalar. Then T (kA) = (kA) + (kA)T = kA + kAT = k(A + AT ) = kT (A). Thus, T is a linear transformation. Note that if A is any skew-symmetric matrix, then T (A) = A + AT = A + (−A) = 0, so Ker(T ) consists precisely of the 2 × 2 skew-symmetric matrices. matrices take the These 0 −a 0 −1 form , for a constant a, and thus a basis for Ker(T ) is given by , and Ker(T ) is a 0 1 0 1-dimensional. Consequently, T is not one-to-one. Therefore, by Proposition 6.4.14, T also fails to be onto. In fact, by the Rank-Nullity Theorem, Rng(T ) must be 3-dimensional. A typical element of the range of T takes the form a b a b a c 2a b+c T = + = . c d c d b d c+b 2d The characterizing feature of this matrix is that it is symmetric. So Rng(T ) consists of all 2 × 2 symmetric matrices, and hence a basis for Rng(T ) is 1 0 0 1 0 0 , , . 0 0 1 0 0 1 9. YES. We can verify that T respects addition and scalar multiplication as follows: T respects addition: Let (a1 , b1 , c1 ) and (a2 , b2 , c2 ) belong to R3 . Then T ((a1 , b1 , c1 ) + (a2 , b2 , c2 )) = T (a1 + a2 , b1 + b2 , c1 + c2 ) = (a1 + a2 )x2 + (2(b1 + b2 ) − (c1 + c2 ))x + (a1 + a2 − 2(b1 + b2 ) + (c1 + c2 )) = [a1 x2 + (2b1 − c1 )x + (a1 − 2b1 + c1 )] + [a2 x2 + (2b2 − c2 )x + (a2 − 2b2 + c2 )] = T ((a1 , b1 , c1 )) + T ((a2 , b2 , c2 )). T respects scalar multiplication: Let (a, b, c) belong to R3 and let k be a scalar. Then T (k(a, b, c)) = T (ka, kb, kc) = (ka)x2 + (2kb − kc)x + (ka − 2kb + kc) = k(ax2 + (2b − c)x + (a − 2b + c)) = kT ((a, b, c)). Thus, T is a linear transformation. Now, (a, b, c) belongs to Ker(T ) if and only if a = 0, 2b − c = 0, and a − 2b + c = 0. These equations collectively require that a = 0 and 2b = c. Setting c = 2t, we find that b = t. Hence, (a, b, c) belongs to Ker(T ) if and only if (a, b, c) has the form (0, t, 2t) = t(0, 1, 2). Hence, {(0, 1, 2)} is a basis for Ker(T ), which is therefore 1-dimensional. Hence, T is not one-to-one. By Proposition 6.4.14, T is also not onto. In fact, the Rank-Nullity Theorem implies that Rng(T ) must be 2-dimensional. It is spanned by {T (1, 0, 0), T (0, 1, 0), T (0, 0, 1)} = {x2 + 1, 2x − 2, −x + 1}, (c)2017 Pearson Education. Inc. 437 but the last two polynomials are proportional to each other. Omitting the polynomial 2x − 2 (this is an arbitrary choice; we could have omitted −x + 1 instead), we arrive at a basis for Rng(T ): {x2 + 1, −x + 1}. 10. YES. We can verify that T respects addition and scalar multiplication as follows: T respects addition: Let (x1 , x2 , x3 ) and (y1 , y2 , y3 ) be vectors in R3 . Then T ((x1 , x2 , x3 ) + (y1 , y2 , y3 )) = T (x1 + y1 , x2 + y2 , x3 + y3 ) 0 (x1 + y1 ) − (x2 + y2 ) + (x3 + y3 ) = −(x1 + y1 ) + (x2 + y2 ) − (x3 + y3 ) 0 0 y1 − y2 + y3 0 x1 − x2 + x3 + = −x1 + x2 − x3 0 −y1 + y2 − y3 0 = T ((x1 , x2 , x3 )) + T ((y1 , y2 , y3 )). T respects scalar multiplication: Let (x1 , x2 , x3 ) belong to R3 and let k be a scalar. Then 0 (kx1 ) − (kx2 ) + (kx3 ) T (k(x1 , x2 , x3 )) = T (kx1 , kx2 , kx3 ) = −(kx1 ) + (kx2 ) − (kx3 ) 0 0 x1 − x2 + x3 =k −x1 + x2 − x3 0 = kT ((x1 , x2 , x3 )). Thus, T is a linear transformation. Now, (x1 , x2 , x3 ) belongs to Ker(T ) if and only if x1 − x2 + x3 = 0 and −x1 + x2 − x3 = 0. Of course, the latter equation is equivalent to the former, so the kernel of T consists simply of ordered triples (x1 , x2 , x3 ) with x1 − x2 + x3 = 0. Setting x3 = t and x2 = s, we have x1 = s − t, so a typical element of Ker(T ) takes the form (s − t, s, t), where s, t ∈ R. Extracting the free variables, we find a basis for Ker(T ): {(1, 1, 0), (−1, 0, 1)}. Hence, Ker(T ) is 2-dimensional. By the Rank-Nullity Theorem, Rng(T ) must be 1-dimensional. In fact, Rng(T ) consists precisely of the 0 −1 set of 2 × 2 skew-symmetric matrices, with basis . Since M2 (R) is 4-dimensional, T fails to be 1 0 onto. 11. We have T (x, y, z) = (−x + 8y, 2x − 2y − 5z). 12. We have T (x, y) = (−x + 4y, 2y, 3x − 3y, 3x − 3y, 2x − 6y). x x x 5x T (2) = (−1, 5, 0, −2) = − , , 0, −x . 2 2 2 2 a b 14. For an arbitrary 2 × 2 matrix , if we write c d a b 1 0 0 1 1 0 1 1 =r +s +t +u , c d 0 1 1 0 0 0 0 0 13. We have T (x) = we can solve for r, s, t, u to find r = d, s = c, t = a − b + c − d, (c)2017 Pearson Education. Inc. u = b − c. 438 Thus, T a c b d 1 0 0 1 1 0 1 1 + (a − b + c − d) + (b − c) =T d +c 0 1 1 0 0 0 0 0 1 0 0 1 1 0 1 1 = dT + cT + (a − b + c − d)T + (b − c)T 0 1 1 0 0 0 0 0 = d(2, −5) + c(0, −3) + (a − b + c − d)(1, 1) + (b − c)(−6, 2) = (a − 7b + 7c + d, a + b − 4c − 6d). 15. For an arbitrary element ax2 + bx + c in P2 (R), if we write ax2 + bx + c = r(x2 − x − 3) + s(2x + 5) + 6t = rx2 + (−r + 2s)x + (−3r + 5s + 6t), 1 5 a − 12 b + 16 c. Thus, we can solve for r, s, t to find r = a, s = 12 (a + b), and t = 12 1 1 5 2 2 T (ax + bx + c) = T a(x − x − 3) + (a + b)(2x + 5) + a − b + c 2 2 2 1 1 5 c 2 T (6) a− b+ = aT (x − x − 3) + (a + b)T (2x + 5) + 2 12 12 6 1 1 5 c 12 6 −2 1 0 1 =a + (a + b) + a− b+ 6 18 −4 −1 2 −2 2 12 12 6 −a − 5b + 2c 2a − 2b + c = . − 52 a − 32 b + c − 12 a − 17 2 b + 3c 16. Since dim[P5 (R)] = 6 and dim[M2 (R)] = 4 = dim[Rng(T )] (since T is onto), the Rank-Nullity Theorem gives dim[Ker(T )] = 6 − 4 = 2. 17. Since T is one-to-one, dim[Ker(T )] = 0, so the Rank-Nullity Theorem gives dim[Rng(T )] = dim[M2×3 (R)] = 6. 18. We must have n = 8, since dim[M42 (R)] = 8. 19. We know that dim[Ker(T )] + dim[Rng(T )] = dim[V ]. In this case, dim[V ] = 5. Since T is not onto, we must have dim[Rng(T )] < 4. Therefore, dim[Rng(T )] ∈ {0, 1, 2, 3}, from which it follows that dim[Ker(T )] ∈ {5, 4, 3, 2}. 20. Note that T (1) = (1, 1) = 1 · (1, 0) + 1 · (0, 1), T (x) = (0, 1) = 0 · (1, 0) + 1 · (0, 1), T (x2 ) = (0, 1) = 0 · (1, 0) + 1 · (0, 1), T (x3 ) = (0, 1) = 0 · (1, 0) + 1 · (0, 1). Thus, [T ]C B = 1 1 0 1 0 1 0 1 . (c)2017 Pearson Education. Inc. 439 ⎤ 2 ⎢ 0 ⎥ ⎥ Since [v]B = ⎢ ⎣ −1 ⎦, by Theorem 6.5.4, we have 2 ⎡ [T (v)]C = 1 1 0 1 0 1 0 1 ⎤ 2 ⎢ 0 ⎥ 2 ⎥= ⎢ . ⎣ −1 ⎦ −1 2 ⎡ Therefore, T (v) = 2 · (1, 0) + (−1) · (0, 1) = (2, −1). 21. Note that T (0, 1, 0) = T (0, 0, 1) = T (1, 0, 0) = Thus, 0 0 1 −3 0 −1 3 0 0 −1 5 0 = 0 · E21 + (−3) · E22 + 0 · E11 + 1 · E12 , = (−1) · E21 + 0 · E22 + 0 · E11 + 3 · E12 , = 5 · E21 + 0 · E22 + 0 · E11 + (−1) · E12 . ⎤ 0 −1 5 ⎢ −3 0 0 ⎥ ⎥. ⎢ [T ]C B =⎣ 0 0 0 ⎦ 1 3 −1 ⎡ ⎡ ⎤ 1 Since [v]B = ⎣ −2 ⎦, by Theorem 6.5.4, we have −2 ⎤ ⎤ ⎡ ⎤ −8 0 −1 5 ⎡ 1 ⎥ ⎢ ⎢ −3 0 0 ⎥ ⎥ ⎣ −2 ⎦ = ⎢ −3 ⎥ . [T (v)]C = ⎢ ⎣ 0 ⎦ ⎣ 0 0 0 ⎦ −2 −3 1 3 −1 ⎡ Thus, T (v) = −8 · E21 + (−3) · E22 + 0 · E11 + (−3) · E12 = 0 −3 −8 −3 . 22. Note that ⎤ 4 ⎢ 0 ⎥ ⎥ T (1, 1) = ⎢ ⎣ −2 ⎦ = 7 · (0, 0, 0, 1) + (−2) · (0, 0, 1, 0) + 0 · (0, 1, 0, 0) + 4 · (1, 0, 0, 0), 7 ⎡ ⎤ 1 ⎢ −1 ⎥ ⎥ T (1, 0) = ⎢ ⎣ −2 ⎦ = 5 · (0, 0, 0, 1) + (−2) · (0, 0, 1, 0) + (−1) · (0, 1, 0, 0) + 1 · (1, 0, 0, 0). 5 ⎡ (c)2017 Pearson Education. Inc. 440 Thus, ⎤ 7 5 ⎢ −2 −2 ⎥ ⎥. ⎢ [T ]C B =⎣ 0 −1 ⎦ 4 1 ⎡ Next we compute [v]B by writing (−2, 4) = c1 (1, 1) + c2 (1, 0). A short calculation shows that c1 = 4 and 4 , by Theorem 6.5.4, we have c2 = −6. Thus, since [v]B = −6 ⎡ ⎤ ⎤ ⎡ −2 7 5 ⎢ 4 ⎥ ⎢ −2 −2 ⎥ 4 ⎥ ⎢ ⎥ [T (v)]C = ⎢ ⎣ 0 −1 ⎦ −6 = ⎣ 6 ⎦ . 10 4 1 Thus, T (v) = −2 · (0, 0, 0, 1) + 4 · (0, 0, 1, 0) + 6 · (0, 1, 0, 0) + 10 · (1, 0, 0, 0) = (10, 6, 4, −2). 23. Note that T (x2 + x) = 2x + 1 = (−1) · 1 + 2 · (1 + x), T (1) = 0 = 0 · 1 + 0 · (1 + x), T (x) = 1 = 1 · 1 + 0 · 1 + x. Thus, [T ]C B = −1 2 0 0 1 0 . 2 Next we compute [v]B by writing −3 + x2 = c1 · (x ⎡ + x) ⎤ + c2 · 1 + c3 · x. A short calculation shows that 1 c1 = 1, c2 = −3, and c3 = −1. Thus, since [v]B = ⎣ −3 ⎦, by Theorem 6.5.4, we have −1 ⎡ ⎤ 1 −1 0 1 ⎣ −2 −3 ⎦ = [T (v)]C = . 2 0 0 2 −1 Thus, T (v) = −2 · 1 + 2 · (1 + x) = 2x. 24. To see that T1 + T2 is a linear transformation, we must verify that it respects addition and scalar multiplication: T1 + T2 respects addition: Let v1 and v2 belong to V . Then we have (T1 + T2 )(v1 + v2 ) = T1 (v1 + v2 ) + T2 (v1 + v2 ) = [T1 (v1 ) + T1 (v2 )] + [T2 (v1 ) + T2 (v2 )] = [T1 (v1 ) + T2 (v1 )] + [T1 (v2 ) + T2 (v2 )] = (T1 + T2 )(v1 ) + (T1 + T2 )(v2 ), where we have used the linearity of T1 and T2 individually in the second step. (c)2017 Pearson Education. Inc. 441 T1 + T2 respects scalar multiplication: Let v belong to V and let k be a scalar. Then we have (T1 + T2 )(kv) = T1 (kv) + T2 (kv) = kT1 (v) + kT2 (v) = k[T1 (v) + T2 (v)] = k(T1 + T2 )(v), as required. There is no particular relationship between Ker(T1 ), Ker(T2 ), and Ker(T1 + T2 ). 25. FALSE. For instance, consider T1 : R → R defined by T1 (x) = x, and consider T2 : R → R defined by T2 (x) = −x. Both T1 and T2 are linear transformations, and both of them are onto. However, (T1 + T2 )(x) = T1 (x) + T2 (x) = x + (−x) = 0, so Rng(T1 + T2 ) = {0}, which implies that T1 + T2 is not onto. 26. FALSE. For instance, consider T1 : R → R defined by T1 (x) = x, and consider T2 : R → R defined by T2 (x) = −x. Both T1 and T2 are linear transformations, and both of them are one-to-one. However, (T1 + T2 )(x) = T1 (x) + T2 (x) = x + (−x) = 0, so Ker(T1 + T2 ) = R, which implies that T1 + T2 is not one-to-one. 27. Assume that c1 T (v1 + c2 T (v2 ) + · · · + cn T (vn ) = 0. We wish to show that c1 = c2 = · · · = cn = 0. To do this, use the linearity of T to rewrite the above equation as T (c1 v1 + c2 v2 + · · · + cn vn ) = 0. Now, since Ker(T ) = {0}, we conclude that c1 v1 + c2 v2 + · · · + cn vn = 0. Since {v1 , v2 , . . . , vn } is a linearly independent set, we conclude that c1 = c2 = · · · = cn = 0, as required. 28. Assume that V1 ∼ = V2 and V2 ∼ = V3 . Then there exist isomorphisms T1 : V1 → V2 and T2 : V2 → V3 . Since the composition of two linear transformations is a linear transformation (Theorem 6.4.2), we have a linear transformation T2 T1 : V1 → V3 . Moreover, since both T1 and T2 are one-to-one and onto, T2 T1 is also one-to-one and onto (see Problem 38 in Section 6.4). Thus, T2 T1 : V1 → V3 is an isomorphism. Hence, V1 ∼ = V3 , as required. 29. We first show that T is a linear transformation: T respects addition: Let A and B belong to Mn (R). Then T (A + B) = S −1 (A + B)S = S −1 AS + S −1 BS = T (A) + T (B), and so T respects addition. T respects scalar multiplication: Let A belong to Mn (R), and let k be a scalar. Then T (kA) = S −1 (kA)S = k(S −1 AS) = kT (A), and so T respects scalar multiplication. (c)2017 Pearson Education. Inc. 442 Next, we verify that T is both one-to-one and onto (of course, in view of Proposition 6.4.14, it is only necessary to confirm one of these two properties, but we will nonetheless verify them both): T is one-to-one: Assume that T (A) = 0n . That is, S −1 AS = 0n . Left multiplying by S and right multiplying by S −1 on both sides of this equation yields A = S0n S −1 = 0n . Hence, Ker(T ) = {0n }, and so T is one-toone. T is onto: Let B be an arbitrary matrix in Mn (R). Then T (SBS −1 ) = S −1 (SBS −1 )S = (S −1 S)B(S −1 S) = In BIn = B, and hence, B belongs to Rng(T ). Since B was an arbitrary element of Mn (R), we conclude that Rng(T ) = Mn (R). That is, T is onto. Chapter 7 Solutions Solutions to Section 7.1 True-False Review: (a): FALSE. If v = 0, then Av = λv = 0, but by definition, an eigenvector must be a nonzero vector. (b): TRUE. When we compute det(A − λI) for an upper or lower triangular matrix A, the determinant is the product of the entries lying along the main diagonal of A − λI: det(A − λI) = (a11 − λ)(a22 − λ) . . . (ann − λ). The roots of this characteristic equation are precisely the values a11 , a22 , . . . , ann along the main diagonal of the matrix A. (c): TRUE. The eigenvalues of a matrix are precisely the set of roots of its characteristic equation. Therefore, two matrices A and B that have the same characteristic equation will have the same eigenvalues. (d): TRUE. The characteristic equation of an n × n matrix A, det(A − λI) is a polynomial of degree n i the indeterminate λ. Since such a polynomial always possesses n roots (with possible repeated or complex roots) by the Fundamental Theorem of Algebra, the statement is true. (e): TRUE. Geometrically, all nonzero points v = (x, y) in R2 are oriented in a different direction from the origin after a 90◦ rotation than they are initially. Therefore, the vectors v and Av are not parallel. 0 0 (f ): FALSE. Many examples of this can be found. As a simple one, consider A = and B = 0 0 0 1 . We have det(A − λI) = (−λ)2 = λ2 = det(B − λI). Note that every nonzero vector in R2 is an 0 0 a eigenvector of A corresponding to λ = 0. However, only vectors of the form with a = 0 are eigenvectors 0 of B. Therefore, A and B do not have precisely the same set of eigenvectors. In this case, every eigenvector of B is also an eigenvector of A, but not conversely. (g): FALSE. This is not true, in general, when the linear combination formed involves eigenvectors cor 1 0 responding to different eigenvalues. For example, let A = , with eigenvalues λ = 1 and λ = 2. 0 2 (c)2017 Pearson Education. Inc. 443 It is easy to see that corresponding eigenvectors to these eigenvalues are, respectively, v1 = 0 . However, note that v2 = 1 1 A(v1 + bf v 2 ) = , 2 1 0 and which is not of the form λ(v1 + v2 ), and therefore v1 + v2 is not an eigenvector of A. As a more trivial illustration, note that if v is an eigenvector of A, then 0v is a linear combination of {v} that is no longer an eigenvector of A. (h): TRUE. This is basically a fact about roots of polynomials. Complex roots of real polynomials always occur in complex conjugate pairs. Therefore, if λ = a + ib (b = 0) is an eigenvalue of A, then so is λ = a − ib. (i): TRUE. We are given that Av = λv for a nonzero eigenvector v. Then A3 v = A2 (Av) = A2 (λv) = λ(A2 v) = λA(Av) = λA(λv) = λ2 (Av) = λ2 (λv) = λ3 v, which shows that λ3 is an eigenvalue of A3 . Problems: 1. We have −7 −9 Av = as required. 1 2. Av = 2 ⎡ 3 2 1 1 = 4 4 =4 1 1 ⎡ 2 4 2 9 = 4 18 = 2v, = λv. ⎤⎡ ⎤ ⎤ ⎡ ⎤ 1 −2 −6 2 6 2 2 −5 ⎦ ⎣ 1 ⎦ = ⎣ 3 ⎦ = 3 ⎣ 1 ⎦ = λv. 3. Av = ⎣ −2 2 1 8 −1 −3 −1 ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 1 4 c1 + 4c2 4. Since v = c1 ⎣ 0 ⎦ + c2 ⎣ −3 ⎦ = ⎣ −3c2 ⎦, it follows that −3 0 −3c1 ⎤ ⎡ ⎤ ⎤ ⎡ ⎡ ⎤⎡ −2c1 − 8c2 c1 + 4c2 1 4 1 c1 + 4c2 ⎦ = −2 ⎣ −3c2 ⎦ = λv. 6c2 1 ⎦ ⎣ −3c2 ⎦ = ⎣ Av = ⎣ 3 2 −3c1 6c1 −3c1 3 4 −1 5. We have ⎤ ⎡ ⎤ ⎤ ⎡ ⎤⎡ 10c1 c1 6 −1 0 c1 6 0 ⎦ ⎣ −4c1 ⎦ = ⎣ −40c1 ⎦ = 10 ⎣ −4c1 ⎦ = 10v, Av = ⎣ −16 c2 10c2 c2 −4 −1 10 ⎡ as required. 6. We have Av1 = −5 1 2 −4 −2 1 = 12 −6 = −6 −2 1 = −6v1 , which shows that λ1 = −6 is an eigenvalue of A. Furthermore, −5 2 1 −3 1 = = −3 = −3v2 , Av2 = 1 −4 1 −3 1 (c)2017 Pearson Education. Inc. 444 which shows that λ2 = −3 is an eigenvalue of A. Thus, the eigenvalues of A are λ1 = −6 and λ2 = −3. 7. We have Av1 = 4 2 1 3 1 −2 = 2 −4 =2 1 −2 = 2v1 , which shows that λ1 = 2 is an eigenvalue of A. Furthermore, 4 1 1 5 1 Av2 = = =5 = 5v2 , 2 3 1 5 1 which shows that λ2 = 5 is an eigenvalue of A. Thus, the eigenvalues of A are λ1 = 2 and λ2 = 5. 8. The only vectors that are mapped into a scalar multiple of themselves under a reflection in the x-axis are those vectors that either point along the x-axis, or that point along the y-axis. Hence, the eigenvectors are of the form (a, 0) or (0, b) where a and b are arbitrary nonzero real numbers. A vector that points along the x-axis will have neither its magnitude nor its direction altered by a reflection in the x-axis. Hence, the eigenvectors of the form (a, 0) correspond to the eigenvalue λ = 1. A vector of the form (0, b) will be mapped into the vector (0, −b) = −1(0, b) under a reflection in the x-axis. Consequently, the eigenvectors of the form (0, b) correspond to the eigenvalue λ = −1. 9. Any vectors lying along the line y = x are unmoved by the action of T , and hence, any vector (t, t) with t = 0 is an eigenvector with corresponding eigenvalue λ = 1. On the other hand, any vector lying along the line y = −x will be reflected across the line y = x, thereby experiencing a 180◦ change of direction. Therefore, any vector (t, −t) with t = 0 is an eigenvector with corresponding eigenvalue λ = −1. All vectors that do not lie on the line y = x or the line y = −x are not eigenvectors of this linear transformation. 10. If θ = 0, π, there are no vectors that are mapped into scalar multiples of themselves under the rotation, and consequently, there are no real eigenvalues and eigenvectors in this case. If θ = 0, then every vector is mapped onto itself under the rotation, therefore λ = 1, and every nonzero vector in R2 is an eigenvector. If θ = π, then every vector is mapped onto its negative under the rotation, therefore λ = −1, and once again, every nonzero vector in R2 is an eigenvector. 11. Any vectors lying on the y-axis are unmoved by the action of T , and hence, any vector (0, y, 0) with y not zero is an eigenvector of T with corresponding eigenvalue λ = 1. On the other hand, any vector lying in the xz-plane, say (x, 0, z) is transformed under T to (0, 0, 0). Thus, any vector (x, 0, z) with x and z not both zero is an eigenvector with corresponding eigenvalue λ = 0. 5−λ −4 = 0 ⇐⇒ λ2 + 2λ − 3 = 0 ⇐⇒ (λ + 3)(λ − 1) = 0 12. det(A − λI) = 0 ⇐⇒ 8 −7 − λ ⇐⇒ λ = −3 or λ = 1. 8 −4 v1 0 = If λ = −3 then (A − λI)v = 0 assumes the form v2 0 8 −4 1 =⇒ 8v1 − 4v2 = 0 =⇒ v1 = 2 v2 . If we let v2 = r ∈ R, then the solution set of this equation is {( 12 r, r) : r ∈ R}, so the eigenvectors corresponding to λ = −3 are v = r( 12 , 1), where r ∈ R (and r = 0). 0 4 −4 v1 If λ = 1 then (A − λI)v = 0 assumes the form = v2 0 8 −8 =⇒ v1 = v2 . If we let v2 = t ∈ R, then v1 = t and the solution set of this system is {(t, t) : t ∈ R}, so the eigenvectors corresponding to λ = 1 are v = t(1, 1), where t ∈ R (and t = 0). 1−λ 6 = 0 ⇐⇒ λ2 + 2λ − 15 = 0 13. det(A − λI) = 0 ⇐⇒ 2 −3 − λ ⇐⇒ (λ − 3)(λ + 5) = 0 ⇐⇒ λ = 3 or λ = −5. (c)2017 Pearson Education. Inc. 445 0 −2 6 v1 = 0 v2 2 −6 =⇒ v1 − 3v2 = 0. If we let v2 = r ∈ R, then the solution set of this system is {(3r, r) : r ∈ R} so the eigenvectors corresponding to λ = 3 are v = r(3, 1) where r ∈ R. 0 6 6 v1 = If λ = −5 then (A − λI)v = 0 assumes the form 0 v2 2 2 =⇒ v1 + v2 = 0. If we let v2 = s ∈ R, then the solution set of this system is {(−s, s) : s ∈ R} so the eigenvectors corresponding to λ = −5 are v = s(−1, 1) where s ∈ R. 7−λ 4 14. det(A − λI) = 0 ⇐⇒ = 0 ⇐⇒ λ2 − 10λ + 25 = 0 −1 3−λ ⇐⇒ (λ − 5)2 = 0 ⇐⇒ λ = 5 of multiplicity two. 0 2 4 v1 = If λ = 5 then (A − λI)v = 0 assumes the form v2 0 −1 −2 =⇒ v1 + 2v2 = 0 =⇒ v1 = −2v2 . If we let v2 = t ∈ R, then the solution set of this system is {(−2t, t) : t ∈ R} so the eigenvectors corresponding to λ = 5 are v = t(−2, 1) where t ∈ R. 2−λ 0 = 0 ⇐⇒ (2 − λ)2 = 0 ⇐⇒ λ = 2 of multiplicity two. 15. det(A − λI) = 0 ⇐⇒ 0 2−λ 0 0 0 v1 If λ = 2 then (A − λI)v = 0 assumes the form = 0 v2 0 0 Thus, if we let v1 = s and v2 = t where s, t ∈ R, then the solution set of this system is {(s, t) : s, t ∈ R} so the eigenvectors corresponding to λ = 2 are v = s(1, 0) + t(0, 1) where s, t ∈ R. 7−λ 3 = 0 ⇐⇒ λ2 − 8λ + 25 = 0, from which the quadratic formula 16. det(A − λI) = 0 ⇐⇒ −6 1−λ supplies the eigenvalues: λ = 4 ± 3i. 0 3 − 3i 3 v1 = . The given If λ = 4 + 3i then (A − λI)v = 0 assumes the form 0 v2 −6 −3 − 3i system of equations reduces to (3−3i)v1 +3v2 = 0, or after dividing by 3, (1−i)v1 +v2 = 0. Setting v1 = t ∈ R, we have v2 = −t(1 − i), so the eigenvectors corresponding to λ = 4 + 3i are v = (t, −t(1 − i)) = t(1, −(1 − i)), where t ∈ R, with t = 0. If λ = 4 − 3i then we can apply Theorem 7.1.8 and use our calculations above to conclude that the corresponding eigenvectors in this case have the form v = t(1, −(1 + i)), where t ∈ R, with t = 0. −2 − λ −6 = 0 ⇐⇒ λ2 − 2λ + 10 = 0, from which the quadratic formula 17. det(A − λI) = 0 ⇐⇒ 3 4−λ supplies the eigenvalues: λ = 1 ± 3i. 0 −3 − 3i −6 v1 = . The given If λ = 1 + 3i then (A − λI)v = 0 assumes the form v2 0 3 3 − 3i system of equations reduces to 3v1 +(3−3i)v2 = 0, or after dividing by 3, v1 +(1−i)v2 = 0. Setting v2 = t ∈ R, we have v1 = −t(1 − i), so the eigenvectors corresponding to λ = 1 + 3i are v = (−t(1 − i), t) = t(−(1 − i), 1), where t ∈ R, with t = 0. If λ = 1 − 3i then we can apply Theorem 7.1.8 and use our calculations above to conclude that the corresponding eigenvectors in this case have the form v = t(−(1 + i), 1), where t ∈ R, with t = 0. 3−λ −2 = 0 ⇐⇒ λ2 − 2λ + 5 = 0 ⇐⇒ λ = 1 ± 2i. 18. det(A − λI) = 0 ⇐⇒ 4 −1 − λ 0 2 + 2i −2 v1 If λ = 1 − 2i then (A − λI)v = 0 assumes the form = v2 0 4 −2 + 2i =⇒ (1 + i)v1 − v2 = 0. If we let v1 = s ∈ C, then the solution set of this system is {(s, (1 + i)s) : s ∈ C} so If λ = 3 then (A − λI)v = 0 assumes the form (c)2017 Pearson Education. Inc. 446 the eigenvectors corresponding to λ = 1 − 2i are v = s(1, 1 + i) where s ∈ C. By Theorem 7.1.8, since the entries of A are real, λ = 1 + 2i has corresponding eigenvectors of the form v = t(1, 1 − i) where t ∈ C. 2−λ 3 = 0 ⇐⇒ (2 − λ)2 = −9 ⇐⇒ λ = 2 ± 3i. 19. det(A − λI) = 0 ⇐⇒ −3 2−λ 0 3i 3 v1 = If λ = 2 − 3i then (A − λI)v = 0 assumes the form v2 0 −3 3i =⇒ v2 = −iv1 . If we let v1 = t ∈ C, then the solution set of this system is {(t, −it) : t ∈ C} so the eigenvectors corresponding to λ = 2 − 3i are v = t(1, −i) where t ∈ C. By Theorem 7.1.8, since the entries of A are real, λ = 2 + 3i has corresponding eigenvectors of the form v = r(1, i) where r ∈ C. 10 − λ −12 8 = 0 ⇐⇒ (λ − 2)3 = 0 ⇐⇒ λ = 2 of multiplicity three. 0 2−λ 0 20. det(A − λI) = 0 ⇐⇒ −8 12 −6 − λ ⎤ ⎡ ⎡ ⎤⎡ ⎤ 8 −12 8 v1 0 0 0 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ If λ = 2 then (A − λI)v = 0 assumes the form ⎣ 0 −8 12 −8 v3 0 =⇒ 2v1 − 3v2 + 2v3 = 0. Thus, if we let v2 = 2s and v3 = t where s, t ∈ R, then the solution set of this system is {(3s − t, 2s, t) : s, t ∈ R} so the eigenvectors corresponding to λ = 2 are v = s(3, 2, 0) + t(−1, 0, 1) where s, t ∈ R. 3−λ 0 0 2−λ −1 = 0 ⇐⇒ (λ − 1)(λ − 3)2 = 0 ⇐⇒ λ = 1 or λ = 3 of 21. det(A − λI) = 0 ⇐⇒ 0 1 −1 2−λ multiplicity two. ⎤ ⎡ ⎡ ⎤⎡ ⎤ 2 0 0 v1 0 1 −1 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ If λ = 1 then (A − λI)v = 0 assumes the form ⎣ 0 1 −1 1 v3 0 =⇒ v1 = 0 and v2 − v3 = 0. Thus, if we let v3 = s where s ∈ R, then the solution set of this system is {(0, s, s) : s ∈ R} so the eigenvectors corresponding v = s(0, 1) where ⎤ 1,⎡ ⎤ s ∈ R. ⎡ to λ = 1 are ⎤⎡ 0 0 0 v1 0 If λ = 3 then (A − λI)v = 0 assumes the form ⎣ 0 −1 −1 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ 0 1 −1 −1 v3 =⇒ v1 = 0 and v2 + v3 = 0. Thus, if we let v3 = t where t ∈ R, then the solution set of this system is {(0, −t, t) : t ∈ R} so the eigenvectors corresponding to λ = 3 are v = t(0, −1, 1) where t ∈ R. 1−λ 0 0 = 0 ⇐⇒ (λ − 1)3 = 0 ⇐⇒ λ = 1 of multiplicity three. 3−λ 2 22. det(A − λI) = 0 ⇐⇒ 0 2 −2 −1 − λ ⎤⎡ ⎤ ⎡ ⎤ ⎡ v1 0 0 0 0 2 2 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ If λ = 1 then (A − λI)v = 0 assumes the form ⎣ 0 2 −2 −2 0 v3 =⇒ v1 = 0 and v2 + v3 = 0. Thus, if we let v3 = s where s ∈ R, then the solution set of this system is {(0, −s, s) : s ∈ R} so the eigenvectors corresponding to λ = 1 are v = s(0, −1, 1) where s ∈ R. 6−λ 3 −4 = 0 ⇐⇒ (λ − 1)(λ + 1)(λ − 3) = 0 ⇐⇒ λ = −1, λ = 1, −2 − λ 2 23. det(A − λI) = 0 ⇐⇒ −5 0 0 −1 − λ or λ = 3. (c)2017 Pearson Education. Inc. 447 ⎤ ⎡ ⎤⎡ ⎤ 7 3 −4 v1 0 2 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ If λ = −1 then (A − λI)v = 0 assumes the form ⎣ −5 −1 0 0 0 v3 0 =⇒ v1 = v3 − v2 and 4v2 − 3v3 = 0. Thus, if we let v3 = 4r where r ∈ R, then the solution set of this system is {(r, 3r, 4r) : r ∈ R} so the eigenvectors corresponding to λ = −1 ⎡ ⎤ ⎡are v ⎤= r(1, ⎡ 3, 4) ⎤ where r ∈ R. 5 3 −4 v1 0 2 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ If λ = 1 then (A − λI)v = 0 assumes the form ⎣ −5 −3 0 0 −2 v3 0 =⇒ 5v1 + 3v2 = 0 and v3 = 0. Thus, if we let v2 = −5s where s ∈ R, then the solution set of this system is {(3s, −5s, 0) : s ∈ R} so the eigenvectors corresponding to λ = 1 ⎤are ⎡ ⎡ v =⎤s(3,⎡−5, 0) ⎤ where s ∈ R. 3 3 −4 v1 0 2 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ If λ = 3 then (A − λI)v = 0 assumes the form ⎣ −5 −5 0 0 −4 v3 0 =⇒ v1 + v2 = 0 and v3 = 0. Thus, if we let v2 = t where t ∈ R, then the solution set of this system is {(−t, t, 0) : t ∈ R} so the eigenvectors corresponding to λ = 3 are v = t(−1, 1, 0) where t ∈ R. 7−λ −8 6 = 0 ⇐⇒ (λ + 1)3 = 0 ⇐⇒ λ = −1 of multiplicity −9 − λ 6 24. det(A − λI) = 0 ⇐⇒ 8 0 0 −1 − λ three. ⎤ ⎡ ⎡ ⎤⎡ ⎤ 8 −8 6 v1 0 If λ = −1 then (A − λI)v = 0 assumes the form ⎣ 8 −8 6 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ 0 0 0 v3 0 =⇒ 4v1 − 4v2 + 3v3 = 0. Thus, if we let v2 = r and v3 = 4s where r, s ∈ R, then the solution set of this system is {(r −3s, r, 4s) : r, s ∈ R} so the eigenvectors corresponding to λ = −1 are v = r(1, 1, 0)+s(−3, 0, 4) where r, s ∈ R. −λ 1 −1 0 = 0 ⇐⇒ (λ − 1)(λ − 2)2 = 0 ⇐⇒ λ = 1 or λ = 2 of 25. det(A − λI) = 0 ⇐⇒ 0 2 − λ 2 −1 3−λ multiplicity two. ⎤ ⎡ ⎡ ⎤⎡ ⎤ −1 1 −1 v1 0 1 0 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ If λ = 1 then (A − λI)v = 0 assumes the form ⎣ 0 2 −1 2 v3 0 =⇒ v2 = 0 and v1 + v3 = 0. Thus, if we let v3 = r where r ∈ R, then the solution set of this system is {(−r, 0, r) : r ∈ R} so the eigenvectors corresponding to λ = 1 are v = r(−1, r ∈ R. ⎤ 0, ⎡ ⎤⎡ ⎡ 1) where ⎤ −2 1 −1 v1 0 0 0 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ If λ = 2 then (A − λI)v = 0 assumes the form ⎣ 0 2 −1 1 v3 0 =⇒ 2v1 − v2 + v3 = 0. Thus, if we let v1 = s and v3 = t where s, t ∈ R, then the solution set of this system is {(s, 2s + t, t) : s, t ∈ R} so the eigenvectors corresponding to λ = 2 are v = s(1, 2, 0) + t(0, 1, 1) where s, t ∈ R. 1−λ 0 0 −λ 1 = 0 ⇐⇒ (1 − λ)(1 + λ2 ) = 0 ⇐⇒ λ = 1 or λ = ±i. 26. det(A − λI) = 0 ⇐⇒ 0 0 −1 −λ ⎤ ⎡ ⎡ ⎤⎡ ⎤ 0 0 0 v1 0 1 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ If λ = 1 then (A − λI)v = 0 assumes the form ⎣ 0 −1 0 −1 −1 v3 0 =⇒ −v2 + v3 = 0 and −v2 − v3 = 0. The solution set of this system is {(r, 0, 0) : r ∈ C} so the eigenvectors ⎡ (c)2017 Pearson Education. Inc. 448 corresponding to λ = 1 are v = r(1, 0, 0) where r⎡∈ C. ⎤ ⎡ ⎤ ⎤⎡ 1+i 0 0 v1 0 i 1 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ If λ = −i then (A − λI)v = 0 assumes the form ⎣ 0 0 0 −1 i v3 =⇒ v1 = 0 and −v2 + iv3 = 0. The solution set of this system is {(0, si, s) : s ∈ C} so the eigenvectors corresponding to λ = −i are v = s(0, i, 1) where s ∈ C. By Theorem 7.1.8, since the entries of A are real, λ = i has corresponding eigenvectors of the form v = t(0, −i, 1) where t ∈ C. 0 −1 = 0 ⇐⇒ (λ + 2)(λ2 + 4λ + 5) = 0 ⇐⇒ λ = −2 or −3 − λ ⎤ ⎡ ⎡ ⎤⎡ ⎤ 0 1 0 v1 0 If λ = −2 then (A − λI)v = 0 assumes the form ⎣ 1 1 −1 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ 1 3 −1 v3 0 =⇒ v2 = 0 and v1 − v3 = 0. Thus, if we let v3 = r where r ∈ C, then the solution set of this system is {(r, 0, r) : r ∈ C} so the eigenvectors corresponding to⎡ λ = −2 are v = r(1, ⎤ 0,⎡1) where ⎤ r⎡∈ C.⎤ −i 1 0 v1 0 −1 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ If λ = −2 + i then (A − λI)v = 0 assumes the form ⎣ 1 1 − i 1 3 −1 − i v3 0 −2 + i −1 − 2i v3 = 0 and v2 + v3 = 0. Thus, if we let v3 = 5s where s ∈ C, then the solution =⇒ v1 + 5 5 set of this system is {((2 − i)s, (1 + 2i)s, 5s) : s ∈ C} so the eigenvectors corresponding to λ = −2 + i are v = s(2 − i, 1 + 2i, 5) where s ∈ C. By Theorem 7.1.8, since the entries of A are real, λ = −2 − i has corresponding eigenvectors of the form v = t(2 + i, 1 − 2i, 5) where t ∈ C. −2 − λ 1 27. det(A − λI) = 0 ⇐⇒ 1 λ = −2 ± i. 1 −1 − λ 3 3 0 = 0 ⇐⇒ λ(λ − 2)(λ − 4) = 0 ⇐⇒ λ = 0, λ = 2, or λ = 4. 3−λ ⎤⎡ ⎤ ⎡ ⎤ ⎡ v1 2 −1 3 0 1 0 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ If λ = 0 then (A − λI)v = 0 assumes the form ⎣ 3 2 −1 3 0 v3 =⇒ v1 + 2v2 − 3v3 = 0 and −5v2 + 9v3 = 0. Thus, if we let v3 = 5r where r ∈ R, then the solution set of this system is {(−3r, 9r, 5r) : r ∈ R} so the eigenvectors 0 are⎤v = r(−3, 9, 5) where r ∈ R. ⎡ corresponding ⎤ ⎡ to ⎤λ = ⎡ 0 −1 3 v1 0 If λ = 2 then (A − λI)v = 0 assumes the form ⎣ 3 −1 0 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ 2 −1 1 v3 0 =⇒ v1 − v3 = 0 and v2 − 3v3 = 0. Thus, if we let v3 = s where s ∈ R, then the solution set of this system is {(s, 3s, s) : s ∈ R} so the eigenvectors corresponding to λ = 2 are⎤v⎡= s(1, ⎤ 3, 1) ⎡ ⎡ where ⎤ s ∈ R. −2 −1 3 v1 0 0 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ If λ = 4 then (A − λI)v = 0 assumes the form ⎣ 3 −3 2 −1 −1 v3 0 =⇒ v1 − v3 = 0 and v2 − v3 = 0. Thus, if we let v3 = t where t ∈ R, then the solution set of this system is {(t, t, t) : t ∈ R} so the eigenvectors corresponding to λ = 4 are v = t(1, 1, 1) where t ∈ R. 2−λ 28. det(A − λI) = 0 ⇐⇒ 3 2 −1 1−λ −1 5−λ 29. det(A − λI) = 0 ⇐⇒ 0 0 0 5−λ 0 0 0 = 0 ⇐⇒ (λ − 5)3 = 0 ⇐⇒ λ = 5 of multiplicity three. 5−λ (c)2017 Pearson Education. Inc. 449 ⎤ ⎡ ⎤⎡ ⎤ 0 0 0 v1 0 If λ = 5 then (A − λI)v = 0 assumes the form ⎣ 0 0 0 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ 0 0 0 v3 0 Thus, if we let v1 = r, v2 = s, and v3 = t where r, s, t ∈ R, then the solution set of this system is {(r, s, t) : r, s, t ∈ R} so the eigenvectors corresponding to λ = 5 are v = r(1, 0, 0) + s(0, 1, 0) + t(0, 0, 1) where r, s, t ∈ R. That is, every nonzero vector in R3 is an eigenvector of A corresponding to λ = 5. ⎡ −λ 2 2 2 = 0 ⇐⇒ (λ − 4)(λ + 2)2 = 0 ⇐⇒ λ = 4 or λ = −2 of multiplicity 30. det(A − λI) = 0 ⇐⇒ 2 −λ 2 2 −λ two. ⎤ ⎡ ⎡ ⎤⎡ ⎤ −4 2 2 v1 0 2 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ If λ = 4 then (A − λI)v = 0 assumes the form ⎣ 2 −4 2 2 −4 v3 0 =⇒ v1 − v3 = 0 and v2 − v3 = 0. Thus, if we let v3 = r where r ∈ R, then the solution set of this system is {(r, r, r) : r ∈ R} so the eigenvectors corresponding r(1, 1, r ∈ R. ⎡ to λ = 4 are ⎡ 1) where ⎤ ⎡v = ⎤ ⎤ 2 2 2 0 v1 If λ = −2 then (A − λI)v = 0 assumes the form ⎣ 2 2 2 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ v3 2 2 2 0 =⇒ v1 + v2 + v3 = 0. Thus, if we let v2 = s and v3 = t where s, t ∈ R, then the solution set of this system is {(−s − t, s, t) : s, t ∈ R} so the eigenvectors corresponding to λ = −2 are v = s(−1, 1, 0) + t(−1, 0, 1) where s, t ∈ R. 1−λ 2 3 4 4 3−λ 2 1 4 3 2 31. det(A − λI) = 0 ⇐⇒ = 0 ⇐⇒ λ − 14λ − 32λ = 0 4 5 6 − λ 7 7 6 5 4−λ 2 ⇐⇒ λ (λ − 16)(λ + 2) = 0 ⇐⇒ λ = 16, λ = −2, ⎡or λ = 0 of multiplicity two. ⎤ ⎤⎡ ⎤ ⎡ v1 −15 2 3 4 0 ⎥ ⎢ ⎢ ⎥ ⎢ 4 −13 2 1 ⎥ ⎥ ⎢ v2 ⎥ = ⎢ 0 ⎥ If λ = 16 then (A − λI)v = 0 assumes the form ⎢ ⎦ ⎣ ⎣ ⎦ ⎣ 0 ⎦ 4 5 −10 7 v3 0 7 6 5 −12 v4 =⇒ v1 − 1841v3 + 2078v4 = 0, v2 + 82v3 − 93v4 = 0, and 31v3 − 35v4 = 0. Thus, if we let v4 = 31r where r ∈ R, then the solution set of this system is {(17r, 13r, 35r, 31r) : r ∈ R} so the eigenvectors corresponding to λ = 16 are v = r(17, 13, 35, 31) where r ∈ R. ⎡ ⎤⎡ ⎤ ⎤ ⎡ v1 3 2 3 4 0 ⎢ 4 5 2 1 ⎥ ⎢ v2 ⎥ ⎢ 0 ⎥ ⎥⎢ ⎥ ⎥ ⎢ If λ = −2 then (A − λI)v = 0 assumes the form ⎢ ⎣ 4 5 8 7 ⎦ ⎣ v3 ⎦ = ⎣ 0 ⎦ 7 6 5 6 0 v4 =⇒ v1 + v4 = 0, v2 − v4 = 0, and v3 + v4 = 0. Thus, if we let v4 = s where s ∈ R, then the solution set of this system is {(−s, s, −s, s) : s ∈ R} so the eigenvectors corresponding to λ = −2 are v = s(−1, 1, −1, 1) where s ∈ R. ⎤⎡ ⎤ ⎡ ⎤ ⎡ v1 1 2 3 4 0 ⎢ 4 3 2 1 ⎥ ⎢ v2 ⎥ ⎢ 0 ⎥ ⎥⎢ ⎥ ⎥ ⎢ If λ = 0 then (A − λI)v = 0 assumes the form ⎢ ⎣ 4 5 6 7 ⎦ ⎣ v3 ⎦ = ⎣ 0 ⎦ 7 6 5 4 0 v4 =⇒ v1 − v3 − 2v4 = 0 and v2 + 2v3 + 3v4 = 0. Thus, if we let v3 = a and v4 = b where a, b ∈ R, then the solution set of this system is {(a + 2b, −2a − 3b, a, b) : a, b ∈ R} so the eigenvectors corresponding to λ = 0 are v = a(1, −2, 1, 0) + b(2, −3, 0, 1) where a, b ∈ R. (c)2017 Pearson Education. Inc. 450 −λ 1 0 0 −1 −λ 0 0 32. det(A − λI) = 0 ⇐⇒ = 0 ⇐⇒ (λ2 + 1)2 = 0 ⇐⇒ λ = ±i, where each root is of 0 −λ −1 0 0 0 1 −λ multiplicity two. ⎤ ⎤⎡ ⎡ ⎤ ⎡ v1 i 1 0 0 0 ⎥ ⎢ ⎢ −1 i 0 ⎥ ⎢ 0 ⎥ ⎥ ⎢ v2 ⎥ ⎢ 0 ⎥ If λ = −i then (A − λI)v = 0 assumes the form ⎢ ⎣ 0 0 i −1 ⎦ ⎣ v3 ⎦ = ⎣ 0 ⎦ 0 0 0 1 i v4 =⇒ v1 − iv2 = 0 and v3 + iv4 = 0. Thus, if we let v2 = r and v4 = s where r, s ∈ C, then the solution set of this system is {(ir, r, −is, s) : r, s ∈ C} so the eigenvectors corresponding to λ = −i are v = r(i, 1, 0, 0) + s(0, 0, −i, 1) where r, s ∈ C. By Theorem 7.1.8, since the entries of A are real, λ = i has corresponding eigenvectors v = a(−i, 1, 0, 0) + b(0, 0, i, 1) where a, b ∈ C. 33. This matrix is lower triangular, and therefore, the eigenvalues appear along the main diagonal of the matrix: λ = 1 + i, 1 − 3i, 1. Note that the eigenvalues do not occur in complex conjugate pairs, but this does not contradict Theorem 7.1.8 because the matrix does not consist entirely of real elements. 1−λ −1 34. (a). p(λ) = det(A − λI2 ) = = λ2 − 5λ + 6. 2 4−λ (b). 1 −1 1 −1 1 0 −5 +6 2 4 2 4 0 1 −1 −5 −5 5 6 0 0 0 = + + = = 02 . 10 14 −10 −20 0 6 0 0 A2 − 5A + 6I2 = 1 −1 2 4 (c). Using part (b) of this problem: A2 − 5A + 6I2 = 02 ⇐⇒ A−1 (A2 − 5A + 6I2 ) = A−1 · 02 ⇐⇒ A − 5I2 + 6A−1 = 02 ⇐⇒ 6A−1 = 5I2 − A 1 1 0 1 −1 5 − ⇐⇒ A−1 = 0 1 2 4 6 2 1 1 4 1 3 6 , or A−1 = ⇐⇒ A−1 = . − 13 16 6 −2 1 1−λ 2 = 0 ⇐⇒ λ2 + λ − 6 = 0 ⇐⇒ (λ − 2)(λ + 3) = 0 ⇐⇒ λ = 2 or 35. (a). det(A − λI2 ) = 2 −2 − λ λ = −3. 1 2 1 2 1 0 (b). ∼ ∼ = B. 2 −2 0 −6 0 1 1−λ 0 = 0 ⇐⇒ (λ − 1)2 = 0 ⇐⇒ λ = 1 of multiplicity two. Matrices A det(B − λI2 ) = 0 ⇐⇒ 0 1−λ and B do not have the same eigenvalues. 36. A(3v1 − v2 ) = 3Av1 − Av2 = 3(2v1 ) − (−3v2 ) = 6v1 + 3v2 1 2 6 6 12 =6 +3 = + = . −1 1 −6 3 −3 (c)2017 Pearson Education. Inc. 451 37. (a). Let a, b, c ∈ R. If v = av1 + bv2 + cv3 , then 5, 0, 3) = a(1, −1, 1) + b(2, 1, 3) ⎧ + c(−1, −1, 2), or ⎪ ⎨ a + 2b − c = 5 (5, 0, 3) = (a + 2b − c, −a + b − c, a + 3b + 2c). The last equality results in the system: ⎪ ⎩ −a + b − c = 0 a + 3b + 2c = 3. This system has the solution a = 2, b = 1, and c = −1. Consequently, v = 2v1 + v2 − v3 . (b). Using part (a): Av = A(2v1 + v2 − v3 ) = 2Av1 + Av2 − Av3 = 2(2v1 ) + (−2v2 ) − (3v3 ) = 4v1 − 2v2 − 3v3 = 4(1, −1, 1) − 2(2, 1, 3) − 3(−1, −1, 2) = (3, −3, −8). 38. A(c1 v1 + c2 v2 + c3 v3 ) = A(c1 v1 ) + A(c2 v2 ) + A(c3 v3 ) = c1 (Av1 ) + c2 (Av2 ) + c3 (Av3 ) = c1 (λv1 ) + c2 (λv2 ) + c3 (λv3 ) = λ(c1 v1 + c2 v2 + c3 v3 ). Thus, c1 v1 + c2 v2 + c3 v3 is an eigenvector of A corresponding to the eigenvalue λ. 39. Recall that the determinant of an upper (lower) triangular matrix is just the product of its main diagonal elements. Let A be an n × n upper (lower) triangular matrix. It follows that A − λIn is an upper (lower) triangular matrix with main diagonal element aii − λ, i = 1, 2, . . . , n. Consequently, det(A − λIn ) = 0 ⇐⇒ n - (aii − λ) = 0. i=1 This implies that λ = a11 , a22 , . . . , ann . 40. Any scalar λ such that det(A − λI) = 0 is an eigenvalue of A. Therefore, if 0 is an eigenvalue of A, then det(A − 0 · I) = 0, or det(A) = 0, which implies that A is not invertible. On the other hand, if 0 is not an eigenvalue of A, then det(A − 0 · I) = 0, or det(A) = 0, which implies that A is invertible. 41. A is invertible, so A−1 exists. Also, λ is an eigenvalue of A so that Av = λv. Thus, 1 A−1 (Av) = A−1 (λv) =⇒ (A−1 A)v = λA−1 v =⇒ In v = λA−1 v =⇒ v = λA−1 v =⇒ v = A−1 v. λ 1 Therefore is an eigenvalue of A−1 provided that λ is an eigenvalue of A. λ 42. By assumption, we have Av = λv and Bv = μv. (a). Therefore, (AB)v = A(Bv) = A(μv) = μ(Av) = μ(λv) = (λμ)v, which shows that v is an eigenvector of AB with corresponding eigenvalue λμ. (b). Also, (A + B)v = Av + Bv = λv + μv = (λ + μ)v, which shows that v is an eigenvector of A + B with corresponding eigenvalue λ + μ. 43. Recall that a matrix and its transpose have the same determinant. Thus, det(A − λIn ) = det([A − λIn ]T ) = det(AT − λIn ). Since A and AT have the same characteristic polynomial, it follows that both matrices also have the same eigenvalues. (c)2017 Pearson Education. Inc. 452 44. (a). v = r + is is an eigenvector with eigenvalue λ = a + bi, b = 0 =⇒ Av = λv =⇒ A(r + is) = (a + bi)(r + is) = (ar − bs) + i(as + br) =⇒ Ar = ar − bs and As = as + br. Now if r = 0, then A0 = a0 − bs =⇒ 0 = 0 − bs =⇒ 0 = bs =⇒ s = 0 since b = 0. This would mean that v = 0 so v could not be an eigenvector. Thus, it must be that r = 0. Similarly, if s = 0, then r = 0, and again, this would contradict the fact that v is an eigenvector. Hence, it must be the case that r = 0 and s = 0. (b). As in part (a), Ar = ar − bs and As = as + br. Let c1 , c2 ∈ R. Then if c1 r + c2 s = 0, (44.1) we have A(c1 r + c2 s) = 0 =⇒ c1 Ar + c2 As = 0 =⇒ c1 (ar − bs) + c2 (as + br) = 0. Hence, (c1 a + c2 b)r + (c2 a − c1 b)s = 0 =⇒ a(c1 r + c2 s) + b(c2 r − c1 s) = 0 =⇒ b(c2 r − c1 s) = 0 where we have used (39.1). Since b = 0, we must have c2 r − c1 s = 0. Combining this with (44.1) yields c1 = c2 = 0. Therefore, it follows that r and s are linearly independent vectors. 45. λ1 = 2, v = r(−1, 1). λ2 = 5, v = s(1, 2). 46. λ1 = −2 (multiplicity two), v = r(1, 1, 1). λ2 = −5, v = s(20, 11, 14). 47. λ1 = 3 (multiplicity two), v = r(1, 0, −1) + s(0, 1, −1). λ2 = 6, v = t(1, 1, 1). √ √ √ √ √ √ √ √ 48. λ1 = 3 − 6, v = r( 6, −1 + 6, −5 + 6). λ2 = 3 + 6, v = s( 6, 1 + 6, 5 + 6), λ3 = −2, v = t(−1, 3, 0). 49. λ1 = 0, v = r(2, 2, 1). λ2 = 3i, v = s(−4 − 3i, 5, −2 + 6i), λ3 = −3i, v = t(−4 + 3i, 5, −2 − 6i). 50. λ1 = −1 (multiplicity four), v = a(−1, 0, 0, 1, 0) + b(−1, 0, 1, 0, 0) + c(−1, 0, 0, 0, 1) + d(−1, 1, 0, 0, 0). Solutions to Section 7.2 True-False Review: (a): FALSE. There have to be n linearly independent eigenvectors. (b): TRUE. The eigenspace Eλ is equal to the null space of the n × n matrix A − λI, and this null space is a subspace of Rn . (c): TRUE. The dimension of an eigenspace never exceeds the algebraic multiplicity of the corresponding eigenvalue. (d): TRUE. Eigenvectors corresponding to distinct eigenspaces are linearly independent. Therefore if we choose one (nonzero) vector from each distinct eigenspace, the chosen vectors will form a linearly independent set. (e): TRUE. Since each eigenvalue of the matrix A occurs with algebraic multiplicity 1, we can simply choose one eigenvector from each eigenspace to obtain a basis of eigenvectors for A. Thus, A is nondefective. (f ): FALSE. Many examples will show that this statement is false, including the n × n identity matrix In for n ≥ 2. The matrix In is not defective, and yet, has λ = 1 occurring with algebraic multiplicity n. Problems: −7 − λ 1. det(A − λI) = 0 ⇐⇒ −3 λ = −7 (with multiplicity 2). 0 = 0 ⇐⇒ (−7 − λ)2 = 0, so taht we obtain the eigenvalue −7 − λ (c)2017 Pearson Education. Inc. 453 0 0 0 v1 = . The solution set of this 0 v2 −3 0 system is {(0, r) : r ∈ R}, so the eigenspace corresponding to λ = −7 is E1 = {v ∈ R2 : v = r(0, 1), r ∈ R}. A basis for E1 is {(0, 1)}, and dim[E1 ] = 1. We conclude that A is defective since it does not have a complete set of eigenvectors. 1−λ 4 2. det(A − λI) = 0 ⇐⇒ = 0 ⇐⇒ λ2 − 4λ − 5 = 0 ⇐⇒ (λ − 5)(λ + 1) = 0 ⇐⇒ λ = 5 or 2 3−λ λ = −1. 0 −4 4 v1 = =⇒ v1 − v2 = 0. The solution If λ1 = 5 then (A − λI)v = 0 assumes the form 0 v2 2 −2 set of this system is {(r, r) : r ∈ R}, so the eigenspace corresponding to λ1 = 5 is E1 = {v ∈ R2 : v = r(1, 1), r ∈ R}. A basis for E1 is {(1, 1)}, and dim[E 1 ] = 1. 0 2 4 v1 = =⇒ v1 + 2v2 = 0. The solution If λ2 = −1 then (A − λI)v = 0 assumes the form 0 v2 2 4 set of this system is {(−2s, s) : s ∈ R}, so the eigenspace corresponding to λ2 = −1 is E2 = {v ∈ R2 : v = s(−2, 1), s ∈ R}. A basis for E2 is {(−2, 1)}, and dim[E2 ] = 1. A complete set of eigenvectors for A is given by {(1, 1), (−2, 1)}, so A is nondefective. 3−λ 0 3. det(A − λI) = 0 ⇐⇒ = 0 ⇐⇒ (3 − λ)2 = 0 ⇐⇒ λ = 3 of multiplicity two. 0 3−λ 0 0 0 v1 = . If λ1 = 3 then (A − λI)v = 0 assumes the form v2 0 0 0 The solution set of this system is {(r, s) : r, s ∈ R}, so the eigenspace corresponding to λ1 = 3 is E1 = {v ∈ R2 : v = r(1, 0) + s(0, 1), r, s ∈ R}. A basis for E1 is {(1, 0), (0, 1)}, and dim[E1 ] = 2. A is nondefective. 1−λ 2 = 0 ⇐⇒ (λ − 3)2 = 0 ⇐⇒ λ = 3 of multiplicity two. 4. det(A − λI) = 0 ⇐⇒ −2 5−λ 0 −2 2 v1 = =⇒ v1 − v2 = 0. The solution If λ1 = 3 then (A − λI)v = 0 assumes the form 0 v2 −2 2 set of this system is {(r, r) : r ∈ R}, so the eigenspace corresponding to λ1 = 3 is E1 = {v ∈ R2 : v = r(1, 1), r ∈ R}. A basis for E1 is {(1, 1)}, and dim[E1 ] = 1. A is defective since it does not have a complete set of eigenvectors. 5−λ 5 = 0 ⇐⇒ λ2 − 4λ + 5 = 0 ⇐⇒ λ = 2 ± i. 5. det(A − λI) = 0 ⇐⇒ −2 −1 − λ 0 3+i 5 v1 = =⇒ −2v1 +(−3+i)v2 = If λ1 = 2 − i then (A−λI)v = 0 assumes the form 0 v2 −2 −3 + i 0. The solution set of this system is {((−3 + i)r, 2r) : r ∈ C}, so the eigenspace corresponding to λ1 = 2 − i is E1 = {v ∈ C2 : v = r(−3 + i, 2), r ∈ C}. A basis for E1 is {(−3 + i, 2)}, and dim[E1 ] = 1. If λ2 = 2 + i then from Theorem 7.1.8, the eigenvectors corresponding to λ2 = 2 + i are v = s(−3 − i, 2) where s ∈ C, so the eigenspace corresponding to λ2 = 2 + i is E2 = {v ∈ C2 : v = s(−3 − i, 2), s ∈ C}. A basis for E2 is {(−3 − i, 2)}, and dim[E2 ] = 1. A is nondefective since it has a complete set of eigenvectors, namely {(−3 + i, 2), (−3 − i, 2)}. 3−λ −4 −1 −1 − λ −1 = 0 ⇐⇒ (λ + 2)(λ − 3)2 = 0 ⇐⇒ λ = −2 or λ = 3 of 6. det(A − λI) = 0 ⇐⇒ 0 0 −4 2−λ multiplicity two. If λ = −7 then (A − λI)v = 0 assumes the form (c)2017 Pearson Education. Inc. 454 ⎤ ⎡ ⎤⎡ ⎤ 5 −4 −1 v1 0 1 −1 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦. If λ1 = −2 then (A − λI)v = 0 assumes the form ⎣ 0 0 −4 4 v3 0 The solution set of this system is {(r, r, r) : r ∈ R}, so the eigenspace corresponding to λ1 = −2 is E1 = {v ∈ R3 : v = r(1, 1, 1), r ∈ R}. A basis for E1 is⎡{(1, 1, 1)}, and⎤dim[E ] = 1.⎡ 1⎤ ⎡ ⎤ 0 −4 −1 v1 0 If λ2 = 3 then (A − λI)v = 0 assumes the form ⎣ 0 −4 −1 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ =⇒ 4v2 + v3 = 0. The 0 −4 −1 v3 0 solution set of this system is {(s, t, −4t) : s, t ∈ R}, so the eigenspace corresponding to λ2 = 3 is E2 = {v ∈ R3 : v = s(1, 0, 0) + t(0, 1, −4), s, t ∈ R}. A basis for E2 is {(1, 0, 0), (0, 1, −4)}, and dim[E2 ] = 2. A complete set of eigenvectors for A is given by {(1, 1, 1), (1, 0, 0), (0, 1, −4)}, so A is nondefective. 4−λ 0 0 2−λ −3 = 0 ⇐⇒ (λ + 1)(λ − 4)2 = 0 ⇐⇒ λ = −1 or λ = 4 of 7. det(A − λI) = 0 ⇐⇒ 0 0 −2 1−λ multiplicity two. ⎤ ⎡ ⎡ ⎤⎡ ⎤ 5 0 0 v1 0 3 −3 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ If λ1 = −1 then (A − λI)v = 0 assumes the form ⎣ 0 0 −2 2 v3 0 =⇒ v1 = 0 and v2 − v3 = 0. The solution set of this system is {(0, r, r) : r ∈ R}, so the eigenspace corresponding to λ1 = −1 is E1 = {v ∈ R3 : v = r(0, 1, 1), r ∈ R}. A basis for E1 is {(0, 1, 1)}, and dim[E1 ] = 1. ⎤ ⎡ ⎤⎡ ⎡ ⎤ 0 0 0 v1 0 If λ2 = 4 then (A − λI)v = 0 assumes the form ⎣ 0 −2 −3 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ =⇒ 2v2 + 3v3 = 0. 0 −2 −3 v3 0 The solution set of this system is {(s, 3t, −2t) : s, t ∈ R}, so the eigenspace corresponding to λ2 = 4 is E2 = {v ∈ R3 : v = s(1, 0, 0) + t(0, 3, −2), s, t ∈ R}. A basis for E2 is {(1, 0, 0), (0, −1, 1)}, and dim[E2 ] = 2. A complete set of eigenvectors for A is given by {(1, 0, 0), (0, 3, −2), (0, 1, 1)}, so A is nondefective. 3−λ 1 0 5−λ 0 = 0 ⇐⇒ (λ − 4)3 = 0 ⇐⇒ λ = 4 of multiplicity three. 8. det(A − λI) = 0 ⇐⇒ −1 0 0 4−λ ⎤ ⎡ ⎡ ⎤⎡ ⎤ −1 1 0 v1 0 If λ1 = 4 then (A − λI)v = 0 assumes the form ⎣ −1 1 0 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ 0 0 0 v3 0 =⇒ v1 − v2 = 0 and v3 ∈ R. The solution set of this system is {(r, r, s) : r, s ∈ R}, so the eigenspace corresponding to λ1 = 4 is E1 = {v ∈ R3 : v = r(1, 1, 0) + s(0, 0, 1), r, s ∈ R}. A basis for E1 is {(1, 1, 0), (0, 0, 1)}, and dim[E1 ] = 2. A is defective since it does not have a complete set of eigenvectors. 3−λ 0 0 −λ −4 = 0 ⇐⇒ (λ − 3)(λ2 + 16) = 0 ⇐⇒ λ = 3 or λ = ±4i. 9. det(A − λI) = 0 ⇐⇒ 2 1 4 −λ ⎤ ⎡ ⎤ ⎡ ⎤⎡ 0 0 0 v1 0 If λ1 = 3 then (A − λI)v = 0 assumes the form ⎣ 2 −3 −4 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ 0 1 4 −3 v3 =⇒ 11v1 − 25v3 = 0 and 11v2 − 2v3 . The solution set of this system is {(25r, 2r, 11r) : r ∈ C}, so the eigenspace corresponding to λ1 = 3 is E1 = {v ∈ C3 : v = r(25, 2, 11), r ∈ C}. A basis for E1 is {(25, 2, 11)}, and dim[E1 ] = 1. ⎡ (c)2017 Pearson Education. Inc. 455 ⎤ ⎡ ⎤⎡ ⎤ 3 + 4i 0 0 v1 0 2 4i −4 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ If λ2 = −4i then (A − λI)v = 0 assumes the form ⎣ 1 4 4i v3 0 =⇒ v1 = 0 and iv2 − v3 = 0. The solution set of this system is {(0, s, is) : s ∈ C}, so the eigenspace corresponding to λ2 = −4i is E2 = {v ∈ C3 : v = s(0, 1, i), s ∈ C}. A basis for E2 is {(0, 1, i)}, and dim[E2 ] = 1. If λ3 = 4i then from Theorem 7.1.8, the eigenvectors corresponding to λ3 = 4i are v = t(0, 1, −i) where t ∈ C, so the eigenspace corresponding to λ3 = 4i is E3 = {v ∈ C3 : v = t(0, 1, −i), t ∈ C}. A basis for E3 is {(0, 1, −i)}, and dim[E3 ] = 1. A complete set of eigenvectors for A is given by {(25, 2, 11), (0, 1, i), (0, 1, −i)}, so A is nondefective. 4−λ 1 6 −λ −7 = 0 ⇐⇒ (λ + 3)(λ − 2)2 = 0 ⇐⇒ λ = −3 or λ = 2 of 10. det(A − λI) = 0 ⇐⇒ −4 0 0 −3 − λ multiplicity two. ⎤ ⎡ ⎡ ⎤⎡ ⎤ 7 1 6 v1 0 If λ1 = −3 then (A − λI)v = 0 assumes the form ⎣ −4 3 −7 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ 0 0 0 v3 0 =⇒ v1 + v3 = 0 and v2 − v3 = 0. The solution set of this system is {(−r, r, r) : r ∈ R}, so the eigenspace corresponding to λ1 = −3 is E1 = {v ∈ R3 : v = r(−1, 1, 1), r ∈ R}. A basis for E1 is {(−1, 1, 1)}, and dim[E1 ] = 1. ⎤ ⎡ ⎡ ⎤⎡ ⎤ 2 1 6 v1 0 If λ2 = 2 then (A − λI)v = 0 assumes the form ⎣ −4 −2 −7 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ 0 0 5 v3 0 =⇒ 2v1 + v2 = 0 and v3 = 0. The solution set of this system is {(−s, 2s, 0) : s ∈ R}, so the eigenspace corresponding to λ2 = 2 is E2 = {v ∈ R3 : v = s(−1, 2, 0), s ∈ R}. A basis for E2 is {(−1, 2, 0)}, and dim[E2 ] = 1. A is defective because it does not have a complete set of eigenvectors. 2−λ 0 0 2−λ 0 = 0 ⇐⇒ (λ − 2)3 = 0 ⇐⇒ λ = 2 of multiplicity three. 11. det(A − λI) = 0 ⇐⇒ 0 0 0 2−λ ⎤ ⎡ ⎤⎡ ⎡ ⎤ 0 0 0 v1 0 If λ1 = 2 then (A − λI)v = 0 assumes the form ⎣ 0 0 0 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦. The solution set of this 0 0 0 v3 0 system is {(r, s, t) : r, s, t ∈ R}, so the eigenspace corresponding to λ1 = 2 is E1 = {v ∈ R3 : v = r(1, 0, 0) + s(0, 1, 0) + t(0, 0, 1), r, s, t ∈ R}. A basis for E1 is {(1, 0, 0), (0, 1, 0), (0, 0, 1)}, and dim[E1 ] = 3. A is nondefective since it has a complete set of eigenvectors. 7−λ −8 6 = 0 ⇐⇒ (λ + 1)3 = 0 ⇐⇒ λ = −1 of multiplicity −9 − λ 6 12. det(A − λI) = 0 ⇐⇒ 8 0 0 −1 − λ three. ⎤ ⎡ ⎤ ⎡ ⎤⎡ 8 −8 6 v1 0 If λ1 = −1 then (A − λI)v = 0 assumes the form ⎣ 8 −8 6 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ =⇒ 4v1 − 4v2 + 3v3 = 0. 0 0 0 0 v3 The solution set of this system is {(r − 3s, r, 4s) : r, s ∈ R}, so the eigenspace corresponding to λ1 = −1 is E1 = {v ∈ R3 : v = r(1, 1, 0) + s(−3, 0, 4), r, s ∈ R}. A basis for E1 is {(1, 1, 0), (−3, 0, 4)}, and dim[E1 ] = 2. A is defective since it does not have a complete set of eigenvectors. ⎡ (c)2017 Pearson Education. Inc. 456 2−λ 13. det(A − λI) = 0 ⇐⇒ 2 2 multiplicity two. −1 −1 = 0 ⇐⇒ (λ − 2)λ2 = 0 ⇐⇒ λ = 2 or λ = 0 of −1 − λ ⎤ ⎡ ⎡ ⎤⎡ ⎤ 0 2 −1 v1 0 If λ1 = 2 then (A − λI)v = 0 assumes the form ⎣ 2 −1 −1 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ 2 3 −3 v3 0 =⇒ 4v1 − 3v3 = 0 and 2v2 − v3 = 0. The solution set of this system is {(3r, 2r, 4r) : r ∈ R}, so the eigenspace corresponding to λ1 = 2 is E1 = {v ∈ R3 : v = r(3, 2, 4), r ∈ R}. A basis for E1 is {(3, 2, 4)}, and dim[E1 ] = 1. ⎤ ⎡ ⎤ ⎡ ⎤⎡ 2 2 −1 v1 0 If λ2 = 0 then (A − λI)v = 0 assumes the form ⎣ 2 1 −1 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ 2 3 −1 v3 0 =⇒ 2v1 − v3 = 0 and v2 = 0. The solution set of this system is {(s, 0, 2s) : s ∈ R}, so the eigenspace corresponding to λ2 = 0 is E2 = {v ∈ R3 : v = s(1, 0, 2), s ∈ R}. A basis for E2 is {(1, 0, 2)}, and dim[E2 ] = 1. A is defective because it does not have a complete set of eigenvectors. 1−λ −1 2 −1 − λ 2 = 0 ⇐⇒ (λ − 2)λ2 = 0 ⇐⇒ λ = 2 or λ = 0 of 14. det(A − λI) = 0 ⇐⇒ 1 1 −1 2−λ multiplicity two. ⎤ ⎡ ⎤ ⎡ ⎤⎡ −1 −1 2 v1 0 If λ1 = 2 then (A − λI)v = 0 assumes the form ⎣ 1 −3 2 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ 0 1 −1 0 v3 =⇒ v1 − v3 = 0 and v2 − v3 = 0. The solution set of this system is {(r, r, r) : r ∈ R}, so the eigenspace corresponding to λ1 = 2 is E1 = {v ∈ R3 : v = r(1, 1, 1), r ∈ R}. A basis for E1 is {(1, 1, 1)}, and dim[E1 ] = 1. ⎤ ⎡ ⎤⎡ ⎡ ⎤ 1 −1 2 v1 0 If λ2 = 0 then (A − λI)v = 0 assumes the form ⎣ 1 −1 2 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ =⇒ v1 − v2 + 2v3 = 0. 1 −1 2 v3 0 The solution set of this system is {(s − 2t, s, t) : s, t ∈ R}, so the eigenspace corresponding to λ2 = 0 is E2 = {v ∈ R3 : v = s(1, 1, 0) + t(−2, 0, 1), s, t ∈ R}. A basis for E2 is {(1, 1, 0), (−2, 0, 1)}, and dim[E2 ] = 2. A is nondefective because it has a complete set of eigenvectors. 2−λ 3 0 −λ 1 = 0 ⇐⇒ (λ − 2)3 = 0 ⇐⇒ λ = 2 of multiplicity three. 15. det(A − λI) = 0 ⇐⇒ −1 −2 −1 4 − λ ⎤ ⎡ ⎡ ⎤⎡ ⎤ 0 3 0 v1 0 If λ1 = 2 then (A − λI)v = 0 assumes the form ⎣ −1 −2 1 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ −2 −1 2 v3 0 =⇒ v1 − v3 = 0 and v2 = 0. The solution set of this system is {(r, 0, r) : r ∈ R}, so the eigenspace corresponding to λ1 = 2 is E1 = {v ∈ R3 : v = r(1, 0, 1), r ∈ R}. A basis for E1 is {(1, 0, 1)}, and dim[E1 ] = 1. A is defective since it does not have a complete set of eigenvectors. −λ −1 −1 16. det(A − λI) = 0 ⇐⇒ −1 −λ −1 = 0 ⇐⇒ (λ + 2)(λ − 1)2 = 0 ⇐⇒ λ = −2 or λ = 1 of multiplicity −1 −1 −λ two. 2 1−λ 3 (c)2017 Pearson Education. Inc. 457 ⎤ ⎡ ⎤⎡ ⎤ 2 −1 −1 v1 0 2 −1 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ If λ1 = −2 then (A − λI)v = 0 assumes the form ⎣ −1 −1 −1 2 v3 0 =⇒ v1 − v3 = 0 and v2 − v3 = 0. The solution set of this system is {(r, r, r) : r ∈ R}, so the eigenspace corresponding to λ1 = −2 is E1 = {v ∈ R3 : v = r(1, 1, 1), r ∈ R}. A basis for E1 is {(1, 1, 1)}, and dim[E1 ] = 1. ⎤ ⎡ ⎡ ⎤⎡ ⎤ −1 −1 −1 v1 0 If λ2 = 1 then (A − λI)v = 0 assumes the form ⎣ −1 −1 −1 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ =⇒ v1 + v2 + v3 = 0. −1 −1 −1 v3 0 The solution set of this system is {(−s − t, s, t) : s, t ∈ R}, so the eigenspace corresponding to λ2 = 1 is E2 = {v ∈ R3 : v = s(−1, 1, 0) + t(−1, 0, 1), s, t ∈ R}. A basis for E2 is {(−1, 1, 0), (−1, 0, 1)}, and dim[E2 ] = 2. A is nondefective because it has a complete set of eigenvectors. ⎡ 17. (λ − 4)(λ + 1) = 0 ⇐⇒ λ = 4 or λ = −1. Since A has two distinct eigenvalues, it has two linearly independent eigenvectors and is, therefore, nondefective. 18. (λ − 1)2 = 0 ⇐⇒ λ = 1 of multiplicity two. 0 5 5 v1 = =⇒ v1 + v2 = 0. The solution v2 0 −5 −5 set of this system is {(−r, r) : r ∈ R}, so the eigenspace corresponding to λ1 = 1 is E1 = {v ∈ R2 : v = r(−1, 1), r ∈ R}. A basis for E1 is {(1, 1)}, and dim[E1 ] = 1. A is defective since it does not have a complete set of eigenvectors. If λ1 = 1 then (A − λI)v = 0 assumes the form 19. λ2 − 4λ + 13 = 0 ⇐⇒ λ = 2 ± 3i. Since A has two distinct eigenvalues, it has two linearly independent eigenvectors and is, therefore, nondefective. 20. (λ − 2)2 (λ + 1) = 0 ⇐⇒ λ = −1 or λ = 2 of multiplicity two. To determine whether A is nondefective, all we require is the dimension of the eigenspace⎡corresponding ⎤to⎡λ = 2. ⎤ ⎡ ⎤ −1 −3 1 v1 0 If λ = 2 then (A − λI)v = 0 assumes the form ⎣ −1 −3 1 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ =⇒ −v1 − 3v2 + v3 = 0. −1 −3 1 v3 0 The solution set of this system is {(−3s + t, s, t) : s, t ∈ R}, so the eigenspace corresponding to λ = 2 is E = {v ∈ R3 : v = s(−3, 1, 0) + t(1, 0, 1), s, t ∈ R}. Since dim[E] = 2, A is nondefective. 21. (λ − 3)3 = 0 ⇐⇒ λ = 3 of multiplicity three. ⎡ ⎤ ⎡ ⎤ ⎤⎡ −4 2 2 v1 0 If λ = 3 then (A − λI)v = 0 assumes the form ⎣ −4 2 2 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ =⇒ −2v1 + v2 + v3 = 0 and 0 −4 2 2 v3 v2 = 0. The solution set of this system is {(r, 2r − s, s) : r, s ∈ R}, so the eigenspace corresponding to λ = 3 is E = {v ∈ R3 : v = r(1, 2, 0) + s(0, −1, 1), r, s ∈ R}. A is defective since it does not have a complete set of eigenvectors. 22. Since A is upper triangular, we note that det(A − λI) = (4 − λ)3 (−λ)2 , 3) and λ2 = 0 ⎤(of from which we obtain the eigenvalues λ1 = 4 (of multiplicity ⎤ ⎡ multiplicity ⎡ 2). ⎡ ⎤ v1 0 0 1 0 0 0 ⎢ ⎢ 0 ⎥ ⎢ 0 0 1 ⎥ 0 0 ⎥ ⎥ ⎥ ⎢ v2 ⎥ ⎢ ⎢ ⎥ ⎥ ⎢ ⎢ ⎥ 0 0 ⎥ ⎢ v3 ⎥ = ⎢ If λ1 = 4 then (A − λI)v = 0 assumes the form ⎢ 0 0 0 ⎢ 0 ⎥, from which ⎦ ⎣ ⎣ ⎣ 0 0 0 −4 ⎦ 0 ⎦ 1 v4 0 0 0 0 0 −4 v5 (c)2017 Pearson Education. Inc. 458 we compute quickly that v2 = v3 = v4 = v5 = 0, leaving only v1 as a free variable. Thus, a solution set for this system is {(r, 0, 0, 0, 0) : r ∈ R}, so the eigenspace corresponding to λ1 = 4 is E1 = {v ∈ R5 : v = r(1, 0, 0, 0, 0), r ∈ R}. A basis for E1 is {(1, 0, 0, 0, 0)} and dim[E1 ] = 1. We conclude that A is defective already since the eigenvalue λ1 occurred with multiplicity 3, but only generated one linearly independent eigenvector. It is thus impossible to form a complete set of eigenvectors. Note: We could also consider the eigenvalue λ2 = 0 and go through a similar analysis to that done above, but since we already know that A is defective, it is unnecessary to carry out this consideration. 23. In Problem 1 in this section, we saw that there is only one eigenspace, E1 , corresponding to λ = −7. A basis for this eigenspace was determined to be {(0, 1)} in R2 , so that this eigenspace consists of the y-axis in the xy-plane R2 . 2−λ 1 = 0 ⇐⇒ (λ − 1)(λ − 5) = 0 ⇐⇒ λ = 1 or λ = 5. 24. det(A − λI) = 0 ⇐⇒ 3 4−λ 0 1 1 v1 If λ1 = 1 then (A − λI)v = 0 assumes the form = =⇒ v1 + v2 = 0. The eigenspace v2 0 3 3 2 corresponding to λ1 = 1 is E1 = {v ∈ R : v = r(−1, forE1 is {(−1, 1)}. 1), r ∈ R}. A basis 0 −3 1 v1 = =⇒ 3v1 − v2 = 0. The If λ2 = 5 then (A − λI)v = 0 assumes the form 0 v2 3 −1 2 eigenspace corresponding to λ2 = 5 is E2 = {v ∈ R : v = s(1, 3), s ∈ R}. A basis for E2 is {(1, 3)}. v2 3 2 1 E1 -2 -1 -1 E2 1 2 v1 Figure 0.0.72: Figure for Problem 24 2−λ 25. det(A − λI) = 0 ⇐⇒ 0 3 = 0 ⇐⇒ (2 − λ)2 = 0 ⇐⇒ λ = 2 of multiplicity two. 2−λ 0 0 3 v1 If λ1 = 2 then (A − λI)v = 0 assumes the form = =⇒ v1 ∈ R and v2 = 0. The v2 0 0 0 eigenspace corresponding to λ1 = 2 is E1 = {v ∈ R2 : v = r(1, 0), r ∈ R}. A basis for E1 is {(1, 0)}. 0 = 0 ⇐⇒ (5 − λ)2 = 0 ⇐⇒ λ = 5 of multiplicity two. 5−λ 0 0 0 v1 If λ1 = 5 then (A − λI)v = 0 assumes the form = =⇒ v1 , v2 ∈ R. The eigenspace v2 0 0 0 corresponding to λ1 = 5 is E1 = {v ∈ R2 : v = r(1, 0) + s(0, 1), r, s ∈ R}. A basis for E1 is {(1, 0), (0, 1)}. 5−λ 26. det(A − λI) = 0 ⇐⇒ 0 (c)2017 Pearson Education. Inc. 459 v2 E1 v1 Figure 0.0.73: Figure for Problem 25 v2 E1 is the whole of R2 v1 Figure 0.0.74: Figure for Problem 26 3−λ 27. det(A − λI) = 0 ⇐⇒ 1 −1 multiplicity two. −1 −1 = 0 ⇐⇒ (λ − 5)(λ − 2)2 = 0 ⇐⇒ λ = 5 or λ = 2 of 3−λ ⎤ ⎡ ⎡ ⎤⎡ ⎤ −2 1 −1 v1 0 If λ1 = 5 then (A − λI)v = 0 assumes the form ⎣ 1 −2 −1 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ −1 −1 −2 v3 0 =⇒ v1 + v3 = 0 and v2 + v3 = 0. The eigenspace corresponding to λ1 = 5 is E1 = {v ∈ R3 : v = r(1, 1, −1), r ∈ R}. A basis for ⎤ ⎡ ⎡ E1 is {(1, 1, −1)}. ⎤⎡ ⎤ 1 1 −1 v1 0 1 −1 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ =⇒ v1 + v2 − v3 = 0. If λ2 = 2 then (A − λI)v = 0 assumes the form ⎣ 1 −1 −1 1 v3 0 The eigenspace corresponding to λ2 = 2 is E2 = {v ∈ R3 : v = s(−1, 1, 0) + t(1, 0, 1), s, t ∈ R}. A basis for E2 is {(−1, 1, 0), (1, 0, 1)}. 1 3−λ −1 0 = 0 ⇐⇒ (λ + 2)3 = 0 ⇐⇒ λ = −2 of multiplicity 2 −2 − λ ⎤ ⎡ ⎤⎡ ⎡ ⎤ −1 1 0 v1 0 If λ1 = −2 then (A − λI)v = 0 assumes the form ⎣ −1 1 2 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ =⇒ v1 − v2 = 0 and 0 0 0 v3 0 v3 = 0. The eigenspace corresponding to λ1 = −2 is E1 = {v ∈ R3 : v = r(1, 1, 0), r ∈ R}. A basis for E1 is {(1, 1, 0)}. −3 − λ 28. det(A − λI) = 0 ⇐⇒ −1 0 three. 1 −1 − λ 0 (c)2017 Pearson Education. Inc. 460 v3 E2 (1, 0, 1) (-1, 1, 0) v2 E1 V1 (1, 1, -1) Figure 0.0.75: Figure for Problem 27 v3 E1 v2 V1 (1, 1, 0) Figure 0.0.76: Figure for Problem 28 ⎤ ⎡ ⎤ ⎤⎡ 1 −2 3 v1 0 29. (a). If λ1 = 1 then (A − λI)v = 0 assumes the form ⎣ 1 −2 3 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ 0 1 −2 3 v3 =⇒ v1 − 2v2 + 3v3 = 0. The eigenspace corresponding to λ1 = 1 is E1 = {v ∈ R3 : v = r(2, 1, 0) + s(−3, 0, 1), r, s ∈ R}. A basis for E1 is {(2, 1, 0), (−3, 0, 1)}. Now apply the Gram-Schmidt process where v1 = (−3, 0, 1), and v2 = (2, 1, 0). Let u1 = v1 so that v2 , u1 = (2, 1, 0), (−3, 0, 1) = 2(−3) + 1 · 0 + 0 · 1 = −6 and ||u1 ||2 = (−3)2 + 02 + 12 = 10. v2 , u1 6 1 u1 = (2, 1, 0) + (−3, 0, 1) = (1, 5, 3). u2 = v2 − 2 ||u1 || 10 5 Thus, {(−3, 0, 1), (1, 5, 3)} is an orthogonal basis for E1 . ⎤ ⎡ ⎡ ⎤⎡ ⎤ −1 −2 3 v1 0 (b). If λ2 = 3 then (A − λI)v = 0 assumes the form ⎣ 1 −4 3 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ =⇒ v1 − v2 = 0 and 1 −2 1 v3 0 v2 − v3 = 0. The eigenspace corresponding to λ2 = 3 is E2 = {v ∈ R3 : v = r(1, 1, 1), r ∈ R}. A basis for E2 is {(1, 1, 1)}. To determine the orthogonality of the vectors, consider the following inner products: (−3, 0, 1), (1, 1, 1) = −3 + 0 + 1 = −2 = 0 and (1, 5, 3), (1, 1, 1) = 1 + 5 + 3 = 9 = 0. Thus, the vectors in E1 are not orthogonal to the vectors in E2 . ⎤ ⎡ ⎡ ⎤⎡ ⎤ −1 −1 1 v1 0 1 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ 30. (a). If λ1 = 2 then (A − λI)v = 0 assumes the form ⎣ −1 −1 1 1 −1 v3 0 =⇒ v1 + v2 − v3 = 0. The eigenspace corresponding to λ1 = 2 is E1 = {v ∈ R3 : v = r(−1, 1, 0) + s(1, 0, 1), r, s ∈ R}. A basis for E1 is {(−1, 1, 0), (1, 0, 1)}. ⎡ (c)2017 Pearson Education. Inc. 461 Now apply the Gram-Schmidt process where v1 = (1, 0, 1), and v2 = (−1, 1, 0). Let u1 = v1 so that v2 , u1 = (−1, 1, 0), (1, 0, 1) = −1 · 1 + 1 · 0 + 0 · 1 = −1 and ||u1 ||2 = 12 + 02 + 12 = 2. 1 v2 , u1 1 u1 = (−1, 1, 0) + (1, 0, 1) = (−1, 2, 1). u2 = v2 − ||u1 ||2 2 2 Thus, {(1, 0, 1), (−1, 2, 1)} is an orthogonal basis for E1 . ⎤ ⎡ ⎡ ⎤⎡ ⎤ 2 −1 1 v1 0 2 1 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ . The eigenspace (b). If λ2 = −1 then (A − λI)v = 0 assumes the form ⎣ −1 1 1 2 v3 0 corresponding to λ2 = −1 is E2 = {v ∈ R3 : v = r(−1, −1, 1), r ∈ R}. A basis for E2 is {(−1, −1, 1)}. To determine the orthogonality of the vectors, consider the following inner products: (1, 0, 1), (−1, −1, 1) = −1 + 0 + 1 = 0 and (−1, 2, 1), (−1, −1, 1) = 1 − 2 + 1 = 0. Thus, the vectors in E1 are orthogonal to the vectors in E2 . 31. We are given that the eigenvalues of A are λ1 = 0 (multiplicity two), and λ2 = a + b + c. There are two cases to consider: λ1 = λ2 or λ1 = λ2 . ⎤ ⎡ ⎤⎡ a b c v1 If λ1 = λ2 then λ1 = 0 is of multiplicity three, and (A − λI)v = 0 assumes the form ⎣ a b c ⎦ ⎣ v2 ⎦ = a b c v3 ⎤ ⎡ 0 ⎣ 0 ⎦, or equivalently, 0 av1 + bv2 + cv3 = 0. (31.1) The only way to have three linearly independent eigenvectors for A is if a = b = c = 0. If λ1 = λ2 then λ1 = 0 is of multiplicity two, and λ2 = a + b + c = 0 are distinct eigenvalues. By Theorem 7.2.11, E1 must have dimension two for A to possess a complete set of eigenvectors. The system for determining the eigenvectors corresponding to λ1 = 0 is once more given by (31.1). Since we can choose two variables freely in (31.1), it follows that there are indeed two corresponding linearly independent eigenvectors. Consequently, A is nondefective in this case. 32. (a). Setting λ = 0 in (7.2.4), we have p(λ) = det(A − λI) = det(A), and in (7.2.5), we have p(λ) = p(0) = bn . Thus, bn = det(A). The value of det(A − λI) is the sum of products of its elements, one taken from each row and each column. Expanding det(A − λI) yields equation (7.2.5). The expression involving λn in p(λ) comes from the product, n (aii − λ), of the diagonal elements. All the remaining products of the determinant have degree not higher i=1 than n − 2, since, if one of the factors of the product is aij , where i = j, then this product cannot contain the factors λ − aii and λ − ajj . Hence, p(λ) = n - (aii − λ) + (terms of degree not higher than n − 2) i=1 so, p(λ) = (−1)n λn + (−1)n−1 (a11 + a22 + · · · + ann )λn−1 + · · · + an . Equating like coefficients from (5.7.5), it follows that b1 = (−1)n−1 (a11 + a22 + · · · + ann ). (b). Letting λ = 0, we have from (7.2.6) that p(0) = n - (λi − 0) or p(0) = i=1 (c)2017 Pearson Education. Inc. n i=1 λi , but from (7.2.5), p(0) = bn , 462 therefore bn = n - λi . Letting λ = 1, we have from (7.2.6) that i=1 p(λ) = n i=1 (λi − λ) = (−1)n n - (λ − λi ) = (−1)n [λn − (λ1 + λ2 + · · · + λn )λn−1 + · · · + bn ]. i=1 Equating like coefficients with (7.2.5), it follows that b1 = (−1)n−1 (λ1 + λ2 + · · · + λn ). (c). From (a), bn = det(A), and from (b), det(A) = λ1 λ2 · · · λn , so det(A) is the product of the eigenvalues of A. From (a), b1 = (−1)n−1 (a11 + a22 + · · · + ann ) and from (b), b1 = (−1)n−1 (λ1 + λ2 + · · · + λn ), thus a11 + a22 + · · · + ann = λ1 + λ2 + · · · + λn . That is, tr(A) is the sum of the eigenvalues of A. 33. The sum of the eigenvalues of A is tr(A) = (−1) + (−3) + 1 = −3 and the product of the eigenvalues of A (obtained using the arrow method for a 3 × 3 determinant) is det(A) = 3 + 32 + 0 − 0 − 16 + 12 = 31. 34. We have det(A) = 19 and tr(A) = 3, so the product of the eigenvalues of A is 19, and the sum of the eigenvalues of A is 3. 35. We have det(A) = −69 and tr(A) = 1, so the product of the eigenvalues of A is -69, and the sum of the eigenvalues of A is 1. 36. We have det(A) = −607 and tr(A) = 24, so the product of the eigenvalues of A is -607, and the sum of the eigenvalues of A is 24. 37. Note that Ei = ∅ since 0 belongs to Ei . Closure under Addition: Let v1 , v2 ∈ Ei . Then A(v1 + v2 ) = Av1 + Av2 = λi v1 + λi v2 = λi (v1 + v2 ) =⇒ v 1 + v 2 ∈ Ei . Closure under Scalar Multiplication: Let c ∈ C and v1 ∈ Ei . Then A(cv1 ) = c(Av1 ) = c(λi v1 ) = λi (cv1 ) =⇒ cv1 ∈ Ei . Thus, by Theorem 4.3.2, Ei is a subspace of C n . 38. The condition c1 v 1 + c 2 v 2 = 0 (38.1) =⇒ A(c1 v1 ) + A(c2 v2 ) = 0 =⇒ c1 Av1 + c2 Av2 = 0 =⇒ c1 (λ1 v1 ) + c2 (λ2 v2 ) = 0. (38.2) Substituting c2 v2 from (38.1) into (38.2) yields (λ1 − λ2 )c1 v1 = 0. Since λ2 = λ1 , we must have c1 v1 = 0, but v1 = 0, so c1 = 0. Substituting into (38.1) yields c2 = 0 also. Consequently, v1 and v2 are linearly independent. 39. Consider c1 v1 + c2 v2 + c3 v3 = 0. (39.1) (c)2017 Pearson Education. Inc. 463 If c1 = 0, then the preceding equation can be written as w1 + w2 = 0, where w1 = c1 v1 and w2 = c2 v2 + c3 v3 . But this would imply that {w1 , w2 } is linearly dependent, which would contradict Theorem 7.2.5 since w1 and w2 are eigenvectors corresponding to different eigenvalues. Consequently, we must have c1 = 0. But then (39.1) implies that c2 = c3 = 0 since {v1 , v2 } is a linearly independent set by assumption. Hence {v1 , v2 , v3 } is linearly independent. 40. λ1 = 1 (multiplicity 3), basis: {(0, 1, 1)}. 41. λ1 = 0 (multiplicity 2), basis: {(−1, 1, 0), (−1, 0, 1)}. λ2 = 3, basis: {(1, 1, 1)}. √ √ √ 42. λ1 = 2, basis: {(1, −2 2, 1)}. λ2 = 0, basis: {(1, 0, −1)}. λ3 = 7, basis: {( 2, 1, 2)}. 43. λ1 = −2, basis: {(2, 1, −4)}. λ2 = 3 (multiplicity 2), basis: {(0, 2, 1), (3, 11, 0)}. 44. λ1 = 0 (multiplicity 2), basis: {(0, 1, 0, −1), (1, 0, −1, 0)}. λ2 = 6, basis: {(1, 1, 1, 1)}. λ3 = −2, basis: {1, −1, 1, −1)}. 45. A has eigenvalues: 3 1 2 a + 8b2 , a+ 2 2 3 1 2 λ2 = a − a + 8b2 , 2 2 λ3 = 0. λ1 = Provided a = ±b, these eigenvalues are distinct, and therefore the matrix is nondefective. If a = b = 0, then the eigenvalue λ = 0 has multiplicity two. A basis for the corresponding eigenspace is {(−1, 0, 1), (−1, 1, 0)}. Since this is two-dimensional, the matrix is nondefective in this case. If a = −b = 0, then the eigenvalue λ = 0 once more has multiplicity two. A basis for the corresponding eigenspace is {(0, 1, 1), (1, 1, 0)}, therefore the matrix is nondefective in this case also. If a = b = 0, then A = 02 , so that λ = 0 (multiplicity three), and the corresponding eigenspace is all of R3 . Hence A is nondefective. 46. If a = b = 0, then A = 03 , which is nondefective. We now assume that at least one of either a or b is nonzero. A has eigenvalues: λ1 = b, λ2 = a − b, and λ3 = 3a + b. Provided a = 0, 2b, −b, the eigenvalues are distinct, and therefore A is nondefective. If a = 0, then the eigenvalue λ = b has multiplicity two. In this case, a basis for the corresponding eigenspace is {(1, 0, 1), (0, 1, 0)}, so that A is nondefective. If a = 2b, then the eigenvalue λ = b has multiplicity two. In this case, a basis for the corresponding eigenspace is {(−2, 1, 0), (−1, 0, 1)}, so that A is nondefective. If a = −b, then the eigenvalue λ = −2b has multiplicity two. In this case, a basis for the corresponding eigenspace is {(0, 1, 1), (1, 0, −1)}, so that A is nondefective. Solutions to Section 7.3 True-False Review: (a): TRUE. The terms “diagonalizable” and “nondefective” are synonymous. The diagonalizability of a matrix A hinges on the ability to form an invertible matrix S with a full set of linearly independent eigenvectors of the matrix as its columns. This, in turn, requires the original matrix to be nondefective. (c)2017 Pearson Education. Inc. 464 (b): TRUE. If we assume that A is diagonalizable, then there exists an invertible matrix S and a diagonal matrix D such that S −1 AS = D. Since A is invertible, we can take the inverse of each side of this equation to obtain D−1 = (S −1 AS)−1 = S −1 A−1 S, and since D−1 is still a diagonal matrix, this equation shows that A−1 is diagonalizable. 1 1 (c): FALSE. For instance, the matrices A = I2 and B = both have eigenvalue λ = 1 (with 0 1 multiplicity 2). However, A and B are not similar. [Reason: If A and B were similar, then S −1 AS = B for some invertible matrix S, but since A = I2 , this would imply that B = I2 , contrary to our choice of B above.] (d): FALSE. An n × n matrix is diagonalizable if and only if it has n linearly independent eigenvectors. Besides, every matrix actually has infinitely many eigenvectors, obtained by taking scalar multiples of a single eigenvector v. (e): TRUE. Assume A is an n × n matrix such that p(λ) = det(A − λI) has no repeated roots. This implies that A has n distinct eigenvalues. Corresponding to each eigenvalue, we can select an eigenvector. Since eigenvectors corresponding to distinct eigenvalues are linearly independent, this yields n linearly independent eigenvectors for A. Therefore, A is nondefective, and hence, diagonalizable. (f ): TRUE. Assuming that A is diagonalizable, then there exists an invertible matrix S and a diagonal matrix D such that S −1 AS = D. Therefore, D2 = (S −1 AS)2 = (S −1 AS)(S −1 AS) = S −1 ASS −1 AS = S −1 A2 S. Since D2 is still a diagonalizable matrix, this equation shows that A2 is diagonalizable. (g): TRUE. Since In−1 AIn = A, A is similar to itself. (h): TRUE. The sum of the dimensions of the eigenspaces of such a matrix is even, and therefore not equal to n. This means we cannot obtain n linearly independent eigenvectors for A, and therefore, A is defective (and not diagonalizable). (i): TRUE. Assume that c1 v1 + c2 v2 = 0. Left-multiplying the matrix A on both sides, we obtain c1 Av1 + c2 Av2 = A0 = 0. Using Av1 = λ1 v1 and Av2 = λ2 v2 , this becomes c1 λ1 v1 + c2 λ2 v2 = 0. On the other hand, multiplying the first equation through by λ1 , we obtain c1 λ1 v1 + c2 λ1 v2 = 0. Subtracting the last two equations, we obtain c2 (λ2 − λ1 )v2 = 0. Since v2 = 0 and λ2 = λ1 , we conclude that c2 = 0. Now the original equation implies that c1 = 0. Since c1 = c2 = 0, we conclude that {v1 , v2 } is linearly independent. (c)2017 Pearson Education. Inc. 465 Problems: 0 = 0 ⇐⇒ (−9 − λ)2 = 0 ⇐⇒ λ = −9 of multiplicity two. −9 − λ 0 0 0 v1 = =⇒ v1 = 0. If we let If λ = −9 then (A − λI)v = 0 assumes the form 0 v2 4 0 v2 = r ∈ R, then the solution set of the linear system is {(0, r) : r ∈ R}, so the eigenvectors corresponding to λ = −9 are v = r(0, 1), where r ∈ R. −9 − λ 1. det(A − λI) = 0 ⇐⇒ 4 A has only one linearly independent eigenvector, so by Theorem 7.3.4, A is not diagonalizable. −1 − λ −2 = 0 ⇐⇒ λ2 − λ − 6 = 0 ⇐⇒ (λ − 3)(λ + 2) = 0 ⇐⇒ λ = 3 or 2. det(A − λI) = 0 ⇐⇒ −2 2−λ λ = −2. A is diagonalizable because it has two distinct eigenvalues. 0 −4 −2 v1 = =⇒ 2v1 + v2 = 0. If we let If λ1 = 3 then (A − λI)v = 0 assumes the form 0 v2 −2 −1 v1 = r ∈ R, then the solution set of this system is {(−r, 2r) : r ∈ R}, so the eigenvectors corresponding to λ1 = 3 are v1 = r(−1, 2) where r ∈ R. 0 1 −2 v1 = =⇒ v1 − 2v2 = 0. If we let If λ2 = −2 then (A − λI)v = 0 assumes the form v2 0 −2 4 v2 = s ∈ R, then the solution set of this system is {(2s, s) : s ∈ R}, so the eigenvectors corresponding to λ2 = −2 are v2 = s(2, 1) where s ∈ R. −1 2 Thus, the matrix S = satisfies S −1 AS = diag(3, −2). 2 1 −7 − λ 4 3. det(A − λI) = 0 ⇐⇒ = 0 ⇐⇒ λ2 + 6λ + 9 = 0 ⇐⇒ (λ + 3)2 = 0 ⇐⇒ λ = −3 of −4 1−λ multiplicity two. 0 −4 4 v1 = =⇒ v1 − v2 = 0. If we let If λ = −3 then (A − λI)v = 0 assumes the form 0 v2 −4 4 v1 = r ∈ R, then the solution set of this system is {(r, r) : r ∈ R}, so the eigenvectors corresponding to λ = −3 are v = r(1, 1) where r ∈ R. A has only one linearly independent eigenvector, so by Theorem 7.3.4, A is not diagonalizable. 1−λ −8 = 0 ⇐⇒ λ2 + 6λ + 9 = 0 ⇐⇒ (λ + 3)2 = 0 ⇐⇒ λ = −3 of 4. det(A − λI) = 0 ⇐⇒ 2 −7 − λ multiplicity two. 0 4 −8 v1 = =⇒ v1 − 2v2 = 0. If we let If λ = −3 then (A − λI)v = 0 assumes the form v2 0 2 −4 v2 = r ∈ R, then the solution set of this system is {(2r, r) : r ∈ R}, so the eigenvectors corresponding to λ = −3 are v = r(2, 1) where r ∈ R. A has only one linearly independent eigenvector, so by Theorem 7.3.4, A is not diagonalizable. −λ 4 = 0 ⇐⇒ λ2 + 16 = 0 ⇐⇒ λ = ±4i. A is diagonalizable because it has 5. det(A − λI) = 0 ⇐⇒ −4 −λ two distinct eigenvalues. 0 4i 4 v1 = =⇒ v1 − iv2 = 0. If we let If λ = −4i then (A − λI)v = 0 assumes the form 0 v2 −4 4i v2 = r ∈ C, then the solution set of this system is {(ir, r) : r ∈ C}, so the eigenvectors corresponding to λ = −4i are v = r(i, 1) where r ∈ C. Since the entries of A are real, it follows from Theorem 7.1.8 that v2 = (−i, 1) is an eigenvector corresponding to λ = 4i. (c)2017 Pearson Education. Inc. 466 i −i satisfies S −1 AS = diag(−4i, 4i). 1 1 1−λ 0 0 = 0 ⇐⇒ (1 − λ)(λ + 4)(λ − 4) = 0 ⇐⇒ λ = 1, λ = −4 or 3−λ 7 6. det(A − λI) = 0 ⇐⇒ 0 1 1 −3 − λ λ = 4. ⎤ ⎡ ⎡ ⎤⎡ ⎤ 0 0 0 v1 0 7 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ If λ = 1 then (A − λI)v = 0 assumes the form ⎣ 0 2 1 1 −4 v3 0 =⇒ 2v1 − 15v3 = 0 and 2v2 + 7v3 = 0. If we let v3 = 2r where r ∈ R, then the solution set of this system is {(15r, −7r, 2r) : r ∈ R} so the eigenvectors corresponding to λ ⎤ v1 ⎡= r(15, ⎡ ⎤= ⎡ 1 are ⎤ −7, 2) where r ∈ R. 5 0 0 v1 0 If λ = −4 then (A − λI)v = 0 assumes the form ⎣ 0 7 7 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ 1 1 1 v3 0 =⇒ v1 = 0 and v2 + v3 = 0. If we let v2 = s ∈ R, then the solution set of this system is {(0, s, −s) : s ∈ R} so the eigenvectors corresponding to λ = −4 are⎡ v2 = s(0, 1, −1)⎤where s⎤∈ R.⎡ ⎡ ⎤ −3 0 0 v1 0 7 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ If λ = 4 then (A − λI)v = 0 assumes the form ⎣ 0 −1 1 1 −7 v3 0 =⇒ v1 = 0 and v2 − 7v3 = 0. If we let v3 = t ∈ R, then the solution set of this system is {(0, 7t, t) : t ∈ R} so the eigenvectors corresponding to λ⎤= 4 are v3 = t(0, 7, 1) where t ∈ R. ⎡ 15 0 0 1 ⎦ satisfies S −1 AS = diag(1, 4, −4). Thus, the matrix S = ⎣ −7 7 2 1 −1 1−λ −2 0 = 0 ⇐⇒ (λ + 1)3 = 0 ⇐⇒ λ = −1 of multiplicity three. −3 − λ 0 7. det(A − λI) = 0 ⇐⇒ 2 2 −2 −2 − λ ⎤ ⎡ ⎤⎡ ⎡ ⎤ 2 −2 0 v1 0 If λ = −1 then (A − λI)v = 0 assumes the form ⎣ 2 −2 0 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ =⇒ v1 − v2 = 0 and 2 −2 0 v3 0 v3 ∈ R. If we let v2 = r ∈ R and v3 = s ∈ R, then the solution set of this system is {(r, r, s) : r, s ∈ R}, so the eigenvectors corresponding to λ = −1 are v1 = r(1, 1, 0) and v2 = s(0, 0, 1) where r, s ∈ R. A has only two linearly independent eigenvectors, so by Theorem 7.3.4, A is not diagonalizable. −λ −2 −2 8. det(A − λI) = 0 ⇐⇒ −2 −λ −2 = 0 ⇐⇒ (λ − 2)2 (λ + 4) = 0 ⇐⇒ λ = −4 or λ = 2 of multiplicity −2 −2 −λ two. ⎤ ⎡ ⎤ ⎡ ⎤⎡ 4 −2 −2 v1 0 4 −2 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ If λ = −4 then (A − λI)v = 0 assumes the form ⎣ −2 0 −2 −2 4 v3 =⇒ v1 − v3 = 0 and v2 − v3 = 0. If we let v3 = r ∈ R, then the solution set of this system is {(r, r, r) : r ∈ R} so the eigenvectors corresponding to λ = −4 are⎡ v1 = r(1, 1, 1) where ⎤ R. ⎡ ⎤⎡ r ∈ ⎤ −2 −2 −2 v1 0 If λ = 2 then (A − λI)v = 0 assumes the form ⎣ −2 −2 −2 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ =⇒ v1 + v2 + v3 = 0. If −2 −2 −2 v3 0 we let v2 = s ∈ R and v3 = t ∈ R, then the solution set of this system is {(−s − t, s, t) : s, t ∈ R}, so two linearly independent eigenvectors corresponding to λ = 2 are v2 = s(−1, 1, 0) and v3 = t(−1, 0, 1). Thus, the matrix S = (c)2017 Pearson Education. Inc. 467 ⎡ ⎤ 1 −1 −1 0 1 ⎦ satisfies S −1 AS = diag(−4, 2, 2). Thus, the matrix S = ⎣ 1 1 1 0 4 4 = 0 ⇐⇒ λ2 (λ−3) = 0 ⇐⇒ λ = 3, or λ = 0 of multiplicity 4−λ ⎤ ⎡ ⎡ ⎤⎡ ⎤ −5 1 4 v1 0 If λ = 3 then (A − λI)v = 0 assumes the form ⎣ −2 −2 4 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ −2 1 1 v3 0 =⇒ v1 − v3 = 0 and v2 − v3 = 0. If we let v3 = r ∈ R, then the solution set of this system is {(r, r, r) : r ∈ R}, so the eigenvectors corresponding to λ = 3 are ⎡ v1 = r(1, 1, 1)⎤where r⎤ ∈ R. ⎡ ⎡ ⎤ −2 1 4 v1 0 If λ = 0 then (A − λI)v = 0 assumes the form ⎣ −2 1 4 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ =⇒ −2v1 + v2 + 4v3 = 0. If −2 1 4 v3 0 we let v1 = s ∈ R and v3 = t ∈ R, then the solution set of this system is {(s, 2s − 4t, t) : s, t ∈ R}, so two linearly independent eigenvectors corresponding to λ = 0 are v2 = s(1, 2, 0) and v3 = t(0, −4, 1). ⎡ ⎤ 1 1 0 Thus, the matrix S = ⎣ 1 2 −4 ⎦ satisfies S −1 AS = diag(3, 0, 0). 1 0 1 −2 − λ 9. det(A−λI) = 0 ⇐⇒ −2 −2 two. 2−λ 10. det(A − λI) = 0 ⇐⇒ 0 2 multiplicity two. 1 1−λ 1 0 0 = 0 ⇐⇒ (λ − 2)(λ − 1)2 = 0 ⇐⇒ λ = 2 or λ = 1 of 1−λ ⎤ ⎡ ⎤⎡ ⎡ ⎤ 1 0 0 v1 0 0 0 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ =⇒ v1 = v2 = 0 and If λ = 1 then (A − λI)v = 0 assumes the form ⎣ 0 2 −1 0 v3 0 v3 ∈ R. If we let v3 = r ∈ R then the solution set of this system is {(0, 0, r) : r ∈ R}, so there is only one corresponding linearly independent eigenvector. Hence, by Theorem 7.3.4, A is not diagonalizable. 4−λ 0 0 −1 − λ −1 = 0 ⇐⇒ (λ2 + 1)(λ − 4) = 0 ⇐⇒ λ = 4, or λ = ±i. 11. det(A − λI) = 0 ⇐⇒ 3 0 2 1−λ ⎤ ⎡ ⎡ ⎤⎡ ⎤ 0 0 0 v1 0 If λ = 4 then (A − λI)v = 0 assumes the form ⎣ 3 −5 −1 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ 0 2 −3 v3 0 =⇒ 6v1 − 17v3 = 0 and 2v2 − 3v3 = 0. If we let v3 = 6r ∈ C, then the solution set of this system is {(17r, 9r, 6r) : r ∈ C}, so the eigenvectors corresponding to λ = 4 are v1⎤=⎡r(17,⎤9, 6)⎡where ⎡ ⎤ r ∈ C. 4−i 0 0 v1 0 −1 − i −1 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ =⇒ v1 = 0 and If λ = i then (A − λI)v = 0 assumes the form ⎣ 3 0 2 1−i v3 0 2v2 + (1 − i)v3 = 0. If we let v3 = −2s ∈ C, then the solution set of this system is {(0, (1 − i)s, 2s) : s ∈ C}, so the eigenvectors corresponding to λ = i are v2 = s(0, 1 − i, 2) where s ∈ C. Since the entries of A are real, v3 = t(0, 1 + i, 2) where⎡ t ∈ C are the eigenvectors corresponding to λ = −i by Theorem 7.1.8. ⎤ 17 0 0 Thus, the matrix S = ⎣ 9 1 − i 1 + i ⎦ satisfies S −1 AS = diag(4, i, −i). 6 2 2 0 1−λ −1 (c)2017 Pearson Education. Inc. 468 −λ 2 12. det(A − λI) = 0 ⇐⇒ −2 −λ 1 2 −1 −2 = 0 ⇐⇒ λ(λ2 + 9) = 0 ⇐⇒ λ = 0, or λ = ±3i. −λ ⎤ ⎡ ⎡ ⎤⎡ ⎤ 0 2 −1 v1 0 If λ = 0 then (A − λI)v = 0 assumes the form ⎣ −2 0 −2 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ 1 2 0 v3 0 =⇒ v1 + v3 = 0 and 2v2 − v3 = 0. If we let v3 = 2r ∈ C, then the solution set of this system is {(−2r, r, 2r) : r ∈ C}, so the eigenvectors corresponding to λ = 0 are ⎡ ⎤ ⎡v1 =⎤r(−2, ⎡ 1, 2) ⎤ where r ∈ C. 3i 2 −1 v1 0 If λ = −3i then (A − λI)v = 0 assumes the form ⎣ −2 3i −2 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ 1 2 3i v3 0 =⇒ 5v1 +(−4+3i)v3 = 0 and 5v2 +(2+6i)v3 = 0. If we let v3 = 5s ∈ C, then the solution set of this system is {((4−3i)s, (−2−6i)s, 5s) : s ∈ C}, so the eigenvectors corresponding to λ = −3i are v2 = s(4−3i, −2−6i, 5) where s ∈ C. Since the entries of A are real, v3 = t(4 + 3i, −2 + 6i, 5) where t ∈ C are the eigenvectors corresponding to λ = 3i ⎡ by Theorem 7.1.8. ⎤ −2 4 + 3i 4 − 3i Thus, the matrix S = ⎣ 1 −2 + 6i −2 − 6i ⎦ satisfies S −1 AS = diag(0, 3i, −3i). 2 5 5 1−λ −2 0 1−λ 0 = 0 ⇐⇒ (λ − 3)2 (λ + 1) = 0 ⇐⇒ λ = −1 or λ = 3 of 13. det(A − λI) = 0 ⇐⇒ −2 0 0 3−λ multiplicity two. ⎤ ⎡ ⎤ ⎡ ⎤⎡ 2 −2 0 v1 0 2 0 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ If λ = −1 then (A − λI)v = 0 assumes the form ⎣ −2 0 0 0 4 v3 =⇒ 2v1 − v2 = 0 and v3 = 0. If we let v1 = r ∈ R, then the solution set of this system is {(r, r, 0) : r ∈ R} so the eigenvectors corresponding to λ = −1 are v⎡1 = r(1, 1, 0) where ⎤ ⎡ r ∈⎤R. ⎡ ⎤ −2 −2 0 v1 0 If λ = 3 then (A − λI)v = 0 assumes the form ⎣ −2 −2 0 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ =⇒ v1 + v2 = 0 and 0 0 0 v3 0 v3 ∈ R. If we let v2 = s ∈ R and v3 = t ∈ R, then the solution set of this system is {(−s, s, t) : s, t ∈ R}, so the eigenvectors corresponding to λ ⎤ = 3 are v2 = s(−1, 1, 0) and v3 = t(0, 0, 1). ⎡ 1 −1 0 1 0 ⎦ satisfies S −1 AS = diag(−1, 3, 3). Thus, the matrix S = ⎣ 1 0 0 1 −1 − λ 1 0 0 0 −1 − λ 0 0 = 0 ⇐⇒ (−1 − λ)3 (1 − λ) = 0 ⇐⇒ λ = −1 14. det(A − λI) = 0 ⇐⇒ 0 0 −1 − λ 0 0 0 0 1−λ of multiplicity three and λ = 1 of multiplicity one. ⎤ ⎤⎡ ⎤ ⎡ ⎡ v1 0 0 1 0 0 ⎢ 0 ⎥ ⎢ 0 0 0 0 ⎥ ⎢ v2 ⎥ ⎥ ⎥⎢ ⎥ ⎢ If λ = −1 then (A − λI)v = 0 assumes the form ⎢ ⎣ 0 0 0 0 ⎦ ⎣ v3 ⎦ = ⎣ 0 ⎦ =⇒ v2 = v4 = 0. 0 0 0 0 2 v4 If we let v1 = r and v3 = s be free variables (r, s ∈ R), then the solution set of the linear system is {(r, 0, s, 0) : r, s ∈ R}, so the eigenvectors corresponding to λ = −1 are v = r(1, 0, 0, 0) + s(0, 0, 1, 0), where r, s ∈ R. (c)2017 Pearson Education. Inc. 469 ⎤⎡ ⎤ ⎤ ⎡ v1 −2 1 0 0 0 ⎢ ⎥ ⎢ 0 −2 ⎥ ⎢ 0 0 ⎥ ⎥ ⎢ v2 ⎥ = ⎢ 0 ⎥ =⇒ v1 = v2 = If λ = 1 then (A − λI)v = 0 assumes the form ⎢ ⎦ ⎣ ⎣ 0 ⎦ ⎣ 0 −2 0 0 ⎦ v3 0 0 0 0 0 v4 v3 = 0. If we let v4 = t, (t ∈ R), then the solution set of the linear system is {(0, 0, 0, t) : t ∈ R}, so the eigenvectors corresponding to λ = 1 are v = t(0, 0, 0, 1), where t ∈ R. We have collected only three linearly independent eigenvectors for the 4 × 4 matrix A, so by Theorem 7.3.4, A is not diagonalizable. ⎡ 15. Since all rows of A are multiples of the first row, this matrix is not invertible, and we see that λ = 0 must be an eigenvalue. ⎤⎡ ⎤ ⎡ ⎤ ⎡ v1 1 2 3 4 0 ⎢ −1 −2 −3 −4 ⎥ ⎢ v2 ⎥ ⎢ 0 ⎥ ⎥⎢ ⎥ =⇒ v1 + 2v2 + ⎥=⎢ If λ = 0 then (A − λI)v = 0 assumes the form ⎢ ⎣ 2 4 6 8 ⎦ ⎣ v3 ⎦ ⎣ 0 ⎦ −2 −4 −6 −8 0 v4 3v3 + 4v4 = 0. If we let v2 = r, v3 = s, and v4 = t, where r, s, t ∈ R, then v1 = −2r − 3s − 4t. Thus, the solution set of the linear system is {(−2r − 3s − 4t, r, s, t) : r, s, t ∈ R}, so the eigenvectors corresponding to λ = 0 are v = r(−2, 1, 0, 0) + s(−3, 0, 1, 0) + t(−4, 0, 0, 1), where r, s, t ∈ R. Thus, we obtain three linearly independent eigenvectors, v1 = (−2, 1, 0, 0), v2 = (−3, 0, 1, 0), and v3 = (−4, 0, 0, 1). Since A is a 4 × 4 matrix, we know that A has four eigenvalues (including multiplicities). Moreover, by Problem 32 in Section 7.2, we know that the sum of the four eigenvalues of A is tr(A) = 1+(−2)+6+(−8) = −3. Since λ = 0 occurs with multiplicity three, we must⎡have the final eigenvalue ⎤ ⎡ λ =⎤−3. ⎡ ⎤ v1 4 2 3 4 0 ⎢ ⎥ ⎢ −1 ⎥ ⎢ 1 −3 −4 ⎥ ⎥ ⎢ v2 ⎥ = ⎢ 0 ⎥. A rowIf λ = −3 then (A − λI)v = 0 assumes the form ⎢ ⎦ ⎣ ⎣ 2 ⎦ ⎣ 4 9 8 0 ⎦ v3 −2 −4 −6 −5 0 v4 ⎤ ⎡ 1 −1 3 4 ⎢ 0 1 − 32 −2 ⎥ ⎥. Setting v4 = t, by echelon form for the coefficient matrix of this linear system is ⎢ ⎣ 0 0 1 1 ⎦ 0 0 0 0 back substitution we obtain v3 = −t, v2 = 12 t, and then v1 = − 12 t. Thus, the solution set of the linear system is {(− 12 t, 12 t, −t, t) : t ∈ R}, so we obtain one linearly independent eigenvector corresponding to λ = −3, namely v4 = (− 12 , 12 , −1, 1). Altogether, we have obtained four linearly independent eigenvectors of A, which implies that A is diagonalizable. We construct S by placing the four eigenvectors we have obtained in the columns: ⎡ ⎤ −2 −3 −4 − 21 1 ⎥ ⎢ 1 0 0 2 ⎥. S=⎢ ⎣ 0 1 0 −1 ⎦ 0 0 1 1 This matrix satisfies S −1 AS = diag(0, 0, 0, −3). 16. λ1 = 2 (multiplicity 2), basis for eigenspace: {(−3, 1, 0), (3, 0, 1)}. λ2 = 1, basis for eigenspace: {(1, 2, 2)}. ⎡ ⎤ −3 3 1 Set S = ⎣ 1 0 2 ⎦. Then S −1 AS = diag(2, 2, 1). 0 1 2 17. λ1 = 0 (multiplicity 2), basis for eigenspace: {(0, 1, 0, −1), (1, 0, −1, 0)}. (c)2017 Pearson Education. Inc. 470 for eigenspace: ⎤ {(1, 1, 1, 1)}. λ3 = 10, basis for eigenspace: {(−1, 1, −1, 1)}. λ2 = 2, basis ⎡ 0 1 1 −1 ⎢ 1 0 1 1 ⎥ −1 ⎥ Set S = ⎢ ⎣ 0 −1 1 −1 ⎦. Then S AS = diag(0, 0, 2, 10). −1 0 1 1 1 4 18. The given system can be written as x = Ax, where A = . 2 3 A has eigenvalues λ1 = −1, λ2 = 5 with corresponding linearly independent eigenvectors v1 = (−2, 1) and −2 1 , then S −1 AS = diag(−1, 5), therefore, under the transformation v2 = (1, 1). If we set S = 1 1 x = Sy, the given system of differential equations simplifies to y1 y2 = −1 0 0 5 y1 y2 . Hence, y1 = −y1 and y2 = 5y2 . Integrating these equations, we obtain y1 (t) = c1 e−t , y2 (t) = c2 e5t . Returning to the original variables, we have −2c1 e−t + c2 e5t −2 1 c1 e−t = . x = Sy = c2 e5t c1 e−t + c2 e5t 1 1 Consequently, x1 (t) = −2c1 e−t + c2 e5t and x2 (t) = c1 e−t + c2 e5t . 6 −2 . 19. The given system can be written as x = Ax, where A = −2 6 A has eigenvalues λ1 = 4, λ2 = 8 withcorresponding linearly independent eigenvectors v1 = (1, 1) and 1 −1 , then S −1 AS = diag(4, 8), therefore, under the transformation v2 = (−1, 1). If we set S = 1 1 x = Sy, the given system of differential equations simplifies to y1 y2 = 4 0 0 8 y1 y2 . Hence, y1 = 4y1 and y2 = 8y2 . Integrating these equations, we obtain y1 (t) = c1 e4t , y2 (t) = c2 e8t . Returning to the original variables, we have c1 e4t − c2 e8t 1 −1 c1 e4t . = x = Sy = c2 e8t c1 e4t + c2 e8t 1 1 Consequently, x1 (t) = c1 e4t − c2 e8t and x2 (t) = c1 e4t + c2 e8t . 20. The given system can be written as x = Ax, where A = 6 −1 −5 2 (c)2017 Pearson Education. Inc. . 471 A has eigenvalues λ1 = 1 and λ2 = 7 with corresponding linearly independent eigenvectors v1 = (1, 5) 1 −1 , then S −1 AS = diag(1, 7). Therefore, under the transformation and v2 = (−1, 1). If we set S = 5 1 x = Sy, the given system of differential equations simplifies to y1 1 0 y1 = . 0 7 y2 y2 Hence, y1 = y1 and y2 = 7y2 . Integrating these equations, we obtain y1 (t) = c1 et , y2 (t) = c2 e7t . Returning to the original variables, we have 1 −1 c1 et c1 et − c2 e7t x = Sy = = . 5 1 c2 e7t 5c1 et + c2 e7t Consequently, x1 (t) = c1 et − c2 e7t and x2 (t) = 5c1 et + c2 e7t . −12 −7 21. The given system can be written as x = Ax, where A = . 16 10 A has eigenvalues λ1 = 2, λ2= −4 withcorresponding linearly independent eigenvectors v1 = (1, −2) and 1 7 , then S −1 AS = diag(2, −4), therefore, under the transformation v2 = (7, −8). If we set S = −2 −8 x = Sy, the given system of differential equations simplifies to 2 0 y1 y1 = . y2 0 −4 y2 Hence, y1 = 2y1 and y2 = −4y2 . Integrating these equations, we obtain y1 (t) = c1 e2t , y2 (t) = c2 e−4t . Returning to the original variables, we have 1 7 c1 e2t c1 e2t + 7c2 e−4t x = Sy = = . −2 −8 c2 e−4t −2c1 e2t − 8c2 e−4t Consequently, x1 (t) = c1 e2t + 7c2 e−4t and x2 (t) = −2c1 e2t − 8c2 e−4t . 0 1 22. The given system can be written as x = Ax, where A = . −1 0 A has eigenvalues λ1 = i, λ2 = −i with corresponding linearly independent eigenvectors v1 = (1, i) and 1 1 , then S −1 AS = diag(i, −i), therefore, under the transformation v2 = (1, −i). If we set S = i −i x = Sy, the given system of differential equations simplifies to i 0 y1 y1 = . 0 −i y2 y2 Hence, y1 = iy1 and y2 = −iy2 . Integrating these equations, we obtain y1 (t) = c1 eit , y2 (t) = c2 e−it . (c)2017 Pearson Education. Inc. 472 Returning to the original variables, we have c1 eit + c2 e−it 1 1 c1 eit = . x = Sy = c2 e−it i(c1 eit − c2 e−it ) i −i Consequently, x1 (t) = c1 eit + c2 e−it and x2 (t) = i(c1 eit − c2 e−it ). Using Euler’s formula, these expressions can be written as x1 (t) = (c1 + c2 ) cos t + i(c1 − c2 ) sin t, x2 (t) = i(c1 − c2 ) cos t − (c1 + c2 ) sin t, or equivalently, x1 (t) = a cos t + b sin t, x2 (t) = b cos t − a sin t, where a = c1 + c2 , and b = i(c1 − c2 ). ⎡ ⎤ 3 −4 −1 23. The given system can be written as x = Ax, where A = ⎣ 0 −1 −1 ⎦. 0 −4 2 = 3 with correA has eigenvalue λ1 = −2 with corresponding eigenvector v1 = (1, 1, 1), and eigenvalue λ2 ⎡ ⎤ 1 1 0 1 ⎦, sponding linearly independent eigenvectors v2 = (1, 0, 0) and v3 = (0, 1, −4). If we set S = ⎣ 1 0 1 0 −4 then S −1 AS = diag(−2, 3, 3), therefore, under the transformation x = Sy, the given system of differential equations simplifies to ⎡ ⎤ ⎡ ⎤ ⎤⎡ y1 −2 0 0 y1 ⎣ y2 ⎦ = ⎣ 0 3 0 ⎦ ⎣ y2 ⎦ . y3 0 0 3 y3 Hence, y1 = −2y1 , y2 = 3y2 , and y3 = 3y3 . Integrating these equations, we obtain y1 (t) = c1 e−2t , y2 (t) = c2 e3t , y3 (t) = c3 e3t . Returning to the original variables, we have ⎤ ⎡ ⎤ ⎡ ⎤⎡ 1 1 0 c1 e−2t c1 e−2t + c2 e3t 1 ⎦ ⎣ c2 e3t ⎦ = ⎣ c1 e−2t + c3 e3t ⎦ . x = Sy = ⎣ 1 0 c3 e3t c1 e−2t − 4c3 e3t 1 0 −4 Consequently, x1 (t) = c1 e−2t + c2 e3t , x2 (t) = c1 e−2t + c3 e3t , and c1 e−2t − 4c3 e3t . ⎡ ⎤ 1 1 −1 1 ⎦. 24. The given system can be written as x = Ax, where A = ⎣ 1 1 −1 1 1 A has eigenvalue λ1 = −1 with corresponding eigenvector v1 = (−1, 1, −1), and eigenvalue λ⎡2 = 2 with corre⎤ −1 0 1 0 ⎦, sponding linearly independent eigenvectors v2 = (0, 1, 1) and v3 = (1, 0, −1). If we set S = ⎣ 1 1 −1 1 −1 then S −1 AS = diag(−1, 2, 2), therefore, under the transformation x = Sy, the given system of differential equations simplifies to ⎡ ⎤ ⎡ ⎤ ⎤⎡ y1 −1 0 0 y1 ⎣ y2 ⎦ = ⎣ 0 2 0 ⎦ ⎣ y2 ⎦ . y3 0 0 2 y3 (c)2017 Pearson Education. Inc. 473 Hence, y1 = −y1 , y2 = 2y2 , and y3 = 2y3 . Integrating these equations, we obtain y1 (t) = c1 e−t , y2 (t) = c2 e2t , y3 (t) = c3 e2t . Returning to the original variables, we have ⎤ ⎡ ⎤ ⎡ ⎤⎡ −1 0 1 c1 e−t −c1 e−t + c3 e2t ⎦. c1 e−t + c2 e2t 0 ⎦ ⎣ c2 e2t ⎦ = ⎣ x = Sy = ⎣ 1 1 2t −1 1 −1 c3 e −c1 e−t + c2 e2t − c3 e2t Consequently, x1 (t) = −c1 e−t + c3 e2t , x2 (t) = c1 e−t + c2 e2t , and −c1 e−t + (c2 − c3 )e2t . 25. A2 = (SDS −1 )(SDS −1 ) = SD(S −1 S)DS −1 = SDIn DS −1 = SD2 S −1 . We now use mathematical induction to establish the general result. Suppose that for k = m > 2 that Am = SDm S −1 . Then Am+1 = AAm = (SDS −1 )(SDm S −1 = SD(S −1 S)Dm S −1 ) = SDm+1 S −1 . It follows by mathematical induction that Ak = SDk S −1 for k = 1, 2, . . . 26. Let A = diag(a1 , a2 , . . . , an ) and let B = diag(b1 , b2 , . . . , bn ). Then from the index form of the matrix product, n , 0, if i = j, (AB)ij = aik bkj = aii bij = if i = j. a i bi , k=1 Consequently, AB = diag(a1 b1 , a2 b2 , . . . , an bn ). Applying this result to the matrix D = diag(λ1 , λ2 , . . . , λk ), it follows directly that Dk = diag(λk1 , λk2 , . . . , λkn ). 27. The matrix A has eigenvalues λ1 = 5, λ 2 = −1, with corresponding eigenvectors v1 = (1, −3) and 1 2 , then S −1 AS = D, where D = diag(5, −1). v2 = (2, −3). Thus, if we set S = −3 −3 Equivalently, A = SDS −1 . It follows from the results of the previous two examples that −127 −84 1 2 125 0 −1 − 23 3 3 −1 A = SD S = = , 1 378 251 −3 −3 0 −1 1 3 whereas 5 5 A = SD S −1 = 1 2 −3 −3 3125 0 0 −1 −1 1 − 23 1 3 = −3127 −2084 9378 6251 . 28. (a). This is self-evident from matrix multiplication. Another perspective on this is that when we multiply a matrix B on the left by a diagonal matrix√D, the √ ith√diagonal element √ ith row of B gets multiplied by the of D. Thus, if we multiply the ith row of D, λ , by the ith diagonal element of D, i √ √ √ λi , the result in √ √ the ith row of the product is λi λi = λi . Therefore, D D = D, which means that D is a square root of D. (b). We have √ √ √ √ √ (S DS −1 )2 = (S DS −1 )(S DS −1 ) = S( DI D)S −1 = SDS −1 = A, as required. (c)2017 Pearson Education. Inc. 474 (c). We begin by diagonalizing A. We have det(A − λI) = det 6−λ −3 −2 7−λ = (6 − λ)(7 − λ) − 6 = λ2 − 13λ + 36 = (λ − 4)(λ − 9), so the eigenvalues of A are λ = 4 and λ = 9. An eigenvector of A corresponding to λ = 4 is 2 . Thus, we can form eigenvector of A corresponding to λ = 9 is −3 S= √ We take D = 2 0 0 3 root of A is given by 1 1 2 −3 and D= 4 0 . A fast computation shows that S √ √ A = S DS −1 = −1 12/5 −3/5 = 0 9 , and an . 3/5 2/5 1/5 −1/5 −2/5 13/5 1 1 . By part (b), one square . Directly squaring this result confirms this matrix as a square root of A. 29. (a). Show: A ∼ A. The identity matrix, I, is invertible, and I = I −1 . Since IA = AI =⇒ A = I −1 AI, it follows that A ∼ A. (b). Show: A ∼ B ⇐⇒ B ∼ A. A ∼ B =⇒ there exists an invertible matrix S such that B = S −1 AS =⇒ A = SBS −1 = (S −1 )−1 BS −1 . But S −1 is invertible since S is invertible. Consequently, B ∼ A. (c). Show A ∼ B and B ∼ C ⇐⇒ A ∼ C. A ∼ B =⇒ there exists an invertible matrix S such that B = S −1 AS; moreover, B ∼ C =⇒ there exists an invertible matrix P such that C = P −1 BP . Thus, C = P −1 BP = P −1 (S −1 AS)P = (P −1 S −1 )A(SP ) = (SP )−1 A(SP ) where SP is invertible. Therefore A ∼ C. 30. Let A ∼ B mean that A is similar to B. Therefore, we must show: A ∼ B ⇐⇒ AT ∼ B T . A ∼ B =⇒ there exists an invertible matrix S such that B = S −1 AS =⇒ B T = (S −1 AS)T = S T AT (S −1 )T = S T AT (S T )−1 . S T is invertible because S is invertible, [since det(S) = det(S T )]. Thus, AT ∼ B T . 31. We are given that Av = λv and B = S −1 AS. B(S −1 v) = (S −1 AS)(S −1 v) = S −1 A(SS −1 )v = S −1 AIv = S −1 Av = S −1 (λv) = λ(S −1 v). Hence, S −1 v is an eigenvector of B corresponding to the eigenvalue λ. 32. (a). S −1 AS = diag(λ1 , λ2 , . . . , λn ) =⇒ det(S −1 AS) = λ1 λ2 · · · λn =⇒ det(A) det(S −1 ) det(S) = λ1 λ2 · · · λn =⇒ det(A) = λ1 λ2 · · · λn . Since all eigenvalues are nonzero, it follows that det(A) = 0. Consequently, A is invertible. −1 (b). S −1 AS = diag(λ1 , λ2 , . . . , λn ) =⇒ [S −1 AS] = [diag(λ1 , λ2 , . . . , λn )]−1 1 1 1 =⇒ S −1 A−1 (S −1 )−1 = diag , ,..., λ1 λ 2 λ n 1 1 1 . , ,..., =⇒ S −1 A−1 S = diag λ1 λ2 λn (c)2017 Pearson Education. Inc. 475 33. (a). S −1 AS = diag(λ1 , λ2 , . . . , λn ) =⇒ (S −1 AS)T = [diag(λ1 , λ2 , . . . , λn )]T =⇒ S T AT (S −1 )T = diag(λ1 , λ2 , . . . , λn ) =⇒ S T AT (S T )−1 = diag(λ1 , λ2 , . . . , λn ). Since we have that Q = (ST )−1 , this implies that Q−1 AT Q = diag(λ1 , λ2 , . . . , λn ). (b). Let MC = [v1 , v2 , v3 , . . . , vn ] where MC denotes the matrix of cofactors of S. We see from part (a) that AT is nondefective, which means it possesses a complete set of eigenvectors. Also from part (a), Q−1 AT Q = diag(λ1 , λ2 , . . . , λn ) where Q = (S T )−1 , so AT Q = Q diag(λ1 , λ2 , . . . , λn ) (33.1) If we let MC denote the matrix of cofactors of S, then T MC adj(S) MCT MC MCT =⇒ Q = S −1 = =⇒ (S T )−1 = = =⇒ (S −1 )T = . det(S) det(S) det(S) det(S) det(S) Substituting this result into Equation (33.1), we obtain MC MC AT = diag(λ1 , λ2 , . . . , λn ) det(S) det(S) T =⇒ A MC = MC diag(λ1 , λ2 , . . . , λn ) =⇒ AT [v1 , v2 , v3 , . . . , vn ] = [v1 , v2 , v3 , . . . , vn ] diag(λ1 , λ2 , . . . , λn ) =⇒ [AT v1 , AT v2 , AT v3 , . . . , AT vn ] = [λ1 v1 , λ2 v2 , λ3 v3 , . . . , λn vn ] =⇒ AT vi = λvi for i ∈ {1, 2, 3, . . . , n}. Hence, the column vectors of MC are linearly independent eigenvectors of AT . −2 − λ 4 = 0 ⇐⇒ (λ + 3)(λ − 2) = 0 ⇐⇒ λ1 = −3 or λ2 = 2. 34. det(A − λI) = 0 ⇐⇒ 1 1−λ If λ1 = −3 the corresponding eigenvectors are of the form v1 = r(−4, 1) where r ∈ R. If λ2 = 2 the corresponding eigenvectors are of the form v2 = s(1, 1) where s ∈R. −4 1 Thus, a complete set of eigenvectors is {(−4, 1), (1, 1)} so that S = . If MC denotes the matrix 1 1 1 −1 . Consequently, from Problem 33, (1, −1) is an eigenvector of cofactors of S, then MC = −1 −4 corresponding to λ = −3 and (−1, −4) is an eigenvector corresponding to λ = 2 for the matrix AT . λ 1 −1 35. S AS = Jλ ⇐⇒ AS = SJλ ⇐⇒ A[v1 , v2 ] = [v1 , v2 ] 0 λ =⇒ [Av1 , Av2 ] = [λv1 , v1 + λv2 ] =⇒ Av1 = λv1 and Av2 = v1 + λv2 =⇒ (A − λI)v1 = 0 and (A − λI)v2 = v1 . 2−λ 1 36. det(A − λI) = 0 ⇐⇒ = 0 ⇐⇒ (λ − 3)2 = 0 ⇐⇒ λ1 = 3 of multiplicity two. −1 4−λ If λ1 = 3 the corresponding eigenvectors are of the v1 = r(1, 1) where r ∈ R. Consequently, A does not have a complete set of eigenvectors, so itis a defective matrix. 3 1 By the preceding problem, J3 = is similar to A. Hence, there exists S = [v1 , v2 ] such that 0 3 −1 S AS = J3 . From the first part of the problem, we can let v1 = (1, 1). Now consider (A − λI)v2 = v1 where v1 = (a, b) for a, b ∈ R. Upon substituting, we obtain −1 1 a 1 = =⇒ −a + b = 1. −1 1 b 1 1 b−1 1 −1 Thus, S takes the form S = where b ∈ R, and if b = 0, then S = . 1 b 1 0 (c)2017 Pearson Education. Inc. 476 ⎡ 37. λ S −1 AS = ⎣ 0 0 1 λ 0 ⎤ ⎡ 0 λ 1 ⎦ ⇐⇒ A[v1 , v2 , v3 ] = [v1 , v2 , v3 ] ⎣ 0 λ 0 1 λ 0 ⎤ 0 1 ⎦ λ ⇐⇒ [Av1 , Av2 , Av3 ] = [λv1 , v1 + λv2 , v2 + λv3 ] ⇐⇒ Av1 = λv1 , Av2 = v1 + λv2 , and Av3 = v2 + λv3 ⇐⇒ (A − λI)v1 = 0, (A − λI)v2 = v1 , and (A − λI)v3 = v2 . ⎡ ⎤ n n , , ⎣ci 38. (a). From (7.3.15), c1 f1 + c2 f2 + · · · + cn fn = 0 ⇐⇒ sji ej ⎦ = 0 ⇐⇒ n n , , j=1 i=1 sji ci ej = 0 ⇐⇒ i=1 n , j=1 sji ci = 0, j = 1, 2, . . . , n, since {ei } is a linearly independent set. The i=1 latter equation is just the component form of the linear system Sc = 0. Since {fi } is a linearly independent set, the only solution to this system is the trivial solution c = 0, so det(S) = 0. Consequently, S is invertible. (b). From (7.3.14) and (7.3.15), we have T (fk ) = n , bik i=1 n , sji ej = j=1 n n , , j=1 sji bik ej , k = 1, 2, . . . , n. i=1 Replacing i with j and j with i yields ⎡ ⎤ n n , , ⎣ T (fk ) = sij bjk ⎦ ei , k = 1, 2, . . . , n. i=1 (∗) j=1 (c). From (7.3.15) and (7.3.13), we have T (fk ) = n , sjk T (ej ) = j=1 that is, n , j=1 sjk n , aij ei , i=1 ⎡ ⎤ n n , , ⎣ T (fk ) = aij sjk ⎦ ei , k = 1, 2, . . . , n. i=1 (∗∗) j=1 (d). Subtracting (∗) from (∗∗) yields ⎡ ⎤ n n , , ⎣ (sij bjk − aij sjk )⎦ ei = 0. i=1 j=1 Thus, since {e1 } is a linearly independent set, n , j=1 sij bjk = n , aij sjk , i = 1, 2, . . . , n. j=1 But this is just the index form of the matrix equation SB = AS. Multiplying both sides of the preceding equation on the left by S −1 yields B = S −1 AS. (c)2017 Pearson Education. Inc. 477 Solutions to Section 7.4 True-False Review: (a): TRUE. In the definition of the matrix exponential function eAt , we see that powers of the matrix A must be computed: (At)2 (At)3 eAt = In + (At) + + + .... 2! 3! In order to do this, A must be a square matrix. (b): TRUE. We see this by plugging in t = 1 into the definition of the matrix exponential function. All terms containing A3 , A4 , A5 , . . . must be zero, leaving us with the result given in this statement. (c): FALSE. The inverse of the matrix exponential function eAt is the matrix exponential function e−At , and this will exist for all square matrices A, not just invertible ones. (d): TRUE. The matrix exponential function eAt converges to a matrix the same size as A, for all t ∈ R. This is asserted, but not proven, directly beneath Definition 7.4.1. (e): FALSE. The correct statement is (SDS −1 )k = SDk S −1 . The matrices S and S −1 on the right-hand side of this equation do not get raised to the power k. (f ): FALSE. According to Property 1 of the Matrix Exponential Function, we have (eAt )2 = (eAt )(eAt ) = e2At . Problems: 1. We have det(A − λI) = 0 ⇐⇒ 1−λ 0 2 = 0 ⇐⇒ λ2 − 4λ + 3 = 0 ⇐⇒ λ = 1 or λ = 3. 3−λ 1 . 0 −2 2 1 Eigenvalue λ = 3: We have A − 3I = , so we can choose the eigenvector v2 = . 0 0 1 1 1 1 0 We form the matrices S = and D = . From Theorem 7.4.3, we have 0 1 0 3 t t 0 1 −1 e e3t − et 1 1 e = . eAt = SeDt S −1 = 0 e3t 0 1 0 e3t 0 1 Eigenvalue λ = 1: We have A − I = 0 0 2 2 , so we can choose the eigenvector v1 = 2. We have det(A − λI) = 0 ⇐⇒ 3−λ 1 Eigenvalue λ = 4: We have A − 4I = 1 = 0 ⇐⇒ λ2 − 6λ + 8 = 0 ⇐⇒ λ = 2 or λ = 4. 3−λ −1 1 1 −1 , so we can choose the eigenvector v1 = (c)2017 Pearson Education. Inc. 1 1 . 478 1 1 1 , so we can choose the eigenvector v2 = . 1 1 −1 1 4 0 and D = . From Theorem 7.4.3, we have −1 0 2 Eigenvalue λ = 2: We have A − 2I = We form the matrices S = 1 1 e4t 0 0 e2t det(A − λI) = 0 ⇐⇒ −λ −2 eAt = SeDt S −1 = 1 1 1 −1 3. We have −2i −2 Eigenvalue λ = 2i: We have A − 2iI = 1/2 1/2 1/2 −1/2 = 1 4t 2t 2 (e + e 1 4t 2t 2 (e − e ) 1 4t 2t 2 (e − e ) 1 4t 2t 2 (e + e ) . 2 = 0 ⇐⇒ λ2 + 4 = 0 ⇐⇒ λ = ±2i. −λ 2 −2i 1 i , so we can choose the eigenvector v1 = . Eigenvalue λ = −2i: By taking the conjugate of the eigenvector obtained above, we can choose v2 = We form the matrices S = e At = Se Dt S −1 = 1 i 1 −i 1 i 1 −i 2i 0 and D = e2it 0 0 e−2it 0 −2i 1 −i . . From Theorem 7.4.3, we have 1/2 1/2 −i/2 i/2 = cos 2t sin 2t − sin 2t cos 2t . The simplification in the last step uses the well-known identities cos x = 12 (eix + e−ix ) and sin x = − 2i (eix − e−ix ). 4. We have det(A − λI) = 0 ⇐⇒ −1 − λ −3 3 = 0 ⇐⇒ λ2 + 2λ + 10 = 0 ⇐⇒ λ = −1 ± 3i. −1 − λ Eigenvalue λ = −1 + 3i: We have A − (−1 + 3i)I = 1 . i −3i 3 −3 −3i , so we can choose the eigenvector v1 = Eigenvalue λ = −1 − 3i: By taking the conjugate of the eigenvector obtained above, we can choose v2 = 1 . −i 1 1 −1 + 3i 0 We form the matrices S = and D = . From Theorem 7.4.3, we have i −i 0 −1 − 3i e At = Se Dt S −1 = 1 i 1 −i e(−1+3i)t 0 0 e(−1−3i)t 1/2 −i/2 1/2 i/2 =e −t cos 3t sin 3t − sin 3t cos 3t . The simplification in the last step uses the well-known identities cos x = 12 (eix + e−ix ) and sin x = − 2i (eix − e−ix ). Alternative Solution: Apply Problem 5 below. (c)2017 Pearson Education. Inc. 479 5. We have det(A − λI) = 0 ⇐⇒ a−λ −b b = 0 ⇐⇒ λ2 − 2aλ + a2 + b2 = 0 ⇐⇒ λ = a ± bi. a−λ Eigenvalue λ = a + bi: We have A − (a + bi)I = −bi −b b −bi , so we can choose the eigenvector v1 = 1 i . Eigenvalue λ = a − bi: By taking the conjugate of the eigenvector obtained above, we can choose v2 = 1 . −i 1 1 a + bi 0 We form the matrices S = and D = . From Theorem 7.4.3, we have i −i 0 a − bi (a+bi)t 0 1/2 −i/2 cos bt sin bt 1 1 e At Dt −1 at =e . e = Se S = 1/2 i/2 − sin bt cos bt i −i 0 e(a−bi)t The simplification in the last step uses the well-known identities cos x = 21 (eix + e−ix ) and sin x = − 2i (eix − e−ix ). 6. We have det(A − λI) = 0 ⇐⇒ 3−λ 1 0 −2 −λ 0 −2 −2 = 0 ⇐⇒ (3 − λ)(λ2 − 3λ + 2) = 0 ⇐⇒ λ = 1 or λ = 2 or λ = 3. 3−λ ⎡ ⎤ ⎡ ⎤ 2 −2 −2 1 Eigenvalue λ = 1: We have A − I = ⎣ 1 −1 −2 ⎦, and we may choose the vector v1 = ⎣ 1 ⎦ in 0 0 2 0 nullspace(A − I) as an eigenvector corresponding to λ = 1. ⎤ ⎡ ⎤ ⎡ 1 −2 −2 2 Eigenvalue λ = 2: We have A − 2I = ⎣ 1 −2 −2 ⎦, and we may choose the vector v2 = ⎣ 1 ⎦ in 0 0 1 0 nullspace(A − 2I) as an eigenvector corresponding to λ = 2. ⎡ ⎤ ⎡ ⎤ 0 −2 −2 1 Eigenvalue λ = 3: We have A − 3I = ⎣ 1 −3 −2 ⎦, and we may choose the vector v3 = ⎣ 1 ⎦ in 0 0 0 −1 nullspace(A − 3I) as an eigenvector corresponding to λ = 3. ⎡ ⎤ ⎡ ⎤ 1 2 1 1 0 0 1 ⎦ and D = ⎣ 0 2 0 ⎦. From Theorem 7.4.3, we have We form the matrices S = ⎣ 1 1 0 0 −1 0 0 3 eAt = SeDt S −1 ⎤⎡ ⎤ ⎤⎡ ⎡ −1 2 1 1 0 0 1 2 1 0 ⎦ 1 ⎦ ⎣ 0 2 0 ⎦ ⎣ 1 −1 =⎣ 1 1 0 0 3 0 0 −1 0 0 −1 ⎡ t t ⎤ t t t 2t e (2e − 1) −2e (e − 1) −e (e − 1) −et (et − 2) −et (e2t − 1) ⎦ . = ⎣ et (et − 1) 0 0 e3t (c)2017 Pearson Education. Inc. 480 7. We have det(A − λI) = 0 ⇐⇒ 6−λ 8 4 −2 −2 − λ −2 −1 −2 = 0 ⇐⇒ (λ − 2)2 (λ − 1) = 0 ⇐⇒ λ = 2 or λ = 1. 1−λ ⎡ ⎤ ⎡ ⎤ 4 −2 −1 1 Eigenvalue λ = 2: We have A − 2I = ⎣ 8 −4 −2 ⎦, and we may choose two vectors, v1 = ⎣ 2 ⎦ and 4 −2 −1 0 ⎡ ⎤ 1 v2 = ⎣ 0 ⎦ in nullspace(A − 2I) as an eigenvector corresponding to λ = 2. 4 ⎡ ⎤ 5 −2 −1 Eigenvalue λ = 1: We have A − I = ⎣ 8 −3 −2 ⎦. The reduced row-echelon form of this matrix is 4 −2 0 ⎡ ⎤ ⎡ ⎤ 1 1 0 −1 ⎣ 0 1 −2 ⎦, so we make choose v3 = ⎣ 2 ⎦ in nullspace(A − I) as an eigenvector corresponding to 1 0 0 0 λ = 1. ⎡ 1 We form the matrices S = ⎣ 2 0 1 0 4 ⎤ ⎡ 1 2 2 ⎦ and D = ⎣ 0 1 0 0 2 0 ⎤ 0 0 ⎦. From Theorem 7.4.3, we have 1 eAt = SeDt S −1 ⎤⎡ ⎤ ⎡ ⎤ ⎡ 2t 0 0 4 −3/2 −1 1 1 1 e 0 ⎦ = ⎣ 2 0 2 ⎦ ⎣ 0 e2t 0 ⎦ ⎣ 1 −1/2 t −4 2 1 0 0 e 0 4 1 ⎤ ⎡ t t t t t t e (5e − 4) 2e (1 − e ) e (1 − e ) = ⎣ 8et (et − 1) et (4 − 3et ) 2et (1 − et ) ⎦ . 4et (et − 1) 2et (1 − et ) et (c)2017 Pearson Education. Inc. 481 8. Note that 1 1 eAt = I + At + (At)2 + · · · + (At)k + . . . 2! k! ⎤ ⎡ ⎡ 1 0 ... 0 d1 t 0 . . . ⎢ 0 1 . . . 0 ⎥ ⎢ 0 d2 t . . . ⎢ ⎥ ⎢ =⎢ . . . +⎢ . .. .. . . ... ⎥ ⎣ .. .. ⎦ ⎣ .. . . 0 0 ... 1 ⎡ ∞ 1 k k=0 k! (d1 t) ⎢ 0 ⎢ =⎢ .. ⎣ . 0 ... 0 ∞ .. . 0 ⎤ ed1 t ⎢ 0 ⎢ =⎢ . ⎣ .. 0 ed2 t .. . ... ... .. . 0 0 .. . 0 0 ... edn t ⎡ (d1 t)k ⎥ 0 1 ⎢ ⎥ ⎢ ⎥ + ··· + ⎢ .. k! ⎣ ⎦ . dn t ... ... .. . ... 1 k k=0 k! (d2 t) 0 ⎡ 0 ⎤ 0 0 .. . ∞ ... ... .. . 0 0 .. . 0 ... (dn t)k 0 ⎤ 0 0 .. . 0 (d2 t)k .. . ⎥ ⎥ ⎥ ⎦ 1 k k=1 k! (dn t) ⎥ ⎥ ⎥ ⎦ = diag(ed1 t , ed2 t , . . . , edn t ). 9. Using Problem 8, we have e−3t 0 0 e5t eλt ⎢ 0 ⎢ eλIn t = ⎢ . ⎣ .. 0 eλt .. . ... ... .. . 0 0 .. . 0 0 ... eλt 0 0 b 0 eAt = 10. We have ⎡ and e−At = ⎤ ⎡ e3t 0 0 . e−5t 1 ⎥ ⎢ 0 ⎥ ⎢ ⎥ = eλt ⎢ .. ⎦ ⎣ . 0 ... 1 ... .. . . . . 0 0 ... ⎤ 0 0 ⎥ ⎥ λt .. ⎥ = e In . . ⎦ 1 11. (a). We have BC = 2 (b). We have C = 0 0 b 0 a 0 0 a 0 0 b 0 = e Ct = I + Ct = = 0 0 0 0 1 0 0 1 0 0 ab 0 = 0 0 b 0 a 0 0 a = CB. = 02 . Therefore, + 0 0 b 0 t= 1 0 bt 1 (c). Using the results in the first two parts of this problem, it follows that at at 0 1 bt e e eAt = e(B+C)t = eBt eCt = = 0 eat 0 1 0 (c)2017 Pearson Education. Inc. . btebt eat . ⎤ ⎥ ⎥ ⎥ ⎦ 482 a 0 0 b 0 ab and C = , so that A = B + C. Now BC = = CB, so 0 a −b 0 −ab 0 Property (1) of the matrix exponential function implies that eAt = e(B+C)t = eBt eCt . Now at 0 1 0 e Bt at =e , e = 0 eat 0 1 12. Define B = by Problem 8. To determine eCt , observe first that 4 0 −b2 0 b3 b 2 3 4 , C = , C = C = 0 −b2 0 −b3 0 In general, we have C 2n =b 2n (−1)n 0 0 (−1)n and C 2n+1 =b 2n+1 for each integer n ≥ 0. Therefore, from Definition 7.4.1, we have ∞ (−1)n 2n+1 ∞ (−1)n 2n n=0 (2n)! b n=0 (2n+1)! b Ct e = ∞ (−1)n 2n+1 ∞ (−1)n 2n − n=0 (2n+1)! b n=0 (2n)! b Thus, eAt = eBt eCt = eat cos bt − sin bt 0 b4 0 (−1)n+1 (−1)n 0 −b5 0 , C = cos bt sin bt − sin bt cos bt = sin bt cos bt 0 b5 5 ,... . . 13. Direct computation shows that A2 = 02 . Therefore, 1 0 1 1 1+t At + t= e = I + At = 0 1 −1 −1 −t t 1−t . 14. Direct computation shows that A2 = 02 . Therefore, 1 0 −3 9 −3t + 1 9t + t= . eAt = I + At = 0 1 −1 3 −t 3t + 1 ⎡ ⎤ −4 8 −4 2 −1 ⎦ and A3 = 03 . Therefore, 15. Direct computation shows that A2 = ⎣ −1 2 −4 2 ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 1 0 0 −1 −6 −5 −4 8 −4 t2 2 −1 ⎦ eAt = ⎣ 0 1 0 ⎦ + ⎣ 0 −2 −1 ⎦ t + ⎣ −1 2 0 0 1 1 2 3 2 −4 2 ⎤ ⎡ 1 − t − 2t2 −6t + 4t2 −5t − 2t2 −t2 /2 1 − 2t + t2 −t − t2 /2 ⎦ . =⎣ 2 t+t 2t − 2t2 1 + 3t + t2 ⎡ ⎤ 0 0 0 16. Direct computation shows that A2 = ⎣ 0 0 0 ⎦ and A3 = 03 . Therefore, 1 0 0 ⎤ ⎡ ⎤ ⎤ ⎡ ⎡ ⎤ ⎡ 1 0 0 0 0 0 0 0 0 1 0 0 2 t 1 0 ⎦. eAt = ⎣ 0 1 0 ⎦ + ⎣ 1 0 0 ⎦ t + ⎣ 0 0 0 ⎦ = ⎣ t 2 2 0 0 1 0 1 0 1 0 0 t /2 t 1 (c)2017 Pearson Education. Inc. 483 ⎡ 0 ⎢ 0 2 ⎢ 17. Direct computation shows that A = ⎣ 0 0 0 0 0 0 1 0 0 0 ⎤ ⎡ 0 0 ⎢ 1 ⎥ ⎥ , A3 = ⎢ 0 ⎣ 0 0 ⎦ 0 0 0 0 0 0 0 0 0 0 ⎤ 1 0 ⎥ ⎥, and A4 = 04 . Therefore, 0 ⎦ 0 (At)2 (At)3 eAt = I4 + At + + 2 ⎤ ⎡6 ⎤ ⎡ ⎤ ⎤ ⎡ ⎡ 0 0 t2 /2 0 t 0 0 0 0 0 t3 /6 0 1 0 0 0 ⎢ ⎢ 0 1 0 0 ⎥ ⎢ 0 0 t 0 ⎥ ⎢ 0 0 0 t2 /2 ⎥ 0 ⎥ ⎥+⎢ ⎥ ⎥+⎢ ⎥+⎢ 0 0 0 =⎢ ⎣ 0 0 1 0 ⎦ ⎣ 0 0 0 t ⎦ ⎣ 0 0 0 ⎦ 0 0 ⎦ ⎣ 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 ⎡ ⎤ 2 3 1 t t /2 t /6 ⎢ 0 1 t t2 /2 ⎥ ⎥. =⎢ ⎣ 0 0 1 t ⎦ 0 0 0 1 18. It is easy to see (and can be formally verified by induction on k) that the matrix Ak contains all zero entries except for the (k + i, i) entry for 1 ≤ i ≤ n − k. In particular, An = An+1 = An+2 = · · · = 0n . Therefore, by Definition 7.4.1, we have eAt = In + At + ⎡ 1 ⎢ t ⎢ ⎢ t2 ⎢ 2!3 =⎢ ⎢ t3! ⎢ .. ⎢ . ⎣ tn−1 (n−1)! (At)2 (At)3 (At)n−1 + + ··· + 2! 3! (n − 1)! ⎤ 0 0 0 ... 0 1 0 0 ... 0 ⎥ ⎥ t 1 0 ... 0 ⎥ ⎥ ⎥. t2 t 1 . . . 0 ⎥ 2! ⎥ .. .. .. . .. . . . ⎥ . . . ⎦ tn−2 tn−3 tn−4 . . . 1 (n−2)! (n−3)! (n−4)! 19. Assume that A0 is m × m and B0 is n × n. Note that two matrices must be square in order for these Ak0 0 A0 t B0 t k . Therefore, and e to make sense. The key point is that A = e 0 B0k (At)3 (At)2 eAt = I + At + + + ... 2! 3! 2 2 3 3 t t A0 0 A0 0 Im 0 A0 0 + t+ = + + ... 3 0 In 0 B0 0 B02 0 B 2! 3! 0 2 3 0 t) 0 t) + (A3! + ... 0 Im + A0 t + (A2! = 2 3 0 t) 0 t) 0 In + B0 t + (B2! + (B3! + ... At 0 0 e . = 0 e B0 t Solutions to Section 7.5 True-False Review: (c)2017 Pearson Education. Inc. 484 (a): TRUE. This is immediate from Definition 7.5.1. (b): TRUE. The roots of p(λ) = λ3 + λ are λ = 0 and λ = ±i. Therefore, the matrix has complex eigenvalues. But Theorem 7.5.4 indicates that a real matrix that is symmetric must have only real eigenvalues. (c): FALSE. The zero matrix is a real, symmetric matrix for which both v1 and v2 are eigenvectors. (d): TRUE. This is the statement of the Principal Axes Theorem (Theorem 7.5.6). (e): TRUE. This is a direct application of Lemma 7.5.9. (f ): TRUE. Assuming that A and B are orthogonal matrices, then A and B are invertible, and hence AB is invertible. Moreover, (AB)−1 = B −1 A−1 = B T AT = (AB)T , so that AB meets the requirements of orthogonality spelled out in Definition 7.5.1. (g): TRUE. This is essentially the content of Theorem 7.5.3, since a set of orthogonal unit vectors are precisely a set of orthonormal vectors. (h): TRUE. Let S be the matrix consisting of a complete set of orthonormal eigenvectors of A. Then S T AS = diag(λ1 , λ2 , . . . , λn ) = D, where the λi are the corresponding eigenvalues. Then S is an orthogonal matrix: S T = S −1 . Hence, A = SDS T and AT = (SDS T )T = SDT S T = SDS T = A, so A is symmetric. Problems: 4 = 0 ⇐⇒ λ2 + 4λ − 21 = 0 ⇐⇒ (λ + 7)(λ − 3) = 0 ⇐⇒ λ = −5 − λ 0 8 4 v1 = =⇒ 8v1 + 4v2 = 0. Therefore, If λ = −7 then (A − λI)v = 0 assumes the form v2 0 4 2 2 v1 1 is a unit = −√ , √ v1 = (−1, 2) is an eigenvector of A corresponding to λ = −7, and w1 = ||v1 || 5 5 eigenvector corresponding to λ = −7. 0 −2 4 v1 = =⇒ −2v1 + 4v2 = 0. If λ = 3 then (A − λI)v = 0 assumes the form v2 0 4 −8 1 v2 2 is a = √ ,√ Therefore, v2 = (2, 1) is an eigenvector of A corresponding to λ = 3, and w2 = ||v2 || 5 5 ⎡ ⎤ 1 2 √ −√ ⎢ 5 5 ⎥ T unit eigenvector corresponding to λ = 3. Thus, S = ⎣ 2 1 ⎦ and S AS = diag(−7, 3). √ √ 5 5 2−λ 2 = 0 ⇐⇒ λ2 − λ − 6 = 0 ⇐⇒ (λ + 2)(λ − 3) = 0 ⇐⇒ λ = −2 or 2. det(A − λI) = 0 ⇐⇒ 2 −1 − λ λ = 3. 0 4 2 v1 = =⇒ 2v1 + v2 = 0. v1 = (1, −2), is If λ = −2 then (A − λI)v = 0 assumes the form v 0 2 1 2 2 v1 1 is a unit eigenvector corresponding an eigenvector corresponding to λ = −2 and w1 = = √ , −√ ||v1 || 5 5 to λ = −2. 0 −1 2 v1 If λ = 3 then (A − λI)v = 0 assumes the form = =⇒ v1 + 2v2 = 0. v2 = (2, 1), 2 −4 v2 0 v2 2 1 is a unit eigenvector corresponding = √ ,√ is an eigenvector corresponding to λ = 3 and w2 = ||v2 || 5 5 1−λ 1. det(A − λI) = 0 ⇐⇒ 4 −7 or λ = 3. (c)2017 Pearson Education. Inc. 485 to λ = 3. ⎡ 2 ⎤ √ 5 ⎥ T 1 ⎦ and S AS = diag(−2, 3). √ 5 4−λ 6 = 0 ⇐⇒ λ2 − 13λ = 0 ⇐⇒ λ(λ − 13) = 0 ⇐⇒ λ = 0 or λ = 13. 3. det(A − λI) = 0 ⇐⇒ 6 9−λ 4 6 v1 0 = =⇒ 2v1 + 3v2 = 0. v1 = (−3, 2), is If λ = 0 then (A − λI)v = 0 assumes the form 0 6 9 v2 v1 3 2 is a unit eigenvector corresponding = −√ , √ an eigenvector corresponding to λ = 0 and w1 = ||v1 || 13 13 to λ = 0. 0 −9 6 v1 If λ = 13 then (A − λI)v = 0 assumes the form = =⇒ −3v1 + 2v2 = 0. v2 0 6 −4 v2 2 3 is a unit eigenvector = √ ,√ v2 = (2, 3), is an eigenvector corresponding to λ = 13 and w2 = ||v2 || 13 13 corresponding ⎡ to 3λ = 13.2 ⎤ √ −√ ⎢ 13 13 ⎥ T Thus, S = ⎣ 2 3 ⎦ and S AS = diag(0, 13). √ √ 13 13 1−λ 2 = 0 ⇐⇒ (λ − 1)2 − 4 = 0 ⇐⇒ (λ + 1)(λ − 3) = 0 ⇐⇒ λ = −1 or 4. det(A − λI) = 0 ⇐⇒ 2 1−λ λ = 3. 0 2 2 v1 = =⇒ v1 + v2 = 0. v1 = (−1, 1), is If λ = −1 then (A − λI)v = 0 assumes the form 0 v 2 2 2 v1 1 1 is a unit eigenvector corresponding = −√ , √ an eigenvector corresponding to λ = −1 and w1 = ||v1 || 2 2 to λ = −1. 0 −2 2 v1 If λ = 3 then (A − λI)v = 0 assumes the form = =⇒ v1 − v2 = 0. v2 = (1, 1), is v2 0 2 −2 1 v2 1 an eigenvector corresponding to λ = 3 and w2 = is a unit eigenvector corresponding to = √ ,√ ||v2 || 2 2 λ = 3. ⎡ 1 ⎤ 1 √ −√ ⎢ 2 2 ⎥ T Thus, S = ⎣ 1 ⎦ and S AS = diag(−1, 3). 1 √ √ 2 2 −λ 0 3 0 = 0 ⇐⇒ (λ + 2)(λ − 3)(λ + 3) = 0 ⇐⇒ λ = −3, λ = −2, or 5. det(A − λI) = 0 ⇐⇒ 0 −2 − λ 3 0 −λ λ = 3. ⎤ ⎡ ⎡ ⎤⎡ ⎤ 3 0 3 v1 0 If λ = −3 then (A − λI)v = 0 assumes the form ⎣ 0 1 0 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ =⇒ v1 + v3 = 0 and v2 = 0. 3 0 3 v3 0 v1 1 1 is a unit v1 = (−1, 0, 1), is an eigenvector corresponding to λ = −3 and w1 = = − √ , 0, √ ||v1 || 2 2 eigenvector corresponding to λ = −3. 1 √ ⎢ 5 Thus, S = ⎣ 2 −√ 5 (c)2017 Pearson Education. Inc. 486 ⎤ ⎡ ⎤⎡ ⎤ 3 v1 0 0 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ =⇒ v1 = v3 = 0 and v2 ∈ R. 2 v3 0 v2 v2 = (0, 1, 0), is an eigenvector corresponding to λ = −2 and w2 = = (0, 1, 0) is a unit eigenvector ||v2 || corresponding to λ = −2. ⎤ ⎤ ⎡ ⎤⎡ ⎡ −3 0 3 v1 0 0 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ =⇒ v1 − v3 = 0 and If λ = 3 then (A − λI)v = 0 assumes the form ⎣ 0 −5 0 3 0 −3 v3 v3 1 1 is a unit = √ , 0, √ v2 = 0. v3 = (1, 0, 1), is an eigenvector corresponding to λ = 3 and w3 = ||v3 || 2 2 eigenvector⎡corresponding to λ = 3. 1 1 ⎤ −√ 0 √ ⎢ 2 2 ⎥ ⎢ ⎥ 0 1 0 ⎥ and S T AS = diag(−3, −2, 3). Thus, S = ⎢ ⎣ 1 1 ⎦ √ 0 √ 2 2 1−λ 2 1 4−λ 2 = 0 ⇐⇒ λ3 − 6λ2 = 0 ⇐⇒ λ2 (λ − 6) = 0 ⇐⇒ λ = 6 or 6. det(A − λI) = 0 ⇐⇒ 2 1 2 1−λ λ = 0 of multiplicity two. ⎤ ⎡ ⎤⎡ ⎡ ⎤ 1 2 1 v1 0 If λ = 0 then (A − λI)v = 0 assumes the form ⎣ 2 4 2 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ =⇒ v1 + 2v2 + v3 = 0. 1 2 1 v3 0 v1 = (−1, 0, 1) and v2 = (−2, 1, 0) are linearly independent eigenvectors corresponding to λ = 0. v1 and v2 are not orthogonal since v1 , v2 = 2 = 0, so we will use the Gram-Schmidt procedure. Let u1 = v1 = (−1, 0, 1), so ⎡ 2 If λ = −2 then (A − λI)v = 0 assumes the form ⎣ 0 3 0 0 0 v2 , u1 2 u1 = (−2, 1, 0) − (−1, 0, 1) = (−1, 1, −1). ||u1 ||2 2 1 1 1 u1 u2 1 1 and w2 = are orthonormal eigenvectors Now w1 = = − √ , 0, √ = −√ , √ , √ ||u1 || ||u2 || 2 2 3 3 3 corresponding to λ = 0. ⎤ ⎡ ⎤⎡ ⎡ ⎤ −5 2 1 v1 0 2 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ =⇒ v1 − v3 = 0 and If λ = 6 then (A − λI)v = 0 assumes the form ⎣ 2 −2 1 2 −5 v3 0 2 1 v3 1 is v2 − 2v3 = 0. v3 = (1, 2, 1), is an eigenvector corresponding to λ = 6 and w3 = = √ ,√ ,√ ||v3 || 6 6 6 a unit eigenvector corresponding to ⎡ ⎤ λ = 6. 1 1 1 √ −√ −√ ⎢ 2 3 6 ⎥ ⎢ 1 2 ⎥ ⎢ √ √ ⎥ 0 Thus, S = ⎢ ⎥ and S T AS = diag(0, 0, 6). ⎢ 3 6 ⎥ ⎣ 1 1 ⎦ 1 √ √ −√ 2 3 6 2−λ 0 0 3−λ 1 = 0 ⇐⇒ (λ − 2)2 (λ − 4) = 0 ⇐⇒ λ = 4 or λ = 2 of 7. det(A − λI) = 0 ⇐⇒ 0 0 1 3−λ u2 = v2 − (c)2017 Pearson Education. Inc. 487 multiplicity two. ⎤ ⎤⎡ ⎡ ⎤ 0 0 0 v1 0 If λ = 2 then (A − λI)v = 0 assumes the form ⎣ 0 1 1 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ =⇒ v2 + v3 = 0 and 0 1 1 v3 0 linearly independent eigenvectors corresponding to λ = 2, and v1 ∈ R. v1 = (1, 0, 0) and v2 = (0, −1, 1) are v1 v2 1 1 are unit eigenvectors corresponding to λ = 2. w1 w1 = = (1, 0, 0) and w2 = = 0, − √ , √ ||v1 || ||v2 || 2 2 and w2 are also orthogonal because w1 , w2 =⎡0. ⎤ ⎡ ⎤⎡ ⎤ −2 0 0 v1 0 1 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ =⇒ v1 = 0 and If λ = 4 then (A − λI)v = 0 assumes the form ⎣ 0 −1 0 1 −1 v3 0 v3 1 1 is a v2 − v3 = 0. v3 = (0, 1, 1), is an eigenvector corresponding to λ = 4 and w3 = = 0, √ , √ ||v3 || 2 2 unit eigenvector corresponding⎤ to λ = 4. ⎡ 1 0 0 1 ⎥ 1 ⎢ √ ⎥ ⎢ 0 −√ Thus, S = ⎢ and S T AS = diag(2, 2, 4). 2 2 ⎥ ⎣ ⎦ 1 1 √ √ 0 2 2 −λ 1 0 0 = 0 ⇐⇒ (λ − 1)2 (λ + 1) = 0 ⇐⇒ λ = −1 or λ = 1 of 8. det(A − λI) = 0 ⇐⇒ 1 −λ 0 0 1−λ multiplicity two. ⎤ ⎡ ⎤⎡ ⎡ ⎤ −1 1 0 v1 0 If λ = 1 then (A − λI)v = 0 assumes the form ⎣ 1 −1 0 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ =⇒ v1 − v2 = 0 and 0 0 0 v3 0 1, 0) and v = (0, 0, 1) are linearly independent eigenvectors corresponding to λ = 1, and v3 ∈ R. v1 = (1, 2 v1 v2 1 1 w1 = = √ , √ , 0 and w2 = = (0, 0, 1) are unit eigenvectors corresponding to λ = 1. w1 ||v1 || ||v2 || 2 2 and w2 are also orthogonal because w1 , w2 = 0. ⎤ ⎡ ⎡ ⎤⎡ ⎤ 1 1 0 v1 0 If λ = −1 then (A − λI)v = 0 assumes the form ⎣ 1 1 0 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ =⇒ v1 + v2 = 0 and v3 = 0. 0 0 2 v3 0 v3 1 1 v3 = (−1, 1, 0), is an eigenvector corresponding to λ = −1 and w3 = = − √ , √ , 0 is a unit ||v3 || 2 2 eigenvector⎡corresponding to λ = −1. 1 ⎤ 1 0 √ −√ ⎢ 2 2 ⎥ ⎢ 1 ⎥ and S T AS = diag(−1, 1, 1). Thus, S = ⎢ √1 0 √ ⎥ ⎦ ⎣ 2 2 0 1 0 1−λ 1 −1 1−λ 1 = 0 ⇐⇒ (λ − 2)2 (λ + 1) = 0 ⇐⇒ λ = 2 of multiplicity two 9. det(A − λI) = 0 ⇐⇒ 1 −1 1 1−λ λ = −1. ⎤ ⎡ ⎤ ⎡ ⎤⎡ −1 1 −1 v1 0 1 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ =⇒ v1 − v2 + v3 = 0. If λ = 2 then (A − λI)v = 0 assumes the form ⎣ 1 −1 0 −1 1 −1 v3 ⎡ (c)2017 Pearson Education. Inc. 488 v1 = (1, 1, 0) and v2 = (−1, 0, 1) are linearly independent eigenvectors corresponding to λ = 2. v1 and v2 are not orthogonal since v1 , v2 = −1 = 0, so we will use the Gram-Schmidt procedure. Let u1 = v1 = (1, 1, 0), so v2 , u1 1 1 1 u2 = v2 − u1 = (−1, 0, 1) + (1, 1, 0) = − , , 1 . ||u1 ||2 2 2 2 u1 u2 1 1 1 1 2 are orthonormal eigenvectors Now w1 = = √ , √ , 0 and w2 = = −√ , √ , √ ||u1 || ||u || 2 2 6 6 6 2 corresponding to λ = 2. ⎤ ⎤ ⎡ ⎤⎡ ⎡ 2 1 −1 v1 0 1 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ =⇒ v1 − v3 = 0 and If λ = −1 then (A − λI)v = 0 assumes the form ⎣ 1 2 −1 1 2 0 v3 v3 1 1 1 v2 + v3 = 0. v3 = (1, −1, 1), is an eigenvector corresponding to λ = −1 and w3 = = √ , −√ , √ ||v3 || 3 3 3 is a unit eigenvector corresponding⎤ to λ = −1. ⎡ 1 1 1 √ √ −√ ⎢ 2 6 3 ⎥ ⎢ 1 1 1 ⎥ ⎢ √ √ −√ ⎥ Thus, S = ⎢ ⎥ and S T AS = diag(2, 2, −1). ⎢ 2 6 3 ⎥ ⎣ 1 ⎦ 2 √ √ 0 6 3 1−λ 0 −1 1−λ 1 = 0 ⇐⇒ (1 − λ)(λ + 1)(λ − 2) = 0 ⇐⇒ λ = −1, λ = 1, or 10. det(A − λI) = 0 ⇐⇒ 0 −1 1 −λ λ = 2. ⎤ ⎡ ⎡ ⎤⎡ ⎤ 2 0 −1 v1 0 1 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ =⇒ 2v1 − v3 = 0 and If λ = −1 then (A − λI)v = 0 assumes the form ⎣ 0 2 −1 1 1 v3 0 v1 1 1 2 2v2 +v3 = 0. v1 = (1, −1, 2), is an eigenvector corresponding to λ = −1 and w1 = = √ , −√ , √ ||v1 || 6 6 6 is a unit eigenvector corresponding to λ = −1. ⎡ ⎤ ⎤⎡ ⎡ ⎤ 0 0 −1 v1 0 1 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ =⇒ v1 − v2 = 0 and If λ = 1 then (A − λI)v = 0 assumes the form ⎣ 0 0 −1 1 −1 v3 0 v2 1 1 v3 = 0. v2 = (1, 1, 0), is an eigenvector corresponding to λ = 1 and w2 = = √ , √ , 0 is a unit ||v2 || 2 2 eigenvector corresponding to λ = 1. ⎤ ⎤ ⎡ ⎤⎡ ⎡ −1 0 −1 v1 0 1 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ =⇒ v1 + v3 = 0 and If λ = 2 then (A − λI)v = 0 assumes the form ⎣ 0 −1 0 −1 1 −2 v3 1 1 v3 1 = −√ , √ , √ v2 − v3 = 0. v3 = (−1, 1, 1), is an eigenvector corresponding to λ = 2 and w3 = ||v3 || 3 3 3 is a unit eigenvector corresponding⎤ to λ = 2. ⎡ 1 1 1 √ √ −√ ⎢ 6 2 3 ⎥ ⎢ 1 1 ⎥ ⎢ − √1 √ √ ⎥ Thus, S = ⎢ ⎥ and S T AS = diag(−1, 1, 2). ⎢ 6 2 3 ⎥ ⎣ 2 1 ⎦ √ √ 0 6 3 (c)2017 Pearson Education. Inc. 489 3−λ 11. det(A − λI) = 0 ⇐⇒ 3 4 or λ = 8. 4 0 = 0 ⇐⇒ (λ − 3)(λ + 2)(λ − 8) = 0 ⇐⇒ λ = −2, λ = 3, 3−λ ⎤ ⎡ ⎡ ⎤⎡ ⎤ 5 3 4 v1 0 If λ = −2 then (A − λI)v = 0 assumes the form ⎣ 3 5 0 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ =⇒ 4v1 + 5v3 = 0 and 4v2 − 4 0 5 v3 0 v1 1 3 4 3v3 = 0. v1 = (−5, 3, 4), is an eigenvector corresponding to λ = −2 and w1 = = −√ , √ , √ ||v1 || 2 5 2 5 2 is a unit eigenvector corresponding to λ = −2. ⎡ ⎤ ⎡ ⎤⎡ ⎤ 0 3 4 v1 0 If λ = 3 then (A − λI)v = 0 assumes the form ⎣ 3 0 0 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ =⇒ v1 = 0 and 3v2 + 4v3 = 0. 4 0 0 v3 0 v2 4 3 is a unit eigenvector v2 = (0, −4, 3), is an eigenvector corresponding to λ = 3 and w2 = = 0, − , ||v2 || 5 5 corresponding to λ = 3. ⎤ ⎡ ⎡ ⎤⎡ ⎤ −5 3 4 v1 0 0 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ =⇒ 4v1 − 5v3 = 0 and If λ = 8 then (A − λI)v = 0 assumes the form ⎣ 3 −5 4 0 −5 v3 0 v3 1 3 4 4v2 − 3v3 = 0. v3 = (5, 3, 4), is an eigenvector corresponding to λ = 8 and w3 = = √ , √ , √ ||v3 || 2 5 2 5 2 is a unit eigenvector corresponding to λ = 8. ⎡ ⎤ 1 1 −√ 0 √ ⎢ 2 2 ⎥ ⎢ 3 3 ⎥ 4 ⎢ √ √ ⎥ − Thus, S = ⎢ ⎥ and S T AS = diag(−2, 3, 8). 5 5 2 ⎥ ⎢ 5 2 ⎣ 3 4 4 ⎦ √ √ 5 5 2 5 2 3 3−λ 0 2 = 0 ⇐⇒ (λ + 5)2 (λ − 1) = 0 ⇐⇒ λ = −5 of 2 −3 − λ ⎤ ⎡ ⎤⎡ ⎡ ⎤ −4 2 2 v1 0 2 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ =⇒ v1 − v3 = 0 and If λ = 1 then (A − λI)v = 0 assumes the form ⎣ 2 −4 2 2 −4 v3 0 v1 1 1 1 is a v2 − v3 = 0. v1 = (1, 1, 1) is an eigenvector corresponding to λ = 1 and w1 = = √ ,√ ,√ ||v1 || 3 3 3 unit eigenvector corresponding to λ = 1. ⎤ ⎡ ⎤⎡ ⎡ ⎤ 2 2 2 v1 0 If λ = −5 then (A − λI)v = 0 assumes the form ⎣ 2 2 2 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ =⇒ v1 + v2 + v3 = 0. 2 2 2 v3 0 v2 = (−1, 1, 0) and v3 = (−1, 0, 1) are linearly independent eigenvectors corresponding to λ = −5. v2 and v3 are not orthogonal since v2 , v3 = −1 = 0, so we will use the Gram-Schmidt procedure. Let u2 = v2 = (−1, 1, 0), so −3 − λ 2 12. det(A − λI) = 0 ⇐⇒ 2 multiplicity two, or λ = 1. u3 = v3 − 2 −3 − λ 2 v3 , u2 1 u2 = (−1, 0, 1) − (−1, 1, 0) = ||u2 ||2 2 (c)2017 Pearson Education. Inc. 1 1 − ,− ,1 . 2 2 490 1 1 2 u2 u3 1 1 are orthonormal eigenvectors = − √ , √ , 0 and w3 = = −√ , −√ , √ ||u2 || ||u3 || 2 2 6 6 6 corresponding ⎤ ⎡ to λ = −5. 1 1 1 √ −√ −√ ⎢ 3 2 6 ⎥ ⎢ 1 1 1 ⎥ ⎢ √ √ −√ ⎥ Thus, S = ⎢ ⎥ and S T AS = diag(1, −5, −5). ⎢ 3 2 6 ⎥ ⎣ 1 2 ⎦ √ √ 0 3 6 −λ 1 1 1 = 0 ⇐⇒ (λ + 1)2 (λ − 2) = 0 ⇐⇒ λ = −1 of multiplicity two, or 13. det(A − λI) = 0 ⇐⇒ 1 −λ 1 1 −λ λ = 2. ⎤⎡ ⎤ ⎤ ⎡ ⎡ v1 1 1 1 0 If λ = −1 then (A − λI)v = 0 assumes the form ⎣ 1 1 1 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ =⇒ v1 + v2 + v3 = 0. 1 1 1 0 v3 v1 = (−1, 0, 1) and v2 = (−1, 1, 0) are linearly independent eigenvectors corresponding to λ = −1. v1 and v2 are not orthogonal since v1 , v2 = 1 = 0, so we will use the Gram-Schmidt procedure. Let u1 = v1 = (−1, 0, 1), so 1 v2 , u1 1 1 . u2 = v2 − u1 = (−1, 1, 0) − (−1, 0, 1) = − , 1, − ||u1 ||2 2 2 2 u1 u2 1 1 1 2 1 √ √ √ √ √ Now w1 = and w2 = are orthonormal eigenvectors = − = − , 0, , ,− ||u1 || ||u2 || 2 2 6 6 6 corresponding to λ = −1. ⎤ ⎤ ⎡ ⎤⎡ ⎡ −2 1 1 v1 0 1 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ =⇒ v1 − v3 = 0 and If λ = 2 then (A − λI)v = 0 assumes the form ⎣ 1 −2 0 1 1 −2 v3 v3 1 1 1 is a = √ ,√ ,√ v2 − v3 = 0. v3 = (1, 1, 1), is an eigenvector corresponding to λ = 2 and w3 = ||v3 || 3 3 3 unit eigenvector corresponding to ⎤λ = 2. ⎡ 1 1 1 √ −√ −√ ⎢ 2 6 3 ⎥ ⎢ 2 1 ⎥ ⎢ √ √ ⎥ 0 Thus, S = ⎢ ⎥ and S T AS = diag(−1, −1, 2). ⎢ 6 3 ⎥ ⎣ 1 1 ⎦ 1 √ √ −√ 2 6 3 eigenvectors v1 = (1, 1) and v2 = (−1, 1). 14. A has eigenvalues λ1 = 4 and λ2 = −2 with corresponding 1 1 Therefore, a set of principal axes is √ (1, 1), √ (−1, 1) . Relative to these principal axes, the quadratic 2 2 form reduces to 4y12 − 2y22 . Now w2 = 15. A has eigenvalues λ1 = 7 and λ2 = 3 with corresponding eigenvectors v1 = (1, 1) and v2 = (1, −1). 1 1 Therefore, a set of principal axes is √ (1, 1), √ (1, −1) . Relative to these principal axes, the quadratic 2 2 form reduces to 7y12 + 3y22 . 16. A has eigenvalue λ = 2 of multiplicity two with corresponding linearly independent eigenvectors v1 = (1, 0, −1) and v2 = (0, 1, 1). Using the Gram-Schmidt procedure, an orthogonal basis in this eigenspace (c)2017 Pearson Education. Inc. 491 is {u1 , u2 } where 1 1 u1 = (1, 0, −1), u2 = (0, 1, 1) + (1, 0, −1) = (1, 2, 1). 2 2 1 1 An orthonormal basis for the eigenspace is √ (1, 0, −1), √ (1, 2, 1) . The remaining eigenvalue of A is 2 6 a set of principal axes for the given quadratic λ = −1,with eigenvector v3 = (−1, 1, −1). Consequently, 1 1 1 form is √ (1, 0, −1), √ (1, 2, 1), √ (−1, 1, −1) . Relative to these principal axes, the quadratic form 2 6 3 reduces to 2y12 + 2y22 − y32 . 17. A has eigenvalue λ = 0 of multiplicity two with corresponding linearly independent eigenvectors v1 = (0, 1, 0, −1) and v2 = (1, 0, −1, 0). Notice that these are orthogonal, hence we do not need to apply the Gram-Schmidt procedure. The remaining eigenvalues of A are λ = 4 and λ = 8, with corresponding eigenvectors v3 = (−1, 1, −1, 1) and v4 = (1, 1, 1, 1), respectively. Consequently, a set of principal axes for the given quadratic form is 1 1 1 1 √ (0, 1, 0, −1), √ (1, 0, −1, 0), (−1, 1, −1, 1), (1, 1, 1, 1) . 2 2 2 2 Relative to these principal axes, the quadratic form reduces to 4y32 + 8y42 . a b 18. A = where a, b, c ∈ R. b c a−λ c Consider det(A − λI) = 0 ⇐⇒ = 0 ⇐⇒ λ2 − (a + c)λ + (ac − b2 ) = 0 c b−λ (a + c) ± (a − c)2 + 4b2 ⇐⇒ λ = . Now A has repeated eigenvalues 2 a 0 ⇐⇒ (a − c)2 + 4b2 = 0 ⇐⇒ a = c and b = 0 (since a, b, c ∈ R) ⇐⇒ A = 0 a ⇐⇒ A = aI2 ⇐⇒ A is a scalar matrix. 19. (a). A is a real n × n symmetric matrix =⇒ A possesses a complete set of eigenvectors =⇒ A is similar to a diagonal matrix =⇒ there exists an invertible matrix S such that S −1 AS = diag(λ, λ, λ, . . . , λ) where λ occurs n times. But diag(λ, λ, λ, . . . , λ) = λIn , so S −1 AS = λIn =⇒ AS = S(λIn ) =⇒ AS = λS =⇒ A = λSS −1 =⇒ A = λIn =⇒ A is a scalar matrix (b). Theorem: Let A be a nondefective n × n matrix. If λ is an eigenvalue of multiplicity n, then A is a scalar matrix. Proof: The proof is the same as that in part (a) since A has a complete set of eigenvectors. 20. Since real eigenvectors of A that correspond to distinct eigenvalues are orthogonal, it must be the case that if y = (y1 , y2 ) corresponds to λ2 where Ay = λ2 y, then x, y = 0 =⇒ (1, 2), (y1 , y2 ) = 0 =⇒ y1 + 2y2 = 0 =⇒ y1 = −2y2 =⇒ y = (−2y2 , y2 ) = y2 (−2, 1). Consequently, (−2, 1) is an eigenvector corresponding to λ2 . 21. (a). Let A be a real symmetric 2 × 2 matrix with two distinct eigenvalues, λ1 and λ2 , where v1 = (a, b) is an eigenvector corresponding to λ1 . Since real eigenvectors of A that correspond to distinct eigenvalues are orthogonal, it follows that if v2 = (c, d) corresponds to λ2 where av2 = λ2 v2 , then (c)2017 Pearson Education. Inc. 492 v1 , v2 = 0 =⇒ (a, b), (c, d) = 0, that is, ac + bd = 0. By inspection, we see that v2 = (−b, a). An orthonormal set of eigenvectors for A is √ ⎡ 1 a ⎢ a 2 + b2 Thus, if S = ⎣ 1 √ b a 2 + b2 √ 1 1 √ (a, b), (−b, a) . a 2 + b2 a 2 + b2 ⎤ 1 b a 2 + b2 ⎥ ⎦, then S T AS = diag(λ1 , λ2 ). 1 √ a a 2 + b2 −√ T −1 (b). S T AS = diag(λ1 , λ2 ) =⇒ AS = S diag(λ 1 , λ2 ), since S = S . Thus, 1 a −b λ1 0 a b A = S diag(λ1 , λ2 )S T = 2 a 0 λ2 −b a (a + b2 ) b 1 a −b λ1 a λ1 b = 2 a −λ2 b λ2 a (a + b2 ) b 2 2 1 λ1 a + λ2 b ab(λ1 − λ2 ) . = 2 (a + b2 ) ab(λ1 − λ2 ) λ1 b2 + λ2 a2 22. A is a real symmetric 3 × 3 matrix with eigenvalues λ1 and λ2 of multiplicity two. (a). Let v1 = (1, −1, 1) be an eigenvector of A that corresponds to λ1 . Since real eigenvectors of A that correspond to distinct eigenvalues are orthogonal, it must be the case that if v = (a, b, c) corresponds to λ2 where Av = λ2 v, then v1 , v = 0 =⇒ (1, −1, 1), (a, b, c) = 0 =⇒ a−b+c = 0 =⇒ v = r(1, 1, 0)+s(−1, 0, 1) where r and s are free variables. Consequently, v2 = (1, 1, 0) and v3 = (−1, 0, 1) are linearly independent eigenvectors corresponding to λ2 . Thus, {(1, 1, 0), (−1, 0, 1)} is a basis for E2 . v2 and v3 are not orthogonal since v2 , v3 = −1 = 0, so we will apply the Gram-Schmidt procedure to v2 and v3 . Let u2 = v2 = (1, 1, 0) and v3 , u2 1 u3 = v3 − u2 = (−1, 0, 1) + (1, 1, 0) = ||u2 ||2 2 1 1 − , ,1 . 2 2 v1 1 1 1 is a unit eigenvector corresponding to λ1 , and = √ , −√ , √ ||v 3 3 3 1 || u2 u3 1 1 1 1 2 are orthonormal eigenvectors corresponding = √ , √ , 0 , w3 = = −√ , √ , √ w2 = ||u2 || ||u3 || 2 2 6 6 6 to λ2 . ⎡ ⎤ 1 1 1 √ √ −√ ⎢ 3 2 6 ⎥ ⎢ 1 1 ⎥ ⎢ − √1 √ √ ⎥ Consequently, S = ⎢ ⎥ and S T AS = diag(λ1 , λ2 , λ2 ). ⎢ 3 2 6 ⎥ ⎣ 2 ⎦ 1 √ √ 0 3 6 Now, w1 = (b). Since S is an orthogonal matrix, S T AS = diag(λ1 , λ2 , λ2 ) =⇒ AS = S diag(λ1 , λ2 , λ3 ) =⇒ A = S diag(λ1 , λ2 , λ3 )S T =⇒ (c)2017 Pearson Education. Inc. 493 ⎡ 1 √ ⎢ 3 ⎢ ⎢ − √1 A=⎢ ⎢ 3 ⎣ 1 √ 3 ⎡ 1 √ ⎢ 3 ⎢ ⎢ − √1 =⎢ ⎢ 3 ⎣ 1 √ 3 1 √ 2 1 √ 2 0 1 √ 2 1 √ 2 0 ⎤ ⎤ ⎡ 1 1 1 1 √ √ −√ −√ ⎡ ⎤ ⎢ 6 ⎥ λ1 0 3 3 3 ⎥ 0 ⎥ ⎢ 1 1 ⎥ 1 ⎥ ⎢ √ ⎥ ⎣ 0 λ2 0 ⎦ ⎢ √ √ 0 ⎥ ⎥ ⎥ ⎢ 6 ⎥ 2 2 0 0 λ2 ⎣ 1 1 2 ⎦ 2 ⎦ √ √ √ −√ 6 6 6 6 ⎤⎡ ⎤ λ1 λ1 λ1 1 √ √ −√ −√ ⎡ ⎢ 3 3 3 ⎥ 6 ⎥ λ1 + 2λ2 ⎥ ⎢ ⎥ 1 λ2 λ2 ⎢ √ ⎥ 1⎣ √ ⎥ √ 0 −λ ⎥⎢ ⎥= 1 + λ2 ⎥ 3 6 ⎥⎢ 2 2 λ 1 − λ2 2 ⎦⎣ λ2 λ 2λ2 ⎦ √ √2 √ −√ 6 6 6 6 −λ1 + λ2 λ1 + 2λ2 −λ1 + λ2 ⎤ λ 1 − λ2 −λ1 + λ2 ⎦ . λ1 + 2λ2 23. (a). Let v1 , v2 ∈ Cn and recall that v1T v2 = [v1 , v2 ]. [Av1 , v2 ] = (Av1 )T v2 = (v1T AT )v2 = v1T (−A)v2 = −v1T (A)v2 = −v1T (Av2 ) = −v1T Av2 = [−v1 , Av2 ]. Thus, Av1 , v2 = −v1 , Av2 . (23.1) (b). Let v1 be an eigenvector corresponding to the eigenvalue λ1 , so Av1 = λ1 v1 . (23.2) Taking the inner product of (23.2) with v1 yields Av1 , v1 = λ1 v1 , v1 , that is Av1 , v1 = λ1 ||v1 ||2 . (23.3) Taking the complex conjugate of (23.3) gives Av1 , v1 = λ1 ||v1 ||2 , that is v1 , Av1 = λ1 ||v1 ||2 . (23.4) Adding (23.3) and (23.4), and using (23.1) with v2 = v1 yields (λ1 + λ1 )||v1 ||2 = 0. But ||v1 ||2 = 0, so λ1 + λ1 = 0, or equivalently, λ1 = −λ1 , which means that all nonzero eigenvalues of A are pure imaginary. 24. Let A be an n × n real skew-symmetric matrix where n is odd. Since A is real, the characteristic equation, det(A − λI) = 0, has real coefficients, so its roots come in conjugate pairs. By Problem 23, all nonzero solutions of det(A − λI) = 0 are pure imaginary, hence when n is odd, zero will be one of the eigenvalues of A. −λ 4 −4 25. det(A − λI) = 0 ⇐⇒ −4 −λ −2 = 0 ⇐⇒ λ3 + 36λ = 0 ⇐⇒ λ = 0 or λ = ±6i. 4 2 −λ ⎤ ⎤ ⎡ ⎤⎡ ⎡ 0 4 −4 v1 0 If λ = 0 then (A − λI)v = 0 assumes the form ⎣ −4 0 −2 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ =⇒ 2v1 + v3 = 0 and 0 4 2 0 v3 v2 − v3 = 0. If we let v3 = 2r ∈ C, then the solution set of this system is {(−r, 2r, 2r) : r ∈ C} so the eigenvectors corresponding to λ = 0 are v1 = r(−1, ⎡ 2, 2) where r⎤∈⎡ C. ⎤ ⎡ ⎤ 6i 4 −4 v1 0 If λ = −6i then (A−λI)v = 0 assumes the form ⎣ −4 6i −2 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ =⇒ 5v1 +(−2+6i)v3 = 0 4 2 6i v3 0 and 5v2 + (4 + 3i)v3 = 0. If we let v3 = 5s ∈ C, then the solution set of this system is {(2 − 6i)s, (−4 − 3i)s, 5s : s ∈ C}, so the eigenvectors corresponding to λ = −6i are v2 = s(2 − 6i, −4 − 3i, 5) (c)2017 Pearson Education. Inc. 494 where s ∈ C. By Theorem 7.1.8, since A is a matrix with real entries, the eigenvectors corresponding to λ = 6i are of the form v3 = t(2 + 6i, −4 + 3i, 5) where t ∈ C. −λ −1 −6 √ 5 = 0 ⇐⇒ −λ3 − 62λ = 0 ⇐⇒ λ = 0 or λ = ± 62i. 26. det(A − λI) = 0 ⇐⇒ 1 −λ 6 −5 −λ ⎤ ⎡ ⎡ ⎤⎡ ⎤ 0 −1 −6 v1 0 0 5 ⎦ ⎣ v2 ⎦ = ⎣ 0 ⎦ =⇒ v3 = t, v2 = −6t, v1 = If λ = 0 then (A − λI)v = 0 assumes the form ⎣ 1 6 −5 0 v3 0 −5t, where t ∈ C. Thus, the solution set of the system is {(−5t, −6t, t) : t ∈ C}, so the eigenvectors corresponding to λ = 0 are v = t(−5, −6, 1), where t ∈ C. For the other eigenvalues, it is best to use technology to generate the corresponding eigenvectors. 27. A is a real n×n orthogonal matrix ⇐⇒ A−1 = AT ⇐⇒ AT A = In . Suppose that A = [v1 , v2 , v3 , . . . , vn ]. Since the ith row of AT is equal to viT , the matrix multiplication assures us that the ij th entry of AT A is equal to viT , vi . Thus from the equality AT A = In = [δij ], an n × n matrix A = [v1 , v2 , v3 , . . . , vn ] is orthogonal if an only if viT , vi = δij , that is, if and only if the columns (rows) of A, {v1 , v2 , v3 , . . . , vn }, form an orthonormal set of vectors. Solutions to Section 7.6 True-False Review: (a): TRUE. See Remark 1 following Definition 7.6.1. (b): TRUE. Each Jordan block corresponds to a cycle of generalized eigenvectors, and each such cycle contains exactly one eigenvector. By construction, the eigenvectors (and generalized eigenvectors) are chosen to form a linearly independent set of vectors. (c): FALSE. For example, in a diagonalizable n × n matrix, the n linearly independent eigenvectors can be arbitrarily placed in the columns of the matrix S. Thus, an ample supply of invertible matrices S can be constructed. 0 1 (d): FALSE. For instance, if J1 = J2 = , then J1 and J2 are in Jordan canonical form, but 0 0 0 2 is not in Jordan canonical form. J1 + J 2 = 0 0 (e): TRUE. This is simply a restatement of the definition of a generalized eigenvector. (f ): FALSE. The number of Jordan blocks corresponding to λ in the Jordan canonical form of A is the number of linearly independent eigenvectors of A corresponding to λ, which is dim[Eλ ], the dimension of the eigenspace corresponding to λ, not dim[Kλ ]. (g): TRUE. This is the content of Theorem 7.6.8. 1 1 , then J1 and J2 are in Jordan canonical form, but (h): FALSE. For instance, if J1 = J2 = 0 1 1 2 J1 J2 = is not in Jordan canonical form. 0 1 (i): TRUE. If we place the vectors in a cycle of generalized eigenvectors of A (see Equation (7.6.3)) in the columns of the matrix S formulated in this section in the order they appear in the cycle, then the corresponding columns of the matrix S −1 AS will form a Jordan block. (c)2017 Pearson Education. Inc. 495 (j): TRUE. The assumption here is that all Jordan blocks have size 1 × 1, which precisely says that the Jordan canonical form of A is a diagonal matrix. This means that A is diagonalizable. (k): TRUE. Suppose that S −1 AS = B and that J is a Jordan canonical form of A. So there exists an invertible matrix T such that T −1 AT = J. Then B = S −1 AS = S −1 (T JT −1 )S = (T −1 S)−1 J(T −1 S), and hence, (T −1 S)B(T −1 S)−1 = J, which shows that J is also a Jordan canonical form of B. 0 1 0 (l): FALSE. For instance, if J = and r = 2, then J is in Jordan canonical form, but rJ = 0 0 0 is not in Jordan canonical form. Problems: 1. There is just one Jordan canonical form (disregarding rearrangements of the Jordan blocks): ⎡ ⎤ −4 0 0 ⎣ 0 0 0 ⎦. 0 0 9 2. There are 3 possible Jordan canonical forms: ⎡ ⎤ ⎡ 1 0 0 1 ⎣ 0 1 0 ⎦, ⎣ 0 0 0 1 0 1 1 0 3. There are 4 possible Jordan canonical forms: ⎤ ⎤ ⎡ ⎡ 1 1 0 0 1 0 0 0 ⎢ 0 1 0 0 ⎥ ⎢ 0 1 0 0 ⎥ ⎥ ⎥ ⎢ ⎢ ⎣ 0 0 3 0 ⎦, ⎣ 0 0 3 0 ⎦, 0 0 0 3 0 0 0 3 ⎤ 0 0 ⎦, 1 ⎡ ⎤ 0 1 ⎦. 1 1 ⎣ 0 0 1 1 0 0 1 0 0 0 0 3 0 ⎤ 0 0 ⎥ ⎥, 1 ⎦ 3 2 ⎢ 0 ⎢ ⎢ 0 ⎢ ⎣ 0 0 1 2 0 0 0 0 0 2 0 0 0 0 1 2 0 0 0 1 2 0 ⎤ 0 0 ⎥ ⎥ 0 ⎥ ⎥, 0 ⎦ 2 ⎡ 1 ⎢ 0 ⎢ ⎣ 0 0 1 1 0 0 0 0 3 0 ⎤ 0 0 ⎥ ⎥. 1 ⎦ 3 2 ⎢ 0 ⎢ ⎢ 0 ⎢ ⎣ 0 0 1 2 0 0 0 0 1 2 0 0 0 0 1 2 0 ⎤ 0 0 ⎥ ⎥ 0 ⎥ ⎥ 1 ⎦ 2 ⎡ 1 ⎢ 0 ⎢ ⎣ 0 0 4. There are 7 possible Jordan canonical forms: ⎡ 2 ⎢ 0 ⎢ ⎢ 0 ⎢ ⎣ 0 0 0 2 0 0 0 0 0 2 0 0 0 0 0 2 0 ⎡ 2 ⎢ 0 ⎢ ⎢ 0 ⎢ ⎣ 0 0 ⎡ ⎤ 0 0 ⎥ ⎥ 0 ⎥ ⎥, 0 ⎦ 2 1 2 0 0 0 0 1 2 0 0 2 ⎢ 0 ⎢ ⎢ 0 ⎢ ⎣ 0 0 1 2 0 0 0 0 0 2 0 0 0 0 0 2 0 ⎤ 0 0 ⎥ ⎥ 0 ⎥ ⎥, 1 ⎦ 2 0 0 0 2 0 ⎡ ⎤ 0 0 ⎥ ⎥ 0 ⎥ ⎥, 0 ⎦ 2 ⎡ 2 ⎢ 0 ⎢ ⎢ 0 ⎢ ⎣ 0 0 1 2 0 0 0 0 1 2 0 0 5. There are 10 possible Jordan canonical forms: (c)2017 Pearson Education. Inc. ⎡ 2 ⎢ 0 ⎢ ⎢ 0 ⎢ ⎣ 0 0 ⎡ ⎤ 0 0 ⎥ ⎥ 0 ⎥ ⎥, 0 ⎦ 2 1 2 0 0 0 0 1 2 0 0 0 0 0 2 0 ⎤ 0 0 ⎥ ⎥ 0 ⎥ ⎥ 0 ⎦ 2 2 0 496 ⎡ 3 ⎢ 0 ⎢ ⎢ 0 ⎢ ⎢ 0 ⎢ ⎣ 0 0 ⎡ 3 ⎢ 0 ⎢ ⎢ 0 ⎢ ⎢ 0 ⎢ ⎣ 0 0 0 3 0 0 0 0 0 0 3 0 0 0 0 0 0 3 0 0 0 0 0 0 9 0 1 3 0 0 0 0 0 1 3 0 0 0 0 0 1 3 0 0 0 0 0 1 9 0 ⎤ 0 0 ⎥ ⎥ 0 ⎥ ⎥, 0 ⎥ ⎥ 0 ⎦ 9 ⎤ 0 0 ⎥ ⎥ 0 ⎥ ⎥, 0 ⎥ ⎥ 0 ⎦ 9 ⎡ 3 ⎢ 0 ⎢ ⎢ 0 ⎢ ⎢ 0 ⎢ ⎣ 0 0 ⎡ 3 ⎢ 0 ⎢ ⎢ 0 ⎢ ⎢ 0 ⎢ ⎣ 0 0 ⎡ 3 ⎢ 0 ⎢ ⎢ 0 ⎢ ⎢ 0 ⎢ ⎣ 0 0 1 3 0 0 0 0 0 0 3 0 0 0 0 0 0 3 0 0 0 0 0 0 9 0 0 3 0 0 0 0 0 0 3 0 0 0 0 0 0 3 0 0 0 0 0 0 9 0 1 3 0 0 0 0 0 1 3 0 0 0 0 0 0 3 0 0 0 0 0 0 9 0 ⎤ 0 0 ⎥ ⎥ 0 ⎥ ⎥, 0 ⎥ ⎥ 0 ⎦ 9 ⎤ 0 0 ⎥ ⎥ 0 ⎥ ⎥, 0 ⎥ ⎥ 1 ⎦ 9 ⎤ 0 0 ⎥ ⎥ 0 ⎥ ⎥, 0 ⎥ ⎥ 1 ⎦ 9 ⎡ 3 ⎢ 0 ⎢ ⎢ 0 ⎢ ⎢ 0 ⎢ ⎣ 0 0 ⎡ 3 ⎢ 0 ⎢ ⎢ 0 ⎢ ⎢ 0 ⎢ ⎣ 0 0 ⎡ 3 ⎢ 0 ⎢ ⎢ 0 ⎢ ⎢ 0 ⎢ ⎣ 0 0 1 3 0 0 0 0 0 0 3 0 0 0 0 0 1 3 0 0 0 0 0 0 9 0 1 3 0 0 0 0 0 0 3 0 0 0 0 0 0 3 0 0 0 0 0 0 9 0 1 3 0 0 0 0 0 1 3 0 0 0 0 0 1 3 0 0 0 0 0 1 9 0 ⎤ 0 0 ⎥ ⎥ 0 ⎥ ⎥, 0 ⎥ ⎥ 0 ⎦ 9 ⎤ 0 0 ⎥ ⎥ 0 ⎥ ⎥, 0 ⎥ ⎥ 1 ⎦ 9 ⎤ 0 0 ⎥ ⎥ 0 ⎥ ⎥. 0 ⎥ ⎥ 1 ⎦ 9 ⎡ 3 ⎢ 0 ⎢ ⎢ 0 ⎢ ⎢ 0 ⎢ ⎣ 0 0 ⎡ 3 ⎢ 0 ⎢ ⎢ 0 ⎢ ⎢ 0 ⎢ ⎣ 0 0 1 3 0 0 0 0 0 1 3 0 0 0 0 0 0 3 0 0 0 0 0 0 9 0 1 3 0 0 0 0 0 0 3 0 0 0 0 0 1 3 0 0 0 0 0 0 9 0 ⎤ 0 0 ⎥ ⎥ 0 ⎥ ⎥, 0 ⎥ ⎥ 0 ⎦ 9 ⎤ 0 0 ⎥ ⎥ 0 ⎥ ⎥, 0 ⎥ ⎥ 1 ⎦ 9 6. Since λ = 2 occurs with multiplicity 4, it can give rise to the following possible Jordan block sizes: (a) 4 (b) 3,1 (c) 2,2 (d) 2,1,1 (e) 1,1,1,1 Likewise, λ = 6 occurs with multiplicity 4, so it can give rise to the same five possible Jordan block sizes. Finally, λ = 8 occurs with multiplicity 3, so it can give rise to three possible Jordan block sizes: (a) 3 (b) 2,1 (c) 1,1,1 Since the block sizes for each eigenvalue can be independently determined, we have 5 · 5 · 3 = 75 possible Jordan canonical forms. 7. Since λ = 2 occurs with multiplicity 4, it can give rise to the following possible Jordan block sizes: (a) 4 (b) 3,1 (c) 2,2 (d) 2,1,1 (e) 1,1,1,1 Next, λ = 5 occurs with multiplicity 6, so it can give rise to the following possible Jordan block sizes: (a) 6 (b) 5,1 (c) 4,2 (d) 4,1,1 (e) 3,3 (f) 3,2,1 (c)2017 Pearson Education. Inc. 497 (g) 3,1,1,1 (h) 2,2,2 (i) 2,2,1,1 (j) 2,1,1,1,1 (k) 1,1,1,1,1,1 There are 5 possible Jordan block sizes corresponding to λ = 2 and 11 possible Jordan block sizes corresponding to λ = 5. Multiplying these results, we have 5 · 11 = 55 possible Jordan canonical forms. 8. Since (A − 5I)2 = 0, no cycles of generalized eigenvectors corresponding to λ = 5 can have length greater than 2, and hence, only Jordan block sizes 2 or less are possible. Thus, the possible block sizes under this restriction (corresponding to λ = 5) are: 2,2,2 2,2,1,1 2,1,1,1,1 1,1,1,1,1,1 There are four such. There are still five possible block size lists corresponding to λ = 2. Multiplying these results, we have 5 · 4 = 20 possible Jordan canonical forms under this restriction. 9. (a). ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 0 0 0 0 0 0 0 0 0 λ1 0 λ1 1 λ1 1 ⎢ 0 λ1 0 ⎢ ⎢ 0 0 ⎥ 0 0 ⎥ 0 0 ⎥ ⎢ ⎥ ⎢ 0 λ1 0 ⎥ ⎢ 0 λ1 1 ⎥ ⎢ 0 ⎥ ⎢ ⎥ ⎢ 0 λ1 0 0 λ1 0 0 λ1 0 0 ⎥,⎢ 0 0 ⎥,⎢ 0 0 ⎥ ⎢ ⎥, ⎣ 0 0 0 λ2 0 ⎦ ⎣ 0 0 0 λ2 0 ⎦ ⎣ 0 0 0 λ2 0 ⎦ 0 0 0 0 λ2 0 0 0 0 λ2 0 0 0 0 λ2 ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ 0 0 0 0 0 0 0 0 0 λ1 1 λ1 1 λ1 0 ⎢ ⎢ ⎢ 0 λ1 0 0 0 ⎥ 0 0 ⎥ 0 0 ⎥ ⎥ ⎥ ⎢ 0 λ1 0 ⎥ ⎢ 0 λ1 1 ⎢ ⎥ ⎥,⎢ 0 ⎥,⎢ 0 ⎢ 0 0 λ 0 0 0 λ 0 λ 0 0 0 0 1 1 1 ⎥ ⎥ ⎢ ⎥ ⎢ ⎢ ⎣ 0 0 0 λ2 1 ⎦ 0 0 λ2 1 ⎦ ⎣ 0 0 0 λ2 1 ⎦ ⎣ 0 0 0 0 0 λ2 0 0 0 0 λ2 0 0 0 0 λ2 (b). The assumption that (A − λ1 I)2 = 0 implies that there can be no Jordan blocks corresponding to λ1 of size 3 × 3 (or greater). Thus, the only possible Jordan canonical forms for this matrix now are ⎤ ⎡ ⎤ ⎡ λ1 1 λ1 0 0 0 0 0 0 0 ⎢ ⎢ 0 λ1 0 0 0 ⎥ 0 0 ⎥ ⎥ ⎢ 0 λ1 0 ⎥ ⎢ ⎢ ⎥ ⎢ 0 0 λ1 0 0 λ1 0 0 ⎥,⎢ 0 0 ⎥ ⎥, ⎢ ⎣ 0 0 0 λ2 0 ⎦ ⎣ 0 0 0 λ2 0 ⎦ 0 0 0 0 λ2 0 0 0 0 λ2 ⎡ ⎤ ⎡ ⎤ λ1 0 0 0 0 0 0 0 λ1 1 ⎢ 0 λ1 0 ⎢ 0 0 ⎥ 0 0 ⎥ ⎢ ⎥ ⎢ 0 λ1 0 ⎥ ⎢ 0 ⎥ ⎥ ⎢ 0 0 λ 0 λ 0 0 0 0 , 1 1 ⎢ ⎥ ⎥ ⎢ ⎣ 0 0 0 λ2 1 ⎦ ⎣ 0 0 0 λ2 1 ⎦ 0 0 0 0 λ2 0 0 0 0 λ2 10. The assumption that (A − λI)3 = 0 implies no Jordan blocks of size greater than 3 × 3 are possible. The fact that (A − λI)2 = 0 implies that there is at least one Jordan block of size 3 × 3. Thus, the possible (c)2017 Pearson Education. Inc. 498 block size combinations for a 6 × 6 matrix with eigenvalue λ of multiplicity 6 and no blocks of size greater than 3 × 3 with at least one 3 × 3 block are: 3,3 3,2,1 3,1,1,1 Thus, there are 3 possible Jordan canonical forms. (We omit the list itself; it can be produced simply from the list of block sizes above.) 11. The eigenvalues of the matrix with this characteristic polynomial are λ = 4, 4, −6. The possible Jordan canonical forms in this case are therefore: ⎤ ⎡ ⎤ ⎡ 4 0 0 4 1 0 ⎣ 0 4 0 ⎦,⎣ 0 4 0 ⎦. 0 0 −6 0 0 −6 12. The eigenvalues of the matrix with this characteristic polynomial are λ = 4, 4, 4, −1, −1. The possible Jordan canonical forms in this case are therefore: ⎡ 4 ⎢ 0 ⎢ ⎢ 0 ⎢ ⎣ 0 0 ⎡ 4 ⎢ 0 ⎢ ⎢ 0 ⎢ ⎣ 0 0 0 4 0 0 0 0 4 0 0 0 ⎤ ⎡ 4 0 0 0 ⎢ 0 0 0 0 ⎥ ⎥ ⎢ ⎢ 4 0 0 ⎥ ⎥,⎢ 0 ⎣ 0 ⎦ 0 −1 0 0 0 0 −1 ⎤ ⎡ 4 0 0 0 ⎢ 0 0 0 0 ⎥ ⎥ ⎢ ⎢ 4 0 0 ⎥ ⎥,⎢ 0 0 −1 1 ⎦ ⎣ 0 0 0 0 −1 1 4 0 0 0 1 4 0 0 0 ⎤ ⎡ 4 0 0 0 ⎢ 0 0 0 0 ⎥ ⎥ ⎢ ⎢ 4 0 0 ⎥ ⎥,⎢ 0 ⎣ 0 ⎦ 0 −1 0 0 0 −1 0 ⎤ ⎡ 4 0 0 0 ⎢ 0 0 0 0 ⎥ ⎥ ⎢ ⎢ 4 0 0 ⎥ ⎥,⎢ 0 0 −1 1 ⎦ ⎣ 0 0 0 0 −1 ⎤ 1 0 0 0 4 1 0 0 ⎥ ⎥ 0 4 0 0 ⎥ ⎥, 0 0 −1 0 ⎦ 0 0 0 −1 ⎤ 1 0 0 0 4 1 0 0 ⎥ ⎥ 0 4 0 0 ⎥ ⎥. 0 0 −1 1 ⎦ 0 0 0 −1 13. The eigenvalues of the matrix with this characteristic polynomial are λ = −2, −2, −2, 0, 0, 3, 3. The possible Jordan canonical forms in this case are therefore: ⎤ ⎡ −2 0 0 0 0 0 0 −2 0 0 ⎢ 0 −2 ⎢ 0 0 0 0 0 ⎥ 0 ⎥ ⎢ 0 −2 ⎢ ⎢ 0 ⎢ 0 −2 0 0 0 0 ⎥ 0 −2 ⎥ ⎢ 0 ⎢ ⎥ ⎢ 0 0 0 0 0 0 0 ⎥,⎢ 0 0 ⎢ ⎢ 0 ⎢ 0 ⎢ 0 0 0 0 0 0 ⎥ 0 0 ⎥ ⎢ 0 ⎢ ⎣ 0 0 0 0 0 3 0 ⎦ ⎣ 0 0 0 0 0 0 0 0 0 3 0 0 0 ⎡ 0 0 0 0 0 0 0 ⎤ ⎤ ⎡ −2 0 0 0 0 0 0 0 0 0 ⎢ 0 0 0 0 0 ⎥ 0 0 0 ⎥ ⎥ ⎥ ⎢ 0 −2 ⎥ ⎢ 0 −2 0 0 0 0 ⎥ 0 0 0 ⎥ ⎢ 0 ⎥ ⎢ 0 0 0 0 0 0 ⎥ 1 0 0 ⎥ ⎥, ⎥,⎢ 0 ⎢ 0 0 0 0 0 0 ⎥ 0 0 ⎥ ⎥ ⎥ ⎢ 0 0 0 0 0 3 1 ⎦ 0 3 0 ⎦ ⎣ 0 0 0 0 0 0 0 3 0 0 3 ⎤ ⎡ ⎤ ⎡ ⎤ −2 1 0 0 0 0 0 −2 1 0 0 0 0 0 −2 0 0 0 0 0 0 ⎢ ⎢ ⎢ 0 −2 0 0 0 0 0 ⎥ 0 0 0 0 0 ⎥ 0 0 0 0 0 ⎥ ⎥ ⎢ 0 −2 ⎥ ⎢ 0 −2 ⎢ ⎥ ⎢ ⎥ ⎢ ⎥ ⎢ 0 0 −2 0 0 0 0 ⎥ 0 −2 0 0 0 0 ⎥ ⎢ 0 0 −2 0 0 0 0 ⎥ ⎢ 0 ⎢ ⎥ ⎢ ⎢ ⎢ 0 0 0 0 1 0 0 ⎥ 0 0 0 0 0 0 ⎥ 0 0 0 1 0 0 ⎥ ⎥,⎢ 0 ⎥,⎢ 0 ⎢ ⎥, ⎥ ⎢ 0 ⎥ ⎢ 0 ⎢ 0 ⎥ 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 ⎥ ⎢ ⎥ ⎢ ⎢ ⎥ ⎣ ⎦ ⎣ ⎦ ⎣ 0 0 0 0 0 0 3 0 ⎦ 0 0 0 0 0 3 0 0 0 0 0 3 1 0 0 0 0 0 0 3 0 0 0 0 0 0 3 0 0 0 0 0 0 3 ⎡ (c)2017 Pearson Education. Inc. 499 ⎡ −2 1 0 0 ⎢ 0 −2 0 0 ⎢ ⎢ 0 0 −2 0 ⎢ ⎢ 0 0 0 0 ⎢ ⎢ 0 0 0 0 ⎢ ⎣ 0 0 0 0 0 0 0 0 ⎡ −2 1 0 0 ⎢ 0 −2 1 0 ⎢ ⎢ 0 0 −2 0 ⎢ ⎢ 0 0 0 0 ⎢ ⎢ 0 0 0 0 ⎢ ⎣ 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 3 0 0 0 0 0 0 0 1 0 0 0 3 0 0 ⎤ ⎡ −2 1 0 0 ⎢ 0 0 ⎥ ⎥ ⎢ 0 −2 ⎢ 0 −2 0 ⎥ ⎥ ⎢ 0 ⎥ 0 0 0 ⎥,⎢ ⎢ 0 ⎢ 0 0 0 ⎥ ⎥ ⎢ 0 0 0 1 ⎦ ⎣ 0 0 0 0 3 ⎤ ⎡ 0 −2 1 0 ⎢ 0 −2 0 ⎥ 1 ⎥ ⎢ ⎢ 0 −2 0 ⎥ ⎥ ⎢ 0 ⎥ 0 0 0 0 ⎥,⎢ ⎢ ⎢ 0 0 0 0 ⎥ ⎥ ⎢ 0 0 0 ⎦ ⎣ 0 0 0 0 3 ⎤ ⎡ −2 1 0 0 0 0 0 ⎢ 1 0 0 0 0 ⎥ ⎥ ⎢ 0 −2 ⎢ 0 −2 0 0 0 0 ⎥ ⎥ ⎢ 0 ⎥ 0 0 0 1 0 0 ⎥,⎢ ⎢ 0 ⎢ 0 0 0 0 0 0 ⎥ ⎥ ⎢ 0 0 0 0 0 3 1 ⎦ ⎣ 0 0 0 0 0 0 0 3 ⎤ ⎡ −2 1 0 0 0 0 0 ⎢ 0 −2 1 0 0 0 0 ⎥ ⎥ ⎢ ⎢ 0 −2 0 0 0 0 ⎥ ⎥ ⎢ 0 ⎥ 0 0 0 0 0 0 0 ⎥,⎢ ⎢ ⎢ 0 0 0 0 0 0 0 ⎥ ⎥ ⎢ 0 0 0 0 3 1 ⎦ ⎣ 0 0 0 0 0 0 0 3 ⎤ 0 0 0 0 0 0 0 0 ⎥ ⎥ 0 0 0 0 ⎥ ⎥ 0 0 0 0 ⎥ ⎥, 0 0 0 0 ⎥ ⎥ 0 0 3 0 ⎦ 0 0 0 3 ⎤ 0 0 0 0 0 0 0 0 ⎥ ⎥ 0 0 0 0 ⎥ ⎥ 0 1 0 0 ⎥ ⎥. 0 0 0 0 ⎥ ⎥ 0 0 3 1 ⎦ 0 0 0 3 14. The eigenvalues of the matrix with this characteristic polynomial are λ = −2, −2, 6, 6, 6, 6, 6. The possible Jordan canonical forms in this case are therefore: ⎡ −2 0 ⎢ 0 −2 ⎢ ⎢ 0 0 ⎢ ⎢ 0 0 ⎢ ⎢ 0 0 ⎢ ⎣ 0 0 0 0 ⎡ ⎢ ⎢ ⎢ ⎢ ⎢ ⎢ ⎢ ⎢ ⎣ 0 0 6 0 0 0 0 −2 1 0 0 −2 0 0 0 6 0 0 0 0 0 0 0 0 0 0 0 0 ⎤ ⎡ ⎤ ⎡ 0 −2 −2 0 0 0 0 0 0 ⎢ 0 −2 0 0 0 0 0 ⎥ ⎢ 0 0 ⎥ ⎥ ⎢ ⎥ ⎢ ⎢ ⎢ 0 ⎥ 0 6 0 0 0 0 ⎥ ⎥ ⎢ 0 ⎥ ⎢ 0 ⎥ ⎥ ⎢ 0 ⎥,⎢ 0 0 0 6 1 0 0 ⎥,⎢ ⎢ 0 ⎥ ⎢ ⎢ 0 ⎥ ⎢ 0 0 0 0 6 0 0 ⎥ ⎥ ⎢ 0 ⎦ ⎣ 0 ⎦ ⎣ 0 0 0 0 0 0 6 0 6 0 0 0 0 0 0 0 6 ⎤ ⎡ −2 0 0 0 0 −2 0 0 0 0 0 0 ⎢ 0 −2 0 0 0 0 0 ⎥ ⎢ 0 −2 0 0 0 ⎥ ⎢ ⎢ ⎢ ⎢ 0 0 6 1 0 0 6 1 0 0 0 ⎥ ⎥ ⎢ 0 ⎢ ⎥ ⎢ 0 ⎢ 0 0 0 6 1 0 0 6 1 0 0 , ⎥ ⎢ ⎢ ⎥ ⎢ 0 ⎢ 0 0 0 0 6 0 0 0 6 0 0 ⎥ ⎢ ⎢ ⎣ 0 0 0 0 0 0 0 0 0 6 0 ⎦ ⎣ 0 0 0 0 0 0 0 0 0 0 0 0 6 ⎤ ⎡ ⎡ −2 0 0 0 0 −2 0 0 0 0 0 0 ⎢ 0 −2 0 0 0 0 0 ⎥ ⎢ 0 −2 0 0 0 ⎥ ⎢ ⎢ ⎢ ⎢ 0 0 6 1 0 0 6 1 0 0 0 ⎥ ⎥ ⎢ 0 ⎢ ⎥,⎢ 0 ⎢ 0 0 0 6 1 0 0 6 1 0 0 ⎥ ⎢ ⎢ ⎥ ⎢ 0 ⎢ 0 0 0 0 6 0 0 0 6 1 0 ⎥ ⎢ ⎢ ⎣ 0 0 0 0 0 0 0 0 0 6 0 ⎦ ⎣ 0 0 0 0 0 0 0 6 0 0 0 0 0 ⎤ ⎡ ⎤ ⎡ −2 −2 1 0 0 0 0 0 0 0 0 0 ⎢ 0 −2 0 0 0 0 0 ⎥ ⎢ 0 0 0 0 0 ⎥ ⎥ ⎢ ⎥ ⎢ ⎢ ⎢ 0 6 1 0 0 0 ⎥ 0 0 0 0 ⎥ ⎥ ⎢ 0 ⎥ ⎢ 0 ⎥ ⎥ ⎢ 0 0 6 0 0 0 ⎥,⎢ 6 0 0 0 ⎥,⎢ 0 ⎢ 0 ⎢ ⎥ ⎢ 0 0 0 6 0 0 ⎥ 0 6 0 0 ⎥ ⎢ 0 ⎥ ⎢ 0 ⎦ ⎣ 0 ⎦ ⎣ 0 0 0 0 0 6 0 0 0 6 0 0 0 0 0 0 0 0 6 0 0 0 6 0 0 0 6 0 0 0 ⎡ 0 0 0 0 6 0 0 0 0 0 0 0 6 0 (c)2017 Pearson Education. Inc. ⎤ 0 0 0 0 0 0 −2 0 0 0 0 0 ⎥ ⎥ 0 6 1 0 0 0 ⎥ ⎥ 0 0 6 0 0 0 ⎥ ⎥, 0 0 0 6 1 0 ⎥ ⎥ 0 0 0 0 6 0 ⎦ 0 0 0 0 0 6 ⎤ 0 0 0 0 ⎥ ⎥ 0 0 ⎥ ⎥ 0 0 ⎥ ⎥ 0 0 ⎥ ⎥ 6 1 ⎦ 0 6 ⎤ 0 0 0 0 ⎥ ⎥ 0 0 ⎥ ⎥ 0 0 ⎥ ⎥ 1 0 ⎥ ⎥ 6 1 ⎦ 0 6 ⎤ 1 0 0 0 0 0 −2 0 0 0 0 0 ⎥ ⎥ 0 6 1 0 0 0 ⎥ ⎥ 0 0 6 0 0 0 ⎥ ⎥, 0 0 0 6 1 0 ⎥ ⎥ 0 0 0 0 6 0 ⎦ 0 0 0 0 0 6 500 ⎡ −2 1 ⎢ 0 −2 ⎢ ⎢ 0 0 ⎢ ⎢ 0 0 ⎢ ⎢ 0 0 ⎢ ⎣ 0 0 0 0 ⎡ −2 1 ⎢ 0 −2 ⎢ ⎢ 0 0 ⎢ ⎢ 0 0 ⎢ ⎢ 0 0 ⎢ ⎣ 0 0 0 0 0 0 0 0 6 1 0 6 0 0 0 0 0 0 0 0 0 0 6 1 0 6 0 0 0 0 0 0 ⎤ ⎤ ⎡ −2 1 0 0 0 0 0 0 0 0 ⎥ ⎢ 0 0 0 ⎥ ⎥ ⎢ 0 −2 0 0 0 0 0 ⎥ ⎢ ⎥ 0 6 1 0 0 0 ⎥ 0 0 0 ⎥ ⎢ 0 ⎥ ⎢ ⎥ 0 0 6 1 0 0 ⎥ 1 0 0 ⎥,⎢ 0 ⎥ ⎢ 0 0 0 6 0 0 ⎥ 6 0 0 ⎥ ⎥ ⎥ ⎢ 0 0 0 0 0 6 1 ⎦ 0 6 0 ⎦ ⎣ 0 0 0 0 0 0 0 6 0 0 6 ⎤ ⎤ ⎡ −2 1 0 0 0 0 0 0 0 0 ⎥ ⎢ 0 0 0 ⎥ ⎥ ⎢ 0 −2 0 0 0 0 0 ⎥ ⎥ ⎢ 0 0 6 1 0 0 0 0 0 0 ⎥ ⎥ ⎥ ⎢ ⎥ ⎢ 0 0 0 6 1 0 0 1 0 0 ⎥ , ⎥ ⎥ ⎢ ⎥ ⎢ 0 0 0 0 6 1 0 6 1 0 ⎥ ⎥ ⎥ ⎢ 0 0 0 0 6 1 ⎦ 0 6 0 ⎦ ⎣ 0 0 0 0 0 0 0 6 0 0 6 15. Of the Jordan canonical forms in Problem 14, we are asked here to find the ones that contain exactly five Jordan blocks, since there is a correspondence between the Jordan blocks and the linearly independent eigenvectors. There are three: ⎡ ⎤ ⎤ ⎡ ⎤ ⎡ −2 0 0 0 0 0 0 −2 0 0 0 0 0 0 −2 1 0 0 0 0 0 ⎢ 0 −2 0 0 0 0 0 ⎥ ⎢ 0 −2 0 0 0 0 0 ⎥ ⎢ 0 −2 0 0 0 0 0 ⎥ ⎥ ⎥ ⎢ ⎥ ⎢ ⎢ ⎢ ⎢ ⎢ 0 0 6 1 0 0 0 ⎥ 0 6 1 0 0 0 ⎥ 0 6 1 0 0 0 ⎥ ⎥ ⎥ ⎢ 0 ⎥ ⎢ 0 ⎢ ⎢ ⎢ ⎢ 0 0 0 6 1 0 0 ⎥ 0 0 6 0 0 0 ⎥ 0 0 6 0 0 0 ⎥ ⎥ ⎥,⎢ 0 ⎥,⎢ 0 ⎢ ⎢ ⎢ ⎢ 0 0 0 0 6 0 0 ⎥ 0 0 0 6 1 0 ⎥ 0 0 0 6 0 0 ⎥ ⎥ ⎥ ⎢ 0 ⎥ ⎢ 0 ⎢ ⎣ 0 0 0 0 0 6 0 ⎦ 0 0 0 0 6 0 ⎦ ⎣ 0 0 0 0 0 6 0 ⎦ ⎣ 0 0 0 0 0 0 0 6 0 0 0 0 0 0 6 0 0 0 0 0 0 6 0 0 1 0 16. Many examples are possible here. Let A = . The only eigenvalue of A is 0. The vector 0 1 v= is not an eigenvector since Av = = λv. However A2 = 02 , so every vector is a generalized 1 0 eigenvector of A corresponding to λ = 0. ⎡ ⎤ 0 1 0 17. Many examples are possible here. Let A = ⎣ 0 0 0 ⎦. The only eigenvalue of A is 0. The vector 0 0 0 ⎡ ⎤ ⎡ ⎤ 0 1 v = ⎣ 1 ⎦ is not an eigenvector since Av = ⎣ 0 ⎦ = λv. However, A2 = 03 , so every vector is a generalized 0 0 eigenvector of A corresponding to λ = 0. 18. Since the given matrix is already a diagonal matrix, it is already in Jordan canonical form. Thus, we can simply take J = A and S = I2 . 19. The characteristic polynomial is det(A − λI) = det 4−λ −4 4 12 − λ = λ2 − 16λ + 64 = (λ − 8)2 , with roots λ = 8, 8. Let us determine the eigenvectors corresponding to this eigenvalue: (c)2017 Pearson Education. Inc. 501 Eigenvalue λ = 8: Consider nullspace(A − 8I) = nullspace −4 −4 4 4 , which is equivalent to the equation −4x + 4y = 0. Therefore, if we set y = t, then x = t. Thus, setting t = 1 1 . Hence, we see that A is not diagonalizable, and therefore, we obtain the eigenvector 1 8 1 JCF(A) = . 0 8 To determine the matrix S, we must find a cycle of generalized eigenvectors of length 2. Therefore, it suffices 0 2 to find a vector v in R such that (A − 8I)v = 0. Many choices are possible here. We take v = . 1 4 Then, (A − 8I)v = . Thus, we have 4 4 0 S= . 4 1 20. The characteristic polynomial is 1−λ det(A − λI) = det −1 1 3−λ with roots λ = 2, 2. We have A − 2I = = (1 − λ)(3 − λ) + 1 = λ2 − 4λ + 4 = (λ − 2)2 , −1 −1 1 1 ∼ 1 −1 0 0 . Because there is only one unpivoted column in this latter matrix, we only have one eigenvector for A. Hence, A is not diagonalizable, and therefore 2 1 JCF(A) = . 0 2 To determine the matrix S, we must find a cycle of generalized eigenvectors of length 2. Therefore, itsuffices 0 2 . Then to find a vector v in R such that (A − 2I)v = 0. Many choices are possible here. We take v = 1 1 (A − 2I)v = . Thus, we have 1 1 0 S= . 1 1 21. The characteristic polynomial is ⎡ ⎤ 1−λ 1 1 1−λ 1 ⎦ = (1 − λ) (1 − λ)2 − 1 = (1 − λ)(λ2 − 2λ) = λ(1 − λ)(λ − 2), det(A − λI) = det ⎣ 0 0 1 1−λ with roots λ = 0, 1, 2. Since A is a 3×3 matrix with three distinct eigenvalues, it is diagonalizable. Therefore, it’s Jordan canonical form is simply a diagonal matrix with the eigenvalues as its diagonal entries: ⎤ ⎡ 0 0 0 JCF(A) = ⎣ 0 1 0 ⎦ . 0 0 2 (c)2017 Pearson Education. Inc. 502 To determine the invertible matrix S, we must find eigenvectors associated with each eigenvalue. Eigenvalue λ = 0: Consider ⎡ ⎤ 1 1 1 nullspace(A) = nullspace ⎣ 0 1 1 ⎦ , 0 1 1 ⎡ ⎤ 1 1 1 and this latter matrix can be row reduced to ⎣ 0 1 1 ⎦. The equations corresponding to the rows of this 0 0 0 matrix are x + y + z = 0 and y + z = 0. Setting z = t, then y = −t and x = 0. With t = 1 this gives us the eigenvector (0, −1, 1). Eigenvalue λ = 1: Consider ⎡ 0 nullspace(A − I) = nullspace ⎣ 0 0 1 0 1 ⎤ 1 1 ⎦. 0 By inspection, we see that z = 0 and y = 0 are required, but x = t is free. Thus, an eigenvector associated with λ = 1 may be chosen as (1, 0, 0). Eigenvalue λ = 2: Consider ⎡ ⎤ −1 1 1 1 ⎦, nullspace(A − 2I) = nullspace ⎣ 0 −1 0 1 −1 ⎡ ⎤ 1 −1 −1 1 −1 ⎦. Setting z = t, we have y = t and x = 2t. Thus, with t = 1 which can be row-reduced to ⎣ 0 0 0 0 we obtain the eigenvector (2, 1, 1) associated with λ = 2. Placing the eigenvectors obtained as the columns of S (with columns corresponding to the eigenvalues of JCF(A) above), we have ⎤ ⎡ 0 1 2 S = ⎣ −1 0 1 ⎦ . 1 0 1 22. We can get the characteristic polynomial by using cofactor expansion along the second column as follows: ⎡ 5−λ nullspace(A−λI) = det ⎣ 1 1 0 4−λ 0 ⎤ −1 −1 ⎦ = (4−λ) [(5 − λ)(3 − λ) + 1] = (4−λ)(λ2 −8λ+16) = (4−λ)(λ−4)2 , 3−λ with roots λ = 4, 4, 4. ⎡ ⎤ 1 0 −1 We have A − 4I = ⎣ 1 0 −1 ⎦, and so vectors (x, y, z) in the nullspace of this matrix must satisfy 1 0 −1 x − z = 0. Setting z = t and y = s, we have x = t. Hence, we obtain two linearly independent eigenvectors (c)2017 Pearson Education. Inc. 503 ⎡ ⎤ ⎡ ⎤ 1 0 of A corresponding to λ = 4: ⎣ 0 ⎦ and ⎣ 1 ⎦. Therefore, JCF(A) contains exactly two Jordan blocks. 1 0 This uniquely determines JCF(A), up to a rearrangement of the Jordan blocks: ⎡ ⎤ 4 1 0 JCF(A) = ⎣ 0 4 0 ⎦ . 0 0 4 To determine the matrix S, we must seek a generalized eigenvector. It is easy to verify that (A − 4I)2 = 03 , so every nonzero vector v is a generalized eigenvector. We must choose one such that (A − ⎡ 4I)v ⎤ = 0 in 1 order to form a cycle of length 2. There are many choices here, but let us choose v = ⎣ 0 ⎦. Then 0 ⎡ ⎤ 1 (A − 4I)v = ⎣ 1 ⎦. Notice that this is an eigenvector of A corresponding to λ = 4. To complete the matrix 1 S, we will need a second independent eigenvector. Again, there are a multitude of choices. Let us ⎡ linearly ⎤ 0 choose the eigenvector ⎣ 1 ⎦ found above. Thus, 0 ⎡ ⎤ 1 1 0 S = ⎣ 1 0 1 ⎦. 1 0 0 23. We will do cofactor expansion along the first column of the matrix to obtain the characteristic polynomial: ⎡ ⎤ 4−λ −4 5 4−λ 2 ⎦ det(A − λI) = det ⎣ −1 −1 2 4−λ = (4 − λ)(λ2 − 8λ + 12) + (−4)(4 − λ) − 10 − (−8 − 5(4 − λ)) = (4 − λ)(λ2 − 8λ + 12) + (2 − λ) = (4 − λ)(λ − 2)(λ − 6) + (2 − λ) = (λ − 2) [(4 − λ)(λ − 6) − 1] = (λ − 2)(−λ2 + 10λ − 25) = −(λ − 2)(λ − 5)2 , with eigenvalues λ = 2, 5, 5. Since λ = 5 is a repeated eigenvalue, ⎤we consider this eigenvalue first.⎡ We must consider the nullspace ⎡ ⎤ −1 −4 5 1 1 −2 2 ⎦, which can be row-reduced to ⎣ 0 3 −3 ⎦, a matrix that has of the matrix A − 5I = ⎣ −1 −1 −1 2 −1 0 0 0 only one unpivoted column, and hence λ = 5 only yields one linearly independent eigenvector. Thus, ⎤ ⎡ 2 0 0 JCF(A) = ⎣ 0 5 1 ⎦ . 0 0 5 (c)2017 Pearson Education. Inc. 504 To determine an invertible matrix S, we first proceed to find a cycle of eigenvectors of length 2 corresponding to λ = 5. Therefore, we must find a vector v in R3 such that (A − 5I)v = 0, but (A − 5I)2 v =⎡0 (in ⎤ order 1 0 18 −18 , so if we set v = ⎣ 0 ⎦, then that v is a generalized eigenvector). Note that (A − 5I)2 = 0 9 −9 0 ⎡ ⎤ −1 (A − 5I)v = ⎣ −1 ⎦ and (A − 5I)2 v = 0. Thus, we obtain the cycle −1 ⎧⎡ ⎤ ⎡ ⎤⎫ 1 ⎬ ⎨ −1 ⎣ −1 ⎦ , ⎣ 0 ⎦ . ⎩ ⎭ −1 0 Next, corresponding to ⎤ λ = 2, we must find an eigenvector. We need to ⎡find a nonzero ⎤vector (x, y, z) ⎡ 2 −4 5 1 −2 −2 2 2 ⎦, and this latter matrix can be row-reduced to ⎣ 0 0 1 ⎦. The middle in nullspace⎣ −1 −1 2 2 0 0 0 row requires that z = 0, and if we set y = t, then x = 2t. Thus, by using t = 1, we obtain the eigenvector (2, 1, 0). Thus, we can form the matrix ⎤ ⎡ 2 −1 1 S = ⎣ 1 −1 0 ⎦ . 0 −1 0 24. We are given that λ = −5 occurs with multiplicity 2 as a root of the characteristic polynomial of A. To search for corresponding eigenvectors, we consider ⎤ ⎡ −1 1 0 1 ⎦ nullspace(A + 5I) = nullspace ⎣ − 12 21 , 2 1 1 1 −2 2 −2 ⎡ ⎤ 1 −1 0 0 1 ⎦. Since there is only one unpivoted column in this row-echelon and this matrix row-reduces to ⎣ 0 0 0 0 form of A, the eigenspace corresponding to λ = −5 is only one-dimensional. Thus, based on the eigenvalues λ = −5, −5, −6, we already know that ⎡ ⎤ −5 1 0 0 ⎦. JCF(A) = ⎣ 0 −5 0 0 −6 Next, we seek a cycle of generalized eigenvectors of length 2 corresponding to λ = −5. The cycle 2 takes the form + 5I)v, v}, ⎤ where v is a vector such that (A + 5I) v = 0. We readily compute that ⎡ {(A 1 1 1 −2 2 2 0 0 ⎦. An obvious vector that is killed by (A + 5I)2 (although other choices are (A + 5I)2 = ⎣ 0 1 1 −⎡2 12⎤ 2 ⎤ ⎡ 0 1 also possible) is v = ⎣ 1 ⎦. Then (A + 5I)v = ⎣ 1 ⎦. Hence, we have a cycle of generalized eigenvectors 1 0 (c)2017 Pearson Education. Inc. 505 corresponding to λ = −5: ⎧⎡ ⎤ ⎡ ⎤⎫ 0 ⎬ ⎨ 1 ⎣ 1 ⎦,⎣ 1 ⎦ . ⎩ ⎭ 0 1 Now consider the eigenspace corresponding to λ = −6. We need only find one eigenvector (x, y, z) in this eigenspace. To do so, we must compute ⎤ ⎡ 0 1 0 nullspace(A + 6I) = nullspace ⎣ − 12 32 12 ⎦ , − 12 12 12 ⎡ ⎤ 1 −3 −1 1 0 ⎦. We see that y = 0 and x − 3y − z = 0, which is equivalent and this matrix row-reduces to ⎣ 0 0 0 0 to x − z = 0. Setting z = t, we have x = t. With t = 1, we obtain the eigenvector (1, 0, 1). Hence, we can form the matrix ⎡ ⎤ 1 0 1 S = ⎣ 1 1 0 ⎦. 0 1 1 25. Because the matrix is upper triangular, the eigenvalues of A appear along the main diagonal: λ = 2, 2, 3. Let us first consider the eigenspace corresponding to λ = 2: We consider ⎡ ⎤ 0 −2 14 1 −7 ⎦ , nullspace(A − 2I) = nullspace ⎣ 0 0 0 0 ⎡ ⎤ 0 1 −7 0 ⎦. There are two unpivoted columns, so this eigenspace is two-dimensional. which row reduces to ⎣ 0 0 0 0 0 Therefore, the matrix A is diagonalizable: ⎤ ⎡ 2 0 0 JCF(A) = ⎣ 0 2 0 ⎦ . 0 0 3 Next, we must determine an invertible matrix S such that S −1 AS is the Jordan canonical form just obtained. Using the row-echelon form of A − 2I obtained above, vectors (x, y, z) in nullspace(A − 2I) must satisfy y − 7z = 0. Setting z = t, y = 7t, and x = s, we obtain the eigenvectors (1, 0, 0) and (0, 7, 1). Next, consider the eigenspace corresponding to λ = 3. We consider ⎡ ⎤ −1 −2 14 0 −7 ⎦ , nullspace(A − 3I) = nullspace ⎣ 0 0 0 −1 ⎡ ⎤ 1 2 −14 1 ⎦. Vectors (x, y, z) in the nullspace of this matrix must satisfy z = 0 which row reduces to ⎣ 0 0 0 0 0 and x + 2y − 14z = 0 or x + 2y = 0. Setting y = t and x = −2t. Hence, setting t = 1 gives the eigenvector (−2, 1, 0). (c)2017 Pearson Education. Inc. 506 Thus, using the eigenvectors obtained above, we obtain the matrix ⎤ ⎡ 1 0 −2 1 ⎦. S=⎣ 0 7 0 1 0 26. We use the characteristic polynomial to determine the eigenvalues of A: ⎡ ⎤ 7−λ −2 2 4−λ −1 ⎦ det(A − λI) = det ⎣ 0 −1 1 4−λ = (4 − λ) [(7 − λ)(4 − λ) + 2] + (7 − λ − 2) = (4 − λ)(λ2 − 11λ + 30) + (5 − λ) = (4 − λ)(λ − 5)(λ − 6) + (5 − λ) = (5 − λ) [1 − (4 − λ)(6 − λ)] = (5 − λ)(λ − 5)2 = −(λ − 5)3 . Hence, the eigenvalues are λ = 5, 5, 5. Let us consider the eigenspace corresponding to λ = 5. We consider ⎡ ⎤ 2 −2 2 nullspace(A − 5I) = nullspace ⎣ 0 −1 −1 ⎦ , −1 1 −1 ⎡ ⎤ 1 −1 1 1 1 ⎦, which contains one unpivoted column. Therefore, the and this latter matrix row-reduces to ⎣ 0 0 0 0 eigenspace corresponding to λ = 5 is one-dimensional. Therefore, the Jordan canonical form of A consists of one Jordan block: ⎡ ⎤ 5 1 0 JCF(A) = ⎣ 0 5 1 ⎦ . 0 0 5 A corresponding invertible matrix S in this case must have columns that consist of one cycle of generalized eigenvectors, which will take the form {(A − 5I)2 v, (A − 5I)v, v}, where v is a generalized eigenvector. Now, we can verify quickly that ⎡ ⎤ ⎡ ⎤ 2 −2 2 2 0 4 2 ⎦ , (A − 5I)3 = 03 . A − 5I = ⎣ 0 −1 −1 ⎦ , (A − 5I)2 = ⎣ 1 0 −1 1 −1 −1 0 −2 we The fact that (A − 5I)3 = 03 means that every nonzero vector v is a generalized eigenvector. ⎡ Hence, ⎤ 1 simply choose v such that (A − 5I)2 v = 0. There are many choices. Let us take v = ⎣ 0 ⎦. Then 0 ⎡ ⎤ ⎡ ⎤ 2 2 (A − 5I)v = ⎣ 0 ⎦ and (A − 5I)2 v = ⎣ 1 ⎦. Thus, we have the cycle of generalized eigenvectors −1 −1 ⎧⎡ ⎤ ⎡ ⎤ ⎡ ⎤⎫ 2 2 1 ⎬ ⎨ ⎣ 1 ⎦,⎣ 0 ⎦,⎣ 0 ⎦ . ⎩ ⎭ −1 −1 0 (c)2017 Pearson Education. Inc. 507 Hence, we have ⎡ ⎤ 2 2 1 0 0 ⎦. S=⎣ 1 −1 −1 0 27. Because the matrix is upper triangular, the eigenvalues of A appear along the main diagonal: λ = −1, −1, −1. Let us consider the eigenspace corresponding to λ = −1. We consider ⎡ ⎤ 0 −1 0 0 −2 ⎦ , nullspace(A + I) = nullspace ⎣ 0 0 0 0 and it is straightforward to see that the nullspace here consists precisely of vectors that are multiples of (1, 0, 0). Because only one linearly independent eigenvector was obtained, the Jordan canonical form of this matrix consists of only one Jordan block: ⎡ ⎤ −1 1 0 1 ⎦. JCF(A) = ⎣ 0 −1 0 0 −1 A corresponding invertible matrix S in this case must have columns that consist of one cycle of generalized eigenvectors, which will take the form {(A + I)2 v, (A + I)v, v}. Here, we have ⎡ ⎤ ⎡ ⎤ 0 0 2 0 −1 0 0 −2 ⎦ A+I =⎣ 0 and (A + I)3 = 03 . and (A + I)2 = ⎣ 0 0 0 ⎦ 0 0 0 0 0 0 Therefore, every nonzero vector is a generalized eigenvector. We⎡ wish ⎤ to choose a vector ⎡v such⎤ that 0 0 (A + I)2 v = 0. There are many choices, but we will choose v = ⎣ 0 ⎦. Then (A + I)v = ⎣ −2 ⎦ and 1 0 ⎡ ⎤ 2 (A + I)2 v = ⎣ 0 ⎦. Hence, we form the matrix S as follows: 0 ⎤ 2 0 0 S = ⎣ 0 −2 0 ⎦ . 0 0 1 ⎡ 28. We use the characteristic polynomial to determine the eigenvalues of A: ⎤ ⎡ 2−λ −1 0 1 ⎢ 0 3−λ −1 0 ⎥ ⎥ det(A − λI) = det ⎢ ⎣ 0 1 1−λ 0 ⎦ 0 −1 0 3−λ = (2 − λ)(3 − λ) [(3 − λ)(1 − λ) + 1] = (2 − λ)(3 − λ)(λ2 − 4λ + 4) = (2 − λ)(3 − λ)(λ − 2)2 , and so the eigenvalues are λ = 2, 2, 2, 3. (c)2017 Pearson Education. Inc. 508 First, consider the eigenspace corresponding to λ = 2. We consider ⎤ ⎡ 0 −1 0 1 ⎢ 0 1 −1 0 ⎥ ⎥, nullspace(A − 2I) = nullspace ⎢ ⎣ 0 1 −1 0 ⎦ 0 −1 0 1 ⎤ ⎡ 0 1 −1 0 ⎥ ⎢ 0 0 1 −1 ⎥. There are two unpivoted columns, and and this latter matrix can be row-reduced to ⎢ ⎣ 0 0 0 0 ⎦ 0 0 0 0 therefore two linearly independent eigenvectors corresponding to λ = 2. Thus, we will obtain two Jordan blocks corresponding to λ = 2, and they necessarily will have size 2 × 2 and 1 × 1. Thus, we are already in a position to write down the Jordan canonical form of A: ⎤ ⎡ 2 1 0 0 ⎢ 0 2 0 0 ⎥ ⎥ JCF(A) = ⎢ ⎣ 0 0 2 0 ⎦. 0 0 0 3 We continue in order to obtain an invertible matrix S such that S −1 AS is in Jordan canonical form. To this end, we see a generalized eigenvector v such that (A − 2I)v = 0 and (A − 2I)2 v = 0. Note that ⎤ ⎤ ⎡ ⎡ 0 −1 0 1 0 −2 1 1 ⎢ 0 ⎢ 0 1 −1 0 ⎥ 0 0 0 ⎥ ⎥ ⎥. and (A − 2I)2 = ⎢ A − 2I = ⎢ ⎦ ⎣ 0 ⎣ 1 −1 0 0 0 0 0 ⎦ 0 −1 0 1 0 −2 1 1 ⎤ ⎡ 0 ⎢ 0 ⎥ ⎥ By inspection, we see that by taking v = ⎢ ⎣ −1 ⎦ (there are many other valid choices, of course), then 1 ⎤ ⎡ 1 ⎢ 1 ⎥ 2 ⎥ (A − 2I)v = ⎢ ⎣ 1 ⎦ and (A − 2I) v = 0. We also need a second eigenvector corresponding to λ = 2 that is 1 linearly independent from (A − 2I)v just obtained.⎡ From ⎤ the row-echelon form of A −⎡ 2I, ⎤we see that all 1 s ⎢ 0 ⎥ ⎢ t ⎥ ⎥ ⎥ ⎢ eigenvectors corresponding to λ = 2 take the form ⎢ ⎣ t ⎦, so for example, we can take ⎣ 0 ⎦. 0 t Next, we consider the eigenspace corresponding to λ = 3. We consider ⎤ ⎡ −1 −1 0 1 ⎢ 0 0 −1 0 ⎥ ⎥. nullspace(A − 3I) = nullspace ⎢ ⎣ 0 1 −2 0 ⎦ 0 −1 0 0 Now, if (x, y, z, w) is an eigenvector corresponding to λ = 3, the last three rows of the matrix imply that y = z = 0. Thus, the first row becomes −x + w = 0. Setting w = t, then x = t, so we obtain eigenvectors in the form (t, 0, 0, t). Setting t = 1 gives the eigenvector (1, 0, 0, 1). (c)2017 Pearson Education. Inc. 509 Thus, we can now form the matrix S such that S −1 AS is the Jordan canonical form we obtained above: ⎤ ⎡ 1 0 1 1 ⎢ 1 0 0 0 ⎥ ⎥ S=⎢ ⎣ 1 −1 0 0 ⎦ . 1 1 0 1 29. From the characteristic polynomial, we have eigenvalues λ = 2, 2, 4, 4. Let us consider the associated eigenspaces. Corresponding to λ = 2, we seek eigenvectors (x, y, z, w) by computing ⎤ ⎡ 0 −4 2 2 ⎢ −2 −2 1 3 ⎥ ⎥ nullspace(A − 2I) = nullspace ⎢ ⎣ −2 −2 1 3 ⎦ , −2 −6 3 5 ⎤ ⎡ 2 2 −1 −3 ⎢ 0 2 −1 −1 ⎥ ⎥. Setting w = 2t and z = 2s, we obtain y = s + t and this matrix can be row-reduced to ⎢ ⎣ 0 0 0 0 ⎦ 0 0 0 0 and x = 2t, so we obtain the eigenvectors (2, 1, 0, 2) and (0, 1, 2, 0) corresponding to λ = 2. Next, corresponding to λ = 4, we seek eigenvectors (x, y, z, w) by computing ⎤ ⎡ −2 −4 2 2 ⎢ −2 −4 1 3 ⎥ ⎥ nullspace(A − 4I) = nullspace ⎢ ⎣ −2 −2 −1 3 ⎦ . −2 −6 3 3 ⎤ ⎡ 1 2 −1 −1 ⎢ 0 2 −1 −1 ⎥ ⎥. Since there is only one unpivoted column, this eigenspace This matrix can be reduced to ⎢ ⎣ 0 0 1 −1 ⎦ 0 0 0 0 is only one-dimensional, despite λ = 4 occurring with multiplicity 2 as a root of the characteristic equation. Therefore, we must seek a generalized eigenvector v such that (A − 4I)v is an eigenvector. This in turn requires that (A − 4I)2 v = 0. We find that ⎤ ⎤ ⎡ ⎡ 4 8 −4 −4 −2 −4 2 2 ⎢ 4 4 ⎢ −2 −4 0 −4 ⎥ 1 3 ⎥ ⎥. ⎥ and (A − 4I)2 = ⎢ A − 4I = ⎢ ⎣ 4 0 ⎣ −2 −2 −1 3 ⎦ 4 −4 ⎦ 4 8 −4 −4 −2 −6 3 3 ⎤ ⎡ ⎤ ⎡ 1 0 ⎢ 0 ⎥ ⎢ 1 ⎥ 2 ⎥ ⎥ ⎢ Note that the vector v = ⎢ ⎣ 0 ⎦ satisfies (A − 4I) v = 0 and (A − 4I)v = ⎣ 1 ⎦. Thus, we have the cycle 1 1 of generalized eigenvectors corresponding to λ = 4 given by ⎧⎡ ⎤⎫ ⎤ ⎡ 1 ⎪ ⎪ ⎪ ⎪ 0 ⎨ ⎢ 1 ⎥ ⎢ 0 ⎥⎬ ⎥ . ⎥,⎢ ⎢ ⎣ 1 ⎦ ⎣ 0 ⎦⎪ ⎪ ⎪ ⎪ ⎭ ⎩ 1 1 (c)2017 Pearson Education. Inc. 510 Hence, we can form the matrix ⎡ 2 ⎢ 1 S=⎢ ⎣ 0 2 0 1 2 0 0 1 1 1 ⎤ 1 0 ⎥ ⎥ 0 ⎦ 1 ⎡ and 2 ⎢ 0 JCF(A) = ⎢ ⎣ 0 0 0 2 0 0 0 0 4 0 ⎤ 0 0 ⎥ ⎥. 1 ⎦ 4 30. Since A is upper triangular, the eigenvalues appear along the main diagonal: λ = 2, 2, 2, 2, 2. Looking at ⎤ ⎡ 0 1 1 1 1 ⎢ 0 0 0 0 1 ⎥ ⎥ ⎢ ⎥ nullspace(A − 2I) = nullspace ⎢ ⎢ 0 0 0 0 1 ⎥, ⎣ 0 0 0 0 1 ⎦ 0 0 0 0 0 we see that the row-echelon form of this matrix will contain two pivots, and therefore, three unpivoted columns. That means that the eigenspace corresponding to λ = 2 is three-dimensional. Therefore, JCF(A) consists of three Jordan blocks. The only list of block sizes for a 5 × 5 matrix with three blocks are (a) 3,1,1 and (b) 2,2,1. In this case, note that ⎤ ⎡ 0 0 0 0 4 ⎢ 0 0 0 0 1 ⎥ ⎥ ⎢ ⎥ (A − 2I)2 = ⎢ ⎢ 0 0 0 0 1 ⎥ = 05 , ⎣ 0 0 0 0 1 ⎦ 0 0 0 0 1 so that it is possible to find a vector v that generates a cycle of generalized eigenvectors of length 3: {(A − 2I)2 v, (A − 2I)v, v}. Thus, JCF(A) contains a Jordan block of size 3 × 3. We conclude that the correct list of block sizes for this matrix is 3,1,1: ⎤ ⎡ 2 1 0 0 0 ⎢ 0 2 1 0 0 ⎥ ⎥ ⎢ ⎥ JCF(A) = ⎢ ⎢ 0 0 2 0 0 ⎥. ⎣ 0 0 0 2 0 ⎦ 0 0 0 0 2 31. Since A is upper triangular, the eigenvalues appear along the main diagonal: λ = 0, 0, 0, 0, 0. Looking at nullspace(A − 0I) = nullspace(A), we see that eigenvectors (x, y, z, u, v) corresponding to λ = 0 must satisfy z = u = v = 0 (since the third row gives 6u = 0, the first row gives u + 4v = 0, and the second row gives z + u + v = 0). Thus, we have only two free variables, and thus JCF(A) will consist of two Jordan blocks. The only list of block sizes for a 5 × 5 matrix with two blocks are (a)4,1 and (b) 3,2. In this case, it is easy to verify that A3 = 0, so that the longest possible cycle of generalized eigenvectors {A2 v, Av, v} has length 3. Therefore, case (b) holds: JCF(A) consists of one Jordan block of size 3 × 3 and one Jordan block of size 2 × 2: ⎤ ⎡ 0 1 0 0 0 ⎢ 0 0 1 0 0 ⎥ ⎥ ⎢ ⎥ JCF(A) = ⎢ ⎢ 0 0 0 0 0 ⎥. ⎣ 0 0 0 0 1 ⎦ 0 0 0 0 0 (c)2017 Pearson Education. Inc. 511 32. Since A is upper triangular, the eigenvalues appear along the main diagonal: λ = 1, 1, 1, 1, 1, 1, 1, 1. Looking at ⎤ ⎡ 0 1 0 1 0 1 0 1 ⎢ 0 0 0 0 0 0 0 0 ⎥ ⎥ ⎢ ⎢ 0 0 0 1 0 1 0 1 ⎥ ⎥ ⎢ ⎢ 0 0 0 0 0 0 0 0 ⎥ ⎥ nullspace(A − I) = nullspace ⎢ ⎢ 0 0 0 0 0 1 0 1 ⎥, ⎥ ⎢ ⎢ 0 0 0 0 0 0 0 0 ⎥ ⎥ ⎢ ⎣ 0 0 0 0 0 0 0 1 ⎦ 0 0 0 0 0 0 0 0 we see that if (a, b, c, d, e, f, g, h) is an eigenvector of A, then b = d = f = h = 0, and a, c, e, and g are free variables. Thus, we have four linearly independent eigenvectors of A, and hence we expect four Jordan blocks. Now, an easy calculation shows that (A − I)2 = 0, and thus, no Jordan blocks of size greater than 2 × 2 are permissible. Thus, it must be the case that JCF(A) consists of four Jordan blocks, each of which is a 2 × 2 matrix: ⎤ ⎡ 1 1 0 0 0 0 0 0 ⎢ 0 1 0 0 0 0 0 0 ⎥ ⎥ ⎢ ⎢ 0 0 1 1 0 0 0 0 ⎥ ⎥ ⎢ ⎢ 0 0 0 1 0 0 0 0 ⎥ ⎥ JCF(A) = ⎢ ⎢ 0 0 0 0 1 1 0 0 ⎥. ⎥ ⎢ ⎢ 0 0 0 0 0 1 0 0 ⎥ ⎥ ⎢ ⎣ 0 0 0 0 0 0 1 1 ⎦ 0 0 0 0 0 0 0 1 33. NOT SIMILAR. We will compute JCF(A) and JCF(B). If they are the same (up to a rearrangement of the Jordan blocks), then A and B are similar; otherwise they are not. Both matrices have eigenvalues λ = 6, 6, 6. Consider ⎡ ⎤ 1 1 0 nullspace(A − 6I) = nullspace ⎣ −1 −1 0 ⎦ . 1 0 0 We see that eigenvectors (x, y, z) corresponding to λ = 6 must satisfy x + y = 0 and x = 0. Therefore, x = y = 0, while z is a free variable. Since we obtain only one free variable, JCF(A) consists of just one Jordan block corresponding to λ = 6: ⎡ ⎤ 6 1 0 JCF(A) = ⎣ 0 6 1 ⎦ . 0 0 6 Next, consider ⎡ ⎤ 0 −1 1 0 0 ⎦. nullspace(B − 6I) = nullspace ⎣ 0 0 0 0 In this case, eigenvectors (x, y, z) corresponding to λ = 6 must satisfy −y + z = 0. Therefore, both x and z are free variables, and hence, JCF(B) consists of two Jordan blocks corresponding to λ = 6: ⎤ ⎡ 6 1 0 JCF(B) = ⎣ 0 6 0 ⎦ . 0 0 6 (c)2017 Pearson Education. Inc. 512 Since JCF(A) = JCF(B), we conclude that A and B are not similar. 34. SIMILAR. We will compute JCF(A) and JCF(B). If they are the same (up to a rearrangement of the Jordan blocks), then A and B are similar; otherwise they are not. Both matrices have eigenvalues λ = 5, 5, 5. (For A, this is easiest to compute by expanding det(A − λI) along the middle row, and for B, this is easiest to compute by expanding det(B − λI) along the second column.) In Problem 26, we computed ⎡ ⎤ 5 1 0 JCF(A) = ⎣ 0 5 1 ⎦ . 0 0 5 Next, consider ⎡ ⎤ −2 −1 −2 1 1 ⎦, nullspace(B − 5I) = nullspace ⎣ 1 1 0 1 ⎡ ⎤ 1 1 1 and this latter matrix can be row-reduced to ⎣ 0 1 0 ⎦, which has only one unpivoted column. Thus, 0 0 0 the eigenspace of B corresponding to λ = 5 is only one-dimensional, and so ⎡ ⎤ 5 1 0 JCF(B) = ⎣ 0 5 1 ⎦ . 0 0 5 Since A and B each had the same Jordan canonical form, they are similar matrices. 35. SIMILAR. We will compute JCF(A) and JCF(B). If they are the same (up to a rearrangement of the Jordan blocks), then A and B are similar; otherwise they are not. Both matrices have eigenvalues λ = −1, −1, 2. For λ = −1, consider ⎡ ⎤ 4 0 4 0 ⎦. nullspace(A + I) = nullspace ⎣ 0 3 −4 0 −4 We see that the eigenvectors (x, y, z) corresponding to λ = −1 must satisfy x + z = 0 and y = 0. Therefore, we can choose z = t as a free variable, from which x = −t. Hence, we only obtain one linearly independent eigenvector (−1, 0, 1) corresponding to λ − 1, from which we conclude that ⎡ ⎤ −1 1 0 JCF(A) = ⎣ 0 −1 0 ⎦ . 0 0 2 Next, consider ⎡ ⎤ 0 −1 3 0 1 ⎦. nullspace(B + I) = nullspace ⎣ 0 0 0 3 In this case, eigenvectors (x, y, z) corresponding to λ = −1 must satisfy y = z = 0. Therefore, only x = t is a free variable, resulting in just one linearly independent eigenvector (1, 0, 0) corresponding to λ = −1. Therefore, we conclude that ⎡ ⎤ −1 1 0 JCF(B) = ⎣ 0 −1 0 ⎦ . 0 0 2 (c)2017 Pearson Education. Inc. 513 Since JCF(A) = JCF(B), we conclude that A and B are similar. 36. The eigenvalues of A are λ = −5, −5, and the eigenspace corresponding to λ = −5 is only onedimensional. Thus, we seek a generalized eigenvector v of A corresponding to λ = −5 such that {(A+5I)v, v} 1 1 is a cycle of generalized eigenvectors. Note that A + 5I = and (A + 5I)2 = 02 . Thus, every −1 −1 1 nonzero vector v is a generalized eigenvector of A corresponding to λ = −5. Let us choose v = . Then 0 1 (A + 5I)v = . Form the matrices −1 1 1 −5 1 S= and J = . −1 0 0 −5 Via the substitution x = Sy, the system x = Ax is transformed into y = Jy. The corresponding equations are y1 = −5y1 + y2 and y2 = −5y2 . The solution to the second equation is y2 (t) = c1 e−5t . Substituting this solution into y1 = −5y1 + y2 gives y1 + 5y1 = c1 e−5t . This is a first order linear equation with integrating factor I(t) = e5t . When we multiply the differential equation for y1 (t) by I(t), it becomes (y1 · e5t ) = c1 . Integrating both sides yields y1 · e5t = c1 t + c2 . Thus, y1 (t) = c1 te−5t + c2 e−5t . Thus, we have y(t) = y1 (t) y2 (t) = c1 te−5t + c2 e−5t c1 e−5t . Finally, we solve for x(t): c1 te−5t + c2 e−5t x(t) = Sy(t) = c1 e−5t c1 te−5t + c2 e−5t + c1 e−5t = −(c1 te−5t + c2 e−5t ) t+1 1 −5t −5t + c2 e . = c1 e −t −1 1 −1 1 0 This is an acceptable answer, or we can write the individual equations comprising the general solution: x1 (t) = c1 te−5t + c2 e−5t + c1 e−5t and x2 (t) = −(c1 te−5t + c2 e−5t ). 37. The eigenvalues of A are λ = −1, −1, and the eigenspace corresponding to λ = −1 is only onedimensional. Thus, we seek a generalized eigenvector v of A corresponding to λ = −1 such that {(A+I)v, v} −2 −2 is a cycle of generalized eigenvectors. Note that A + I = and (A + I)2 = 02 . Thus, every 2 2 (c)2017 Pearson Education. Inc. 514 nonzero vector v is a generalized eigenvector of A corresponding to λ = −1. Let us choose v = −2 (A + I)v = . Form the matrices 2 −2 1 −1 1 S= and J = . 2 0 0 −1 1 0 . Then Via the substitution x = Sy, the system x = Ax is transformed into y = Jy. The corresponding equations are y1 = −y1 + y2 and y2 = −y2 . The solution to the second equation is y2 (t) = c1 e−t . Substituting this solution into y1 = −y1 + y2 gives y1 + y1 = c1 e−t . This is a first order linear equation with integrating factor I(t) = et . When we multiply the differential equation for y1 (t) by I(t), it becomes (y1 · et ) = c1 . Integrating both sides yields y1 · et = c1 t + c2 . Thus, y1 (t) = c1 te−t + c2 e−t . Thus, we have y(t) = y1 (t) y2 (t) = c1 te−t + c2 e−t c1 e−t . Finally, we solve for x(t): x(t) = Sy(t) = −2 2 1 0 c1 te−t + c2 e−t c1 e−t −2(c1 te−t + c2 e−t ) + c1 e−t = 2(c1 te−t + c2 e−t ) −2t + 1 −2 −t −t + c2 e . = c1 e 2t 2 This is an acceptable answer, or we can write the individual equations comprising the general solution: x1 (t) = −2(c1 te−t + c2 e−t ) + c1 e−t and x2 (t) = 2(c1 te−t + c2 e−t ). 38. The eigenvalues of A are λ = −1, −1, 1. The eigenspace corresponding to λ = −1 is ⎡ ⎤ 1 1 0 nullspace(A + I) = nullspace ⎣ 0 1 1 ⎦ , 1 1 0 which is only one-dimensional, spanned by the vector (1, −1, 1). Therefore, we seek a generalized eigenvector v of A corresponding to λ = −1 such that {(A + I)v, v} is a cycle of generalized eigenvectors. Note that ⎡ ⎡ ⎤ ⎤ 1 1 0 1 2 1 A+I =⎣ 0 1 1 ⎦ and (A + I)2 = ⎣ 1 2 1 ⎦ . 1 1 0 1 2 1 (c)2017 Pearson Education. Inc. 515 In order that v be a generalized eigenvector of A corresponding to λ = −1, we should choose that ⎡ v such ⎤ 1 (A + I)2 v = 0 and (A + I)v = 0. There are many valid choices; let us choose v = ⎣ 0 ⎦. Then −1 ⎡ ⎤ 1 (A + I)v = ⎣ −1 ⎦. Hence, we obtain the cycle of generalized eigenvectors corresponding to λ = −1: 1 ⎧⎡ ⎤ ⎡ ⎤⎫ 1 1 ⎬ ⎨ ⎣ −1 ⎦ , ⎣ 0 ⎦ . ⎩ ⎭ 1 −1 Next, consider the eigenspace corresponding to λ = 1. For this, we compute ⎤ −1 1 0 1 ⎦. nullspace(A − I) = nullspace ⎣ 0 −1 1 1 −2 ⎡ ⎡ ⎤ ⎤ 1 1 −1 0 1 −1 ⎦. We find the eigenvector ⎣ 1 ⎦ as a basis for this eigenspace. This can be row-reduced to ⎣ 0 1 0 0 0 Hence, we are ready to form the matrices S and J: ⎡ ⎡ ⎤ 1 1 1 0 1 ⎦ S = ⎣ −1 1 −1 1 ⎡ and ⎤ −1 1 0 J = ⎣ 0 −1 0 ⎦ . 0 0 1 Via the substitution x = Sy, the system x = Ax is transformed into y = Jy. The corresponding equations are y1 = −y1 + y2 , y2 = −y2 , y3 = y3 . The third equation has solution y3 (t) = c3 et , the second equation has solution y2 (t) = c2 e−t , and so the first equation becomes y1 + y1 = c2 e−t . This is a first-order linear equation with integrating factor I(t) = et . When we multiply the differential equation for y1 (t) by I(t), it becomes (y1 · et ) = c2 . Integrating both sides yields y1 · et = c2 t + c1 . Thus, y1 (t = c2 te−t + c1 e−t . Thus, we have ⎡ ⎤ ⎤ ⎡ y1 (t) c2 te−t + c1 e−t ⎦. c2 e−t y(t) = ⎣ y2 (t) ⎦ = ⎣ t y3 (t) c3 e (c)2017 Pearson Education. Inc. 516 Finally, we solve for x(t): ⎤ ⎤⎡ 1 1 1 c2 te−t + c1 e−t ⎦ c2 e−t 0 1 ⎦⎣ x(t) = Sy(t) = ⎣ −1 c3 et 1 −1 1 ⎤ ⎡ c2 te−t + c1 e−t + c2 e−t + c3 et ⎦ −c2 te−t − c1 e−t + c3 et =⎣ c2 te−t + c1 e−t − c2 e−t + c3 et ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 1 t+1 1 = c1 e−t ⎣ −1 ⎦ + c2 e−t ⎣ −t ⎦ + c3 et ⎣ 1 ⎦ . 1 t−1 1 ⎡ 39. The eigenvalues of A are λ = −2, −2, −2. The eigenspace corresponding to λ = −2 is ⎡ ⎤ 0 0 0 nullspace(A + 2I) = nullspace ⎣ 1 −1 −1 ⎦ , −1 1 1 and there are two linearly independent vectors in this nullspace, corresponding to the unpivoted columns of the row-echelon form of this matrix. Therefore, the Jordan canonical form of A is ⎡ ⎤ −2 1 0 0 ⎦. J = ⎣ 0 −2 0 0 −2 To form an invertible matrix S such that S −1 AS = J, we must find a cycle of generalized eigenvectors corresponding to λ = −2 of length 2: {(A + 2I)v, v}. Now ⎡ ⎤ 0 0 0 A + 2I = ⎣ 1 −1 −1 ⎦ and (A + 2I)2 = 03 . −1 1 1 Since (A + 2I)2 = 03 , every nonzero vector in R3 is a generalized eigenvector corresponding to λ = −2. We need ⎡only ⎤find a nonzero vector v⎡ such ⎤that (A + 2I)v = 0. There are many valid choices; let us choose 1 0 v = ⎣ 0 ⎦. Then (A + 2I)v = ⎣ 1 ⎦, an eigenvector of A corresponding to λ = −2. We also need a 0 −1 second ⎡ ⎤ linearly independent eigenvector corresponding to λ = −2. There are many choices; let us choose 1 ⎣ 1 ⎦. Therefore, we can form the matrix 0 ⎡ 0 S=⎣ 1 −1 1 0 0 ⎤ 1 1 ⎦. 0 Via the substitution x = Sy, the system x = Ax is transformed into y = Jy. The corresponding equations are y1 = −2y1 + y2 , y2 = −2y2 , y3 = −2y3 . (c)2017 Pearson Education. Inc. 517 The third equation has solution y3 (t) = c3 e−2t , the second equation has solution y2 (t) = c2 e−2t , and so the first equation becomes y1 + 2y1 = c2 e−2t . This is a first-order linear equation with integrating factor I(t) = e2t . When we multiply the differential equation for y1 (t) by I(t), it becomes (y1 · e2t ) = c2 . Integrating both sides yields y1 · e2t = c2 t + c1 . Thus, y1 (t) = c2 te−2t + c1 e−2t . Thus, we have ⎤ ⎤ ⎡ y1 (t) c2 te−2t + c1 e−2t ⎦. c2 e−2t y(t) = ⎣ y2 (t) ⎦ = ⎣ −2t y3 (t) c3 e ⎡ Finally, we solve for x(t): ⎤ ⎤⎡ 0 1 1 c2 te−2t + c1 e−2t ⎦ c2 e−2t x(t) = Sy(t) = ⎣ 1 0 1 ⎦ ⎣ −1 0 0 c3 e−2t ⎤ ⎡ c2 e−2t + c3 e−2t = ⎣ c2 te−2t + c1 e−2t + c3 e−2t ⎦ −(c2 te−2t + c1 e−2t ) ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 0 1 1 = c1 e−2t ⎣ 1 ⎦ + c2 e−2t ⎣ t ⎦ + c3 e−2t ⎣ 1 ⎦ . −1 −t 0 ⎡ 40. The eigenvalues of A are λ = 4, 4, 4. The eigenspace corresponding to λ = 4 is ⎡ ⎤ 0 0 0 nullspace(A − 4I) = nullspace ⎣ 1 0 0 ⎦ , 0 1 0 and there is only one eigenvector. Therefore, the Jordan canonical form of A is ⎡ ⎤ 4 1 0 J = ⎣ 0 4 1 ⎦. 0 0 4 Next, we need to find an invertible matrix S such that S −1 AS = J. To do this, we must find a cycle of generalized eigenvectors {(A − 4I)2 v, (A − 4I)v, v} of length 3 corresponding to λ = 4. We have ⎡ ⎤ ⎡ ⎤ 0 0 0 0 0 0 A − 4I = ⎣ 1 0 0 ⎦ and (A − 4I)2 = ⎣ 0 0 0 ⎦ and (A − 4I)3 = 03 . 0 1 0 1 0 0 From (A − 4I)3 = 03 , we ⎡ know ⎤ that every nonzero vector is a generalized eigenvector corresponding to 1 λ = 4. We choose v = ⎣ 0 ⎦ (any multiple of the chosen vector v would be acceptable as well). Thus, 0 (c)2017 Pearson Education. Inc. 518 ⎡ ⎤ ⎤ 0 0 (A − 4I)v = ⎣ 1 ⎦ and (A − 4I)2 v = ⎣ 0 ⎦. Thus, we have the cycle of generalized eigenvectors 1 0 ⎡ ⎧⎡ ⎤ ⎡ ⎤ ⎡ ⎤⎫ 0 1 ⎬ ⎨ 0 ⎣ 0 ⎦,⎣ 1 ⎦,⎣ 0 ⎦ . ⎩ ⎭ 1 0 0 Thus, we can form the matrix ⎡ 0 S=⎣ 0 1 0 1 0 ⎤ 1 0 ⎦. 0 Via the substitution x = Sy, the system x = Ax is transformed into y = Jy. The corresponding equations are y1 = 4y1 + y2 , y2 = 4y2 + y3 , y3 = 4y3 . The third equation has solution y3 (t) = c3 e4t , and the second equation becomes y2 − 4y2 = c3 e4t . This is a first-order linear equation with integrating factor I(t) = e−4t . When we multiply the differential equation for y2 (t) by I(t), it becomes (y2 · e−4t ) = c3 . Integrating both sides yields y2 · e−4t = c3 t + c2 . Thus, y2 (t) = c3 te4t + c2 e4t = e4t (c3 t + c2 ). Therefore, the differential equation for y1 (t) becomes y1 − 4y1 = e4t (c3 t + c2 ). This equation is first-order linear with integrating factor I(t) = e−4t . When we multiply the differential equation for y1 (t) by I(t), it becomes (y1 · e−4t ) = c3 t + c2 . Integrating both sides, we obtain y1 · e−4t = c3 Hence, Thus, we have t2 + c2 t + c1 . 2 2 t y1 (t) = e4t c3 + c2 t + c1 . 2 ⎤ ⎤ ⎡ 4t t2 y1 (t) c3 2 + c2 t + c1 e ⎥ ⎢ y(t) = ⎣ y2 (t) ⎦ = ⎣ ⎦. e4t (c3 t + c2 ) 4t y3 (t) c3 e ⎡ (c)2017 Pearson Education. Inc. 519 Finally, we solve for x(t): ⎡ 0 x(t) = Sy(t) = ⎣ 0 1 ⎡ ⎢ =⎣ ⎤ ⎤ ⎡ 4t t2 1 c3 2 + c2 t + c1 e ⎥ ⎢ 0 ⎦⎣ ⎦ e4t (c3 t + c2 ) 4t 0 c3 e ⎤ 4t c3 e 4t ⎥ e (c3 t + c2 ) ⎦ 0 1 0 2 e4t c3 t2 + c2 t + c1 ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 0 0 1 = c1 e4t ⎣ 0 ⎦ + c2 e4t ⎣ 1 ⎦ + c3 e4t ⎣ t ⎦ . 1 t t2 /2 41. The eigenvalues of A are λ = −1, −1, 2. The eigenspace corresponding to λ = −1 is ⎡ ⎤ 4 0 4 0 ⎦. nullspace(A + I) = nullspace ⎣ 0 3 −4 0 −4 We see that the eigenvectors (x, y, z) corresponding to λ = −1 must satisfy x + z = 0 and y = 0. Therefore, we can choose z = t as a free variable, from which x = −t. Hence, we only obtain one linearly independent eigenvector (−1, 0, 1) corresponding to λ = −1, from which we conclude that ⎡ ⎤ −1 1 0 JCF(A) = J = ⎣ 0 −1 0 ⎦ . 0 0 2 To form an invertible matrix S such that S −1 AS = J, we must find a cycle of generalized eigenvectors corresponding to λ = −1 of length 2: {(A + I)v, v}. Now ⎡ ⎤ ⎡ ⎤ 4 0 4 0 0 0 0 ⎦ A+I =⎣ 0 3 and (A + I)2 = ⎣ 0 9 0 ⎦ . −4 0 −4 0 0 0 We need to find v such that (A ⎡ + I)v ⎤ = 0 but (A + I)2 v = 0. There are many valid choices; let ⎡ a vector ⎤ 1 4 us choose v = ⎣ 0 ⎦. Then (A + I)v = ⎣ 0 ⎦, an eigenvector of A corresponding to λ = −1. Thus, we −4 ⎧0⎡ ⎤ ⎡ ⎤⎫ 4 1 ⎨ ⎬ obtain the cycle ⎣ 0 ⎦ , ⎣ 0 ⎦ . ⎩ ⎭ −4 0 Next, we seek an eigenvector corresponding to λ = 2. The eigenspace corresponding to λ = 2 is ⎡ ⎤ 1 0 4 0 ⎦. nullspace(A − 2I) = nullspace ⎣ 0 0 −4 0 −7 We see that the eigenvectors (x, y, z) corresponding to λ = 2 must satisfy x + 4z = 0 and −4x − 7z = 0. We quickly find that x = z = 0. However, y = t is a free variable. Therefore, we obtain the eigenvector (0, 1, 0) corresponding to λ = 2. (c)2017 Pearson Education. Inc. 520 Therefore, we can form the matrix ⎡ 4 S=⎣ 0 −4 1 0 0 ⎤ 0 1 ⎦. 0 Via the substitution x = Sy, the system x = Ax is transformed into y = Jy. The corresponding equations are y1 = −y1 + y2 , y2 = −y2 , y3 = 2y3 . The third equation has solution y3 (t) = c3 e2t , the second equation has solution y2 (t) = c2 e−t , and so the first equation becomes y1 + y1 = c2 e−t . This is a first-order linear equation with integrating factor I(t) = et . We we multiply the differential equation for y1 (t) through by I(t), it becomes (y · et ) = c2 . Integrating both sides yields y1 · et = c2 t + c1 . Thus, y1 (t) = c2 te−t + c1 e−t . Thus, we have ⎤ ⎤ ⎡ y1 (t) c2 te−t + c1 e−t ⎦. c2 e−t y(t) = ⎣ y2 (t) ⎦ = ⎣ 2t y3 (t) c3 e ⎡ Finally, we solve for x(t): ⎤ ⎤⎡ 4 1 0 c2 te−t + c1 e−t ⎦ c2 e−t x(t) = Sy(t) = ⎣ 0 0 1 ⎦ ⎣ c3 e2t −4 0 0 ⎤ ⎡ 4(c2 te−t + c1 e−t ) + c2 e−t ⎦ c3 e2t =⎣ −4(c2 te−t + c1 e−t ) ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 4 4t + 1 0 ⎦ + c3 e2t ⎣ 1 ⎦ . 0 = c1 e−t ⎣ 0 ⎦ + c2 e−t ⎣ −4 −4t 0 ⎡ 42. The eigenvalues of A are λ = −3, −3. The eigenspace corresponding to λ = −3 is only one-dimensional, and therefore the Jordan canonical form of A contains one 2 × 2 Jordan block: −3 1 J= . 0 −3 Next, we look for a cycle of generalized eigenvectors of the form {(A + 3I)v, v}, where v is a generalized 2 1 −1 eigenvector of A. Since (A+3I)2 = = 02 , every nonzero vector in R2 is a generalized eigenvector. 1 −1 1 1 Let us choose v = . Then (A + 3I)v = . Thus, we form the matrix 0 1 S= 1 1 1 0 . (c)2017 Pearson Education. Inc. 521 Via the substitution x = Sy, the system x = Ax is transformed into y = Jy. The corresponding equations are y1 = −3y1 + y2 and y2 = −3y2 . The second equation has solution y2 (t) = c2 e−3t . Substituting this expression for y2 (t) into the differential equation for y1 (t) yields y1 + 3y1 = c2 e−3t . An integrating factor for this first-order linear differential equation is I(t) = e3t . Multiplying the differential equation for y1 (t) by I(t) gives us (y1 · e3t ) = c2 . Integrating both sides, we obtain y1 · e3t = c2 t + c1 . Thus, y1 (t) = c2 te−3t + c1 e−3t . Thus, y(t) = y1 (t) y2 (t) = c2 te−3t + c1 e−3t c2 e−3t . Finally, we solve for x(t): x(t) = Sy(t) = 1 1 1 0 c2 te−3t + c1 e−3t c2 e−3t c2 te−3t + c1 e−3t + c2 e−3t = c2 te−3t + c1 e−3t 1 t+1 −3t −3t + c2 e . = c1 e 1 t Now, we must apply the initial condition: 1 1 0 + c2 . = x(0) = c1 1 0 −1 Therefore, c1 = −1 and c2 = 1. Hence, the unique solution to the given initial-value problem is 1 t+1 −3t −3t +e . x(t) = −e 1 t 43. Let J = JCF(A) = JCF(B). Thus, there exist invertible matrices S and T such that S −1 AS = J Thus, and so and T −1 BT = J. S −1 AS = T −1 BT, B = T S −1 AST −1 = (ST −1 )−1 A(ST −1 ), which implies by definition that A and B are similar matrices. 44. Since the characteristic polynomial has degree 3, we know that A is a 3 × 3 matrix. Moreover, the roots of the characteristic equation has roots λ = 0, 0, 0. Hence, the Jordan canonical form J of A must be one of the three below: ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ 0 0 0 0 1 0 0 1 0 ⎣ 0 0 0 ⎦,⎣ 0 0 0 ⎦,⎣ 0 0 1 ⎦. 0 0 0 0 0 0 0 0 0 (c)2017 Pearson Education. Inc. 522 In all three cases, note that J 3 = 03 . Moreover, there exists an invertible matrix S such that S −1 AS = J. Thus, A = SJS −1 , and so A3 = (SJS −1 )3 = SJ 3 S −1 = S03 S −1 = 03 , which implies that A is nilpotent. 45. (a). Let J be an n × n Jordan block with eigenvalue λ. Then the eigenvalues of J T are λ (with multiplicity n). The matrix J T − λI consists of 1’s on the subdiagonal (the diagonal parallel and directly beneath the main diagonal) and zeros elsewhere. Hence, the null space of J T − λI is one-dimensional (with a free variable corresponding to the right-most column of J T −λI). Therefore, the Jordan canonical form of J T consists of a single Jordan block, since there is only one linearly independent eigenvector corresponding to the eigenvalue λ. However, a single Jordan block with eigenvalue λ is precisely the matrix J. Therefore, JCF(J T ) = J. (b). Let JCF(A) = J. Then there exists an invertible matrix S such that S −1 AS = J. Transposing both sides, we obtain (S −1 AS)T = J T , or S T AT (S −1 )T = J T , or S T AT (S T )−1 = J T . Hence, the matrix AT is similar to J T . However, by applying part (a) to each block in J T , we find that JCF(J T ) = J. Hence, J T is similar to J. By Problem 29 in Section 7.3, we conclude that AT is similar to J. Hence, AT and J have the same Jordan canonical form. However, since JCF(J) = J, we deduce that JCF(AT ) = J = JCF(A), as required. Solutions to Section 7.7 Additional Problems: 1. Since A is lower triangular, its eigenvalues lie along the main diagonal, λ1 = 3 and λ2 = −1. Since A is 2 × 2 with two distinct eigenvalues, A is diagonalizable. To get an invertible matrix S such that S −1 AS = D, we need to find an eigenvector associated with each eigenvector: Eigenvalue λ1 = 3: To get an eigenvector, we consider nullspace(A − 3I) = nullspace and we see that one possible eigenvector is 1 4 0 0 16 −4 , . Eigenvalue λ2 = −1: To get an eigenvector, we consider nullspace(A + I) = nullspace and we see that one possible eigenvector is 0 1 Putting the above results together, we form 1 0 S= 4 1 4 16 0 0 3 0 0 −1 , . and D= (c)2017 Pearson Education. Inc. . 523 2. To compute the eigenvalues, we find the characteristic equation 13 − λ −9 det(A − λI) = det = (13 − λ)(−17 − λ) + 225 = λ2 + 4λ + 4, 25 −17 − λ and the roots of this equation are λ = −2, −2. Eigenvalue λ = 2: We compute nullspace(A − 2I) = nullspace 15 −9 25 −15 , but since there is only one linearly independent solution to the corresponding system (one free variable), the eigenvalue λ = 2 does not have two linearly independent solutions. Hence, A is not diagonalizable. 3. To compute the eigenvalues, we find the characteristic equation ⎡ ⎤ −4 − λ 3 0 ⎦ = (−1 − λ) [(−4 − λ)(5 − λ) + 18] 5−λ 0 det(A − λI) = det ⎣ −6 3 −3 −1 − λ = (−1 − λ)(λ2 − λ − 2) = −(λ + 1)2 (λ − 2), so the eigenvalues are λ1 = −1 and λ2 = 2. Eigenvalue λ1 = −1: To get eigenvectors, we consider ⎡ −3 nullspace(A + I) = nullspace ⎣ −6 3 ⎤ ⎡ ⎤ 3 0 1 −1 0 6 0 ⎦∼⎣ 0 0 0 ⎦. −3 0 0 0 0 There are two free variables, z = t and y = s. From the first equation x = s. Thus, two linearly independent eigenvectors can be obtained corresponding to λ1 = −1: ⎡ ⎤ ⎡ ⎤ 1 0 ⎣ 1 ⎦ ⎣ 0 ⎦. and 0 1 Eigenvalue λ2 = 2: To get an eigenvector, we consider ⎡ −6 nullspace(A − 2I) = nullspace ⎣ −6 3 ⎤ ⎡ ⎤ 3 0 1 −1 −1 3 0 ⎦∼⎣ 0 1 2 ⎦. −3 −3 0 0 0 We let z = t. Then y = −2t from the middle line, and x = −t from the top line. Thus, an eigenvector corresponding to λ2 = 2 may be chosen as ⎡ ⎤ −1 ⎣ −2 ⎦ . 1 Putting the above results together, we form ⎡ ⎤ −1 1 0 S = ⎣ −2 1 0 ⎦ 1 0 1 ⎡ and ⎤ 2 0 0 0 ⎦. D = ⎣ 0 −1 0 0 −1 (c)2017 Pearson Education. Inc. 524 4. To compute the eigenvalues, we find the characteristic equation ⎡ ⎤ 1−λ 1 0 ⎦ = (−2 − λ) [(1 − λ)(5 − λ) + 4] 5−λ 0 det(A − λI) = det ⎣ −4 17 −11 −2 − λ = (−2 − λ)(λ2 − 6λ + 9) = (−2 − λ)(λ − 3)2 , so the eigenvalues are λ1 = 3 and λ2 = −2. Eigenvalue λ1 = 3: To get eigenvectors, we consider ⎡ ⎤ ⎡ ⎤ ⎡ ⎤ ⎡ −2 1 0 −2 1 0 1 −3 −5 1 2 0 ⎦ ∼ ⎣ 17 −11 −5 ⎦ ∼ ⎣ −2 1 0 ⎦∼⎣ 0 nullspace(A−3I) = nullspace ⎣ −4 17 −11 −5 0 0 0 0 0 0 0 ⎤ −3 −5 1 2 ⎦. 0 0 The latter matrix contains only one unpivoted column, so that only one linearly independent eigenvector can be obtained. However, λ1 = 3 occurs with multiplicity 2 as a root of the characteristic equation for the matrix. Therefore, the matrix is not diagonalizable. 5. We are given that the only eigenvalue of A is λ = 2. Eigenvalue λ = 2: To get eigenvectors, we consider ⎡ ⎤ ⎡ ⎤ ⎡ −3 −1 3 1 0 −1 1 2 −4 ⎦ ∼ ⎣ −3 −1 3 ⎦∼⎣ 0 nullspace(A − 2I) = nullspace ⎣ 4 −1 0 1 4 2 −4 0 ⎤ 0 −1 1 0 ⎦. 2 0 We see that only one unpivoted column will occur in a row-echelon form of A − 2I, and thus, only one linearly independent eigenvector can be obtained. Since the eigenvalue λ = 2 occurs with multiplicity 3 as a root of the characteristic equation for the matrix, the matrix is not diagonalizable. 6. We are given that the eigenvalues of A are λ1 = 4 and λ2 = −1. Eigenvalue λ1 = 4: We consider ⎡ 5 nullspace(A − 4I) = nullspace ⎣ 0 10 ⎤ 5 −5 −5 0 ⎦. 5 −10 The middle row tells us that nullspace vectors (x, y, z) must have y = 0. From this information, the first and last rows of the matrix tell us the same thing: x = z. Thus, an eigenvector corresponding to λ1 = 4 may be chosen as ⎡ ⎤ 1 ⎣ 0 ⎦. 1 Eigenvalue λ2 = −1: We consider ⎤ ⎡ 10 5 −5 10 0 ⎦∼⎣ 0 nullspace(A + I) = nullspace ⎣ 0 0 10 5 −5 0 ⎡ (c)2017 Pearson Education. Inc. ⎤ 5 −5 0 0 ⎦. 0 0 525 From the first row, a vector (x, y, z) in the nullspace must satisfy 10x + 5y − 5z = 0. Setting z = t and y = s, we get x = 12 t − 12 s. Hence, the eigenvectors corresponding to λ2 = −1 take the form ( 12 t − 12 s, s, t), and so a basis for this eigenspace is ⎧⎡ ⎤ ⎡ ⎤⎫ −1/2 ⎬ ⎨ 1/2 ⎣ 0 ⎦,⎣ 1 ⎦ . ⎩ ⎭ 1 0 Putting the above results together, we form ⎡ ⎤ 1 1/2 −1/2 1 ⎦ S=⎣ 0 0 1 1 0 ⎡ and ⎤ 4 0 0 0 ⎦. D = ⎣ 0 −1 0 0 −1 11. We will compute the dimension of each of the two eigenspaces associated with the matrix A. For λ1 = 1, we compute as follows: ⎡ ⎤ ⎡ ⎤ 4 8 16 1 2 4 0 8 ⎦ ∼ ⎣ 0 1 1 ⎦, nullspace(A − I) = nullspace ⎣ 4 −4 −4 −12 0 0 0 which has only one unpivoted column. Thus, this eigenspace is 1-dimensional. For λ2 = −3, we compute as follows: ⎡ ⎤ ⎡ 8 8 16 1 1 4 8 ⎦∼⎣ 0 0 nullspace(A + 3I) = nullspace ⎣ 4 −4 −4 −8 0 0 ⎤ 2 0 ⎦, 0 which has two unpivoted columns. Thus, this eigenspace is 2-dimensional. Between the two eigenspaces, we have a complete set of linearly independent eigenvectors. Hence, the matrix A in this case is diagonalizable. Therefore, A is diagonalizable, and we may take ⎡ ⎤ 1 0 0 0 ⎦. J = ⎣ 0 −3 0 0 −3 (Of course, the eigenvalues of A may be listed in any order along the main diagonal of the Jordan canonical form, thus yielding other valid Jordan canonical forms for A.) 12. We will compute the dimension of each of the two eigenspaces associated with the matrix A. For λ1 = −1, we compute as follows: ⎡ ⎤ ⎡ ⎤ 3 1 1 1 0 1 nullspace(A + I) = nullspace ⎣ 2 2 −2 ⎦ ∼ ⎣ 0 1 −2 ⎦ , −1 0 −1 0 0 0 which has one unpivoted column. Thus, this eigenspace is 1-dimensional. For λ2 = 3, we compute as follows: ⎡ ⎤ ⎡ ⎤ −1 1 1 1 −1 −1 1 6 ⎦, nullspace(A − 3I) = nullspace ⎣ 2 −2 −2 ⎦ ∼ ⎣ 0 −1 0 −5 0 0 0 which has one unpivoted column. Thus, this eigenspace is also one-dimensional. (c)2017 Pearson Education. Inc. 526 Since we have only generated two linearly independent eigenvectors from the eigenvalues of A, we know that A is not diagonalizable, and hence, the Jordan canonical form of A is not a diagonal matrix. We must have one 1 × 1 Jordan block and one 2 × 2 Jordan block. To determine which eigenvalue corresponds to the 1 × 1 block and which corresponds to the 2 × 2 block, we must determine the multiplicity of the eigenvalues as roots of the characteristic equation of A. A short calculation shows that λ1 = −1 occurs with multiplicity 2, while λ2 = 3 occurs with multiplicity 1. Thus, the Jordan canonical form of A is ⎡ ⎤ −1 1 0 J = ⎣ 0 −1 0 ⎦ . 0 0 3 13. There are 3 different possible Jordan canonical forms, up to a rearrangement of the Jordan blocks: Case 1: ⎤ −1 0 0 0 ⎢ 0 −1 0 0 ⎥ ⎥. J =⎢ ⎣ 0 0 −1 0 ⎦ 0 0 0 2 ⎡ In this case, the matrix has four linearly independent eigenvectors, and because all Jordan blocks have size 1 × 1, the maximum length of a cycle of generalized eigenvectors for this matrix is 1. Case 2: ⎤ −1 1 0 0 ⎢ 0 −1 0 0 ⎥ ⎥. J =⎢ ⎣ 0 0 −1 0 ⎦ 0 0 0 2 ⎡ In this case, the matrix has three linearly independent eigenvectors (two corresponding to λ = −1 and one corresponding to λ = 2). There is a Jordan block of size 2 × 2, and so a cycle of generalized eigenvectors can have a maximum length of 2 in this case. Case 3: ⎤ −1 1 0 0 ⎢ 0 −1 1 0 ⎥ ⎥. J =⎢ ⎣ 0 0 −1 0 ⎦ 0 0 0 2 ⎡ In this case, the matrix has two linearly independent eigenvectors (one corresponding to λ = −1 and one corresponding to λ = 2). There is a Jordan block of size 3 × 3, and so a cycle of generalized eigenvectors can have a maximum length of 3 in this case. 14. There are 7 different possible Jordan canonical forms, up to a rearrangement of the Jordan blocks: Case 1: ⎡ 4 ⎢ 0 ⎢ J =⎢ ⎢ 0 ⎣ 0 0 0 4 0 0 0 0 0 4 0 0 0 0 0 4 0 ⎤ 0 0 ⎥ ⎥ 0 ⎥ ⎥. 0 ⎦ 4 In this case, the matrix has five linearly independent eigenvectors, and because all Jordan blocks have size 1 × 1, the maximum length of a cycle of generalized eigenvectors for this matrix is 1. (c)2017 Pearson Education. Inc. 527 Case 2: ⎡ 4 ⎢ 0 ⎢ J =⎢ ⎢ 0 ⎣ 0 0 1 4 0 0 0 0 0 4 0 0 0 0 0 4 0 ⎤ 0 0 ⎥ ⎥ 0 ⎥ ⎥. 0 ⎦ 4 In this case, the matrix has four linearly independent eigenvectors, and because the largest Jordan block is of size 2 × 2, the maximum length of a cycle of generalized eigenvectors for this matrix is 2. Case 3: ⎡ 4 ⎢ 0 ⎢ J =⎢ ⎢ 0 ⎣ 0 0 1 4 0 0 0 0 0 4 0 0 0 0 1 4 0 ⎤ 0 0 ⎥ ⎥ 0 ⎥ ⎥. 0 ⎦ 4 In this case, the matrix has three linearly independent eigenvectors, and because the largest Jordan block is of size 2 × 2, the maximum length of a cycle of generalized eigenvectors for this matrix is 2. Case 4: ⎡ 4 ⎢ 0 ⎢ J =⎢ ⎢ 0 ⎣ 0 0 1 4 0 0 0 0 1 4 0 0 0 0 0 4 0 ⎤ 0 0 ⎥ ⎥ 0 ⎥ ⎥. 0 ⎦ 4 In this case, the matrix has three linearly independent eigenvectors, and because the largest Jordan block is of size 3 × 3, the maximum length of a cycle of generalized eigenvectors for this matrix is 3. Case 5: ⎡ 4 ⎢ 0 ⎢ J =⎢ ⎢ 0 ⎣ 0 0 1 4 0 0 0 0 1 4 0 0 0 0 0 4 0 ⎤ 0 0 ⎥ ⎥ 0 ⎥ ⎥. 1 ⎦ 4 In this case, the matrix has two linearly independent eigenvectors, and because the largest Jordan block is of size 3 × 3, the maximum length of a cycle of generalized eigenvectors for this matrix is 3. Case 6: ⎡ 4 ⎢ 0 ⎢ J =⎢ ⎢ 0 ⎣ 0 0 1 4 0 0 0 0 1 4 0 0 0 0 1 4 0 ⎤ 0 0 ⎥ ⎥ 0 ⎥ ⎥. 0 ⎦ 4 In this case, the matrix has two linearly independent eigenvectors, and because the largest Jordan block is of size 4 × 4, the maximum length of a cycle of generalized eigenvectors for this matrix is 4. Case 7: ⎡ 4 ⎢ 0 ⎢ J =⎢ ⎢ 0 ⎣ 0 0 1 4 0 0 0 0 1 4 0 0 0 0 1 4 0 ⎤ 0 0 ⎥ ⎥ 0 ⎥ ⎥. 1 ⎦ 4 (c)2017 Pearson Education. Inc. 528 In this case, the matrix has only one linearly independent eigenvector, and because the largest Jordan block is of size 5 × 5, the maximum length of a cycle of generalized eigenvectors for this matrix is 5. 15. There are 10 different possible Jordan canonical forms, up to a rearrangement of the Jordan blocks: Case 1: ⎡ 6 ⎢ 0 ⎢ ⎢ 0 J =⎢ ⎢ 0 ⎢ ⎣ 0 0 0 6 0 0 0 0 0 0 6 0 0 0 ⎤ 0 0 0 0 0 0 ⎥ ⎥ 0 0 0 ⎥ ⎥. 6 0 0 ⎥ ⎥ 0 −3 0 ⎦ 0 0 −3 In this case, the matrix has six linearly independent eigenvectors, and because all Jordan blocks have size 1 × 1, the maximum length of a cycle of generalized eigenvectors for this matrix is 1. Case 2: ⎡ 6 ⎢ 0 ⎢ ⎢ 0 J =⎢ ⎢ 0 ⎢ ⎣ 0 0 1 6 0 0 0 0 0 0 6 0 0 0 ⎤ 0 0 0 0 0 0 ⎥ ⎥ 0 0 0 ⎥ ⎥. 6 0 0 ⎥ ⎥ 0 −3 0 ⎦ 0 0 −3 In this case, the matrix has five linearly independent eigenvectors (three corresponding to λ = 6 and two corresponding to λ = −3), and because the largest Jordan block is of size 2 × 2, the maximum length of a cycle of generalized eigenvectors for this matrix is 2. Case 3: ⎡ 6 ⎢ 0 ⎢ ⎢ 0 J =⎢ ⎢ 0 ⎢ ⎣ 0 0 1 6 0 0 0 0 0 0 6 0 0 0 ⎤ 0 0 0 0 0 0 ⎥ ⎥ 1 0 0 ⎥ ⎥. 6 0 0 ⎥ ⎥ 0 −3 0 ⎦ 0 0 −3 In this case, the matrix has four linearly independent eigenvectors (two corresponding to λ = 6 and two corresponding to λ = −3), and because the largest Jordan block is of size 2 × 2, the maximum length of a cycle of generalized eigenvectors for this matrix is 2. Case 4: ⎡ 6 ⎢ 0 ⎢ ⎢ 0 J =⎢ ⎢ 0 ⎢ ⎣ 0 0 1 6 0 0 0 0 0 1 6 0 0 0 ⎤ 0 0 0 0 0 0 ⎥ ⎥ 0 0 0 ⎥ ⎥. 6 0 0 ⎥ ⎥ 0 −3 0 ⎦ 0 0 −3 In this case, the matrix has four linearly independent eigenvectors (two corresponding to λ = 6 and two corresponding to λ = −3), and because the largest Jordan block is of size 3 × 3, the maximum length of a cycle of generalized eigenvectors for this matrix is 3. (c)2017 Pearson Education. Inc. 529 Case 5: ⎡ 6 ⎢ 0 ⎢ ⎢ 0 J =⎢ ⎢ 0 ⎢ ⎣ 0 0 1 6 0 0 0 0 0 1 6 0 0 0 ⎤ 0 0 0 0 0 0 ⎥ ⎥ 1 0 0 ⎥ ⎥. 6 0 0 ⎥ ⎥ 0 −3 0 ⎦ 0 0 −3 In this case, the matrix has three linearly independent eigenvectors (one corresponding to λ = 6 and two corresponding to λ = −3), and because the largest Jordan block is of size 4 × 4, the maximum length of a cycle of generalized eigenvectors for this matrix is 4. Case 6: ⎡ 6 ⎢ 0 ⎢ ⎢ 0 J =⎢ ⎢ 0 ⎢ ⎣ 0 0 0 6 0 0 0 0 0 0 6 0 0 0 ⎤ 0 0 0 0 0 0 ⎥ ⎥ 0 0 0 ⎥ ⎥. 6 0 0 ⎥ ⎥ 0 −3 1 ⎦ 0 0 −3 In this case, the matrix has five linearly independent eigenvectors (four corresponding to λ = 6 and one corresponding to λ = −3), and because the largest Jordan block is of size 2 × 2, the maximum length of a cycle of generalized eigenvectors for this matrix is 2. Case 7: ⎡ 6 ⎢ 0 ⎢ ⎢ 0 J =⎢ ⎢ 0 ⎢ ⎣ 0 0 1 6 0 0 0 0 0 0 6 0 0 0 ⎤ 0 0 0 0 0 0 ⎥ ⎥ 0 0 0 ⎥ ⎥. 6 0 0 ⎥ ⎥ 0 −3 1 ⎦ 0 0 −3 In this case, the matrix has four linearly independent eigenvectors (three corresponding to λ = 6 and one corresponding to λ = −3), and because the largest Jordan block is of size 2 × 2, the maximum length of a cycle of generalized eigenvectors for this matrix is 2. Case 8: ⎡ 6 ⎢ 0 ⎢ ⎢ 0 J =⎢ ⎢ 0 ⎢ ⎣ 0 0 1 6 0 0 0 0 0 0 6 0 0 0 ⎤ 0 0 0 0 0 0 ⎥ ⎥ 1 0 0 ⎥ ⎥. 6 0 0 ⎥ ⎥ 0 −3 1 ⎦ 0 0 −3 In this case, the matrix has three linearly independent eigenvectors (two corresponding to λ = 6 and one corresponding to λ = −3), and because the largest Jordan block is of size 2 × 2, the maximum length of a cycle of generalized eigenvectors for this matrix is 2. (c)2017 Pearson Education. Inc. 530 Case 9: ⎡ 6 ⎢ 0 ⎢ ⎢ 0 J =⎢ ⎢ 0 ⎢ ⎣ 0 0 1 6 0 0 0 0 0 1 6 0 0 0 ⎤ 0 0 0 0 0 0 ⎥ ⎥ 0 0 0 ⎥ ⎥. 6 0 0 ⎥ ⎥ 0 −3 1 ⎦ 0 0 −3 In this case, the matrix has three linearly independent eigenvectors (two corresponding to λ = 6 and one corresponding to λ = −3), and because the largest Jordan block is of size 3 × 3, the maximum length of a cycle of generalized eigenvectors for this matrix is 3. Case 10: ⎡ 6 ⎢ 0 ⎢ ⎢ 0 J =⎢ ⎢ 0 ⎢ ⎣ 0 0 1 6 0 0 0 0 0 1 6 0 0 0 ⎤ 0 0 0 0 0 0 ⎥ ⎥ 1 0 0 ⎥ ⎥. 6 0 0 ⎥ ⎥ 0 −3 1 ⎦ 0 0 −3 In this case, the matrix has two linearly independent eigenvectors (one corresponding to λ = 6 and one corresponding to λ = −3), and because the largest Jordan block is of size 4 × 4, the maximum length of a cycle of generalized eigenvectors for this matrix is 4. 16. There are 15 different possible Jordan canonical forms, up to a rearrangement of Jordan blocks: Case 1: ⎡ 2 ⎢ 0 ⎢ ⎢ 0 ⎢ J =⎢ ⎢ 0 ⎢ 0 ⎢ ⎣ 0 0 0 2 0 0 0 0 0 0 0 2 0 0 0 0 ⎤ 0 0 0 0 0 0 0 0 ⎥ ⎥ 0 0 0 0 ⎥ ⎥ 2 0 0 0 ⎥ ⎥. 0 −4 0 0 ⎥ ⎥ 0 0 −4 0 ⎦ 0 0 0 −4 In this case, the matrix has seven linearly independent eigenvectors (four corresponding to λ = 2 and three corresponding to λ = −4), and because all Jordan blocks are size 1 × 1, the maximum length of a cycle of generalized eigenvectors for this matrix is 1. Case 2: ⎡ 2 ⎢ 0 ⎢ ⎢ 0 ⎢ J =⎢ ⎢ 0 ⎢ 0 ⎢ ⎣ 0 0 1 2 0 0 0 0 0 0 0 2 0 0 0 0 ⎤ 0 0 0 0 0 0 0 0 ⎥ ⎥ 0 0 0 0 ⎥ ⎥ 2 0 0 0 ⎥ ⎥. 0 −4 0 0 ⎥ ⎥ 0 0 −4 0 ⎦ 0 0 0 −4 In this case, the matrix has six linearly independent eigenvectors (three corresponding to λ = 2 and three corresponding to λ = −4), and because the largest Jordan block is of size 2 × 2, the maximum length of a cycle of generalized eigenvectors for this matrix is 2. (c)2017 Pearson Education. Inc. 531 Case 3: ⎡ 2 ⎢ 0 ⎢ ⎢ 0 ⎢ J =⎢ ⎢ 0 ⎢ 0 ⎢ ⎣ 0 0 1 2 0 0 0 0 0 0 0 2 0 0 0 0 ⎤ 0 0 0 0 0 0 0 0 ⎥ ⎥ 1 0 0 0 ⎥ ⎥ 2 0 0 0 ⎥ ⎥. 0 −4 0 0 ⎥ ⎥ 0 0 −4 0 ⎦ 0 0 0 −4 In this case, the matrix has five linearly independent eigenvectors (two corresponding to λ = 2 and three corresponding to λ = −4), and because the largest Jordan block is of size 2 × 2, the maximum length of a cycle of generalized eigenvectors for this matrix is 2. Case 4: ⎤ 2 1 0 0 0 0 0 ⎢ 0 2 1 0 0 0 0 ⎥ ⎥ ⎢ ⎢ 0 0 2 0 0 0 0 ⎥ ⎥ ⎢ 0 0 0 ⎥ J =⎢ ⎥. ⎢ 0 0 0 2 ⎥ ⎢ 0 0 0 0 −4 0 0 ⎥ ⎢ ⎣ 0 0 0 0 0 −4 0 ⎦ 0 0 0 0 0 0 −4 ⎡ In this case, the matrix has five linearly independent eigenvectors (two corresponding to λ = 2 and three corresponding to λ = −4), and because the largest Jordan block is of size 3 × 3, the maximum length of a cycle of generalized eigenvectors for this matrix is 3. Case 5: ⎤ 2 1 0 0 0 0 0 ⎢ 0 2 1 0 0 0 0 ⎥ ⎥ ⎢ ⎢ 0 0 2 1 0 0 0 ⎥ ⎢ ⎥ 0 0 0 ⎥ J =⎢ ⎢ 0 0 0 2 ⎥. ⎢ 0 0 0 0 −4 ⎥ 0 0 ⎢ ⎥ ⎣ 0 0 0 0 0 −4 0 ⎦ 0 0 0 0 0 0 −4 ⎡ In this case, the matrix has four linearly independent eigenvectors (one corresponding to λ = 2 and three corresponding to λ = −4), and because the largest Jordan block is of size 4 × 4, the maximum length of a cycle of generalized eigenvectors for this matrix is 4. Case 6: ⎡ 2 ⎢ 0 ⎢ ⎢ 0 ⎢ J =⎢ ⎢ 0 ⎢ 0 ⎢ ⎣ 0 0 0 2 0 0 0 0 0 0 0 2 0 0 0 0 ⎤ 0 0 0 0 0 0 0 0 ⎥ ⎥ 0 0 0 0 ⎥ ⎥ 2 0 0 0 ⎥ ⎥. 0 −4 1 0 ⎥ ⎥ 0 0 −4 0 ⎦ 0 0 0 −4 In this case, the matrix has six linearly independent eigenvectors (four corresponding to λ = 2 and two corresponding to λ = −4), and because the largest Jordan block is of size 2 × 2, the maximum length of a cycle of generalized eigenvectors for this matrix is 2. (c)2017 Pearson Education. Inc. 532 Case 7: ⎡ 2 ⎢ 0 ⎢ ⎢ 0 ⎢ J =⎢ ⎢ 0 ⎢ 0 ⎢ ⎣ 0 0 1 2 0 0 0 0 0 0 0 2 0 0 0 0 ⎤ 0 0 0 0 0 0 0 0 ⎥ ⎥ 0 0 0 0 ⎥ ⎥ 2 0 0 0 ⎥ ⎥. 0 −4 1 0 ⎥ ⎥ 0 0 −4 0 ⎦ 0 0 0 −4 In this case, the matrix has five linearly independent eigenvectors (three corresponding to λ = 2 and two corresponding to λ = −4), and because the largest Jordan block is of size 2 × 2, the maximum length of a cycle of generalized eigenvectors for this matrix is 2. Case 8: ⎡ 2 ⎢ 0 ⎢ ⎢ 0 ⎢ J =⎢ ⎢ 0 ⎢ 0 ⎢ ⎣ 0 0 1 2 0 0 0 0 0 0 0 2 0 0 0 0 ⎤ 0 0 0 0 0 0 0 0 ⎥ ⎥ 1 0 0 0 ⎥ ⎥ 2 0 0 0 ⎥ ⎥. 0 −4 1 0 ⎥ ⎥ 0 0 −4 0 ⎦ 0 0 0 −4 In this case, the matrix has four linearly independent eigenvectors (two corresponding to λ = 2 and two corresponding to λ = −4), and because the largest Jordan block is of size 2 × 2, the maximum length of a cycle of generalized eigenvectors for this matrix is 2. Case 9: ⎡ 2 ⎢ 0 ⎢ ⎢ 0 ⎢ J =⎢ ⎢ 0 ⎢ 0 ⎢ ⎣ 0 0 1 2 0 0 0 0 0 0 1 2 0 0 0 0 ⎤ 0 0 0 0 0 0 0 0 ⎥ ⎥ 0 0 0 0 ⎥ ⎥ 2 0 0 0 ⎥ ⎥. 0 −4 1 0 ⎥ ⎥ 0 0 −4 0 ⎦ 0 0 0 −4 In this case, the matrix has four linearly independent eigenvectors (two corresponding to λ = 2 and two corresponding to λ = −4), and because the largest Jordan block is of size 3 × 3, the maximum length of a cycle of generalized eigenvectors for this matrix is 3. Case 10: ⎡ 2 ⎢ 0 ⎢ ⎢ 0 ⎢ J =⎢ ⎢ 0 ⎢ 0 ⎢ ⎣ 0 0 1 2 0 0 0 0 0 0 1 2 0 0 0 0 ⎤ 0 0 0 0 0 0 0 0 ⎥ ⎥ 1 0 0 0 ⎥ ⎥ 2 0 0 0 ⎥ ⎥. 0 −4 1 0 ⎥ ⎥ 0 0 −4 0 ⎦ 0 0 0 −4 In this case, the matrix has three linearly independent eigenvectors (one corresponding to λ = 2 and two corresponding to λ = −4), and because the largest Jordan block is of size 4 × 4, the maximum length of a cycle of generalized eigenvectors for this matrix is 4. (c)2017 Pearson Education. Inc. 533 Case 11: ⎡ 2 ⎢ 0 ⎢ ⎢ 0 ⎢ J =⎢ ⎢ 0 ⎢ 0 ⎢ ⎣ 0 0 0 2 0 0 0 0 0 0 0 2 0 0 0 0 ⎤ 0 0 0 0 0 0 0 0 ⎥ ⎥ 0 0 0 0 ⎥ ⎥ 2 0 0 0 ⎥ ⎥. 0 −4 1 0 ⎥ ⎥ 0 0 −4 1 ⎦ 0 0 0 −4 In this case, the matrix has five linearly independent eigenvectors (four corresponding to λ = 2 and one corresponding to λ = −4), and because the largest Jordan block is of size 3 × 3, the maximum length of a cycle of generalized eigenvectors for this matrix is 3. Case 12: ⎤ 2 1 0 0 0 0 0 ⎢ 0 2 0 0 0 0 0 ⎥ ⎥ ⎢ ⎢ 0 0 2 0 0 0 0 ⎥ ⎥ ⎢ 0 0 0 ⎥ J =⎢ ⎥. ⎢ 0 0 0 2 ⎥ ⎢ 0 0 0 0 −4 1 0 ⎥ ⎢ ⎣ 0 0 0 0 0 −4 1 ⎦ 0 0 0 0 0 0 −4 ⎡ In this case, the matrix has four linearly independent eigenvectors (three corresponding to λ = 2 and one corresponding to λ = −4), and because the largest Jordan block is of size 3 × 3, the maximum length of a cycle of generalized eigenvectors for this matrix is 3. Case 13: ⎤ 2 1 0 0 0 0 0 ⎢ 0 2 0 0 0 0 0 ⎥ ⎥ ⎢ ⎢ 0 0 2 1 0 0 0 ⎥ ⎥ ⎢ 0 0 0 ⎥ J =⎢ ⎢ 0 0 0 2 ⎥. ⎢ 0 0 0 0 −4 1 0 ⎥ ⎢ ⎥ ⎣ 0 0 0 0 0 −4 1 ⎦ 0 0 0 0 0 0 −4 ⎡ In this case, the matrix has three linearly independent eigenvectors (two corresponding to λ = 2 and one corresponding to λ = −4), and because the largest Jordan block is of size 3 × 3, the maximum length of a cycle of generalized eigenvectors for this matrix is 3. Case 14: ⎤ 2 1 0 0 0 0 0 ⎢ 0 2 1 0 0 0 0 ⎥ ⎥ ⎢ ⎢ 0 0 2 0 0 0 0 ⎥ ⎥ ⎢ 0 0 0 ⎥ J =⎢ ⎥. ⎢ 0 0 0 2 ⎥ ⎢ 0 0 0 0 −4 1 0 ⎥ ⎢ ⎣ 0 0 0 0 0 −4 1 ⎦ 0 0 0 0 0 0 −4 ⎡ In this case, the matrix has three linearly independent eigenvectors (two corresponding to λ = 2 and one corresponding to λ = −4), and because the largest Jordan block is of size 3 × 3, the maximum length of a cycle of generalized eigenvectors for this matrix is 3. (c)2017 Pearson Education. Inc. 534 Case 15: ⎡ 2 ⎢ 0 ⎢ ⎢ 0 ⎢ J =⎢ ⎢ 0 ⎢ 0 ⎢ ⎣ 0 0 1 2 0 0 0 0 0 0 1 2 0 0 0 0 ⎤ 0 0 0 0 0 0 0 0 ⎥ ⎥ 1 0 0 0 ⎥ ⎥ 2 0 0 0 ⎥ ⎥. 0 −4 1 0 ⎥ ⎥ 0 0 −4 1 ⎦ 0 0 0 −4 In this case, the matrix has two linearly independent eigenvectors (one corresponding to λ = 2 and one corresponding to λ = −4), and because the largest Jordan block is of size 4 × 4, the maximum length of a cycle of generalized eigenvectors for this matrix is 4. 1 1 1 0 17. FALSE. For instance, if A = and B = , then we have eigenvalues λA = λB = 1, but 0 1 1 1 0 1 the matrix A − B = is invertible, and hence, zero is not an eigenvalue of A − B. This can also −1 0 be verified directly. 1 1 , then A2 = B 2 = I2 , but the matrices A and B are 18. FALSE. For instance, if A = I2 and B = 0 1 not similar. (Otherwise, there would exist an invertible matrix S such that S −1 AS = B. But since A = I2 this reduces to I2 = B, which is clearly not the case. Thus, no such invertible matrix S exists.) 19. We have Ai v = λi v for each i = 1, 2, . . . , k. Thus, (A1 A2 . . . Ak )v = (A1 A2 . . . Ak−1 )(Ak v) = (A1 A2 . . . Ak−1 )(λk v) = λk (A1 A2 . . . Ak−1 )v = λk (A1 A2 . . . Ak−2 )(Ak−1 v) = λk (A1 A2 . . . Ak−2 )(λk−1 v) = λk−1 λk (A1 A2 . . . Ak−2 )v .. . = λ2 λ3 . . . λk (A1 v) = λ2 λ3 . . . λk (λ1 v) = (λ1 λ2 . . . λk )v, which shows that v is an eigenvector of A1 A2 . . . Ak with corresponding eigenvalue λ1 λ2 . . . λk . 20. We are given that Av = λ1 v and Bv = λ2 v. Note that v = 0 (since v is an eigenvector), and (AB − BA)v = A(Bv) − B(Av) = A(λ2 v) − B(λ1 v) = λ2 (Av) − λ1 (Bv) = λ2 (λ1 v) − λ1 (λ2 v) = 0, which shows that v belongs to the null space of AB − BA. Thus, nullspace(AB − BA) = 0. Therefore, by the Invertible Matrix Theorem, AB − BA cannot be invertible. Chapter 8 Solutions Solutions to Section 8.1 True-False Review: (c)2017 Pearson Education. Inc. 535 (a): TRUE. This is essentially the statement of Theorem 8.1.4. (b): FALSE. As stated in Theorem 8.1.6, if there is any point x0 in I such that W [y1 , y2 , . . . , yn ](x0 ) = 0, then {y1 , y2 , . . . , yn } is linearly dependent on I. (c): FALSE. Many counterexamples are possible. Note that (xD − Dx)(x) = xD(x) − D(x2 ) = x − 2x = (−1)(x). Therefore, xD − Dx = −1. Setting L1 = x and L2 = D, we therefore see that L1 L2 = L2 L1 in this example. (d): TRUE. By assumption, L1 (y1 + y2 ) = L1 (y1 ) + L1 (y2 ) and L1 (cy) = cL1 (y) for all functions y, y1 , y2 . Likewise, L2 (y1 + y2 ) = L2 (y1 ) + L2 (y2 ) and L2 (cy) = cL2 (y) for all functions y, y1 , y2 . Therefore, (L1 + L2 )(y1 + y2 ) = L1 (y1 + y2 ) + L2 (y1 + y2 ) = (L1 (y1 ) + L1 (y2 )) + (L2 (y1 ) + L2 (y2 )) = (L1 (y1 ) + L2 (y1 )) + (L1 (y2 ) + L2 (y2 )) = (L1 + L2 )(y1 ) + (L1 + L2 )(y2 ) and (L1 + L2 )(cy) = L1 (cy) + L2 (cy) = cL1 (y) + cL2 (y) = c(L1 (y) + L2 (y)) = c(L1 + L2 )(y). Therefore, L1 + L2 is a linear differential operator. (e): TRUE. By assumption L(y1 + y2 ) = L(y1 ) + L(y2 ) and L(ky) = kL(y) for all scalars k. Therefore, for all constants c, we have (cL)(y1 + y2 ) = cL(y1 + y2 ) = c(L(y1 ) + L(y2 )) = cL(y1 ) + cL(y2 ) = (cL)(y1 ) + (cL)(y2 ) and (cL)(ky) = c(L(ky)) = c(k(L(y))) = k(cL)(y). Therefore, cL is a linear differential operator. (f ): TRUE. We have L(yp + u) = L(yp ) + L(u) = F + 0 = F. (g): TRUE. We have L(y1 + y2 ) = L(y1 ) + L(y2 ) = F1 + F2 . (h):TRUE. It follows from the results of this section that Ker(L) is a vector space of dimension 2. Therefore any two linearly independent functions in Ker(L) will form a basis for Ker(L). Since L(ex ) = (D2 − 1)(ex ) = D2 (ex ) − ex = ex − ex = 0, and L(e−x ) = (D2 − 1)(e−x ) = D2 (e−x ) − e−x = e−x − e−x = 0 it follows that both ex and e−x are in Ker(L). Further, ex and e−x are not proportional and therefore are linearly independent on any interval. Consequently, {ex , e−x } is a basis for Ker(L). (i): TRUE. It follows from the results of this section that the solution space to the given differential equation is a vector space of dimension 2. Therefore a basis for this solution space will consist of precisely two linearly independent solutions to the differential equation. If we set y1 (x) = ex , andy2 (x) = e2x , then y1 − 3y1 + 2y1 = ex − 3ex + 2ex = 0, (c)2017 Pearson Education. Inc. 536 and y2 − 3y2 + 2y2 = 4e−2x − 6e−2x + 2e−2x = 0. Hence, y1 (x) and y2 (x) are both solutions to the differential equation. Further, they are not proportional and so are linearly independent on any interval. Consequently, {ex , e−2x } is a basis for the solution space of the differential equation. (j): FALSE. It follows from the results of this section that the solution space to the given differential equation is a vector space of dimension 3. Since the given set of vectors contains only two functions, it cannot form a basis for the solution space. Problems: 1. (a). L(y(x)) = (D − x)(2e3x ) = D(2e3x ) − 2xe3x = 6e3x − 2xe3x . (b). L(y(x)) = (D − x)(3 ln x) = D(3 ln x) − 3x ln x = 3 − 3x ln x. x (c). Since L is a linear operator, we have L(2e3x + 3 ln x) = L(2e3x ) + L(3 ln x) = 6e3x − 2xe3x + where we have used the results from (a) and (b). 3 − 3x ln x, x 2. (a). L(y(x)) = (D2 − x2 D + x)(2e3x ) = D2 (2e3x ) − x2 D(2e3x ) + 2xe3x = 18e3x − 6x2 e3x + 2xe3x . 3 (b). L(y(x)) = (D2 − x2 D + x)(3 ln x) = D2 (3 ln x) − x2 D(3 ln x) + x(3 ln x) = − 2 − 3x + 3x ln x. x (c). Since L is a linear operator, we have L(2e3x + 3 ln x) = L(2e3x ) + L(3 ln x) = 18e3x − 6x2 e3x + 2xe3x − 3 − 3x + 3x ln x, x2 where we have used the results from (a) and (b). 3. (a). L(y(x)) = (D3 − 2xD2 )(2e3x ) = D3 (2e3x ) − 2xD2 (2e3x ) = 54e3x − 36xe3x . (b). L(y(x)) = (D3 − 2xD2 )(3 ln x) = D3 (3 ln x) − 2xD2 (3 ln x) = 6 6 + . 3 x x (c). Since L is a linear operator, we have L(2e3x + 3 ln x) = L(2e3x ) + L(3 ln x) = 54e3x − 36xe3x + 6 6 + , 3 x x where we have used the results from (a) and (b). 4. (a). L(y(x)) = (D3 − D + 4)(2e3x ) = D3 (2e3x ) − D(2e3x ) + 8e3x = 54e3x − 6e3x + 8e3x = 56e3x . (b). L(y(x)) = (D3 − D + 4)(3 ln x) = D3 (3 ln x) − 2D(3 ln x) + 12 ln x = 6 3 − + 12 ln x. x3 x (c). Since L is a linear operator, we have L(2e3x + 3 ln x) = L(2e3x ) + L(3 ln x) = 56e3x + where we have used the results from (a) and (b). (c)2017 Pearson Education. Inc. 6 3 − + 12 ln x, 3 x x 537 5. L(y(x)) = (D2 − 4D + 4)(xe2x ) = D2 (xe2x ) − 4D(xe2x ) + 4xe2x = D(e2x + 2xe2x ) − 4(e2x + 2xe2x ) + 4xe2x = 4e2x + 4xe2x − 4(e2x + 2xe2x ) + 4xe2x = 0. Thus, xe2x ∈ Ker(L). 6. L(y(x)) = (x2 D2 + 2xD − 2)(x−2 ) = x2 D2 (x−2 ) + 2xD(x−2 ) − 2x−2 = x2 D(−2x−3 ) − 4x−2 − 2x−2 = 6x−2 − 6x−2 = 0. Thus, f ∈ Ker(L). 7. L(y(x)) = (D2 − x−1 D + 4x2 )(sin(x2 )) = D2 (sin (x2 )) − x−1 D(sin (x2 )) + 4x2 (sin (x2 )) = D(2x cos (x2 )) − x−1 (2x cos (x2 )) + 4x2 (sin (x2 )) = 2[cos (x2 ) − 2x2 sin (x2 )] − 2 cos (x2 ) + 4x2 (sin (x2 )) = 0. Thus, f ∈ Ker(L). 8. L(y(x)) = (D3 + D2 + D + 1)(sin x + cos x) = D3 (sin x + cos x) + D2 (sin x + cos x) + D(sin x + cos x) + (sin x + cos x) = D2 (cos x − sin x) + D(cos x − sin x) + cos x − sin x + sin x + cos x = D(− sin x − cos x) + (− sin x − cos x) + 2 cos x = − cos x + sin x − sin x − cos x + 2 cos x = 0. Thus, f ∈ Ker(L). 9. L(y(x)) = (−D2 + 2D − 1)(xex ) = −D2 (xex ) + 2D(xex ) − xex = −D(ex + xex ) + 2(ex + xex ) − xex = −(ex + ex + xex ) + 2ex + 2xex − xex = 0. Thus, f ∈ Ker(L). 10. L(y(x)) = 0 ⇐⇒ (D − 3x2 )y = 0 ⇐⇒ y − 3x2 y = 0. This linear equation has integrating factor d −x3 2 3 y) = 0, which has a general I = e− 3x dx = e−x , so that the differential equation can be written as (e dx 3 3 solution e−x y = c, that is, y(x) = cex . Consequently, 3 Ker(L) = {y ∈ C 1 (R) : y(x) = cex , c ∈ R}. (c)2017 Pearson Education. Inc. 538 11. L(y(x)) = 0 ⇐⇒ (D2 +1)y = 0 ⇐⇒ y +y = 0. This is a second-order homogeneous linear differential equation, so by Theorem 8.1.4, the solution set to this equation forms a 2-dimensional vector space. A little thought shows that both y1 = cos x and y2 = sin x are solutions to the equation. Since they are linearly independent, these two functions form a basis for the solution set. Therefore, Ker(L) = {a cos x + b sin x : a, b ∈ R}. 12. L(y(x)) = 0 ⇐⇒ (D2 + 2D − 15)y = 0 ⇐⇒ y + 2y − 15y = 0. This is a second-order homogeneous linear differential equation, so by Theorem 8.1.4, the solution set to this equation forms a 2-dimensional vector space. Following the hint given in the text, we try for solutions of the form y(x) = erx . Substituting this into the differential equation, we get r2 erx + 2rerx − 15erx = 0, or erx (r2 + 2r − 15) = 0. It follows that r2 + 2r − 15 = 0. That is (r + 5)(r − 3) = 0, and hence, r = −5 and r = 3 are the solutions. Therefore, we obtain the solutions y1 = e−5x and y2 = e3x . Since they are linearly independent (by computing the Wronskian, for instance), these two functions form a basis for the solution set. Therefore, Ker(L) = {ae−5x + be3x : a, b ∈ R}. 13. Ly = 0 ⇐⇒ (x2 D + x)y = 0 ⇐⇒ x2 y + xy = 0. Dividing by x2 , we can express this as y + x1 y = 0. This differential equation is separable: y1 dy = − x1 dx. Integrating both sides, we obtain ln |y| = − ln |x| + c1 . c3 Therefore |y| = e− ln |x|+c1 = c2 e− ln |x| . Hence, y(x) = c3 e− ln |x| = |x| . Thus, c Ker(L) = :c∈R . |x| 14. We have (L1 L2 )(f ) = L1 (f − 2x2 f ) = (D + 1)(f − 2x2 f ) = f + f − 4xf − 2x2 f − 2x2 f = f + (1 − 2x2 )f − (4x + 2x2 )f, so that Furthermore, L1 L2 = D2 + (1 − 2x2 )D − 2x(2 + x). (L2 L1 )(f ) = L2 (f + f ) = (D − 2x2 )(f + f ) = f + f − 2x2 f − 2x2 f, so that L2 L1 = D2 + (1 − 2x2 )D − 2x2 . Therefore, L1 L2 = L2 L1 . 15. We have (L1 L2 )(f ) = L1 (f + (2x − 1)f ) = (D + x)(f + (2x − 1)f ) = f + xf + 2f + (2x − 1)f + x(2x − 1)f = (D2 + (3x − 1)D + (2x2 − x + 2))(f ), (c)2017 Pearson Education. Inc. 539 so that L1 L2 = D2 + (3x − 1)D + (2x2 − x + 2). Furthermore, (L2 L1 )(f ) = L2 (f + xf ) = (D + (2x − 1))(f + xf ) = f + (2x − 1)f + f + xf + (2x − 1)xf = (D2 + (3x − 1)D + (2x2 − x + 1))(f ), so that L2 L1 = D2 + (3x − 1)D + (2x2 − x + 1). Therefore, L1 L2 = L2 L1 . 16. We have (L1 L2 )(f ) = L1 (D + b1 )f = (D + a1 )(f + b1 f ) = f + [b1 + a1 ]f + (b1 + a1 b1 )f. Thus L1 L2 = D2 + (b1 + a1 )D + (b1 + a1 b1 ). Similarly, L2 L1 = D2 + (a1 + b1 )D + (a1 + b1 a1 ). Thus L1 L2 −L2 L1 = b1 −a1 , which is the zero operator if and only if b1 = a1 which can be integrated directly to obtain b1 = a1 + c2 , where c2 is an arbitrary constant. Consequently we must have L2 = D + [a1 (x) + c2 ]. 17. (D3 + x2 D2 − (sin x)D + ex )y = x3 and 18. (D2 + 4xD − 6x2 )y = x2 sin x y + 4xy − 6x2 y = 0. and y + x2 y − (sin x)y + ex y = 0. 19. x2 and ex are both continuous functions for all x ∈ R and in particular on any interval I containing x0 = 0. Thus, y(x) = 0 is clearly a solution to the given differential equation and also satisfies the initial conditions. Thus, by the existence-uniqueness theorem, y(x) = 0 is the only solution to the initial-value problem. 20. Let a1 , ..., an be functions that are continuous on the interval I. Then, for any x0 in I, the initial-value problem y (n) + a1 (x)y (n−1) + ... + an−1 (x)y + an (x)y = 0, y(x0 ) = 0, y (x0 ) = 0, ..., y (n−1) (x0 ) = 0, has only the trivial solution y(x) = 0. Proof: All of the conditions of the existence-uniqueness theorem are satisfied and y(x) = 0 is a solution; consequently, it is the only solution. 21. Since the differential equation is second order, linear, and homogenous, any basis for the solution space will consist of precisely two linearly independent solutions. We can therefore immediately rule out S1 (too few functions), S2 (too many functions), and S5 (the set is linearly dependent). We now consider the remaining sets of functions. S3 = {e4x , e2x } : Since the two functions in S3 are not proportional, they are linearly independent on any interval. Therefore, S3 will be a basis for the solution space of the differential equation provide y1 (x) = e4x (c)2017 Pearson Education. Inc. 540 and y2 (x) = e2x , are both solutions of the differential equation. Since y1 − 16y1 = 16e4x − 16e4x = 0, it follows that y1 (x) = e4x is a solution to the differential equation. However, y2 − 16y2 = 4e2x − 16e2x = −12e2x = 0, so that y2 (x) = e2x is not a solution to the differential equation. Consequently, S3 is not a basis for the solution space of the differential equation. S4 = {e4x , e−4x }: We have already shown above that y1 (x) = e4x is a solution to the differential equation. Further, if y2 (x) = e−4x , then y2 − 16y2 = 16e−4x − 16e−4x = 0, so that y2 (x) = e−2x is also a solution to the differential equation. Since y1 (x) and y2 (x) are not proportional, they are linearly independent on any interval. Consequently, S4 is a basis for the solution space of the differential equation. 1 4x 1 4x S5 = {cosh 4x, sinh 4x}: Since cosh 4x = e + e−4x , and sinh 4x = e − e−4x , are both linear 2 2 combinations of the solutions e4x , e−4x , it follows that they are themselves solutions. Further since cosh x and sinh 4x are not proportional, they are linearly independent on any interval. Consequently, S5 is a basis for the solution space of the differential equation. 22. Since the differential equation is second order, linear, and homogenous, any basis for the solution space will consist of precisely two linearly independent solutions. We can therefore immediately rule out S1 (too few functions). We now consider the remaining sets of functions. S2 = {x2 , x2 ln x}: Let y1 (x) = x2 . Then, x2 y1 − 3xy1 + 4y1 = x2 (2) − 3x(2x) + 4x2 = 0, so that y1 (x) = x2 is a solution to the differential equation. Further, if y2 (x) = x2 ln x, then y2 (x) = 2x ln x + x, and y2 (x) = 2 ln x + 3, so that x2 y2 − 3xy2 + 4y2 = x2 (2 ln x + 3) − 3x(2x ln x + x) + 4x2 ln x = 0. Hence, y2 (x) = x2 ln x is also a solution to the differential equation. Since y1 (x) and y2 (x) are not proportional, they are linearly independent on any interval. Consequently, S2 is a basis for the solution space of the differential equation. S3 = {2x2 , 3x2 ln x} : The elements of S3 are scalar multiples of the elements of S2 , and therefore are solutions to the differential equation. Further, since they are not proportional, they are linearly independent on (0, ∞). Hence S3 is a basis for the solution space to the differential equation. S4 = {x2 (2 + 3 ln x), x2 (2 − 3 ln x)} : The elements of S4 are linear combinations of the elements of S2 , and therefore are solutions to the differential equation. Further, since they are not proportional, they are linearly independent on (0, ∞). Hence S4 is a basis for the solution space to the differential equation. 23. Given y − 2y − 3y = 0 then r2 − 2r − 3 = 0 =⇒ r ∈ {−1, 3} =⇒ y(x) = c1 e−x + c2 e3x . 24. Given y + 7y + 10y = 0 then r2 + 7r + 10 = 0 =⇒ r ∈ {−5, −2} =⇒ y(x) = c1 e−5x + c2 e−2x . (c)2017 Pearson Education. Inc. 541 25. Given y − 36y = 0 then r2 − 36 = 0 =⇒ r ∈ {−6, 6} =⇒ y(x) = c1 e−6x + c2 e6x . 26. Given y + 4y = 0 then r2 + 4r = 0 =⇒ r ∈ {−4, 0} =⇒ y(x) = c1 e−4x + c2 . 27. Substituting y(x) = erx into the given differential equation yields erx (r3 − 3r2 − r + 3) = 0, so that we will have a solution provided that r satisfies r3 − 3r2 − r + 3 = 0, that is, (r − 1)(r + 1)(r − 3) = 0. Consequently, three solutions to the given differential equation are y1 (x) = ex , y2 (x) = e−x , y3 (x) = e3x . Further, the Wronskian of these solutions is x e e−x e3x x −e−x 3e3x = −16e3x . W [y1 , y2 , y3 ] = e x e e−x 9e3x Since the Wronskian is never zero, the solutions are linearly independent on any interval. Hence the general solution to the differential equation is y(x) = c1 ex + c2 e−x + c3 e3x . 28. Substituting y(x) = erx into the given differential equation yields erx (r3 + 3r2 − 4r − 12) = 0, so that we will have a solution provided that r satisfies r3 + 3r2 − 4r − 12 = 0, that is, (r − 2)(r + 2)(r + 3) = 0. Consequently, three solutions to the given differential equation are y1 (x) = e2x , y2 (x) = e−2x , y3 (x) = e−3x . Further, the Wronskian of these solutions is 2x e e−2x e−3x 2x −2e−2x −3e−3x = −20e−3x . W [y1 , y2 , y3 ] = 2e 2x 4e 4e−2x 9e−3x Since the Wronskian is never zero, the solutions are linearly independent on any interval. Hence the general solution to the differential equation is y(x) = c1 e2x + c2 e−2x + c3 e−3x . 29. Substituting y(x) = erx into the given differential equation yields erx (r3 + 3r2 − 18r − 40) = 0, so that we will have a solution provided r satisfies r3 + 3r2 − 18r − 40 = 0, that is, (r + 5)(r + 2)(r − 4) = 0. Consequently, three solutions to the differential equation are y1 (x) = e−5x , y2 (x) = e−2x , y3 (x) = e4x . Further, the Wronskian of the solution is e−5x e−2x e4x 4e4x = 162e−3x . W [y1 , y2 , y3 ] = −5e−5x −2e−2x −5x −2x 25e 4e 16e4x Since the Wronskian is never zero, the solutions are linearly independent on any interval. Hence the general solution to the differential equation is y(x) = c1 e−5x + c2 e−2x + c3 e4x . 30. Given y − y − 2y = 0 then r3 − r2 − 2r = 0 =⇒ r(r2 − r − 2) = 0 =⇒ r ∈ {−1, 0, 2} =⇒ y(x) = c1 e−x + c2 + c3 e2x . 31. Given y + y − 10y + 8y = 0 then r3 + r2 − 10r + 8 = 0 =⇒ (r − 2)(r − 1)(r + 4) = 0 =⇒ r ∈ {−4, 1, 2} =⇒ y(x) = c1 e−4x + c2 ex + c3 e2x . 32. Given y (iv) − 2y − y + 2y = 0 then r4 − 2r3 − r2 − 2r = 0 =⇒ r(r3 − 2r2 − r − 2) = 0 =⇒ r(r − 2)(r − 1)(r + 1) = 0 =⇒ r ∈ {−1, 0, 1, 2} =⇒ y(x) = c1 e−x + c2 + c3 ex + c4 e2x . 33. Given y (iv) − 13y + 36y = 0 then r4 − 13r2 + 36 = 0 =⇒ (r2 − 9)(r2 − 4) = 0 =⇒ r2 ∈ {4, 9} =⇒ r ∈ {−3, −2, 2, 3} =⇒ y(x) = c1 e−3x + c2 e−2x + c3 e2x + c4 e3x . (c)2017 Pearson Education. Inc. 542 34. Given x2 y + 3xy − 8y = 0, the trial solution gives x2 r(r − 1)xr−2 + 3xrxr−1 − 8xr = 0, or xr [r(r − 1) + 3r − 8] = 0. Therefore, r2 + 2r − 8 = 0, which factors as (r + 4)(r − 2) = 0. Therefore, r = −4 or r = 2. Hence, we obtain the solutions y1 (x) = x−4 and y2 (x) = x2 . Furthermore, W [x−4 , x2 ] = (x−4 )(2x) − (−4x−5 )(x2 ) = 6x−3 = 0, so that {x−4 , x2 } is a linearly independent set of solutions to the given differential equation on (0, ∞). Consequently, from Theorem 8.1.4, the general solution is given by y(x) = c1 x−4 + c2 x2 . 35. Given 2x2 y + 5xy + y = 0, the trial solution gives 2x2 r(r − 1)xr−2 + 5xrxr−1 + xr = 0, or xr [2r(r − 1) + 5r + 1] = 0. Therefore, 2r2 + 3r + 1 = 0, which factors as (2r + 1)(r + 1) = 0. Therefore, 1 r = − 12 and r = −1. Hence, we obtain the solutions y1 (x) = x− 2 and y2 (x) = x−1 . Furthermore, 1 1 1 1 3 1 5 W [x− 2 , x−1 ] = (x− 2 )(− 2 ) − (− x− 2 )(x−1 ) = − x− 2 = 0, x 2 2 1 so that {x− 2 , x−1 } is a linearly independent set of solutions to the given differential equation on (0, ∞). 1 Consequently, from Theorem 8.1.4, the general solution is given by y(x) = c1 x− 2 + c2 x−1 . 36. Substituting y(x) = xr into the given differential equation yields xr [r(r −1)(r −2)+r(r −1)−2r +2] = 0, so that r must satisfy (r − 1)(r − 2)(r + 1) = 0. If follows that three solutions to the differential equation are y1 (x) = x, y2 (x) = x2 , y3 (x) = x−1 . Further, the Wronskian of these solutions is x x2 x−1 W [y1 , y2 , y3 ] = 1 2x −x−2 = 6x−1 . 0 2 2x−3 Since the Wronskian is nonzero on (0, ∞), the solutions are linearly independent on this interval. Consequently, the general solution to the differential equation is y(x) = c1 x + c2 x2 + c3 x−1 . 37. Given x3 y + 3x2 y − 6xy = 0, the trial solution gives x3 r(r − 1)(r − 2)xr−3 + 3x2 r(r − 1)xr−2 − 6xrxr−1 = 0, or xr [r(r − 1)(r − 2) + 3r(r − 1) − 6r] = 0. Therefore, r(r − 1)(r − 2) + 3r(r − 1) − 6r = 0, or r[(r − 1)(r − 2) + 3(r − 1) − 6] = 0. Therefore, r[r2 − 7] =√0, √ √ so that r = 0 or r = ± 7. Hence, we obtain the solutions y1 (x) = 1, y2 (x) = x 7 , and y3 (x) = x− 7 . Furthermore, √ √ 7 1 x √7 x− √ √ √ √ √ 7 − 7 W [1, x , x ]= 0 7x 7−1√ − 7x− 7−1 √ √ √ √ √ 7−2 0 7( 7 − 1)x − 7(− 7 − 1)x− 7−2 √ √ = −7(− 7 − 1)x−3 + 7( 7 − 1)x−3 √ = 14 7x−3 = 0, √ √ so that {1, x 7 , x− 7 } is a linearly independent set of solutions to the given differential equation on (0, ∞). √ √ Consequently, from Theorem 8.1.4, the general solution is given by y(x) = c1 + c2 x 7 + c3 x− 7 . (c)2017 Pearson Education. Inc. 543 38. To determine a particular solution of the form yp (x) = A0 e5x , we substitute this solution into the differential equation: (A0 e5x ) + (A0 e5x ) − 6(A0 e5x ) = 18e5x . Therefore 25A0 e5x + 5A0 e5x − 6A0 e5x = 18e5x , 3 5x e . 4 To obtain the general solution, we need to find the complementary function yc (x), the solution to the associated homogeneous differential equation: y + y − 6y = 0. Seeking solutions of the form y(x) = erx , we obtain r2 erx + rerx − 6erx = 0, or erx (r2 + r − 6) = 0. Therefore, r2 + r − 6 = 0, which factors as (r + 3)(r − 2) = 0. Hence, r = −3 or r = 2. Therefore, we obtain the solutions y1 (x) = e−3x and y2 (x) = e2x . Since W [e−3x , e2x ] = (e−3x )(2e2x ) − (−3e−3x )(e2x ) = 2e−x + 3e−x = 5e−x = 0, which forces A0 = 3/4. Therefore, yp (x) = {e−3x , e2x } is linearly independent. Therefore, the complementary function is yc (x) = c1 e−3x + c2 e2x . By Theorem 8.1.8, the general solution to the differential equation is y(x) = c1 e−3x + c2 e2x + 3e3x . 39. Substituting y(x) = A0 + A1 x + A2 x2 into the differential equation yields (A0 + A1 x + A2 x2 ) + (A0 + A1 x + A2 x2 ) − 2(A0 + A1 x + A2 x2 ) = 4x2 + 5 or (2A2 + A1 − 2A0 ) + (2A2 − 2A1 )x − 2A2 x2 = 4x2 + 5. Equating the powers of x on each side of this equation yields 2A2 + A1 − 2A0 = 5, 2A2 − 2A1 = 0, −2A2 = 4. 11 −2x−2x2 . 2 To obtain the general solution, we need to find the complementary function yc (x), the solution to the associated homogeneous differential equation: y + y − 2y = 0. Seeking solutions of the form y(x) = erx , we obtain r2 erx + rerx − 2erx = 0, or erx (r2 + r − 2) = 0. Therefore, r2 + r − 2 = 0, which factors as (r + 2)(r − 1) = 0. Hence, r = −2 or r = 1. Therefore, we obtain the solutions y1 (x) = e−2x and y2 (x) = ex . Since W [e−2x , ex ] = (e−2x )(ex ) − (−2e−2x )(ex ) = 3e−x = 0, Solving for A0 , A1 , and A2 , we find that A0 = −11/2, A1 = −2, and A2 = −2. Thus, yp (x) = − {e−2x , ex } is linearly independent. Therefore, the complementary function is yc (x) = c1 e−2x + c2 ex . By Theorem 8.1.8, the general solution to the differential equation is y(x) = c1 e−2x + c2 ex − 11 − 2x − 2x2 . 2 40. Substituting y(x) = A0 e2x into the given differential equation yields A0 e2x (8 + 8 − 2 − 2) = 4e2x , so 1 1 that A0 = e2x . Hence, a particular solution to the differential equation is yp (x) = e2x . To determine the 3 3 general solution we need to solve the associated homogeneous differential equation y + 2y − y − 2y = 0. We try for solutions of the form y(x) = erx . Substituting into the homogeneous differential equation gives (c)2017 Pearson Education. Inc. 544 erx (r3 +2r2 −r−2) = 0, so that we choose r to satisfy r3 +2r2 −r−2 = 0, or equivalently, (r+2)(r−1)(r+1) = 0. It follows that three solutions to the differential equation are y1 (x) = e−2x , y2 (x) = ex , y3 (x) = e−x . Further, the Wronskian of these solutions is e−2x ex e−x W [y1 , y2 , y3 ] = −2e−2x ex −e−x = −6e−2x . 4e−2x ex e−x Since the Wronskian is nonzero, the solutions are linearly independent on any interval. Thus, the complementary function is yc (x) = c1 e−2x + c2 ex + c3 e−x . By Theorem 8.1.8, the general solution to the differential equation is 1 y(x) = yc (x) + yp (x) = c1 e−2x + c2 ex + c3 e−x + e2x . 3 41. Substituting y(x) = A0 e−3x into the differential equation yields A0 e−3x (−27 + 9 + 30 + 8) = 24e−3x . 6 6 . Thus, yp (x) = e−3x . 5 5 To obtain the general solution, we need to find the complementary function yc (x), the solution to the associated homogeneous differential equation: y + y − 10y + 8y = 0. Seeking solutions of the form y(x) = erx , we obtain r3 erx + r2 erx − 10rerx + 8erx = 0, or erx (r3 + r2 − 10r + 8) = 0. Therefore, r3 + r2 − 10r + 8 = 0, which has roots r = 1, r = 2, and r = −4. Therefore, we obtain the solutions y1 (x) = ex , y2 (x) = e2x , and y3 (x) = e−4x . Since Therefore, 20A0 = 24. Hence, A0 = x 2x W [e , e , e −4x ex ] = ex ex e2x 2e2x 4e2x e−4x −4e−4x = 30e−x = 0, 16e−4x {ex , e2x , e−4x } is linearly independent. Thus, the complementary function is yc (x) = c1 ex + c2 e2x + c3 e−4x . By Theorem 8.1.8, the general solution to the differential equation is 6 y(x) = c1 ex + c2 e2x + c3 e−4x + e−3x . 5 42. Substituting y(x) = A0 e−x into the differential equation yields A0 e−x (−1 + 5 − 6) = 6e−x . Therefore, −2A0 = 6. Hence, A0 = −3. Thus, yp (x) = −3e−x . To obtain the general solution, we need to find the complementary function yc (x), the solution to the associated homogeneous differential equation: y + 5y + 6y = 0. Seeking solutions of the form y(x) = erx , we obtain r3 erx + 5r2 erx + 6rerx = 0, or erx (r3 + 5r2 + 6r) = 0. Therefore, r3 + 5r2 + 6r = 0, which has roots r = 0, r = −2, and r = −3. Therefore, we obtain the solutions y1 (x) = 1, y2 (x) = e−2x , and y3 (x) = e−3x . Since 1 e−2x e−3x −2x −3x −2x −3e−3x = −6e−5x = 0, W [1, e ,e ] = 0 −2e −2x 0 4e 9e−3x (c)2017 Pearson Education. Inc. 545 {1, e−2x , e−3x } is linearly independent. Therefore, the complementary function is yc (x) = c1 +c2 e−2x +c3 e−3x . By Theorem 8.1.8, the general solution to the differential equation is y(x) = c1 + c2 e−2x + c3 e−3x − 3e−x . 43. Let y1 and y2 belong to C n (I), and let c be a scalar. We must show that L(y1 + y2 ) = L(y1 ) + L(y2 ) and L(cy1 ) = cL(y1 ). We have L(y1 + y2 ) = (Dn + a1 Dn−1 + · · · + an−1 D + an )(y1 + y2 ) = Dn (y1 + y2 ) + a1 Dn−1 (y1 + y2 ) + · · · + an−1 D(y1 + y2 ) + an (y1 + y2 ) = (Dn y1 + a1 Dn−1 y1 + · · · + an−1 Dy1 + an y1 ) + (Dn y2 + a1 Dn−1 y2 + · · · + an−1 Dy2 + an y2 ) = (Dn + a1 Dn−1 + · · · + an−1 D + an )(y1 ) + (Dn + a1 Dn−1 + · · · + an−1 D + an )(y2 ) = L(y1 ) + L(y2 ), and L(cy1 ) = (Dn + a1 Dn−1 + · · · + an−1 D + an )(cy1 ) = Dn (cy1 ) + a1 Dn−1 (cy1 ) + · · · + an−1 D(cy1 ) + an (cy1 ) = cDn (y1 ) + ca1 Dn−1 (y1 ) + · · · + can−1 D(y1 ) + can (y1 ) = c(Dn (y1 ) + a1 Dn−1 (y1 ) + · · · + an−1 D(y1 ) + an (y1 )) = c(Dn + a1 Dn−1 + · · · + an−1 D + an )(y1 ) = cL(y1 ). 44. Prior to the statement of Theorem 8.1.4 it was shown that the set of all solutions forms a vector space. We now show that th dimension of this solution space is n, by constructing a basis. Let y1 , y2 , ..., yn be the unique solutions of the n initial-value problems: (n) a0 (x)yi (n−1) + a1 (x)yi (k−1) yi + ... + an−1 (x)yi + an (x)yi = 0, (x0 ) = δik , k = 1, 2, ..., n, respectively. The Wronskian of these functions at x0 is W [y1 , y2 , ..., yn ](x0 ) = det[In ] = 1 = 0 so that the solutions are linearly independent on I. We now show that they span the solution space. Let u(x) be any solution of the differential equation on I, and suppose that u(x0 ) = u1 , u (x0 ) = u2 , ..., u(n−1) (x0 ) = un , where u1 , u2 , ..., un are constants. It follows that y = u(x) is the unique solution to the initial-value problem: Ly = 0, y(x0 ) = u1 , y (x0 ) = u2 , ..., y (n−1) (x0 ) = un . However, if we define w(x) = u1 y1 (x) + u2 y2 (x) + ... + un yn (x), then w(x) also satisfies this initial value problem . Thus, by uniqueness, we must have u(x) = w(x), that is, u(x) = u1 y1 (x) + u2 y2 (x) + ... + un yn (x). Thus we have shown that {y1 , y2 , ..., yn } forms a basis for the solution space and hence the dimension of this solution space is n. (c)2017 Pearson Education. Inc. 546 45. Consider the linear system c1 y1 (x0 ) + c2 y2 (x0 ) + ... + cn yn (x0 ) c1 y1 (x0 ) + c2 y2 (x0 ) + ... + cn yn (x0 ) .. . (n−1) c 1 y1 (n−1) (x0 ) + c2 y2 (n−1) (x0 ) + ... + cn yn =0 =0 (x0 ) = 0, where we are solving for c1 , c2 , ..., cn . The determinant of the matrix of coefficients of this system is W [y1 , y2 , ..., yn ](x0 ) = 0, so that the system has non-trivial solutions. Let (α1 , α2 , ..., αn ) be one such non-trivial solution for (c1 , c2 , . . . , cn ). Therefore, not all of the αi are zero. Define the function u(x) by u(x) = α1 y1 (x) + α2 y2 (x) + ... + αn yn (x). It follows that y = u(x) satisfies the initial-value problem: a0 y (n) + a1 y (n−1) + ... + an−1 y + an y = 0 y(x0 ) = 0, y (x0 ) = 0, ..., y (n−1) (x0 ) = 0. However, y(x) = 0 also satisfies the above initial value problem and hence, by uniqueness, we must have u(x) = 0, that is, α1 y1 (x) + α2 y2 (x) + ... + αn yn (x) = 0, where not all of the αi are zero. Thus, the functions y1 , y2 , ..., yn are linearly dependent on I. 46. Let vp be a particular solution to Equation (8.1.15), and let v be any solution to this equation. Then T (vp ) = w and T (v) = w. Subtracting these two equations gives T (v) − T (vp ) = 0, or equivalently, since T is a linear transformation, T (v − vp ) = 0. But this latter equation implies that the vector v − vp is an element of Ker(T ) and therefore can be written as v − vp = c1 v1 + c2 v2 + ... + cn vn , for some scalars c1 , c2 , . . . , cn . Hence, v = c1 v1 + c2 v2 + ... + cn vn + vp . Solutions to Section 8.2 True-False Review: (a): FALSE. Even if the auxiliary polynomial fails to have n distinct roots, the differential equation still has n linearly independent solutions. For example, if L = D2 + 2D + 1, then the differential equation Ly = 0 has auxiliary polynomial with (repeated) roots r = −1, −1. Yet we do have two linearly independent solutions y1 (x) = e−x and y2 (x) = xe−x to the differential equation. (b): FALSE. Theorem 8.2.2 only applies to polynomial differential operators. However, in general, many counterexamples can be given. For example, note that (xD − Dx)(x) = xD(x) − D(x2 ) = x − 2x = (−1)(x). (c)2017 Pearson Education. Inc. 547 Therefore, xD − Dx = −1. Setting L1 = x and L2 = D, we therefore see that L1 L2 = L2 L1 in this example. (c): TRUE. This is really just the statement that a polynomial of degree n always has n roots, with multiplicities counted. (d): TRUE. Since 0 is a root of multiplicity four, each term of the polynomial differential operator must contain a factor of D4 , so that any polynomial of degree three or less becomes zero after taking four (or more) derivatives. Therefore, for a homogeneous differential equation of this type, a polynomial of degree three or less must be a solution. (e): FALSE. Note that r = 0 is a root of the auxiliary polynomial, but only of multiplicity 1. The expression c1 + c2 x in the solution reflects r = 0 as a root of multiplicity 2. (f ): TRUE. The roots of the auxiliary polynomial are r = −3, −3, 5i, −5i. The portion of the solution corresponding to the repeated root r = −3 is c1 e−3x +c2 xe−3x , and the portion of the solution corresponding to the complex conjugate pair r = ±5i is c3 cos 5x + c4 sin 5x. (g): TRUE. The roots of the auxiliary polynomial are r = 2 ± i, 2 ± i. The terms corresponding to the first pair 2 ± i are c1 e2x cos x and c2 e2x sin x, and the repeated root gives two more terms: c3 xe2x cos x and c4 xe2x sin x. (h): FALSE. Many counterexamples can be given. For instance, if P (D) = D − 1, then the general solution is y(x) = cex . However, the differential equation (P (D))2 y = (D − 1)2 y has auxiliary equation with roots r = 1, 1 and general solution z(x) = c1 ex + c2 xex = xy(x). Problems: 1. The auxiliary polynomial is P (r) = r2 + 2r − 3 = (r + 3)(r − 1). Therefore, the auxiliary equation has roots r = −3 and r = 1. Therefore, two linearly independent solutions to the given differential equation are y1 (x) = e−3x and y2 (x) = ex . By Theorem 8.1.4, the solution space to this differential equation is 2-dimensional, and hence {e−3x , ex } forms a basis for the solution space. 2. The auxiliary polynomial is P (r) = r2 + 6r + 9 = (r + 3)2 . Therefore, the auxiliary equation has roots r = −3 (with multiplicity 2). Therefore, two linearly independent solutions to the given differential equation are y1 (x) = e−3x and y2 (x) = xe−3x . By Theorem 8.1.4, the solution space to this differential equation is 2-dimensional, and hence {e−3x , xe−3x } forms a basis for the solution space. 3. The auxiliary polynomial is P (r) = r2 − 6r + 25. According to the quadratic equation, the auxiliary equation has roots r = 3 ± 4i. Therefore, two linearly independent solutions to the given differential equation are y1 (x) = e3x cos 4x and y2 (x) = e3x sin 4x. By Theorem 8.1.4, the solution space to this differential equation is 2-dimensional, and hence {e3x cos 4x, e3x sin 4x} forms a basis for the solution space. 4. The general vector in S takes the form y(x) = c(sin 4x + 5 cos 4x). The full solution space for this differential equation is 2-dimensional by Theorem 8.1.4. Note that sin 4x belongs to the solution space and is linearly independent from sin 4x + 5 cos 4x since W [sin 4x + 5 cos 4x, sin 4x] = 20 = 0. (c)2017 Pearson Education. Inc. 548 Therefore, we can extend the basis for S to a basis for the entire solution space with {sin 4x+5 cos 4x, sin 4x}. Many other extensions are also possible, of course. 5. P (r) = r2 + r − 2 = (r + 2)(r − 1) =⇒ two linearly independent solutions to the given differential equation are y1 (x) = ex , y2 (x) = e−2x . According to the Gram-Schmidt process an orthogonal basis for the solution space to the given differential equation is {u1 (x), u2 (x)} where u1 (x) = y1 (x) = ex and 1 −2x x e · e dx x < y2 , u1 > 2(e−1 − 1) x 2 −2x −2x 0 u (x) = e − = e + u2 (x) = y2 (x) − · e e = e−2x − ex . 1 1 2 2 2x ||u1 || e − 1) e(e − 1) e dx 0 6. P (r) = r2 + 4 = 0, when r = ±2i =⇒ two linearly independent solutions to the given differential equation are y1 (x) = cos 2x, y2 (x) = sin 2x. According to the Gram-Schmidt process an orthogonal basis for the solution space to the given differential equation is {u1 (x), u2 (x)} where u1 (x) = y1 (x) = cos 2x and π/4 sin 2x · cos 2x dx < y2 , u1 > 2 0 u (x) = sin 2x − · cos 2x = sin 2x − cos 2x. u2 (x) = y2 (x) − 1 π/4 2 ||u1 ||2 π cos 2x dx 0 7. P (r) = r2 + 1 =⇒ two linearly independent solutions to the given differential equation are y1 ((x) = cos x, π y2 (x) = sin x. Since < y1 , y2 >= 0 cos x sin x dx = 0, {y1 (x), y2 (x)} is an orthogonal basis for the solution √ π π 2 x dx = cos π, and ||y || = sin2 x dx = space to the given differential equation. Further, ||y1 || = 2 0 0 √ 1 1 π, so that an orthonormal basis is √ cos x, √ sin x . π π 8. We have r2 − r − 2 = 0 =⇒ r ∈ {−1, 2} =⇒ y(x) = c1 e−x + c2 e2x . 9. We have r2 − 6r + 9 = 0 =⇒ r ∈ {3, 3} =⇒ y(x) = c1 e3x + c2 xe3x . 10. We have r2 + 6r + 25 = 0 =⇒ r ∈ {−3 − 4i, −3 + 4i} =⇒ y(x) = c1 e−3x cos 4x + c2 e−3x sin 4x. 11. We have (r + 1)(r − 5) = 0 =⇒ r ∈ {−1, 5} =⇒ y(x) = c1 e−x + c2 e5x . 12. We have (r + 2)2 = 0 =⇒ r ∈ {−2, −2} =⇒ y(x) = c1 e−2x + c2 xe−2x . 13. We have r2 − 6r + 34 = 0 =⇒ r ∈ {3 − 5i, 3 + 5i} =⇒ y(x) = c1 e3x cos 5x + c2 e3x sin 5x. 14. We have r2 + 10r + 25 = 0 =⇒ r ∈ {−5, −5} =⇒ y(x) = c1 e−5x + c2 xe−5x . √ √ √ √ 15. We have r2 − 2 = 0 =⇒ r ∈ {− 2, 2} =⇒ y(x) = c1 e− 2x + c2 e 2x . 16. We have r2 + 8r + 20 = 0 =⇒ r ∈ {−4 − 2i, −4 + 2i} =⇒ y(x) = c1 e−4x cos 2x + c2 e−4x sin 2x. 17. We have r2 + 2r + 2 = 0 =⇒ r ∈ {−1 − i, −1 + i} =⇒ y(x) = c1 e−x cos x + c2 e−x sin x. 18. We have (r − 4)(r + 2) = 0 =⇒ r ∈ {−2, 4} =⇒ y(x) = c1 e−2x + c2 e4x . 19. We have r2 − 14r + 58 = 0 =⇒ r ∈ {7 − 3i, 7 + 3i} =⇒ y(x) = c1 e7x cos 3x + c2 e7x sin 3x. 20. We have r3 − r2 + r − 1 = 0 =⇒ r ∈ {1, i, −i} =⇒ y(x) = c1 ex + c2 cos x + c3 sin x. (c)2017 Pearson Education. Inc. 549 21. We have r3 − 2r2 − 4r + 8 = 0 =⇒ r ∈ {−2, 2, 2} =⇒ y(x) = c1 e−2x + (c2 + c3 x)e2x . 22. We have (r − 2)(r2 − 16) = 0 =⇒ r ∈ {2, 4, −4} =⇒ y(x) = c1 e2x + c2 e4x + c3 e−4x . 23. We have (r2 + 2r + 10)2 = 0 =⇒ r ∈ {−1 + 3i, −1 + 3i, −1 − 3i, −1 − 3i} =⇒ y(x) = e−x [c1 cos 3x + c2 sin 3x + x(c4 cos 3x + c4 sin 3x)]. 24. We have (r2 + 4)2 (r + 1) = 0 =⇒ r ∈ {2i, 2i, −2i, −2i, −1} =⇒ y(x) = c1 e−x + c2 cos 2x + c3 sin 2x + x(c4 cos 2x + c5 sin 2x). √ √ −i 3, i 3} =⇒ 25. We have (r2 + 3)(r + 1) = 0√=⇒ r ∈ {−1, −1, √ y(x) = c1 e−x + c2 xe−x + c3 cos ( 3x) + c4 sin ( 3x). 26. We have r2 (r − 1) = 0 =⇒ r ∈ {0, 0, 1} =⇒ y(x) = c1 + c2 x + c3 ex . 27. We have r4 − 8r2 + 16 = (r − 2)2 (r + 2)2 = 0 =⇒ r ∈ {2, 2, −2, −2} =⇒ y(x) = e2x (c1 + c2 x) + e−2x (c3 + c4 x). 28. We have r4 − 16 = 0 =⇒ r ∈ {2, −2, 2i, −2i} =⇒ y(x) = c1 e2x + c2 e−2x + c3 cos 2x + c4 sin 2x. 29. We have r3 +8r2 +22r +20 = 0 =⇒ r ∈ {−2, −3+i, −3−i} =⇒ y(x) = c1 e−2x +e−3x (c2 cos x+c3 sin x). 30. We have r4 − 16r2 + 40r − 25 = 0 =⇒ r ∈ {1, −5, 2 + i, 2 − i} =⇒ y(x) = c1 ex + c2 e−5x + e2x (c3 cos x + c4 sin x). 31. We have (r −1)3 (r2 +9) = 0 =⇒ r ∈ {1, 1, 1, −3i, 3i} =⇒ y(x) = ex (c1 +c2 x+c3 x2 )+c4 cos 3x+c5 sin 3x. 32. We have (r2 − 2r + 2)2 (r2 − 1) = 0 =⇒ r ∈ {1 + i, 1 + i, 1 − i, 1 − i, 1, −1} =⇒ y(x) = ex (c1 cos x + c2 sin x) + xex (c3 cos x + c4 sin x) + c5 e−x + c6 ex . 33. We have (r + 3)(r − 1)(r + 5)3 = 0 =⇒ r ∈ {−3, 1, −5, −5, −5} =⇒ y(x) = c1 e−3x + c2 ex + e−5x (c3 + c4 x + c5 x2 ). 34. We have (r2 + 9)3 = 0 =⇒ r ∈ {3i, 3i, 3i, −3i, −3i, −3i} =⇒ y(x) = c1 cos 3x + c2 sin 3x + x(c3 cos 3x + c4 sin 3x) + x2 (c5 cos 3x + c6 sin 3x). 35. We have r2 − 8r + 16 = 0 =⇒ r ∈ {4, 4} =⇒ y(x) = c1 e4x + c2 xe4x . Now 2 = y(0) = c1 and 7 = y (0) = 4c1 + c2 . Therefore, c1 = 2 and c2 = −1 Hence, the solution to this initial-value problem is y(x) = 2e4x − xe4x . 36. We have r2 − 4r + 5 = 0 =⇒ r ∈ {2 + i, 2 − i} =⇒ y(x) = c1 e2x cos x + c2 e2x sin x. Now 3 = y(0) = c1 and 5 = y (0) = 2c1 + c2 . Therefore, c1 = 3 and c2 = −1. Hence, the solution to this initial-value problem is y(x) = 3e2x cos x − e2x sin x. 37. We have r3 − r2 + r − 1 = 0 =⇒ r ∈ {1, i, −i} =⇒ y(x) = c1 ex + c2 cos x + c3 sin x. Then since y(0) = 0, we have c1 + c2 = 0. Moreover, since y (0) = 1, we have c1 + c3 = 1. Finally, y (0) = 2 implies that c1 − c2 = 2. Solving these equations yields c1 = 1, c2 = −1, c3 = 0. Hence, the solution to this initial-value problem is y(x) = ex − cos x. 38. We have r3 + 2r2 − 4r − 8 = (r − 2)(r + 2)2 = 0 =⇒ r ∈ {2, −2, −2} =⇒ y(x) = c1 e2x + c2 e−2x + c3 xe−2x . Then since y(0) = 0, we have c1 + c2 = 0. Moreover, since y (0) = 6, we have 2c1 − 2c2 + c3 = 6. Further, (c)2017 Pearson Education. Inc. 550 y (0) = 8 implies that 4c1 + 4c2 − 4c3 = 8. Solving these equations yields c1 = 2, c2 = c3 = −2. Hence, the solution to this initial-value problem is y(x) = 2(e2x − e−2x − xe−2x ). 39. Given m > 0 and k > 0, we have auxiliary polynomial P (r) = r2 − 2mr + (m2 + k 2 ) = 0 =⇒ r = m ± ki. Therefore, y(x) = emx (c1 cos kx + c2 sin kx). Differentiating, we obtain y (x) = memx (c1 cos kx + c2 sin kx) + emx (−c1 k sin kx + c2 k cos kx). Now 0 = y(0) = c1 and k = y (0) = c2 k. Therefore, c1 = 0 and c2 = 1. Hence, the solution to this initial-value problem is y(x) = emx sin kx. 40. Given m > 0 and k > 0, we have auxiliary polynomial P (r) = r2 − 2mr + (m2 − k 2 ) = 0 =⇒ r = m ± k. Therefore, the general solution to this differential equation is y(x) = ae(m+k)x +be(m−k)x = emx (aekx +be−kx ). 2 2 Letting a = c1 +c and b = c1 −c in the last equality gives 2 2 y(x) = emx c1 + c2 kx c1 − c2 −kx e + e 2 2 kx e + e−kx ekx − e−kx = emx c1 + c2 2 2 = emx (c1 cosh kx + c2 sinh kx). 41. The auxiliary polynomial is P (r) = r2 + 2cr + k 2 . The quadratic formula supplies the roots of P (r) = 0: √ 2 2 r = −c ± c − k . (a). √If c2 < k 2 , then c2 − k 2 < 0 and the roots above are complex. We can write r = −c ± ωi, where ω = k 2 − c2 . Thus, y(t) = e−ct (c1 cos ωt + c2 sin ωt) and y (t) = e−ct [(ωc2 − cc1 ) cos ωt − (cc2 + ωc1 ) sin ωt] . Using the initial conditions y(0) = y0 and y (0) = 0, we find that c1 = y0 and ωc2 − y0 c = 0, or c2 = yω0 c . Therefore, y0 y(t) = e−ct (ω cos ωt + c sin ωt). ω √ √ (b). Since ω = k 2 − c2 , we find k = ω 2 + c2 (note k is assumed positive, so we take the positive square root). Therefore, continuing from part (a), we have y0 −ct e (ω cos ωt + c sin ωt) ω ω y0 c = k e−ct cos ωt + sin ωt ω k k ω c y0 −ct √ cos ωt + √ sin ωt . =k e ω ω 2 + c2 ω 2 + c2 y(t) = Now since ω √ 2 ω + c2 there exists φ such that cos φ = √ 2 c ω 2 + c2 + c √ 2 ω + c2 and 2 = 1, sin φ = √ (c)2017 Pearson Education. Inc. ω ; ω 2 + c2 551 that is, φ = tan−1 ω c . Hence, y(t) can be written as ky0 −ct ky0 e−ct sin(ωt + φ). y(t) = e (sin φ cos ωt + cos φ sin ωt) = ω ω ky0 e−ct . Since this tends to zero as t → ∞, the vibrations The amplitude is not constant, but is given by ω tend to die out as time goes on. The damping is exponential and the motion is reasonable. 42. Let u(x, y) = ex/α f (ξ), where ξ = βx − αy, and assume that α > 0 and β > 0. (a). Let us compute the partial derivatives of u: ∂u df 1 = ex/α + ex/α f ∂x dx α 1 x/α df ∂ξ =e + ex/α f dξ ∂x α df 1 = ex/α β + ex/α f dξ α and 2 df df ∂ξ β 1 1 ∂2u x/α d f ∂ξ = e β + ex/α β + ex/α + 2 ex/α f 2 2 ∂x dξ ∂x α dξ α dξ ∂x α 2 d f df f β = β 2 ex/α 2 + 2 ex/α + ex/α 2 . dξ α dξ α Moreover, ∂f ∂u = ex/α ∂y ∂y df dξ = ex/α dξ dy df = ex/α (−α) dξ df = −αex/α dξ and From the formulas for df ∂2u x/α ∂ = −αe 2 ∂y ∂y dξ d2 f ∂ξ = −αex/α 2 dξ ∂y 2 d f = −αex/α 2 (−α) dξ 2 d f = α2 ex/α 2 . dξ ∂2u ∂2u and , we have 2 ∂x ∂y 2 2 2 ∂2u ∂2u β df f x/α 2d f 2d f β . + = e + 2 + α + ∂x2 ∂y 2 dξ 2 α dξ α2 dξ 2 (c)2017 Pearson Education. Inc. 552 Now if ∂2u ∂2u + 2 = 0, then the last equation becomes ∂x2 ∂y β2 2 d2 f β df f 2d f + 2 + α =0 + dξ 2 α dξ α2 dξ 2 or (α2 + β 2 ) Therefore, Letting p = d2 f β df f +2 + 2 = 0. dξ 2 α dξ α 2β d2 f df f + + 2 2 = 0. dξ 2 α(α2 + β 2 ) dξ α (α + β 2 ) d2 f df β 1 q , the last equation reduces to + 2p + 2 f = 0. and q = 2 2 2 2 2 α(α + β ) α +β dξ dξ α q (b). The auxiliary equation associated with Equation (8.2.11) is r2 + 2pr + 2 = 0. The quadratic formula α yields the roots q r = −p ± p2 − 2 . α Therefore, 1/2 β α2 + β 2 β2 r=− − 2 2 . ± α(α2 + β 2 ) α2 (α2 + β 2 )2 α (α + β 2 )2 Hence, r= Therefore, −β ± iα = −p ± iq. α(α2 + β 2 ) f (ξ) = e−pξ [A sin qξ + B cos qξ] . Since u(x, y) = ex/α f (ξ), we conclude that u(x, y) = ex/α e−pξ (A sin qξ + B cos qξ) = ex/α−pξ (A sin qξ + B cos qξ). 43. The auxiliary equation is r2 + a1 r + a2 = 0. (a). Depending on whether r1 = r2 , we know that the solution to the differential equation takes one of these two forms: y(x) = c1 er1 x + c2 er2 x or y(x) = er1 x (c1 + c2 x). In order for lim y(x) = 0, we must have that the roots are negative: r1 < 0 and r2 < 0. x→+∞ (b). Complex roots guarantee a solution of the form y(x) = eax (c1 cos bx + c2 sin bx). In order for lim y(x) = 0, we must have a < 0. Therefore, the complex conjugate roots must have negative real x→+∞ part. (c). By the quadratic formula, the roots of the auxiliary equation are −a1 ± a21 − 4a2 r= . 2 (c)2017 Pearson Education. Inc. 553 Case 1: If a21 − 4a2 = 0, then r = − a21 < 0 is a double root and y(x) = erx (c1 + c2 x) goes to zero as x → ∞ (see part (a)). √ −a ± a2 −4a2 Case 2: If a21 − 4a2 > 0, then both roots 1 2 1 are negative; thus, the solution y(x) = c1 er1 x + c2 er2 x goes to zero as x → ∞ (see part (a)). Case 3: If a21 − 4a2 < 0, then the roots of the auxiliary polynomial are complex conjugates with negative real part. Therefore, the conclusion follows from part (b). (d). In this case, the auxiliary equation is r2 + a1 r = 0, with roots r = 0 and r = −a1 . The general solution to the differential equation in this case is y(x) = c1 + c2 e−a1 x . Thus, lim y(x) = c1 , a constant. x→∞ √ √ = ± a2 i. The general (e). In this case, the auxiliary equation is r2 + a2 = 0, with roots √ r = ± −a2√ solution to the differential equation in this case is y(x) = c1 cos a2 x + c2 sin a2 x , which is clearly bounded for all x ∈ R, since the sine and cosine functions are both bounded between ±1 in value. 44. All the roots must have a negative real part. 45. Multiplication of polynomial differential operators is identical to the multiplication of polynomials in general, an operation that is widely known to be commutative: P (D)Q(D) = Q(D)P (D). 46. This is the special case of Problem 47 for m = 1. Therefore, the solution to Problem 47 can be applied to solve this exercise as well. 47. The given equations are linearly independent if and only if m−1 , ck xk eax cos bx + k=0 m , dk xk eax sin bx = 0 =⇒ ck = 0 and dk = 0 k=0 for all k ∈ {0, 1, 2, ..., m}. If we let P (x) = m , ck xk and Q(x) = k=0 m , dk xk k=0 then the first equation can be written as eax [P (x) cos bx + Q(x) sin bx] = 0, which is equivalent to P (x) cos bx + Q(x) sin bx = 0. nπ nπ Now x = where n is an integer implies cos bx = ±1 and sin bx = 0 so P = 0 for all integers b b (2n + 1)π , where n is an integer, then cos bx = 0 and sin bx = ±1 which implies that n. Also if x = b (2n + 1)π = 0 for all integers n. This in turn implies that P (x) and Q(x) are identically zero so ck = 0 Q b and dk = 0 for all k ∈ {0, 1, 2, ..., m}. Thus the given functions are linearly independent. 48. y(x) = c1 e−5x + c2 e−7x + c3 e19x . 49. y(x) = c1 e4x + c2 e−4x + e−2x (c3 cos 3x + c4 sin 3x). (c)2017 Pearson Education. Inc. 554 √ √ √ √ 50. y(x) = e−5x/2 [c1 cos ( 11x/2) + c2 sin ( 11x/2)] + e−3x/2 [c3 cos ( 7x/2) + c4 sin ( 7x/2)]. 51. y(x) = c1 e−4x + c2 cos 5x + c3 sin 5x + x(c4 cos 5x + c5 sin 5x). 52. y(x) = c1 e−3x + c2 cos x + c3 sin x + x(c4 cos x + c5 sin x) + x2 (c6 cos x + c7 sin x). Solutions to Section 8.3 True-False Review: (a): FALSE. Under the given assumptions, we have A1 (D)F1 (x) = 0 and A2 (D)F2 (x) = 0. However, this means that (A1 (D) + A2 (D))(F1 (x) + F2 (x)) = A1 (D)F1 (x) + A1 (D)F2 (x) + A2 (D)F1 (x) + A2 (D)F2 (x) = A1 (D)F2 (x) + A2 (D)F1 (x), which is not necessarily zero. As a specific example, if A1 (D) = D − 1 and A2 (D) = D − 2, then A1 (D) annihilates F1 (x) = ex and A2 (D) annihilates F2 (x) = e2x . However, A1 (D) + A2 (D) = 2D − 3 does not annihilate ex + e2x . (b): FALSE. The annihilator of F (x) in this case is A(D) = Dk+1 , since it takes k + 1 derivatives in order to annihilate xk . (c): TRUE. We apply rule 1 in this section with k = 1, or we can compute directly that (D −a)2 annihilates xeax . (d): FALSE. Some functions cannot be annihilated by a polynomial differential operator. Only those of the forms listed in 1-4 can be annihilated. For example, F (x) = ln x does not have an annihilator. (e): FALSE. For instance, if F (x) = x, A1 (D) = A2 (D) = D, then although A1 (D)A2 (D)F (x) = 0, neither A1 (D) nor A2 (D) annihilates F (x) = x. (f ): FALSE. The annihilator of F (x) = 3 − 5x is D2 , but since r = 0 already occurs twice as a root of the auxiliary equation, the appropriate trial solution here is yp (x) = A0 x2 + A0 x3 . (g): FALSE. The annihilator of F (x) = x4 is D5 , but since r = 0 already occurs three times as a root of the auxiliary equation, the appropriate trial solution here is yp (x) = A0 x3 + A1 x4 + A2 x5 + A3 x6 + A4 x7 . (h): TRUE. The annihilator of F (x) = cos x is D2 + 1, but since r = ±i already occurs once as a complex conjugate pair of roots of the auxiliary equation, the appropriate trial solution is not yp (x) = A0 cos x + B0 sin x; we must multiply by a factor of x to occur for the fact that r = ±i is a pair of roots of the auxiliary equation. Problems: Note: In Problems 1-16, we use the four boxed formulas on pages 518-519 of the text. 1. A(D) = D + 3; (D + 3)(5e−3x ) = −15e−3x + 15e−3x = 0. 2. A(D) = D2 (D − 1); (D − 1)(2ex ) = 0 and D2 (3x) = 0 =⇒ D2 (D − 1)(2ex − 3x) = 0. 3. A(D) = (D2 + 1)(D − 2)2 ; (D2 + 1)(sin x) = 0 and (D − 2)2 (3xe2x ) = 0 =⇒ (D2 + 1)(D − 2)2 (sin x + 3xe2x ) = 0. (c)2017 Pearson Education. Inc. 555 4. A(D) = (D − 7)4 (D2 + 16); (D − 7)4 (x3 e7x ) = 0 and (D2 + 16)(5 sin 4x) = 0 =⇒ (D − 7)4 (D2 + 16)(x3 e7x + 5 sin 4x) = 0. 5. A(D) = D2 + 4D + 5; In the expression 4e−2x sin x, a = −2 and b = 1 so that D2 − 2aD + (a2 + b2 ) = D2 + 4D + 5 =⇒ (D2 + 4D + 5)(4e−2x sin x) = 0. 6. A(D) = (D2 − 2D + 5)(D2 + 4); In the expression ex sin 2x, a = 1 and b = 2, so that D2 − 2aD + (a2 + b2 ) = D2 − 2D + 5 =⇒ (D2 − 2D + 5)(ex sin 2x) = 0. Moreover, in the expression 3 cos 2x, a = 0 and b = 2, so that D2 − 2aD + (a2 + b2 ) = D2 + 4 =⇒ (D2 + 4)(3 cos 2x) = 0. Thus we conclude that (D2 − 2D + 5)(D2 + 4)(ex sin 2x + 3 cos 2x) = 0. 7. A(D) = D3 (D − 4)2 ; Note that D3 (2x2 ) = 0 and (D − 4)2 [(1 − 3x)e4x ] = 0 =⇒ D3 (D − 4)2 ((1 − 3x)e4x + 2x2 ) = 0. 8. A(D) = (D2 − 10D + 26)3 ; In the expression e5x cos x, a = 5 and b = 1 so that D2 − 2aD + (a2 + b2 ) = D2 − 10D + 26 =⇒ (D2 − 10D + 26)3 (e5x (2 − x) cos x) = 0. 9. A(D) = D2 + 6D + 10; In the expression 2e−3x sin x + 7e−3x cos x, a = −3 and b = 1 so that D2 − 2aD + (a2 + b2 ) = D2 + 6D + 10 =⇒ (D2 + 6D + 10)(e−3x (2 sin x + 7 cos x)) = 0. 10. A(D) = (D − 4)2 (D2 − 8D + 41)D2 (D2 + 4D + 5)3 ; Note that (D−4)2 (xe4x ) = 0, (D2 −8D+41)(−2e4x sin 5x) = 0, D2 (3x) = 0, and (D2 +4D+5)3 (x2 e−2x cos x) = 0. Therefore, (D − 4)2 (D2 − 8D + 41)D2 (D2 + 4D + 5)3 (e4x (x − 2 sin 5x) + 3x − x2 e−2x cos x) = 0. 11. A(D) = (D2 + 1)3 ; Here we have applied Rule 3 (page 474) with a = 0, b = 1, and k = 2. 12. A(D) = (D2 + 9)2 ; Here we have applied Rule 3 (page 474) with a = 0, b = 3, and k = 1. 1 + cos 2x 1 13. Following the hint, we write cos2 x = . We use A1 (D) = D to annihilate and we use 2 2 1 A2 (D) = D2 + 4 to annihilate (cos 2x). Hence, A(D) = A1 (D)A2 (D) = D3 + 4D. 2 14. First we simplify 4 2 2 sin x = (sin x) = 1 − cos 2x 2 2 1 1 + cos 4x 3 − 4 cos 2x + cos 4x 1 − 2 cos 2x + = = . 4 2 8 3 1 We use A1 (D) = D to annihilate , A2 (D) = D2 + 4 to annihilate cos 2x, and A3 (D) = D2 + 16 to 8 2 1 2 annihilate cos 4x. Hence, A(D) = A1 (D)A2 (D)A3 (D) = D(D + 4)(D2 + 16) = D5 + 20D3 + 64D. 8 15. To find the annihilator of F (x), let us use the identities cos2 x = 1 + cos 2x 2 and sin2 x = 1 − cos 2x 2 and (c)2017 Pearson Education. Inc. sin 2x = 2 sin x cos x 556 to rewrite the formula for F (x): 1 sin x cos x(1 + cos 2x) 2 1 = (sin 2x)(1 + cos 2x) 4 1 1 = sin 2x + (sin 2x)(cos 2x) 4 4 1 1 = sin 2x + sin 4x. 4 8 F (x) = 1 1 sin 2x is A1 (D) = D2 +4, and the annihilator of sin 4x is A2 (D) = D2 +16. Therefore, 4 8 the annihilator for F (x) is A(D) = A1 (D)A2 (D) = (D2 + 4)(D2 + 16) = D4 + 20D2 + 64. The annihilator of 16. To find the annihilator of F (x), let us use the identities cos2 x = 1 + cos 2x 2 and sin2 x = 1 − cos 2x 2 to rewrite the formula for F (x): F (x) = sin2 x cos2 x cos2 2x 1 − cos 2x 1 + cos 2x 1 + cos 4x = · · 2 2 2 1 2 = (1 − cos 2x)(1 + cos 4x) 8 1 = sin2 2x(1 + cos 4x) 8 1 = (1 − cos 4x)(1 + cos 4x) 16 1 = (1 − cos2 4x) 16 1 1 1 1 = sin2 4x = (1 − cos 8x) = − cos 8x. 16 32 32 32 1 1 is A1 (D) = D, and the annihilator for cos 8x is A2 (D) = D2 + 64. Therefore, the 32 32 annihilator for F (x) is A(D) = A1 (D)A2 (D) = D(D2 + 64) = D3 + 64D. The annihilator for 17. We have P (r) = (r − 1)(r + 2) =⇒ yc (x) = c1 ex + c2 e−2x . Now, A(D) = (D − 3). Operating on the given differential equation with A(D) yields the homogeneous differential equation (D − 3)(D − 1)(D − 2)y = 0, with solution y(x) = c1 ex + c2 e−2x + A0 e3x . Therefore, yp (x) = A0 e3x . Differentiating yp yields yp = 3A0 e3x , yp = 9A0 e3x . Substituting into the given differential equation and simplifying yields 10A0 = 5, so 1 1 1 that A0 = . Hence, yp (x) = e3x , and therefore y(x) = c1 ex + c2 e−2x + e3x . 2 2 2 18. We have P (r) = (r + 5)(r − 2) =⇒ yc (x) = c1 e2x + c2 e−5x . Now, A(D) = (D − 2). Operating on the given differential equation with A(D) yields the homogeneous differential equation (D − 2)2 (D + 5)(D − 2)y = 0, with solution y(x) = c1 e2x + c2 e−5x + A0 xe2x . Therefore, yp (x) = A0 xe2x . Differentiating yp yields yp = A0 e2x (2x + 1), yp = A0 e2x (4x + 4). Substituting into the given differential equation (c)2017 Pearson Education. Inc. 557 yields A0 e2x [4x + 4 + 3(2x + 1) − 10x] = 14e2x , so that A0 = 2. Hence, yp (x) = 2xe2x , and therefore y(x) = c1 e2x + c2 e−5x + 2xe2x . 19. We have P (r) = r2 + 16 =⇒ yc (x) = c1 cos 4x + c2 sin 4x. Now, A(D) = D2 + 1. Operating on the given differential equation with A(D) yields (D2 + 1)(D2 + 16)y = 0, so that y(x) = c1 cos 4x + c2 sin 4x + A0 cos x + B0 sin x. We therefore choose yp (x) = A0 cos x + B0 sin x. Substitution into the given differential 4 equation and simplification yields 15A0 cos x + 15B0 sin x = 4 cos x, so that A0 = , B0 = 0. Hence, 15 4 4 yp (x) = cos x, so that y(x) = c1 cos 4x + c2 sin 4x + cos x. 15 15 20. We have P (r) = (r − 1)2 =⇒ yc (x) = c1 ex + c2 xex . Now, A(D) = (D − 1). Operating on the given differential equation with A(D) yields the homogeneous differential equation (D − 1)3 y = 0, with solution y(x) = c1 ex +c2 xex +A0 x2 ex . Therefore, yp (x) = A0 x2 ex . Differentiating yp yields yp = A0 ex (x2 +2x), yp = A0 ex (x2 +4x+2). Substituting into the given differential equation yields A0 ex [(x2 +4x+2)−2(x2 +2x)+x2 ] = 6ex , so that A0 = 3. Hence, yp (x) = 3x2 ex , and therefore y(x) = c1 ex + c2 xex + 3x2 ex . 21. We have P (r) = (r − 2)(r + 1) =⇒ yc (x) = c1 e2x + c2 e−x . Now, A(D) = D3 . Operating on the given differential equation with A(D) yields the homogeneous differential equation D3 (D − 2)(D + 1)y = 0, with solution y(x) = c1 e2x + c2 e−x + A0 + A1 x + A2 x2 . Therefore, yp (x) = A0 + A1 x + A2 x2 . Differentiating yp yields yp = A1 +2A2 x, yp = 2A2 . Substituting into the given differential equation yields 2A2 −(A1 +2A2 x)− 2(A0 + A1 x + A2 x2 ) = 4x2 − 8x, so that 2A2 − A1 − 2A0 = 0, −2A2 − 2A1 = −8, −2A2 = 4 =⇒A0 = −5, A1 = 6, A2 = −2. Hence, yp (x) = −5 + 6x − 2x2 , and therefore y(x) = c1 e2x + c2 e−x − 5 + 6x − 2x2 . 22. We have P (r) = r2 −1 = (r−1)(r+1) =⇒ yc (x) = c1 ex +c2 e−x . Now, A(D) = (D−2)(D−3). Operating on the given differential equation with A(D) yields the homogeneous differential equation (D −2)(D −3)(D − 1)(D + 1)y = 0, with solution y(x) = c1 ex + c2 e−x + A0 e2x + B0 e3x . Therefore, yp (x) = A0 e2x + B0 e3x . Differentiating yp yields yp = 2A0 e2x + 3B0 e3x , yp = 4A0 e2x + 9B0 e3x . Substituting into the given differential equation yields 4A0 e2x + 9B0 e3x − 2A0 e2x − 3B0 e3x = 3e2x − 8e3x , so that A0 = 1, B0 = −1. Hence, yp (x) = e2x − e3x , and therefore y(x) = c1 ex + c2 e−x + e2x − e3x . 23. We have P (r) = (r + 1)(r − 3) =⇒ yc (x) = c1 e−x + c2 e3x . Now, A(D) = (D2 + 1)(D + 1). Operating on the given differential equation with A(D) yields the homogeneous differential equation (D2 + 1)(D + 1)2 (D − 3)y = 0, with solution y(x) = c1 e−x + c2 e3x + A0 xe−x + A1 cos x + B1 sin x. Therefore, yp (x) = A0 xe−x + A1 cos x + B1 sin x. Differentiating yp yields yp = A0 e−x (1 − x) − A1 sin x + B1 cos x, yp = A0 e−x (x − 2) + A1 cos x + B1 sin x. Substituting into the given differential equation yields A0 e−x (x − 2) + A1 cos x + B1 sin x − 2 [A0 e−x (1 − x) − A1 sin x + B1 cos x] − 3 (A0 xe−x + A1 cos x + B1 sin x) = 4(e−x − 2 cos x) =⇒ −4A0 e−x − 2(2A1 + B1 ) cos x + 2(A1 − 2B1 ) sin x = 4(e−x − 2 cos x) so that A0 = −1, and 8 4 8 4 A1 − 2B1 = 0, 2A1 + B1 = 4 =⇒ A1 = , B1 = . Hence, yp (x) = −xe−x + cos x + sin x, and therefore 5 5 5 5 8 4 y(x) = c1 e−x + c2 e3x − xe−x + cos x + sin x. 5 5 24. We have P (r) = r(r + 3) =⇒ yc (x) = c1 + c2 e−3x . Now, A(D) = D2 (D − 1)2 . Operating on the given differential equation with A(D) yields the homogeneous differential equation D3 (D − 1)2 (D + 3)y = 0, with solution y(x) = c1 + c2 e−3x + A1 x + A2 x2 + B1 ex + B2 xex . Therefore, yp (x) = A1 x + A2 x2 + B1 ex + B2 xex . Differentiating yp yields yp = A1 + 2A2 x + B1 ex + B2 ex (x + 1), yp = 2A2 + B1 ex + B2 ex (x + 2). Substituting into the given differential equation yields 2A2 + B1 ex + B2 ex (x + 2) + 3[A1 + 2A2 x + B1 ex + B2 ex (x + 1)] = 5x + xex , so that 2A2 + 3A1 = 0, 6A2 = 5, 4B1 + 5B2 = 0, 4B2 = 1 =⇒ A1 = −5/9, A2 = 5/6, B1 = −5/16, (c)2017 Pearson Education. Inc. 558 5 5 1 5 5 5 5 B2 = 1/4. Hence, yp (x) = − x + x2 − ex + xex , and therefore y(x) = c1 + c2 e−3x − x + x2 − ex + 9 6 16 4 9 6 16 1 x xe . 4 25. We have P (r) = r2 + 1 =⇒ yc (x) = c1 cos x + c2 sin x. Now, A(D) = D − 1. Operating on the given differential equation with A(D) yields (D − 1)(D2 + 1)y = 0, with solution y(x) = c1 cos x + c2 sin x + A0 ex . Hence, yp (x) = A0 ex . Substitution into the given differential equation yields A0 = 3, so that yp (x) = 3ex . Consequently, y(x) = c1 cos x + c2 sin x + 3ex . 26. We have P (r) = r2 +4r+4 =⇒ yc (x) = c1 e−2x +c2 xe−2x . Now, A(D) = (D+2)2 . Operating on the given differential equation with A(D) yields the homogeneous differential equation (D + 2)4 y = 0, with solution y(x) = c1 e−2x +c2 xe−2x +A0 x2 e−2x +A1 x3 e−2x . Therefore, yp (x) = A0 x2 e−2x +A1 x3 e−2x . Differentiating yp yields yp = A0 e−2x (−2x2 +2x)+A1 e−2x (−2x3 +3x2 ), yp = A0 e−2x (4x2 −8x+2)+A1 e−2x (4x3 −12x2 +6x). Substituting into the given differential equation and simplifying yields 2A0 + 6A1 x = 5x, so that A0 = 5 5 5 0, A1 = . Hence, yp (x) = x3 e−2x , and therefore y(x) = c1 e−2x + c2 xe−2x + x3 e−2x . 6 6 6 27. We have P (r) = r2 + 4 =⇒ yc (x) = c1 cos 2x + c2 sin 2x. Now, A(D) = D2 + 4. Operating on the given differential equation with A(D) yields the homogeneous differential equation (D2 + 4)2 y = 0, with solution y(x) = c1 cos 2x + c2 sin 2x + x(A0 cos 2x + B0 sin 2x). Therefore, yp (x) = x(A0 cos 2x + B0 sin 2x). Differentiating yp yields yp = (A0 cos 2x + B0 sin 2x) + x(−2A0 sin 2x + 2B0 cos 2x), yp = −4A0 sin 2x + 4B0 cos 2x − 4x(A0 cos 2x + B0 sin 2x). Substituting into the given differential equation yields −4A0 sin 2x + 4B0 cos 2x − 4x(A0 cos 2x + B0 sin 2x) + 4x(A0 cos 2x + B0 sin 2x) = 8 sin 2x, so that A0 = −2, B0 = 0. Hence, yp (x) = −2x cos 2x, and therefore y(x) = c1 cos 2x + c2 sin 2x − 2x cos 2x. 28. We have P (r) = (r − 2)(r + 1) =⇒ yc (x) = c1 e2x + c2 e−x . Now, A(D) = D − 2. Operating on the given differential equation with A(D) yields (D−2)2 (D+1) = 0 with general solution y(x) = c1 e2x +c2 e−x +A0 xe2x . We therefore choose yp (x) = A0 xe2x . Differentiating yp yields yp (x) = A0 e2x (2x+1), yp (x) = A0 e2x (4x+4). 5 5 Substituting into the given differential equation and simplifying we find A0 = . Hence, yp (x) = xe2x , and 3 3 5 2x 2x −x so y(x) = c1 e + c2 e + xe . 3 29. We have P (r) = r2 + 2r + 5 =⇒ yc (x) = e−x [c1 cos 2x + c2 sin 2x]. Now, A(D) = D2 + 4. Operating on the given differential equation with A(D) yields (D2 + 4)(D2 + 2D + 5)y = 0 with general solution y(x) = e−x (c1 cos 2x + c2 sin 2x) + A0 cos 2x + B0 sin 2x. We therefore choose yp (x) = A0 cos 2x + B0 sin 2x. Substitution into the given differential equation and simplification yields (A0 + 4B0 ) cos 2x + (−4A0 + B0 ) sin 2x = 12 3 3 sin 2x. Consequently A0 + 4B0 = 0, −4A0 + B0 = 3. This system has a solution A0 = − , B0 = . 17 17 12 12 3 3 Hence, yp (x) = − cos 2x + sin 2x, and so y(x) = e−x (c1 cos 2x + c2 sin 2x) − cos 2x + sin 2x. 17 17 17 17 30. We have P (r) = r3 + 2r2 − 5r − 6 = (r − 2)(r + 3)(r + 1) =⇒ yc (x) = c1 e2x + c2 e−3x + c3 e−x . Now, A(D) = D3 . Operating on the given differential equation with A(D) yields D3 (D − 2)(D + 3)(D + 1)y = 0, with general solution y(x) = c1 e2x + c2 e−3x + c3 e−x + A0 + A1 x + A2 x2 . We therefore choose yp (x) = A0 + A1 x + A2 x2 . Substituting into the given differential equation and simplifying yields 4A2 − 5(A1 + 2A2 x) − 6(A0 + A1 x + A2 x2 ) = 4x2 , so that 4A2 − 5A1 − 6A0 = 0, −10A2 − 6A1 = 0, −6A2 = 4. This system has a solution A0 = − (c)2017 Pearson Education. Inc. 37 , A1 = 27 559 2 37 10 37 10 10 2 2 , A2 = − . Hence, yp (x) = − + x − x2 , and so y(x) = c1 e2x + c2 e−3x + c3 e−x − + x − x2 . 9 3 27 9 3 27 9 3 31. We have P (r) = (r − 1)(r2 + 1) =⇒ yc (x) = c1 ex + c2 cos x + c3 sin x. Now, A(D) = D + 1. Operating on the given differential equation with A(D) yields (D + 1)(D − 1)(D2 + 1)y = 0, with general solution y(x) = c1 ex + c2 cos x + c3 sin x + A0 e−x . We therefore choose yp (x) = A0 e−x . Substitution into the given 9 9 differential equation and simplification yields A0 = − . Hence, yp (x) = − e−x , and so the general solution 4 4 9 −x x is y(x) = c1 e + c2 cos x + c3 sin x − e . 4 32. We have P (r) = (r + 1)3 =⇒ yc (x) = c1 e−x + c2 xe−x + c3 x2 e−x . Now, A(D) = (D − 2)(D + 1). Operating on the given differential equation with A(D) yields (D − 2)(D + 1)4 y = 0, with general solution y(x) = c1 e−x + c2 xe−x + c3 x2 e−x + A0 x3 e−x + A1 e2x . We therefore choose yp (x) = A0 x3 e−x + A1 e2x . Differentiating yp yields yp = A0 e−x (−x3 + 3x2 ) + 2A1 e2x , yp = A0 e−x (x3 − 6x2 + 6x) + 4A1 e2x , yp = A0 e−x (−x3 + 9x2 − 18x + 6) + 8A1 e2x . Substituting into the given differential equation and simplifying we find 6A0 e−x + 27A1 e2x = 2e−x + 3e2x , so 1 1 1 1 that A0 = , A1 = . Hence, yp (x) = x3 e−x + e2x , and therefore, the general solution to the differential 3 9 3 9 1 1 equation is y(x) = c1 e−x + c2 xe−x + c3 x2 e−x + x3 e−x + e2x . 3 9 33. We have P (r) = r2 + 9 =⇒ yc (x) = c1 cos 3x + c2 sin 3x. Now, A(D) = D2 + 4. Operating on the given differential equation with A(D) yields the homogeneous differential equation (D2 + 4)(D2 + 9)y = 0, with solution y(x) = c1 cos 3x + c2 sin 3x + A0 cos 2x + B0 sin 2x. Therefore, yp (x) = A0 cos 2x + B0 sin 2x. Differentiating yp yields yp = −2A0 sin 2x + 2B0 cos 2x, yp = −4(A0 cos 2x + B0 sin 2x). Substituting into the given differential equation yields −4(A0 cos 2x + B0 sin 2x) + 9(A0 cos 2x + B0 sin 2x) = 5 cos 2x, so that A0 = 1, B0 = 0. Hence, yp