5 Change of Basis In many applications, we may need to switch between two or more different bases for a vector space. So it would be helpful to have formulas for converting the components of a vector with respect to one basis into the corresponding components of the vector (or matrix of the operator) with respect to the other basis. The theory and tools for quickly determining these “change of basis formulas” will be developed in these notes. 5.1 Unitary and Orthogonal Matrices Definitions Unitary and orthogonal matrices will naturally arise in change of basis formulas. They are defined as follows: A is a unitary matrix ⇐⇒ A is an invertible matrix with A−1 = A† , and A is an orthogonal matrix ⇐⇒ A is an invertible real matrix with A−1 = AT ⇐⇒ A is a real unitary matrix . Because an orthogonal matrix is simply a unitary matrix with real-valued entries, we will mainly consider unitary matrices (keeping in mind that anything derived for unitary matrices will also hold for orthogonal matrices after replacing A† with AT ). The basic test for determining if a square matrix is unitary is to simply compute A† and see if it is the inverse of A ; that is, see if AA† = I !◮Example 5.1: Let 3 5 A = 4 i 5 Then 3 5 A† = 4 − i 5 9/22/2013 . 4 i 5 3 5 4 5 − i 3 5 Chapter & Page: 5–2 and Change of Basis 3 5 AA† = 4 i 5 So A is unitary. 4 3 i 5 5 3 4 − i 5 5 4 5 − i 3 5 = ··· = I . Obviously, for a matrix to be unitary, it must be square. It should also be fairly clear that, if A is a unitary (or orthogonal) matrix, then so are A∗ , AT , A† and A−1 . ?◮Exercise 5.1: and A−1 . Prove that, if A is a unitary (or orthogonal) matrix, then so are A∗ , AT , A† The term “unitary” comes from the value of the determinant. To see this, first observe that, if A is unitary then I = AA−1 = AA† . Using this and already discussed properties of determinants, we have 1 = det(I) = det(AA† ) = det(A) det(A† ) = det(A) det(A)∗ = |det(A)|2 . Thus, A is unitary H⇒ |det A| = 1 . And since there are only two real numbers which have magnitude 1 , it immediately follows that A is orthogonal H⇒ det A = ±1 . An immediate consequence of this is that if the absolute value of the determinant of a matrix is not 1 , then that matrix cannot be unitary. Why the term “orthogonal” is appropriate will become obvious later. Rows and Columns of Unitary Matrices Let u 11 u 12 u 13 · · · u 1N u 21 u 22 u 23 · · · u 2N U = .. .. .. .. .. . . . . . uN1 uN2 uN3 · · · uN N be a square matrix. By definition, then, u 11 ∗ u 21 ∗ u 12 ∗ u 22 ∗ ∗ † u 13 u 23 ∗ U = .. .. . . ∗ u 1N u 2N ∗ u N 1∗ u N 2∗ u N 3∗ .. . · · · uNN∗ ··· ··· ··· .. . . Unitary and Orthogonal Matrices Chapter & Page: 5–3 More concisely, and [U] jk = u jk Now observe: U† jk = [U]k j ∗ = uk j ∗ . U is unitary ⇐⇒ U† = U−1 ⇐⇒ U† U = I † U U jk = [I] jk ⇐⇒ ⇐⇒ X U† m X ⇐⇒ m [U]mk = δ jk “for all ( j, k)” u m j ∗ u mk = δ jk “for all ( j, k)” jm The righthand side of the last equation is simply the formula for computing the standard matrix inner product of the column matrices u1 j u 1k u u 2j 2k . . , and . . . . uNj uNk and the last line tells us that this inner product = In other words, that line states that u 11 u 21 . , .. uN1 u 12 u 22 . . . uN2 ( 1 if j =k 0 if j 6= k . u 13 u 1N u u 23 2N , .. , . . . , .. . . uN3 uNN is an orthonormal set of column matrices. But these column matrices are simply the columns of our original matrix U . Consequently, the above set of observations (starting with “ U is unitary”) reduces to A square matrix U is unitary if and only if its columns form an orthonormal set of column matrices (using the standard column matrix inner product). You can verify that a similar statement holds using the rows of U instead of its columns. ?◮Exercise 5.2: Show that a square matrix U is unitary if and only if its rows form an orthonormal set of row matrices (using the row matrix inner product). (Hints: Either consider how the above derivation would have changed if we had used UU† instead of U† U , or use the fact just derived along with the fact that U is unitary if and only if UT is unitary.) version: 9/22/2013 Chapter & Page: 5–4 Change of Basis In summary, we have just proven: Theorem 5.1 (The Big Theorem on Unitary Matrices) Let u 11 u 12 u 13 · · · u 1N u 21 u 22 u 23 · · · u 2N U = . .. .. .. .. . . . . . . uN1 uN2 uN3 · · · uN N be a square matrix. Then the following statements are equivalent (that is, if any one statement is true, then all the statements are true). 1. U is unitary. 2. The columns of U form an orthonormal set of column matrices (with respect to the usual matrix inner product). That is, ( X 1 if j = k u m j ∗ u mk = if j 6= k 0 m . 3. The rows of U form an orthonormal set of row matrices (with respect to the usual matrix inner product). That is, ( X 1 if j = k u jm ∗ u km = if j 6= k 0 m . ?◮Exercise 5.3: What is the corresponding “Big Theorem on Orthogonal Matrices”? An Important Consequence and Exercise You can now verify a result that will be important in our change of basis formulas involving orthonormal bases. ?◮Exercise 5.4: Let be a square matrix, and let u 11 u 12 u 13 · · · u 1N u 21 u 22 u 23 · · · u 2N U = . .. .. .. .. . . . . . . uN1 uN2 uN3 · · · uN N S = { e1 , e2 , . . . , e N } and B = { b1 , b2 , . . . , b N } be two sets of vectors (in some vector space) related by b1 b2 . . . b N = e1 e2 . . . e N U . Change of Basis for Vector Components: The General Case (I.e., bj = X Chapter & Page: 5–5 for ek u k j k j = 1, 2, . . . , N .) a: Show that S is orthonormal and U is a unitary matrix H⇒ B is also orthonormal . b: Show that S and B are both orthonormal sets 5.2 U is a unitary matrix . H⇒ Change of Basis for Vector Components: The General Case Given the tools and theory we’ve developed, finding and describing the “most general formulas for changing the basis of a vector space” is disgustingly easy (assuming the space is finite dimensional). So let’s assume we have a vector space V of finite dimension N and with an inner product h · | · i . Let A = { a1 , a2 , . . . , a N } and B = { b1 , b2 , . . . , b N } . be two bases for V , and define the corresponding four N × N matrices MAA , MAB by , MBB a j ak = a j bk = b j bk [MAA ] jk = [MAB ] jk [MBB ] jk and [MBA ] jk = MBA and b j ak , , , . (See the pattern?) These matrices describe how the vectors in A and B are related to each other. Two quick observations should be made about these matrices: 1. The first concerns the relation between MAB and MBA . Observe that ∗ ∗ = [MBA ]k j = MBA † jk [MAB ] jk = a j bk = bk a j . So Likewise, of course, version: 9/22/2013 MAB = MBA † . (5.1a) MBA = MAB † . (5.1b) Chapter & Page: 5–6 2. Change of Basis The second observation is that MAA and MBB greatly simplify if the bases are orthonormal. If A is orthonormal, then [MAA ] jk = a j ak = δ jk = [I] jk . So if A is orthonormal . MAA = I (5.2a) By exactly the same reasoning, it should be clear that if B is orthonormal . MBB = I (5.2b) Now let v be any vector in V and, for convenience, let us denote the components of v with respect to A and B using α j ’s and β j ’s respectively, |viA Remember, this means α1 α 2 = .. . αN v = and X k α k ak = |viB X β1 β 2 = .. . βN βk bk . . (5.3) k Our goal is to find the relations between the α j ’s and the β j ’s so that we can find one set given the other. One set of relations can be found by taking the inner product of v with each a j and using equations (5.3). Doing so: * + * + X X aj v = aj α k ak = a j βk bk k H⇒ H⇒ H⇒ k X X aj v = α k a j ak = βk a j bk k k X X aj v = a j ak α k = a j bk βk k k X X aj v = [MAA ] jk αk = [MAB ] jk βk k k The formulas in the second equation in the last line are simply formulas for the j th entry in the products of MAA and MAB with the column matrices of αk ’s and β’s . So that equation tells us that β1 α1 h a1 | v i β α h a | v i 2 2 2 = MAA . = MAB . .. . . . . . βN αN h aN | v i Change of Basis for Vector Components: The General Case Chapter & Page: 5–7 Recalling what the column matrices of αk ’s and β’s are, we see that this reduces to h a1 | v i h a | v i 2 = MAA |viA = MAB |viB . .. . h aN | v i (5.4) ?◮Exercise 5.5 (semi-optional): Using the same assumptions as were used to derive (5.4), derive that h b1 | v i h b | v i 2 = MBA |viA = MBB |viB . (5.5) . .. h bN | v i Equations (5.4) and (5.5) give the relations between the components of a vector with respect to the two different bases, as well as the relations between these components and the inner products of the vector with each basis vector. Our current main interest is in the relations between the different components (i.e., the rightmost two-thirds of (5.4) and (5.5)). For future reference, let us summarize what the above tells us in that regard. Lemma 5.2 (The Big Lemma on General Change of Bases) Let A = { a1 , a2 , . . . , a N } and B = { b1 , b2 , . . . , b N } be any two bases for an N -dimensional vector space V . Then, for any v in V , and MAA |viA = MAB |viB (5.6a) MBA |viA = MBB |viB (5.6b) where MAA , MAB , MBB and MBA are the four N × N matrices given by , , [MAA ] jk = a j ak [MAB ] jk = a j bk and . [MBB ] jk = b j bk [MBA ] jk = b j ak Naturally, the above formulas simplify considerably when the two bases are orthonormal. That will be of particular interest to us. Before going there, however, let us observe an number of “little results” that immediately follow from the above lemma and its derivation, and which apply whether or not either basis is orthonormal: 1. Solving equations (5.6a) and (5.6b) for |viB yields the change of basis formulas and |viB = MAB −1 MAA |viA |viB = MBB −1 MBA |viA version: 9/22/2013 . Chapter & Page: 5–8 2. Change of Basis Solving equations (5.6a) and (5.6b) for |viA yields the change of basis formulas |viA = MAA −1 MAB |viB and |viA = MBA −1 MBB |viB 3. . From equation (5.4) we have that |viA h a1 | v i h a | v i 2 −1 = MAA .. . h aN | v i . Don’t memorize these formulas. They are nice formulas and can save a little work when dealing with arbitrary matrices. However, since we will attempt to restrict ourselves to orthonormal (or at least orthogonal) bases, we won’t have a great need for these formulas as written. If you must memorize formulas, memorize those in the Big Lemma — they are much more easily memorized. And besides the above change of basis formulas are easily derived from the formulas in that lemma. 5.3 Change of Basis for Vector Components: When the Bases Are Orthonormal The Main Result Now consider how the formulas in the Big Lemma (lemma 5.2 on page 5–7) simplify when the bases A = { a1 , a2 , . . . , a N } and B = { b1 , b2 , . . . , b N } are orthonormal. Then, as noted on page 5–5 (equation set (5.1)), MAA = I and MBB = I . Thus, equation (5.6a) in the Big Lemma reduces to MAB |viB = |viA , (5.7a) and equation (5.6b) in the Big Lemma reduces to |viB = MBA |viA . (5.7b) Equations (5.7a) and (5.7b) are certainly nicer than the original equations in the Big Lemma, but look at what we get when they are combined: |viB = MBA |viA = MBA MAB |viB | {z } |viA That is, |viB = [MBA MAB ] |viB for each v ∈ V , Change of Basis for Vector Components: When the Bases Are Orthonormal Chapter & Page: 5–9 which is only possible if MBA MAB = I , and which, in turn, means that MBA −1 = MAB MAB −1 = MBA and . But remember (equation set (5.1) on page 5–5) that MAB = MBA † MBA = MAB † and . Combining this with the previous line gives us MBA † = MBA −1 MAB † = MAB −1 and . So MBA and its adjoint MAB are unitary matrices. What the above tells us it that all the matrices involved are easily computed via the adjoint once we know one of the matrices. We will make one more small set of observation regarding matrices MAB and MBA and the components of the vectors in basis B with respect to basis A . Remember, since A is an orthonormal basis, the k th component of b j with respect to A is given by h ak | b j i . That is, X X X bj = ak b j ak = ak ak b j = ak [MAB ]k j . k k k This is just the formula for the matrix product b1 b2 · · · b N = a1 a2 · · · a N MAB . In summary, we have the following: Theorem 5.3 (The Big Theorem on Change of Orthonormal Bases) Let V be an N -dimensional vector space with orthonormal bases A = { a1 , a2 , . . . , a N } B = { b1 , b2 , . . . , b N } and . Let MAB and MBA be a pair of N × N matrices which are adjoints of each other and which satisfy any one of the following sets of conditions: 1. [MAB ] jk = a j bk . 2. [MBA ] jk = b j ak . 3. b1 b2 · · · b N = a1 a2 · · · a N MAB . 4. a1 a2 · · · a N = b1 b2 · · · b N MBA . 5. |viA = MAB |viB for each v ∈ V . 6. |viB = MBA |viA for each v ∈ V . Then MAB and MBA are unitary matrices satisfying all of the above conditions, as well as MAB = MBA † = MBA −1 version: 9/22/2013 and MBA = MAB † = MAB −1 . Chapter & Page: 5–10 Change of Basis Finding the Matrices If you keep in mind the equation b1 b2 · · · b N = a1 a2 · · · a N MAB from the Big Theorem (theorem 5.3, above), and have the formulas for the bk ’s in terms of the a j ’s , bk = βk1 a1 + βk2 a2 + · · · + βk N a N for j = 1, 2, . . . , N , then you can obtain the matrix MAB easily by simply noting that each of these equations can be written as βk1 βk2 bk = a1 a2 · · · a N .. . . βk N Comparing the last equation with the one a few lines earlier for b1 b2 · · · b N , it should be clear that the above column matrix on the right must be the k th column in MAB . See, no computations are needed (provided the bases are orthonormal). !◮Example 5.2: basis Let V be a three-dimensional space of traditional vectors with a standard A = { i, j, k } . Let B = { b1 , b2 , b3 } where 1 b1 = √ [i + k] 2 1 , and b2 = √ [i − k] 2 b3 = j “By inspection”, it should be clear that B is also an orthonormal basis for V . Now observe that the formulas for the bk ’s can be rewritten as 1 1 and √ 2 b1 = i j k 0 1 √ 2 , √ 2 b2 = i j k 0 1 −√ 0 b3 = i j k 1 0 2 , which can be written even more consisely as b1 b2 1 √ 2 0 b3 = i j k 1 √ 2 1 √ 2 0 1 −√ 2 0 1 0 . . Traditional Rotated and Flipped Bases Chapter & Page: 5–11 As noted above, we then must have MAB 1 √ 2 = 0 1 √ 2 1 √ 2 0 1 −√ 2 0 1 0 . Multiple Change of Basis Suppose, now, we have three orthonormal bases and A = { a1 , a2 , . . . , a N } , B = { b1 , b2 , . . . , b N } C = { c1 , c2 , . . . , c N } Then, in addition to MBA and MAB we also have MAC , MBA , MCA and MCB , all defined analogously to the way we defined MBA and MAB . Applying the above theorem we see that, for every v in V , MCA |viA = |viC = MCB |viB = MCB MBA |viA Thus, MCA |viA = MCB MBA |viA for every v ∈ V . . From this, it immediately follows that MCA = MCB MBA . (5.8) Remember, this is assuming the bases are all orthonormal. 5.4 Traditional Rotated and Flipped Bases Let us briefly restrict ourselves to a two- or three-dimensional traditional vector space V , with orthonormal bases or version: 9/22/2013 A = { a1 , a2 } and B = { b1 , b2 } A = { a1 , a2 , a3 } and B = { b1 , b2 , b3 } . Chapter & Page: 5–12 Change of Basis Direction Cosines Since we are assuming a space of traditional vectors, the scalars are real numbers, the inner product is the traditional dot product, and the change of bases matrices MAB and MBA are orthogonal and are transposes of each other. In this case, . [MAB ] jk = a j bk = a j · bk = a j kbk k cos θ(a j , bk ) And because the a j ’s and bk ’s are unit vectors, this reduces to where θ jk = angle between a j and bk [MAB ] jk = cos θ jk . These are called the direction cosines relating the two bases. Two-Dimensional Case If V is two dimensional, then the possible geometric relationships between A = { a1 , a2 } and B = { b1 , b2 } are easily sketched: b2 a2 a2 b1 ψ a1 b1 or ψ a1 b2 Clearly, we can only have one of the following two situations: 1. {b1 , b2 } are the vectors that can be obtained by “rotating” {a1 , a2 } through some angle ψ , in which case the matrix relating b1 b2 to a1 a2 is a “rotation matrix” Rψ (see problem K in Homework Handout IV ). 2. {b1 , b2 } are the vectors that can be obtained by “rotating” {a1 , a2 } through some angle ψ , and then “flipping” the direction of the second vector. In this case, the matrix relating b1 b2 to a1 a2 is given by a “rotation matrix” Rψ followed by a “flip the direction of the second vector matrix”. (Equivalently, we could first rotate by a slightly different angle and then flip the direction of the first rotated vector — or do the “flipping” first to either a1 or a2 and then rotate.) or With a little thought, it should be clear that the matrices for the “flips” in the second case are simply " # " # −1 0 1 0 F1 = and F2 = . 0 1 0 −1 Traditional Rotated and Flipped Bases Chapter & Page: 5–13 Clearly, then the matrix MBA can be written as either a single rotation matrix, or as the product of a rotation matrix with one of these “flip” matrices. It’s not at all hard to show that det(any rotation matrix) = 1 and det(either flip matrix) = −1 . Consequently, at least if we are considering orthonormal bases A and B for a two-dimensional, traditional vector space, A and B are rotated images of each other ⇐⇒ det (MBA ) = +1 . Three-Dimensional Case Likewise, if A = { a1 , a2 , a3 } and B = { b1 , b2 , b3 } are any two orthonormal bases for a traditional three-dimensional space, then MBA will either be a matrix for a “rotation by some angle about some vector”, or the product of such a rotation matrix with one of the flip matrices −1 0 0 1 0 0 1 0 0 F1 = 0 1 0 , F2 = 0 −1 0 and F3 = 0 1 0 . 0 0 1 0 0 1 0 0 −1 And, again, it can be shown that (still assuming A and B are orthonormal bases) A and B are rotated images of each other ⇐⇒ det (MBA ) = +1 . Consequently, A and B are both right-handed, or are both left-handed ⇐⇒ det(MBA ) = +1 . It can be shown that the rotation matrix, itself, can be written as a product of three “simple rotation matrices” R3,γ R2,β R3,α where R j,φ corresponds to a rotation of angle φ about the j th basis vector. The angles α , β , and γ are the infamous “Euler angles”. Now turn to pages 139, 140 and 141 in Arfken, Weber and Harris and skim their discussion of rotations and Euler angles. Warnings: 1. There are subtle differences between a matrix for a rotation operator (to be discussed in the next chapter) and the change of basis matrix for “rotated bases”. Both may be called rotation matrices, but may differ in the signs of some of the corresponding entries. 2. I have not carefully checked Arken and Weber’s discussion so I will not guarantee whether their product R3,γ R2,β R3,α is actually what we are calling MBA or if the signs of some of the entries in some of the matrices need to be switched. In other words, I have not verified whether each of Arfken, Weber and Harris’s “rotation matrices” are describing the operation of rotating a vector in space, or describing a change of basis when one basis is a rotation of the other. 3. Different authors use different conventions for the Euler angles — different axes of rotation and different order of operations. Be careful, the computations based on two different conventions may well lead to two different and incompatible results. version: 9/22/2013 Chapter & Page: 5–14 5.5 Change of Basis Sidenote: Vectors Defined by a Transformation Law Suppose we have a huge collection of bases 1 2 3 m m B , B , B , ..., with Bm = bm 1 , b2 , . . . , b1 for our N -dimensional vector space V (for example, this might be the collection of all “rotations” of some favorite orthonormal basis). Then each v in B has components with respect to each of these Bm ’s , m v1 v m N X 2 |viBm = where v = v mj bmj . . . . j=1 v mN Also, for each pair of bases Bm and Bn , there will be a corresponding “change of basis” matrix MBm Bn such that for each v ∈ V |viBm = MBm Bn |viBn . (5.9) This last expression is sometimes called a “transformation law”, even though we are not really dealing with a true transform of vectors here. Now suppose that, in the course of tedious calculations and/or drinking, we obtain, corre sponding to each different basis Bm , a corresponding N -tuple of scalars w1m , w2m , . . . , wmN . The question naturally arrises as to whether w11 , w21 , . . . , w1N , w12 , w22 , . . . , w2N , w13 , w23 , . . . , w3N , ... all describe the same vector w but in the various corresponding bases. Obviously (to us at least) the answer is “yes” if and only if m n w1 w1 w m wn 2 2 . = MBm Bn . for every pair of bases Bm and Bn . (5.10) . . . . wmN wnN This is because we can set w = N X w1j b1j j=1 and use the fact that (5.10) holds to verify that the change of basis formulas (5.9) will give us, for each Bm , the originally given N -tuple of scalars w1m , w2m , . . . , wmN as the components of w with respect to that basis. This w is called the vector defined by the transformation law (5.9). “Volumes” of N -Dimensional Hyper-Parallelpipeds (Part II) 5.6 Chapter & Page: 5–15 “Volumes” of N-Dimensional Hyper-Parallelpipeds (Part II) Let us now continue our discussion from section 3.6 on “volumes of hyper-parallelepipeds”. Recollections from Section 3.6 Recall the situation and notation: In a N -dimensional Euclidean space we have a “hyperparallelepiped” P N generated by a linearly independent set of N vectors { v1 , v2 , . . . , v N } . Our goal is to find a way to compute the N -dimensional volume of this object — denoted by VN (P N ) — using the components of the vk ’s with respect to any basis B = { b1 , b2 , . . . , b N } . For our computations, we are letting VB be the “matrix of components of the v j ’s with respect to basis B ” given by VB = “ |v1 iB |v2 iB |v3 iB v1,1 v 1,2 v · · · |v N iB ” = 1,3 .. . v1,N where, for j = 1, 2, . . . , N , vj = N X v j,k bk v2,1 v2,2 v2,3 .. . v2,N v3,1 v3,2 v3,3 .. . v3,N ··· ··· ··· .. . v N ,1 v N ,2 v N ,3 .. . · · · v N ,N . k=1 Recall, also, that we derived one formula for the volume; namely, VN (P N ) = det(VN ) (5.11) where N is the orthonormal set generated from {v1 , v2 , . . . , v N } via the Gram-Schmidt procedure (see theorem 3.4 on page 3–27). General Component Formulas From lemma 5.2 on page 5–7 on general change of bases formulas, we know MN N v j N = MN B v j B . Since N is orthonormal, however, MN N is just the identity matrix, and the above reduces to version: 9/22/2013 v j v j = M N B N B . Chapter & Page: 5–16 Thus, Change of Basis VN = |v1 iN |v2 iN |v3 iN · · · |v N iN = MN B |v1 iB |v2 iB |v3 iB · · · |v N iB = MN B VB . Combining this with formula (5.11) and a property of determinants (equation (4.3) on page 4–11) then yields VN (P N ) = det(VN ) = det (MN B VB ) = (det MN B ) (det VB ) . Let us simplify matters a little more. Let U be any orthonormal basis. As noted on page 5–11 (equation (5.8)), MN B = MN U MUB . But, since U and N are orthonormal bases of “traditional” vectors, MN U must be an orthogonal matrix (and, thus, have det(MN U ) = ±1 ). Consequently, det (MN B ) = det (MN U MUB ) = det (MN U ) det (MUB ) = ± det (MUB ) . Combining this with our last formula for VN (P N ) , gives us our most general component formula for VN (P N ) (and the next theorem). Theorem 5.4 (general component formula for the volume of a hyper-parallelepiped) Let P N be the hyper-parallelepiped in an generated by a linearly independent set {v1 , v2 , . . . , v N } of traditional vectors. Also, using any basis B for the space spanned by the vk ’s , let VB be the matrix of components of the v j ’s with respect to basis B , VB v1,1 v 1,2 v = 1,3 .. . v1,N v2,1 v2,2 v2,3 .. . v2,N v3,1 v3,2 v3,3 .. . v3,N ··· ··· ··· .. . v N ,1 v N ,2 v N ,3 .. . · · · v N ,N with vk,1 vk,2 vk,3 = |vk iB .. . . vk,N Then, letting VN (P N ) denote the N -dimensional volume of P N , VN (P N ) = |det (MUB ) det (VB )| (5.12) where U is any orthonormal basis. Special Cases B is orthonormal In particular, if B is an orthonormal basis, then we can let U = B in formula (5.12). Because of the orthonormality of B we have det (MUB ) = det (MBB ) = det (I) = 1 , “Volumes” of N -Dimensional Hyper-Parallelpipeds (Part II) Chapter & Page: 5–17 and formula (5.12) reduces to v1,1 v 1,2 v VN (P N ) = |det (VB )| = det 1,3 .. . v v2,1 v2,2 v2,3 .. . v2,N 1,N ?◮Exercise 5.6: basis v N ,1 v N ,2 v N ,3 .. . · · · v N ,N ··· ··· ··· .. . v3,1 v3,2 v3,3 .. . v3,N . (5.13) Let V be a four-dimensional space of traditional vectors with orthonormal {i, j, k, l} and let , v1 = 3i v2 = 2i + 4j , v3 = 8i − 3j + 5k and v4 = 4i + 7j − 2k + 2l . Using formula (5.13), compute the “four-dimensional volume” of the hyper-parallelepiped generated by v1 , v2 , v3 and v4 . (Compare the work done here to that done for the same problem in exercise set 3.15 on page 3–24.) B and P N are “Parallel” Let’s now assume each b j is parallel to v j and “pointing in the same direction”. That is v jk = 0 if j 6= k and vjj ≥ 0 . For “convenience”, let 1vk = vkk so that we can write v j = 1v j b j for j = 1, 2, . . . , N . Then 1v1 0 0 0 1v2 0 0 0 1v3 det (VB ) = det .. .. .. . . . 0 0 0 and formula (5.12) becomes 0 0 0 = 1v1 1v2 1v3 · · · 1v N .. . · · · 1v N ··· ··· ··· .. . VN (P N ) = |det(MUB )| 1v1 1v2 1v3 · · · 1v N where U is any orthonormal basis. version: 9/22/2013 . , (5.14) Chapter & Page: 5–18 Change of Basis Of course, if there is any othogonality, then the above simplifies. In particular, if B is orthogonal, then you can easily show that VN (P N ) = kb1 k kb2 k · · · kb N k 1v1 1v2 1v3 · · · 1v N . If you compare formula (5.14) with formulas (3.11) through (3.11), starting on page 3–28, it should be clear that we’ve already derived geometric formulas for the above |det MUB | when N = 1 , N = 2 and N = 3 . Since “cut and paste” is so easy, we’ll rewrite those geometric formulas: V1 (P1 ) = · · · = kb1 k 1v1 V2 (P2 ) = and V3 (P3 ) = q , kb1 k2 kb2 k2 − (b1 · b2 )2 1v1 1v2 √ , A − B + C 1v1 1v2 1v3 where A = kb1 k2 kb2 k2 kb3 k2 and , B = kb1 k2 (b2 · b3 )2 + kb2 k2 (b1 · b2 )2 + kb3 k2 (b1 · b2 )2 C = 2 (b1 · b2 )2 (b1 · b3 )2 (b2 · b1 )2 (b2 · b3 )2 = 2 = ··· kb1 k2 kb2 k2 , . As we noted in on page 3–29, these formulas reduce even further if the bk ’s are unit vectors, and even further if the set of bk ’s is orthonormal. If you’ve not already done so, then do the next exercise (which is the same as exercise 3.17): ?◮Exercise 5.7: To what do the above formulas for V1 (P1 ) , V2 (P2 ) and V3 (P3 ) reduce a: when the bk ’s are unit vectors. b: when the set of bk is orthonormal.