MATH 304 Linear Algebra Lecture 19: Inner product spaces. Orthogonal sets. The Gram-Schmidt process. Norm The notion of norm generalizes the notion of length of a vector in Rn . Definition. Let V be a vector space. A function α : V → R is called a norm on V if it has the following properties: (i) α(x) ≥ 0, α(x) = 0 only for x = 0 (positivity) (ii) α(r x) = |r | α(x) for all r ∈ R (homogeneity) (iii) α(x + y) ≤ α(x) + α(y) (triangle inequality) Notation. The norm of a vector x ∈ V is usually denoted kxk. Different norms on V are distinguished by subscripts, e.g., kxk1 and kxk2 . Examples. V = Rn , x = (x1 , x2 , . . . , xn ) ∈ Rn . • kxk∞ = max(|x1 |, |x2 |, . . . , |xn |). • kxkp = |x1 |p + |x2 |p + · · · + |xn |p 1/p Examples. V = C [a, b], f : [a, b] → R. • kf k∞ = max |f (x)|. a≤x≤b • kf kp = Z a b |f (x)|p dx 1/p , p ≥ 1. , p ≥ 1. Normed vector space Definition. A normed vector space is a vector space endowed with a norm. The norm defines a distance function on the normed vector space: dist(x, y) = kx − yk. Then we say that a sequence x1 , x2 , . . . converges to a vector x if dist(x, xn ) → 0 as n → ∞. Also, we say that a vector x is a good approximation of a vector x0 if dist(x, x0 ) is small. Inner product The notion of inner product generalizes the notion of dot product of vectors in Rn . Definition. Let V be a vector space. A function β : V × V → R, usually denoted β(x, y) = hx, yi, is called an inner product on V if it is positive, symmetric, and bilinear. That is, if (i) hx, xi ≥ 0, hx, xi = 0 only for x = 0 (positivity) (ii) hx, yi = hy, xi (symmetry) (iii) hr x, yi = r hx, yi (homogeneity) (iv) hx + y, zi = hx, zi + hy, zi (distributive law) An inner product space is a vector space endowed with an inner product. Examples. V = Rn . • hx, yi = x · y = x1 y1 + x2 y2 + · · · + xn yn . • hx, yi = d1 x1 y1 + d2 x2 y2 + · · · + dn xn yn , where d1 , d2 , . . . , dn > 0. • hx, yi = (Dx) · (Dy), where D is an invertible n×n matrix. Remarks. (a) Invertibility of D is necessary to show that hx, xi = 0 =⇒ x = 0. (b) The second example is a particular case of the 1/2 1/2 1/2 third one when D = diag(d1 , d2 , . . . , dn ). Problem. Find an inner product on R2 such that he1 , e1 i = 2, he2 , e2 i = 3, and he1 , e2 i = −1, where e1 = (1, 0), e2 = (0, 1). Let x = (x1 , x2 ), y = (y1 , y2 ) ∈ R2 . Then x = x1 e1 + x2 e2 , y = y1 e1 + y2 e2 . Using bilinearity, we obtain hx, yi = hx1 e1 + x2 e2 , y1 e1 + y2 e2 i = x1 he1 , y1 e1 + y2 e2 i + x2 he2 , y1 e1 + y2 e2 i = x1 y1 he1 , e1 i + x1 y2 he1 , e2 i + x2 y1 he2 , e1 i + x2 y2 he2 , e2 i = 2x1 y1 − x1 y2 − x2 y1 + 3x2 y2 . Examples. V = C [a, b]. Z b f (x)g (x) dx. • hf , g i = a • hf , g i = Z b f (x)g (x)w (x) dx, a where w is bounded, piecewise continuous, and w > 0 everywhere on [a, b]. w is called the weight function. Theorem Suppose hx, yi is an inner product on a vector space V . Then hx, yi2 ≤ hx, xihy, yi for all x, y ∈ V . Proof: For any t ∈ R let vt = x + ty. Then hvt , vt i = hx, xi + 2thx, yi + t 2 hy, yi. The right-hand side is a quadratic polynomial in t (provided that y 6= 0). Since hvt , vt i ≥ 0 for all t, the discriminant D is nonpositive. But D = 4hx, yi2 − 4hx, xihy, yi. Cauchy-Schwarz Inequality: p p |hx, yi| ≤ hx, xi hy, yi. Cauchy-Schwarz Inequality: p p |hx, yi| ≤ hx, xi hy, yi. Corollary 1 |x · y| ≤ |x| |y| for all x, y ∈ Rn . Equivalently, for all xi , yi ∈ R, (x1 y1 + · · · + xn yn )2 ≤ (x12 + · · · + xn2 )(y12 + · · · + yn2 ). Corollary 2 For any f , g ∈ C [a, b], Z b 2 Z b Z 2 f (x)g (x) dx ≤ |f (x)| dx · a a a b |g (x)|2 dx. Norms induced by inner products Theorem Suppose hx, yi is anpinner product on a vector space V . Then kxk = hx, xi is a norm. Proof: Positivity is obvious. Homogeneity: p p p kr xk = hr x, r xi = r 2 hx, xi = |r | hx, xi. Triangle inequality (follows from Cauchy-Schwarz’s): kx + yk2 = hx + y, x + yi = hx, xi + hx, yi + hy, xi + hy, yi ≤ hx, xi + |hx, yi| + |hy, xi| + hy, yi ≤ kxk2 + 2kxk kyk + kyk2 = (kxk + kyk)2 . Examples. • The length of a vector in Rn , p |x| = x12 + x22 + · · · + xn2 , is the norm induced by the dot product x · y = x1 y1 + x2 y2 + · · · + xn yn . • The norm kf k2 = Z a b |f (x)|2 dx 1/2 on the vector space C [a, b] is induced by the inner product Z b f (x)g (x) dx. hf , g i = a Angle Since |hx, yi| ≤ kxk kyk, we can define the angle between nonzero vectors in any vector space with an inner product (and induced norm): hx, yi ∠(x, y) = arccos . kxk kyk Then hx, yi = kxk kyk cos ∠(x, y). In particular, vectors x and y are orthogonal (denoted x ⊥ y) if hx, yi = 0. x+y x y Pythagorean Law: x ⊥ y =⇒ kx + yk2 = kxk2 + kyk2 Proof: kx + yk2 = hx + y, x + yi = hx, xi + hx, yi + hy, xi + hy, yi = hx, xi + hy, yi = kxk2 + kyk2 . Orthogonal sets Let V be an inner product space with an inner product h·, ·i and the induced norm k · k. Definition. A nonempty set S ⊂ V of nonzero vectors is called an orthogonal set if all vectors in S are mutually orthogonal. That is, 0 ∈ / S and hx, yi = 0 for any x, y ∈ S, x 6= y. An orthogonal set S ⊂ V is called orthonormal if kxk = 1 for any x ∈ S. Remark. Vectors v1 , v2 , . . . , vk ∈ V form an orthonormal set if and only if 1 if i = j hvi , vj i = 0 if i 6= j Examples. • V = Rn , hx, yi = x · y. The standard basis e1 = (1, 0, 0, . . . , 0), e2 = (0, 1, 0, . . . , 0), . . . , en = (0, 0, 0, . . . , 1). It is an orthonormal set. • V = R3 , hx, yi = x · y. v1 = (3, 5, 4), v2 = (3, −5, 4), v3 = (4, 0, −3). v1 · v2 = 0, v1 · v3 = 0, v2 · v3 = 0, v1 · v1 = 50, v2 · v2 = 50, v3 · v3 = 25. Thus the set {v1 , v2 , v3 } is orthogonal but not orthonormal. An orthonormal set is formed by normalized vectors w1 = kvv11 k , w2 = kvv22 k , w3 = kvv33 k . • V = C [−π, π], hf , g i = Z π f (x)g (x) dx. −π f1 (x) = sin x, f2 (x) = sin 2x, . . . , fn (x) = sin nx, . . . hfm , fn i = Z π −π sin(mx) sin(nx) dx = π 0 if m = n if m = 6 n Thus the set {f1 , f2 , f3 , . . . } is orthogonal but not orthonormal. It is orthonormal with respect to a scaled inner product Z 1 π f (x)g (x) dx. hhf , g ii = π −π Orthogonality =⇒ linear independence Theorem Suppose v1 , v2 , . . . , vk are nonzero vectors that form an orthogonal set. Then v1 , v2 , . . . , vk are linearly independent. Proof: Suppose t1 v1 + t2 v2 + · · · + tk vk = 0 for some t1 , t2 , . . . , tk ∈ R. Then for any index 1 ≤ i ≤ k we have ht1 v1 + t2 v2 + · · · + tk vk , vi i = h0, vi i = 0. =⇒ t1 hv1 , vi i + t2 hv2 , vi i + · · · + tk hvk , vi i = 0 By orthogonality, ti hvi , vi i = 0 =⇒ ti = 0. Orthonormal bases Let v1 , v2 , . . . , vn be an orthonormal basis for an inner product space V . Theorem Let x = x1 v1 + x2 v2 + · · · + xn vn and y = y1 v1 + y2 v2 + · · · + yn vn , where xi , yj ∈ R. Then (i) hx, yi = x1 y1 + x2 y2 + · · · + xn yn , p (ii) kxk = x12 + x22 + · · · + xn2 . Proof: (ii) follows from (i) when y = x. + * + * n n n n X X X X yj vj xi vi , hx, yi = xi vi , yj vj = i=1 = j=1 n n XX xi yj hvi , vj i = i=1 j=1 j=1 i=1 n X i=1 xi yi . Orthogonal projection Theorem Let V be an inner product space and V0 be a finite-dimensional subspace of V . Then any vector x ∈ V is uniquely represented as x = p + o, where p ∈ V0 and o ⊥ V0 . The component p is the orthogonal projection of the vector x onto the subspace V0 . We have kok = kx − pk = min kx − vk. v∈V0 That is, the distance from x to the subspace V0 is kok. x o p V0 Let V be an inner product space. Let p be the orthogonal projection of a vector x ∈ V onto a finite-dimensional subspace V0 . If V0 is a one-dimensional subspace spanned by a hx, vi vector v then p = v. hv, vi If v1 , v2 , . . . , vn is an orthogonal basis for V0 then hx, v2 i hx, vn i hx, v1 i v1 + v2 + · · · + vn . p= hv1 , v1 i hv2 , v2 i hvn , vn i Indeed, hp, vi i = n X hx, vi i hx, vj i hvj , vi i = hvi , vi i = hx, vi i hv hv j , vj i i , vi i j=1 =⇒ hx−p, vi i = 0 =⇒ x−p ⊥ vi =⇒ x−p ⊥ V0 . The Gram-Schmidt orthogonalization process Let V be a vector space with an inner product. Suppose x1 , x2 , . . . , xn is a basis for V . Let v1 = x1 , hx2 , v1 i v1 , hv1 , v1 i hx3 , v1 i hx3 , v2 i v3 = x3 − v1 − v2 , hv1 , v1 i hv2 , v2 i ................................................. hxn , v1 i hxn , vn−1 i vn = xn − v1 − · · · − vn−1 . hv1 , v1 i hvn−1 , vn−1 i v2 = x2 − Then v1 , v2 , . . . , vn is an orthogonal basis for V . x3 v3 p3 Span(v1 , v2 ) = Span(x1 , x2 ) Any basis x1 , x2 , . . . , xn −→ Orthogonal basis v1 , v2 , . . . , vn Properties of the Gram-Schmidt process: • vk = xk − (α1 x1 + · · · + αk−1 xk−1 ), 1 ≤ k ≤ n; • the span of v1 , . . . , vk is the same as the span of x1 , . . . , xk ; • vk is orthogonal to x1 , . . . , xk−1 ; • vk = xk − pk , where pk is the orthogonal projection of the vector xk on the subspace spanned by x1 , . . . , xk−1 ; • kvk k is the distance from xk to the subspace spanned by x1 , . . . , xk−1 . Normalization Let V be a vector space with an inner product. Suppose v1 , v2 , . . . , vn is an orthogonal basis for V . v1 v2 vn Let w1 = , w2 = ,. . . , wn = . kv1 k kv2 k kvn k Then w1 , w2 , . . . , wn is an orthonormal basis for V . Theorem Any finite-dimensional vector space with an inner product has an orthonormal basis. Remark. An infinite-dimensional vector space with an inner product may or may not have an orthonormal basis. Orthogonalization / Normalization An alternative form of the Gram-Schmidt process combines orthogonalization with normalization. Suppose x1 , x2 , . . . , xn is a basis for an inner product space V . Let v1 = x1 , w1 = v1 kv1 k , v2 = x2 − hx2 , w1 iw1 , w2 = v2 kv2 k , v3 = x3 − hx3 , w1 iw1 − hx3 , w2 iw2 , w3 = v3 kv3 k , ................................................. vn = xn − hxn , w1 iw1 − · · · − hxn , wn−1 iwn−1 , wn = kvvnn k . Then w1 , w2 , . . . , wn is an orthonormal basis for V .