Calculus of Variations Functional Euler’s equation V-1 Maximum and Minimum of Functions V-2 Maximum and Minimum of Functionals V-3 The Variational Notaion V-4 Constraints and Lagrange Multiplier Euler’s equation V-5 Functional Approximate method 1. Method of Weighted Residuals Galerkin method 2. Variational Method Kantorovich Method Raleigh-Ritz Method V-1 Maximum and Minimum of Functions Part A Functional Euler’s equation Maximum and Minimum of functions (a) If f(x) is twice continuously differentiable on [x0 , x1] Nec. Condition for a max. (min.) of f(x) at x [ x0 , x1 ] i.e. is that F x 0 Suff. Condition for a max (min.) of f(x) at x [ x0 , x1 ] are that F x 0 also F x 0 ( F x 0 ) (b) If f(x) over closed domain D. Then nec. and suff. Condition for a max. (min.) of f(x) at x0 2 f that xi x j D D x x0 are that f xi x x0 0 i = 1,2…n and also is a negative infinite . 2 (c) If f(x) on closed domain D If we want to extremize f(x) subject to the constraints gi ( x1 , , xn ) 0 i=1,2,…k ( k < n ) Ex :Find the extrema of f(x,y) subject to g(x,y) =0 (i) 1st method : by direct diff. of g dg g x dx g y dy 0 gx dy dx gy To extremize f df f x dx f y dy 0 gx ( f x f y )dx 0 gy 3 We have fx g y f y gx 0 and g 0 to find (x0,y0) which is to extremize f subject to g = 0 (ii) 2nd method : (Lagrange Multiplier) Let v( x, y, ) f ( x, y ) g ( x, y) extrema of v without any constraint extrema of f subject to g = 0 To extremize v v fx gx 0 x v fy gy 0 y fx g y f y gx 0 v g 0 We obtain the same equations to extrimizing. Where is called The Lagrange Multiplier. 4 V-2 Maximum and Minimum of Functionals 2.1 What are functionals. y ( x) y Functional is function’s function. y1 x1 y ( x) ( x) ( x) y2 x2 2.2 The simplest problem in calculus of variations. Determine y ( x) c 2 [ x , x ] such that the functional: 0 1 I y x F x, y x , y x dx x1 x0 as an extrema 2 where F c over its entire domain , subject to y(x0) = y0 , y(x1) = y1at the end points. 5 x On integrating by parts of the 2nd term [ Fy ( x, y, y) ]xx 1 since 0 d [ Fy ( x, y, y) Fy ( x, y, y)]dx 0 ------- (1) x0 dx x1 ( x0 ) ( x1 ) 0 and since ( x) is arbitrary. d Fy ( x, y, y) Fy ( x, y, y) 0 dx -------- (2) Euler’s Equation Natural B.C’s F y 0 or /and x0 F 0 y x x1 The above requirements are called national b.c’s. 6 V-3 The Variational Notation Variations Imbed u(x) in a “parameter family” of function variation of u(x) is defined as ( x, ) u ( x) ( x) the u ( x) The corresponding variation of F , δF to the order in ε is , since F F ( x y , y ) F ( x, y, y) F F y y y y and x1 I (u ) F ( x, u , u )dx G ( ) x0 Then x1 I F ( x, y, y)dx x0 x1 F ( x, y, y)dx x0 F F y y dx x0 y y x1 7 1 F d F F ydx y y x0 y dx y x x1 x0 Thus a stationary function for a functional is one for which the first variation = 0. F y For the more general cases (a) Several dependent variables. Ex : I F ( x, y, z ; y , z )dx x1 x0 Euler’s Eq. F d F ( )0 y dx y , F d F ( )0 z dx z (b) Several Independent variables. Ex : I F ( x, y, u, ux , u y )dxdy Euler’s Eq. R F F F ( ) ( )0 u x ux y u y 8 (C) High Order. x1 Ex : I F ( x, y, y, y)dx Euler’s Eq. x0 F d F d 2 F ( ) 2 ( ) 0 y dx y dx y Variables Causing more equation. Order Causing longer equation. 9 V-4 Constraints and Lagrange Multiplier Lagrange multiplier Lagrange multiplier can be used to find the extreme value of a multivariate function f subjected to the constraints. Ex : (a) Find the extreme value of I where u ( x1 ) u1 v( x1 ) v1 x1 x0 F ( x, u , v, u x , vx )dx u ( x2 ) u2 v( x2 ) v2 and subject to the constraints G ( x, u , v ) 0 ------------(1) F d F x F d F From I u v dx 0 x 2 1 v dx ux v dx vx -----------(2) 10 Because of the constraints, we don’t get two Euler’s equations. From G G G u v 0 u v Gv v u Gu So (2) Gv F d F F d F I vdx 0 x0 Gu u dx ux v dx vx G F d F G F d F 0 v u dx u x u v dx vx x1 The above equations together with (1) are to be solved for u , v. 11 (b) Simple Isoparametric Problem I To extremize x2 x1 F x, y , y dx , subject to the constraint: (1) J x2 G x, y, y dx const. x1 (2) y (x1) = y1 , y (x2) = y2 Take the variation of two-parameter family: y y y 11 x 22 x (where 1 x and 2 x are some equations which satisfy x x x x 0 ) 1 1 2 1 ) Then , I (1 , 2 ) J (1 , 2 1 x2 x1 x2 x1 2 2 2 F ( x, y 11 22 , y 11 2 2 )dx G ( x, y 11 22 , y 11 22 )dx To base on Lagrange Multiplier Method we can get : 12 I J 1 2 0 0 1 I J 1 2 0 0 2 x2 G d G F d F x1 y dx y y dx y i dx 0 i = 1,2 So the Euler equation is: d F G F G 0 y dx y G d G 0 , is arbitrary numbers. when y dx y The constraint is trivial, we can ignore . 13 Examples Euler’s equation Functional Helmholtz Equation Ex : Force vibration of a membrane. 2u 2 2u 2u c ( 2 2 ) f ( x, y, t ) 2 t x y -----(1) if the forcing function f is of the form f ( x, y, t ) P( x, y ) sin(t ) we may write the steady state disp u in the form (1) u v( x, y ) sin(t ) 2v 2v c ( 2 2 ) 2v p 0 x y 2 14 2 2v 2v 2 c ( ) v p vdxdy 0 -----(2) R x2 y 2 Consider c2 vxx vdxdy c2 vyy vdxdy c2 [(vx v) x vx vx ]dxdy c2 [(vy v) y vy vy ]dxdy R R R Note that (vx v) x vxx v vx vx V Vx i Vy j R (v y v) y v yy v v y v y Vx vx v Vy v y v n cos i sin j Vx Vy ( x v) ( y v) V = x y x y 15 V nds (v v cos v v sin )ds ( V )da x y c2 vxx vdxdy c2vyy vdxdy R R 1 2 c (vx ) 2 dxdy R 2 1 2 2 c v y v sin ds c (v y ) 2 dxdy R 2 c 2 vx v cos ds 1 2 (v 2 )dxdy R 2 R (2) 1 2 c [(vx ) 2 (v y ) 2 ]dxdy R 2 P vdxdy 0 c 2 (vx cos v y sin ) vds c 2 v vds R [ 1 c 2 (v)2 1 2v 2 Pv]dxdy 0 n 2 2 16 Hence : (i) if v f ( x, y ) is given on i.e. v 0 on then the variational problem c2 1 [ (v)2 2v 2 pv]dxdy 0 R 2 2 -----(3) v (ii) if n 0 is given on the variation problem is same as (3) (iii) if v ( s ) is given on n 1 2 1 2 [ R c 2 (v)2 2v 2 pv dxdy c 2 vdx] 0 17 Diffusion Equation Ex : Steady state Heat condition (k T ) f ( x, T ) B.C’s : in T T1 kn T q2 kn T h(T T3 ) on on on D B1 B2 B3 multiply the equation by T , and integrate over the domain D. After integrating by parts, we find the variational problem as follow. T 1 2 [ k (T ) f ( x, T )dT d D 2 T0 1 q2Td B2 with T = T1 on 2 B3 h(T T3 ) 2 d ] 0 B1 18 Poison Equation Ex : Torsion of a primatri Bar 2 2 0 in R on where is the Prandtl stress function and z G y , zy x The variation problem becomes [(D ) 4 ]dxdy 0 2 R with 0 on 19 V-5-1 Approximate Methods (I) Method of Weighted Residuals (MWR) L[u ] 0 in D +homo. b.c’s in B Assume approx. solution n u un Cii i 1 where each trial function i satisfies the b.c’s The residual Rn L[un ] In this method (MWR), Ci are chosen such that Rn is forced to be zero in an average sense. i.e. < wj, Rn > = 0, j = 1,2,…,n where wj are the weighting functions.. 20 (II) Galerkin Method wj are chosen to be the trial functions j hence the trial functions is chosen as members of a complete set of functions. Galerkin method force the residual to be zero w.r.t. an orthogonal complete set. Ex: Torsion of a square shaft 2 2 0 on x a , y a (i) one – term approximation 1 c1 ( x 2 a 2 )( y 2 a 2 ) Ri 2 1 2 2c1[( x a) 2 ( y a) 2 ] 2 1 ( x 2 a 2 )( y 2 a 2 ) 21 a From a a a R11dxdy 0 5 1 c1 8 a2 therefore 1 5 2 2 2 2 ( x a )( y a ) 2 8a the torsional rigidity D1 2G dxdy 0.1388G(2a)4 R the exact value of D is Da 0.1406G (2a) 4 the error is only -1.2% 22 (ii) two – term approximation 2 ( x 2 a 2 )( y 2 a 2 )[c1 c2 ( x 2 y 2 )] By symmetry → even functions R2 2 2 1 ( x 2 a 2 )( y 2 a 2 ) 2 ( x 2 a 2 )( y 2 a 2 )( x 2 y 2 ) From and R R21dxdy 0 R R22 dxdy 0 we obtain c1 1295 1 525 1 , c 2 2216 a 2 4432 a 2 therefore D2 2G 2 dxdy 0.1404G(2a)4 R the error is only -0.14% 23 V-5-2 Variational Methods (I) Kantorovich Method n Assuming the approximate solution as : u Ci ( xn )U i i 1 where Ui is a known function decided by b.c. condition. Ci is a unknown function decided by minimal “I”. Euler Equation of Ci Ex : The torsional problem with a functional “I”. u 2 u 2 I (u ) [( ) ( ) 4u ]dxdy a b x y a b 24 Assuming the one-term approximate solution as : u ( x, y) (b2 y 2 )C ( x) Then, I (C) a b a b {(b2 y 2 )2 [C( x)]2 4 y 2C2 ( x) 4(b2 y 2 )C( x)}dxdy Integrate by y 16 8 16 I (C) [ b5C2 b3C2 b3C]dx a 15 3 3 a Euler’s equation is C( x) 5 5 C( x ) 2b 2 2b 2 where b.c. condition is C( a ) 0 General solution is C( x) A1 cosh( 5x 5x ) A2 sinh( ) 1 2b 2b 25 where A1 1 cosh( 5a ) 2b , A2 0 and cosh( C( x) 1 cosh( 5 2 5 2 x ) b a ) b So, the one-term approximate solution is cosh( u 1 cosh( 5 2 5 2 x ) b (b 2 y 2 ) a ) b 26 (II) Raleigh-Ritz Method This is used when the exact solution is impossible or difficult to obtain. n First, we assume the approximate solution as : u CiU i i 1 Where, Ui are some approximate function which satisfy the b.c’s. Then, we can calculate extreme I . I I (c1 , Ex: Sol : From 1 0 , cn ) y xy x choose c1 ~ cn i.e. , I c1 I 0 cn y (0) y (1) 0 1 2 1 2 y xy x ydx 0 I y xy xy 0 dx 1 2 2 27 Assuming that y x 1 x c1 c2 x c3 x 2 (1)One-term approx y c1 x 1 x c1 x x 2 y c1 1 2 x x 2 2 1 2 2 3 4 2 I c c 1 4 x 4 x c x 2 x x c x x x dx Then, 1 0 1 1 1 2 2 c12 4 c12 1 2 1 1 1 19 c 2 c1 1 2 c1 1 120 12 2 3 2 4 5 6 3 4 I 19 1 0 c1 0 c1 0.263 y (1) 0.263x(1 x) c1 60 12 1 (2)Two-term approx y x(1 x)(c1 c2 x) c1 ( x x 2 ) c2 ( x 2 x3 ) 28 Then y c1 (1 2 x) c2 (2 x 3 x 2 ) 1 2 I (c1 , c2 ) [ c1 1 4 x 4 x 2 2c1c2 2 x 7 x 2 6 x 3 0 2 1 2 3 2 2 3 4 c3 (4 x 12 x 9 x ) c1 x 2 x 4 x5 2c1c2 x 4 2 x5 x 6 2 1 c2 2 ( x5 2 x6 x 7 ) c1 x 2 x3 c2 x3 x 4 ]dx c12 4 1 2 1 7 3 1 1 1 1 2 c1c2 1 2 3 4 5 6 3 2 5 3 7 c12 4 9 1 2 9 c1 c2 3 2 3 5 6 7 8 12 20 19 2 11 107 2 c1 c2 c1 c1c2 c2 120 70 1680 12 20 29 I 0 c1 I 0 c2 19 11 1 c1 c2 60 70 12 11 109 1 c1 c2 70 840 20 0.317 c1 + 0.127 c2 = 0.05 c1 = 0.177 , c2 = 0.173 y(2) (0.177 x 0.173x 2 )(1 x) ( It is noted that the deviation between the successive approxs. y(1) and y(2) is found to be smaller in magnitude than 0.009 over (0,1) ) 30