Lecture 3 A Review of ADI and OperatorSplitting Methods for the Solution of Initial Value Problems 1. Preliminary Remarks Operator-splitting methods have a controversial reputation. Some practitioners see them as particular timediscretization methods, while others view them as iterative methods. They are clearly both, even if in this presentation we focus on the first point of view. Incidentally, let us mention that various numerical methods and algorithms are disguised operator-splitting methods as we will see on a particular example. Their main feature is to take advantage of decomposition properties in the structure of the problem to be solved, some of these properties being obvious, while some others are not. 2. Time-Discretization of Initial Value Problems by ADI and OS Schemes Let us consider the following IVP (flow for the Dyn. Syst. community): (IVP) dφ/dt + A(φ) = 0 in (0, T), φ(0) = φ0. Operator A: V →V and can be multivalued. Suppose that A =ΣJj = 1 Aj (DP) holds. A natural question is: Can we take advantage of (DP) to solve (IVP) ? As we all know here: the answer is yes! 2.1. ADI type schemes We suppose for the moment that J = 2. A first candidate to take advantage of (DP) is provided by the Peaceman-Rachford scheme, namely (with τ > 0 a time-discretization step): (1) φ0 = φ0; for n ≥ 0, φn being known, solve (PRS) (2) (φn + ½ – φn )/(½ τ) + A1(φn + ½) + A2(φn) = 0, (3) (φn + 1 – φn + ½ )/(½ τ) + A1(φn + ½) + A2(φn +1) = 0. (PRS) is of the backward-Euler type for A1 and of the forward-Euler type for A2 on the time interval [tn, tn+½], the situation being reversed on [tn+½, tn+1]. A second candidate is the Douglas-Rachford scheme (of the predictor-corrector type); if J = 2, the DR-scheme reads as follows (1) φ0 = φ0; for n ≥ 0, φn being known, solve (DRS) (2) (φn + ½ – φn )/τ + A1(φn + ½) + A2(φn) = 0, (3) (φn + 1 – φn)/τ + A1(φn + ½) + A2(φn +1) = 0. These schemes, which have been around for about 50 years, have motivated a huge literature, either as methods to approximate the solution of time dependent problems or as iterative methods to solve linear and nonlinear steady state problems in finite or infinite dimension. In the second case a strategy of periodically variable τ is advocated (E. Wachpress and others). Further remarks are in order; among them: ● If V is an Hilbert space and A1 and A2 are monotone operators (possibly multivalued) both schemes are unconditionally stable. ● The most general convergence results in the context of monotone operators in Hilbert spaces are those of P.L. Lions-B. Mercier (1979). ● The D-R scheme can be generalized to decompositions with more than two operators (J > 2). ● The particular case A = – 2, A = A1 + A2 with A1= – ∂2/∂x12 and A2 = – ∂2/∂x22 explains the terminology ADI, but the applicability of (PRS) and (DRS) goes much beyond the above type of decomposition since Aj can be for example the subgradient of a convex functional. ● The above schemes are O(τ), at best, in general; however if A1 and A2 are both linear and commute, (PRS) is O(τ2); T. Arbogast, J. Douglas et al have found a way to make (DRS) second order accurate via a slight modification (if A1, A2 are smooth enough)*. ● (PRS) and (DRS) are not stiff A-stable; indeed, (PRS) is very close to the Crank-Nicolson scheme. ● Several French scientists have contributed to ADI; among them, J. Lieutaud, D. Gabay, P.L. Lions, E.Godlewsky, M.Schatzmann. On the Russian side important contributions by Dyakonov (who passed away recently). What about Spain ? Important contributions by E. Fernandez-Cara et al, Bermudez, Moreno, Pares,….. ● In order to improve the A – stability properties of (PRS) we introduced (in the mid-1980s) the so-called (by us) θ-scheme (the German scientists call it the Fractional Step θ-scheme). The idea is quite simple: (i) Consider θ (0, ½ ) (0 < θ < 1/3, in practice). (ii) Denote (n + )τ by tn+ . (iii) Decompose [tn, tn+1] as [tn, tn+1] = [tn, tn+θ][tn+θ, tn+1– θ][tn+1– θ, tn+1]. Use the above decomposition of the time interval [tn, tn+1] to time-discretize the initial value problem (IVP) by the following variant of the Peaceman-Rachford scheme (where θ* = 1 – 2θ): Description of the θ - Scheme (1) φ0 = φ0; for n ≥ 0, φn being known, solve (2) (φn+θ – φn )/(θτ) + A1(φn+θ) + A2(φn) = 0, (θ-scheme) (3) (φn+1–θ – φn+θ )/(θ*τ) + A1(φn+θ) + A2(φn+1–θ ) = 0, (4) (φn+1 – φn+1–θ)/(θτ) + A1(φn+1) + A2(φn+1–θ ) = 0. Some Properties of the θ-Scheme ● Properties are more problem dependent than for P-R and D-R; basically, for 1/4 < θ < 1/3 the scheme is stiff-A stable. It is only O(τ), but there are situations where θ = 1 – 1/√2 makes it O(τ2) (actually, nearly O(τ3); it would be O(τ3) if √2 = 1.5). ● Quite good to capture steady state solutions (compared to (PRS) and (DRS)). ● Very popular in Germany (R. Rannacher, S.Turek, E. Bänsch, V. Heuveline and others) for the simulation of incompressible viscous flow. Also used in various places (Australia among them) for simulating non-Newtonian fluid flow. 2.2. LIE’S, STRANG’S and MARCHUKYANENKO Schemes Co-existing with ADI, there have been another family of OperatorSplitting schemes going back (cf. Chorin, Hughes, Marsden et al, CPAM 1978) to S. Lie and widely used by CFD, Physics, Chemistry, Russian, etc,,, people. It is related to the Trotter formula in SemiGroup Theory. The basic idea is very simple: suppose that A is linear in (IVP), namely that (IVP) dφ/dt + Aφ = 0 in (0, T), φ(0) = φ0. We have then φ(t) = e – At φ0, and therefore: (FSGR) φ(t + τ) = e – Aτ φ(t), t ≥ 0. We also have (FOSR1) e – A2τ e – A1τ – e – (A 1 + A2)τ = ½ (A2A1 – A1A2)τ2 + O(τ3), and (FOSR2) e – ½A1τ e – A2τ e – ½A1τ – e – (A +A )τ = O(τ3)( = 0 if [A1, A2] = 1 2 0). Relations (FSGR) and (FOSR1) leads to the following OS scheme, known by some as the Lie’s scheme: Descriptions of the Lie’s Scheme (I) Condensed form: (1) φ0 = φ0; n ≥ 0, φn → φn+½ → φn+1 as follows: (LIE’S SCHEME) (2) φn+½ = e – A1τφn, (3) φn+1 = e – A2τφn+½ . The Lie’s scheme is exact if A1 and A2 commute. Descriptions of the Lie’s Scheme (II) Practical and Generalized Form: ( with 0 ≤ , ≤ 1, + = 1) (1) φ0 = φ0; n ≥ 0, φn → φn+½ → φn+1 as follows: dφ/dt + A1[φ, tn + (t – tn)] = 0 on (tn , tn+1), φn+½ = φ(tn+1), (2) φ(tn) = φn, dφ/dt + A2[φ, tn+ + (t – tn)] = 0 on (tn , tn+1), φn+1 = φ(tn+1). (3) φ(tn) = φn+½, Some Properties of the Lie’s Scheme ● Easily generalizable to J > 2 (when simulating particulate flow we may have J ≈ 10). ● Unconditionally stable if the Aj are monotone operators (possibly multivalued). Indeed this scheme is quite robust. ● First order accurate at most, in general. ● Different time and space discretizations can be used to discretize the subproblems (2) and (3) (including closed form solutions). ● The LIE’S scheme is asymptotically inconsistent; when applied as an iterative solver to compute a steady state solution, in general φn and φn+½ converge to different limits whose distance to the exact solution is O(τ) at best [it may happen that ½(φn + φn+½) has better convergence properties]. This makes some practitioners worry (I was one of them) if steady state solutions are required. Description of the Strang’s Scheme (I) Relations (FSGR) and (FOSR2) leads to the following OS scheme, known as the Strang’s scheme: (1) φ0 = φ0; n ≥ 0, φn → φn+½ → φ#n+½ → φn+1 as follows: dφ/dt + A1(φ, t) = 0 on (tn , tn+½), (2) φn+½ = φ(tn+½), φ(tn) = φn, dφ/dt + A2(φ, tn+ ½) = 0 on (0, τ), (3) φ(0) = φn+½, φ#n+½ = φ(τ), Description of the Strang’s Scheme (II) dφ/dt + A1(φ, t) = 0 on (tn+½, tn+1), (4) φn+1 = φ(tn+1). φ(tn+½) = φ#n+½, Properties of the Strang’s Scheme: ● Unconditionally stable if the Aj are monotone operators. ● O( 2) if the Aj are smooth enough. ● Generalizable to J > 2. ● Asymptotically inconsistent but provide steady state solutions whose distance at the exact solution is O( 2). ● The sub-problems (2)-(4) have to be solved by schemes which are themselves second accurate (at least). The following implicit Runge-Kutta scheme of order two (due to J. Cash) is well-suited to such a task: When applied to the solution of dX/dt + f(X, t) = 0 on (t0, tf), (IVP) X(t0) = X0 the scheme reads as follows: (1) X0 = X0; then for m ≥ 0 Xm+θ + θΔt f(Xm+θ, tm+θ ) = Xm (2) Xm+1 – θ = (1 – θ)/θ Xm+θ + (2θ – 1)/θ Xm (3) Xm+1+ θΔt f(Xm+1, tm+1) = Xm+1 – θ (4) The above scheme is stiff A-stable and 2nd order accurate if θ = 1 – 1/√2 (actually, almost 3rd order accurate). ● There exist (e.g., M. Schatzman) variants of the Strang’s scheme which are O( 4) but they are not unconditionally stable. Description of the Marchuk-Yanenko scheme Suppose that to implement the Lie’s scheme we discretize the subinitial value problems using just one step of the backward Euler scheme. We obtain then the following scheme: (1) φ0 = φ0; n ≥ 0, φn → φn+1/J ……→ φn+(J – 1)/J → φn+1 as follows: (2)j (φn+ j /J – φn+(j – 1)/J )/τ + Aj (φn+ j /J, tn+1) = 0, for j = 1, 2, ……., J. The above scheme is known as the Marchuk-Yanenko scheme. If the Aj are are monotone the scheme is unconditionally stable. It is O(τ) at best, robust and relatively easy to implement. 3. ADI and Augmented Lagrangian algorithms (i) Consider the following functional J : V → R; assume that J is differentiable and that (DP) J = J 1 + J 2, J1 and J2 being both differentiable. (ii) Consider now the minimization problem (MP) Minv V J (v). (iii) Suppose that (MP) has a solution denoted by u. We have then: (OC) J’(u) = J1’(u) + J2’(u) = 0. (iv) Problem (MP) is clearly equivalent to (EMP) Min {v, q} W j (v, q), with W = {{v, q}| {v, q} V × V, v – q = 0} and j (v, q) = J1(v) + J2(q). (v) Observe that {u, u} is solution of problem (EMP). (vi) Assuming that V is a real Hilbert space, we associate to (EMP) the following saddle-point problem Find {{u, p}, λ} (V × V) × V such that (SDP) Lr(u, p; μ) ≤ Lr(u, p; λ) ≤ Lr(v, q; λ), {{v, q}, μ} (V × V) × V, where r > 0 and (LA) Lr(v, q; μ) = j(v, q) + ½ r ||v – q||2 + (μ, v – q). (vii) If {{u, p}, λ} is solution of (SDP) then p = u where u is solution of (MP). (viii) To solve (SDP) we use the following Relaxation/Uzawa algorithm ((ALG2) in various related references): (1) {u – 1, λ0} is given in V × V; n ≥ 0, {un – 1, λn} → pn → un → λn + 1 as follows (2) J2’(pn) + r (pn – un – 1) – λn = 0, (3) J1’(un) + r (un – pn) + λn = 0, (4) λn + 1= λn + r (un – pn). Comparing (3) and (4) shows that λn + 1 = – J1’(un) which implies in turn that: AL ADI (a) r (pn – un – 1) + J2’(pn) + J1’(un – 1) = 0, (b) r (un – un – 1) + J2’(pn) + J1’(un ) = 0. We have recovered thus the Douglas-Rachford scheme with = 1/r. One can recover the Peaceman-Rachford scheme by updating λn a 1st time between (2) and (3). 4. A striking application. Let A be a d × d matrix, symmetric and positive definite. We denote by λ1 the smallest eigenvalue of A; we have then: (EVM) with λ1 = Minv SAv.v S = {v| v Rd, ||v|| = 1}. (EVM) is equivalent to (EVM-P) with Is(v) = Minv Rd [½ Av.v + Is(v)] 0 if v S, + ∞ if v Rd \S i.e., Is(.) is the indicator functional of S. If u solves (EVM) we have the following optimality condition: (NOC) Au + ∂IS(u) = 0, with ∂IS(u) a generalized differential of IS at u. To (NOC), we associate the following flow that we time-discretize by the MarchukYanenko Scheme; we obtain thus: u(0) = u0, (NOC-F) du/dt + Au + ∂IS(u) = 0 and then (1) u0 = u0; (M-Y) n ≥ 0; un → un+ ½ → un+1 as follows (2) (un+½ – un)/ + Aun+½ = 0, (3) (un+1– un+½)/ + ∂IS(un+1) = 0. Eq. (2) implies that: (2)’ un+½ = (I + A)–1un Observing that IS = IS, Eq. (3) implies that: un+1 = Arg max v S un+½.v , namely (3)’ un+1= un+½/||un+½||. We have reinvented thus the inverse power method with shift. 5. Another application: Bingham Flow in a Pipe Participating in the Fall of 2005 to a conference on visco-plasticity in Banff (BC), we had the pleasant surprise to discover that the ‘viscoplastic’ community was quite in favor of the Augmented Lagrangian (AL) approach for the numerical simulation of visco-plastic flow with stress yield, like Bingham’s. We are going to discuss thus the AL solution of a simple Bingham flow problem as an illustration of the above methodology: Let Ω be a bounded domain of R2; we denote by Γ the boundary of Ω. We consider then the following problem from the Calculus of Variations u H10(Ω), (BFP) J(u) J(v), v H10(Ω), with J(v) = ½μ ∫Ω|v|2dx + g ∫Ω|v|dx – C ∫Ωvdx. The idea here is to uncouple nonlinearity and derivatives; to do that we will treat v as an additional unknown q and force the relation v – q = 0 by penalty and Lagrange multiplier. To implement the above idea, we proceed as follows: (i) Introduce Q = (L2(Ω))2, W = {{v,q}| v H10(Ω), q Q, q = v }, j(v,q) = ½μ∫Ω|v|2dx + g∫Ω |q|dx – C ∫Ωvdx. (ii) Observe that (BFP) is equivalent to: (BFP-E) {u, p} W, j(u,p) j(v,q), {v,q} W. (iii) Introduce the following augmented Lagrangian Lr (with r > 0) from (H10(Ω) × Q) × Q into R: Lr ({v,q},m) = j(v,q) + ½r ∫Ω|v – q|dx + ∫Ωm.(v – q)dx and observe that if {{u,p},l} is a saddle-point of Lr over the space (H10(Ω) × Q) × Q (i.e., {{u,p},l} (H10(Ω) × Q) × Q, (SDPP) Lr ({u,p},m) Lr ({u,p},l) Lr ({v,q},l), {{v,q},m} (H10(Ω) × Q) × Q, then {u,p} solves (BFP-E) which implies that u solves (BFP) and p = u. (iii) In order to solve (BFP-E) we advocate the following algorithm (a disguised Douglas-Rachford ADI algorithm): u – 1 and l0 are given in H10(Ω) and Q. (1) For n ≥ 0, un – 1 and ln being known, solve (2) then (3) pn Q, Lr ({un – 1 , pn }, ln) Lr ({un – 1 , q }, ln), q Q, un H10(Ω), Lr ({un , pn }, ln) Lr ({v, pn }, ln), v H10(Ω), and finally (4) ln+1 = ln + r ( un – pn ). The convergence follows from, e.g., RG-Le Tallec, 1989. The sub-problems (2) and (3) are simpler that what they look like since 1/r (1 – g / |Xn(x)|) if |Xn(x)| > g, (a) pn ( x ) = 0 if |Xn(x)| g, with Xn = r un – 1 + ln. (b) Problem (3) is equivalent to un H10(Ω), (LVP) (μ + r)∫Ωun.vdx = C∫Ωvdx + ∫Ω(r pn – ln).vdx, v H10(Ω), a ‘simple’ linear problem indeed. 6. More applications Above, we have visualized the behavior of the mixture of solid spherical particles with a Newtonian incompressible viscous fluid in a rotating cylinder, the number of particles being 160, here. When the angular velocity is sufficiently large, the particles cluster in 3 sub-populations, of equal size, approximately (this reminds the formation of Bénard rolls in heated flow). The simulation relies on the Lie’s scheme, with J of the order of 10. We could have illustrated our presentation with many more results from numerical experiments related to many areas of applications. Indeed, new applications of ADI/OS are discovered almost every day (e.g., Monge-Ampère type equations) and we are not sure that the Scientific Community is fully aware of the capabilities of the ADI/OS methods. 7. Some References [1] GLOWINSKI, R. and P. LE TALLEC, Augmented Lagrangians and Operator –Splitting Methods in Nonlinear Mechanics, SIAM, Philadelphia, PA, 1990. [2] GLOWINSKI, R., Finite element methods for incompressible viscous flow. In Handbook of Numerical Analysis, Vol. IX, P.G. Ciarlet and J.L. Lions, eds., North-Holland, Amsterdam, 2003, pp. 3-1176. [3]MARCHUK, G.I., Splitting and alternating direction methods. In Handbook of Numerical Analysis, Vol. I, P.G. Ciarlet and J.L. Lions, eds., North-Holland, Amsterdam, 1990, pp. 197-462. [4] DEAN, E.J. and R.GLOWINSKI, An augmented Lagrangian approach to the numerical solution of the Dirichlet problem for the elliptic Monge-Ampère equation in two dimension, Electronic Transactions in Numerical Analysis, 22, (2006), 71-96. Additional references on ADI and OS can be found in refs. [1]-[3]. In particular, the Chapters 2 and 6 of ref. [2] are almost self-contained introductions to ADI and OS methods. A recent reference concerning inverse problems (4th order) of elliptic nature is: [5] F. DELBOS, J.CH. GILBERT, R. GLOWINSKI & D. SINOQUET, Constrained optimization in seismic reflection tomography: a GaussNewton augmented Lagrangian approach, Geophys. J. International, 164, (2006), 670-684.