AN INTRODUCTION TO MULTIGRID METHODS VIA SUBSPACE CORRECTION - - PowerPoint PPT Presentation

an introduction to multigrid methods via subspace
SMART_READER_LITE
LIVE PREVIEW

AN INTRODUCTION TO MULTIGRID METHODS VIA SUBSPACE CORRECTION - - PowerPoint PPT Presentation

AN INTRODUCTION TO MULTIGRID METHODS VIA SUBSPACE CORRECTION FRAMEWORK LUDMIL ZIKATANOV, DEPARTMENT OF MATHEMATICS, PENN STATE A BSTRACT . We give a simple approach via subspace correction framework on proving the convergence of multigrid and


slide-1
SLIDE 1

AN INTRODUCTION TO MULTIGRID METHODS VIA SUBSPACE CORRECTION FRAMEWORK

LUDMIL ZIKATANOV, DEPARTMENT OF MATHEMATICS, PENN STATE

  • ABSTRACT. We give a simple approach via subspace correction framework on proving the convergence of multigrid and overlapping Schwarz
  • methods. The list of references provided at the end gives some basic references on the subject and on the approach taken here for proving

multigrid convergence.

CONTENTS 1. Introduction 2 2. An introduction to linear and product iterative methods 2 2.1. On the choice of B 3 2.2. Error equation and error transfer operator 3 2.3. Symmetrization of B and convergence 3 2.4. Gauss-Seidel method 4 2.5. Convergence rate 4 2.6. Convergence of Gauss-Seidel (cont) 4 2.7. Convergence – continued 4 2.8. Simple application 5 2.9. Preconditioning and convergent iterations 6 2.10. Some more on linear Iterative Methods 6 2.11. Errors in terms of product of projections 6 3. Subspace corrections 6 3.1. Some idea behind these methods 6 3.2. Variational problem and some notation 7 3.3. Subspace correction method: Algorithm 7 3.4. Error equations 7 3.5. Relationship with the Method of Alternating Projections 8 3.6. The method of alternating projections 8

1

slide-2
SLIDE 2

4. Convergence of subspace correction methods 9 4.1. Convergence result (preliminaries) 9 4.2. Abstract theorem 10 4.3. Outline of the proof 10 5. How to estimate c0 12 5.1. A special decomposition 12 5.2. A multigrid method convergence proof 13 5.3. More general elliptic PDEs 15 5.4. Discretization and subspace splitting 16 5.5. Basic multigrid iteration 16 5.6. Basic facts from Finite element theory 16 5.7. Error estimates 17 5.8. On the addtive methods 17 5.9. Multiplicative overlapping DD-method 18 5.10. Convergence for finite element equations 18 5.11. Convergence via subspace correction framework 18 5.12. Further simplifications: Richardson smoother 19 5.13. Some known results 19 5.14. Stability and convergence 20 Basic references 20

  • 1. INTRODUCTION
  • 2. AN INTRODUCTION TO LINEAR AND PRODUCT ITERATIVE METHODS
  • Problem: Find u ∈ Rn, such that

(1) Au = f,

  • where A ∈ Rn×n is symmetric and positive definite and f ∈ Rn is a given right hand side.
  • Given an initial guess u0 we find the next iterates uk+1 for k = 0, 1, 2, . . . via the recurrence relation:

(2) uk+1 = uk + B(f − Auk), where B ∈ Rn×n is some given matrix (an approximate inverse of A).

2

slide-3
SLIDE 3

2.1. On the choice of B.

  • By convergence, here we mean

(3) ek := u − uk, lim

k→∞ ek = 0.

  • For A being SPD, · = · A. This makes sence, since A defines an inner product on V = Rn and also a norm by:

(u, v)A := Au, vℓ2, u2

A = u, uA.

Let us take u to be the solution to (1). Then by f − Au = 0 we get that (4) u = u + B(f − Au), 2.2. Error equation and error transfer operator.

  • Subtracting the following two equations

uk+1 = uk + B(f − Auk), u = u + B(f − Au) it follows that ek = u − uk = (I − BA)ek−1 = . . . = (I − BA)ke0.

  • A trivial manipulation then gives that a necessary and sufficient condition for convergence: ρ(E) < 1, with E := I − BA.
  • But since we would like to have some idea how fast the convergence is, and we also have that A is SPD, we shall use:

ekA = (I − BA)ke0A ≤ I − BAk

Ae0A.

  • Then everything amounts to estimating norm of G := I − BA. that is for any v ∈ V := Rn we would like to estimate:

(5) v2

A − Gv2 A = ¯

BAv, vA, where (6) ¯ B := Bt + B − BtAB. 2.3. Symmetrization of B and convergence.

  • By the following identity:

Gv2

A = v2 A − ¯

BAv, vA, and we will have convergence, as long as ¯ BA, is positive definite (note that this matrix is symmetric in ·, ·A inner product. ( ¯ BAv, w)A = (A ¯ BAv, w) = (Av, ¯ BAw) = (v, ¯ BAw)A

  • We now take a particular choice of B Gauss-Seidel method and try to estimate the convergence rate

3

slide-4
SLIDE 4

2.4. Gauss-Seidel method.

  • For Gauss-Seidel method we have B := (D − L)−1, where D is the diagonal of A, and −L is the strict lower triangular part of

A.

  • This means nothing else but that A is written as A = D − L − Lt.
  • Let us compute ¯

B for this choice of B: ¯ B := Bt(B−1 + B−t − A)B = Bt(D − L + D − Lt − D + L + Lt)B = BtDB.

  • Hence ¯

B is symmetric and positive definite. and as excersize one can prove that GA < 1, just based on this fact. 2.5. Convergence rate.

  • As we mentioned before, we are interested in an estimate of

G2

A = I − BA2 A = sup v=0

(I − BA)2

A

v2

A

  • Equation (7)gives that ¯

B is invertible. Let us compute ¯ B−1: ¯ B−1 = B−1D−1B−T = (D − L)D−1(D − LT) = A + LD−1LT. 2.6. Convergence of Gauss-Seidel (cont).

  • Hence yet another expression for ¯

B is below: ¯ B = (A + LD−1LT)−1.

  • From before

(I − BA)v2

A = v2 A − ¯

BAv, vA. 2.7. Convergence – continued.

  • After we divide on v2

A and take the sup we obtain that:

(I − BA)2

A = 1 − inf v=0

¯ BAv, vA v, vA = 1 − inf

v=0

A ¯ BAv, v Av, v

  • The expression we have for ¯

B and change of variables w = A1/2v then can be used to obtain that: (I − BA)2

A = 1 − inf w=0

A1/2(A + LD−1LT)−1A1/2w, w w, w .

4

slide-5
SLIDE 5
  • We use now that for any SPD X λmin(X) = 1/λmax(X−1) and change back v = A−1/2w to get that:

(I − BA)2

A = 1 −

1 1 + supv=0

LD−1LT v,v v,vA

.

  • Finally if we denote c0 = supv=0

LD−1LT v,v v,vA

we get that G2

A = 1 −

1 1 + c0 , and that is exactly the contraction number in A norm of the Gauss-Seidel iteration. In applications one needs to give an upper bound on c0 in order to get a convergence result. 2.8. Simple application.

  • Consider A = −D2

h = 1 hdiag(−1, 2, −1). Because L and D look in a very special way, we have that:

LD−1LTv, v = 1 h2

N−1

  • i=2

v2

i .

  • Therefore for c0 we obtain:

c0 = sup

v=0

N−1

i=2 v2 i

h2v, vA .

  • c0 = sup

v=0

N−1

i=2 v2 i

h2v, vA

  • The supremum will be achieved for v being the eigenvector corresponding to a minimal eigenvalue of −h2D2

h, just the left

boundary point being shifted one h to the right. Hence: c0 = 1 4 sin2

1 2(N−1)

.

  • From here is easy to get that G2

A = 1 − ch2, just by using that for x ∈ [0, π/2] we have that 2 πx ≤ sin x ≤ x, we have that:

c0 ∼ N 2 which gives the desired result.

  • For −∆h the same considerations lead to similar estimate.

5

slide-6
SLIDE 6

2.9. Preconditioning and convergent iterations.

  • There is a difference in the terminology: preconditioning and convergent iterations.
  • A good preconditioner for A is a matrix B for which κ(BA) is uniformly bounded. This does not mean that ρ(I − BA) < 1!
  • A good iterator is a matrix B such that I − BA ≤ δ < 1 for some norm · . Note that if δ is uniformly bounded, then for

any eigelue λ(BA) we have |λ(I − BA)| < 1 ⇒ 1 − δ < |λ(BA)| < 1 + δ and hence B is a good preconditioner.

  • This implication can not be reversed. An example is

A =   2 1 1 1 2 1 1 1 2   , B−1 = diag(A) = 2I.

2.10. Some more on linear Iterative Methods. x ← x + Br; r := Ae = b − Ax.

  • In Jacobi B = D−1.
  • In Gauss Seidel B = (D − L)−1.
  • In SOR B = [ω(D − L)]−1, where ω is selected to minimize ρ(I − BA).

2.11. Errors in terms of product of projections.

  • If we denote with Pk the A-orthogonal projection on span(ek), where φk is the k-th basis vector:

Pkv = αk(v)φk, αk = (Aφk, v)/φk2

A,

then ⋆ Jacobi method is an example of additive preconditioner: B = Pk ⋆ Gaus-Seidel and SOR method are product iterative methods, with error transfer operators given by (for Gauss-Seidel ω = 1): ek = Eek, E = (I − ωP1) . . . (I − ωPn).

  • 3. SUBSPACE CORRECTIONS

3.1. Some idea behind these methods.

  • The method of subspace corrections is Divide and Conquer strategy for the solution of a linear equation in a Hilbert space:

⋆ Split the space into sum of subspaces. ⋆ Solve smaller problems on each of the subspaces. ⋆ “Glue up” the solutions of the subspace problems together to get an approximation of the solution in the entire space.

  • Examples: Gauss-Seidel method, Multigrid method, Schwarz alternating method, Domain decomposition methods ...(just to name a few).

6

slide-7
SLIDE 7

3.2. Variational problem and some notation.

  • Variational problem: Find u ∈ V such that

a(u, v) = f(v), ∀v ∈ V.

  • Hilbert space V , and a(·, ·) is symmetric positive definite and continuous bilinear form (defines a scalar product and a norm on V ).
  • Space decomposition: V =

i Vi

  • Approximate subspace problems: ai ≈ a on Vi × Vi and Ti ≈ Pi:

ai(Tiv, vi) = a(v, vi), v ∈ V, vi ∈ Vi. with Ti = Pi if ai = a. 3.3. Subspace correction method: Algorithm.

  • Space decomposition: V =

i Vi

  • Algorithm MSC:

Let u0 ∈ V be given. for ℓ = 1, 2, . . . uℓ−1 = uℓ−1. for i = 1 : J Let ei ∈ Vi solve ai(ei, vi) = f(vi) − a(uℓ−1

i−1, vi)

∀vi ∈ Vi. uℓ−1

i

= uℓ−1

i−1 + ei

endfor uℓ = uℓ−1

J

. endfor 3.4. Error equations.

  • Error transfer operator: u − uℓ−1

i

= (I − Ti)(u − uℓ−1

i−1) and

u − uℓ = E(u − uℓ−1) = . . . = Eℓ(u − u0) where E = (I − TJ)(I − TJ−1) . . . (I − T1).

  • And fpr v ∈ V , Tiv ∈ Vi is defined as the unique solution of

ai([Tiv], vi) = a(v, vi) ∀vi ∈ Vi.

  • Convergence:E < 1?

7

slide-8
SLIDE 8

3.5. Relationship with the Method of Alternating Projections.

  • When all ai(·, ·) = a(·, ·) then Ti = Pi are orthogonal projections and

E = (I − PJ)(I − PJ−1) . . . (I − P1).

  • The equation above relates the MSC and Method of Alternating Projections (estimate of the norm of product of projections).

3.6. The method of alternating projections. Let H be a Hilbert space and Mi ⊂ H be closed subspaces

  • On the product of two projections:

PM2PM1 is a projection if and only if PM2PM1 = PM1PM2, furthermore PM2PM1 = PM1∩M2.

  • A result of von Neumann (1933):

For general closed subspace M1 and M2 lim

k→∞(PM2PM1)k = PM1∩M2.

✑ ✑ ✑ ✑ ✑ ✑ ✑ ✑ ✑ ✑ ✑ ✑ ✑ ✑ ✑ ✰ ✑✑✑✑✑✑✑✑✑✑✑✑✑✑✑ ✸

M1

✛ ✲

M2

w0 = w

❏ ❏ ❏ ❏ ❏ ❏✇

w1

w2

❏ ❏ ❏ ❏ ❏ ❏ ✇

w3

w4

❏ ❏ ❏ ❏ ✇

w5

w6

❏ ❏ ❏ ✇

w7

w8

❏ ❏ ✇

w9

8

slide-9
SLIDE 9
  • Rate of convergence:

(PM2PM1)k − PM1∩M2 = [c(M1, M2)]2k−1 where c(M1, M2) is the cosine of the angle between M1 and M2: c(M1, M2) = sup |(u, v)| uv : u ∈ M1 ∩ (M1 ∩ M2)⊥, v ∈ M2 ∩ (M1 ∩ M2)⊥

  • Fact: c(M1, M2) < 1 if and only if M1 + M2 is closed.

⇒ uniform convergence.

  • 4. CONVERGENCE OF SUBSPACE CORRECTION METHODS

4.1. Convergence result (preliminaries).

  • Assumption on subspace solvers: Tiv2 ≤ ω(Tiv, v),

v ∈ V for some constant ω ∈ (0, 2). LEMMA. Assume that Ti satisfies the above assumption and that Range(Ti) = Vi. Then (1) I − Ti is nonexpansive. (2) Ti, T ∗

i and ¯

Ti have the same kernels: N( ¯ T i) = N(Ti) = N(T ∗

i ).

(3) Ti, T ∗

i and ¯

Ti have the same range: R( ¯ T i) = R(Ti) = R(T ∗

i ) = Vi.

(4) The following inequality holds: 2 − ω ω Tiv2 ≤ v2 − (I − Ti)v2 = ( ¯ Tiv, v). (5) As operators restricted on Vi, the above (1)–(3) are still valid. (6) Ti, T ∗

i and ¯

Ti are all isomorphisms from Vi to itself. (7) ¯ Ti is nonnegative on V and symmetric positive definite on Vi. PROOF. It is obvious that (1) holds. By (1) I − Ti ≤ 1, namely Tiv2 ≤ 2(Tiv, v) = 2(v, T ∗

i ), we see immediately N(T ∗ i ) ⊂ N(Ti). Similarly

N(T ∗

i ) ⊂ N(Ti) since I − T ∗ i = I − Ti. Hence N(Ti) = N(T ∗ i ). It follows from the definition of ¯

Ti that N(Ti) ⊂ N( ¯ T i). Note that by the assumptions on the subspace solvers ( ¯ Tiv, v) = v2 − (I − Ti)v2 = 2(Tiv, v) − Tiv2 ≥ 2 − ω ω Tiv2 which implies N(Ti) ⊃ N( ¯ T i). This completes the proof of (2). (3) follows from (2) by the well-know relations between kernel and range of

  • perators. (4) is already contained in the equation above. (5)–(7) are obvious.

9

slide-10
SLIDE 10

4.2. Abstract theorem. Theorem (Xu and Z. 2003, J. of AMS) E2 ≡ (I − TJ)(I − TJ−1) . . . (I − T1)2 = c0 1 + c0 where ¯ Ti = T ∗

i + Ti − T ∗ i Ti and

c0 = sup

v=1

inf

  • i vi=v

J

  • i=1

(Ti ¯ T −1

i

T ∗

i wi, wi) with wi = J

  • j=i

vj − T −1

i

vi. Assumption: Tiv2 ≤ ω(Tiv, v), ω ∈ (0, 2) = ⇒ c0 > 0. If Ti = Pi, then c0 = sup

v=1

inf

  • i vi=v

J

  • i=1

Piwi2

a with wi = J

  • j=i+1

vj. 4.3. Outline of the proof. The proof is based on a sequence of identities (1) Let E0 = I and Ei = (I − Ti)Ei−1 for i = 1 : J, then v2 − Ev2 =

J

  • i=1

( ¯ T iEi−1v, Ei−1v) = (I

  • ∗(L
  • ∗)−1 ¯

T

  • L
  • −1I

v, v)

where ¯ T

  • = diag( ¯

T1, ¯ T2, . . . , ¯ TJ), I

=

     I I . . . I      , E

=

     I E1 . . . EJ−1      , L

  • =

       I . . . T1 I . . . T1 T2 I . . . . . . . . . . . . ... . . . T1 T2 T3 . . . I        (2) The following identity holds v2 − Ev2 = (T

(S

  • + T
  • ∗T

)−1T

  • ∗v, v)

= ⇒ E2 = 1 − inf

v=1(T (S

  • + T
  • ∗T

)−1T

  • ∗v, v).

10

slide-11
SLIDE 11

where S

  • = (T
  • ∗L
  • − ¯

T

  • ) ¯

T

  • −1(L
  • ∗T
  • − ¯

T

  • ),

T

  • ∗ =

     T ∗

1

T ∗

2

. . . T ∗

J

     : V → V . In fact S

  • =

(T

  • ∗L
  • ¯

T

  • −1 − I
  • )(L
  • ∗T
  • − ¯

T

  • )

= T

  • ∗L
  • ¯

T

  • −1L
  • ∗T
  • − T
  • ∗L
  • − L
  • ∗T
  • + ¯

T

  • =

T

  • ∗L
  • ¯

T

  • −1L
  • ∗T
  • − T
  • ∗T

,

hence I

  • ∗(L
  • ∗)−1 ¯

T

  • L
  • −1I
  • =

I

  • ∗(L
  • ¯

T

  • −1L
  • ∗)−1I
  • =

I

  • ∗T
  • (S + T
  • ∗T

)−1T

  • ∗I
  • =

T

(S + T

  • ∗T

)−1T

  • ∗.

as desired (3) One key observation: for any w ∈ V, ˜ w = (S

  • + T
  • ∗T

)−1T

  • ∗w,

v = T

˜

w: (S

  • ˜

w, ˜ φ) = 0, ∀˜ φ ∈ N(T

) =

⇒ (S

  • ˜

w, ˜ w) = inf

T

˜

v=v(S

  • ˜

v, ˜ v). Indeed, (S

  • ˜

w, ˜ φ) = ((S

  • + T
  • ∗T

− T

  • ∗T

) ˜

w, ˜ φ) = (T

  • ∗w, ˜

φ) − (T

  • ∗T

˜

w, ˜ φ) = 0 Hence for arbitrary ˜ v, we have (S

  • ˜

v, ˜ v) = (S

  • ˜

w, ˜ w) + (S

  • ˜

w − ˜ v, ˜ w − ˜ v), because ˜ w − ˜ v ∈∈ N(T

)

11

slide-12
SLIDE 12

Thus

  • inf

v=1(T (S

  • + T
  • ∗T

)−1T

  • ∗v, v)

−1 = sup

w∈V

(T

(S

  • + T
  • ∗T

)−1T

  • ∗w, w)

(T

(S

  • + T
  • ∗T

)−1T

  • ∗w, T

(S

  • + T
  • ∗T

)−1T

  • ∗w)

= sup

w∈V

inf

T

˜

v=v

(S

  • ˜

v, ˜ v) v2 + 1 = sup

v∈V

inf

T

˜

v=v

(S

  • ˜

v, ˜ v) v2 + 1 ≡ c0 + 1.

  • 5. HOW TO ESTIMATE c0
  • LEMMA. The following identity holds, for all i = 1 : J

(7) (T ∗

i −1 − I)Ti ¯

T −1

i

T ∗

i (T −1 i

− I) = ¯ T −1

i

− I. PROOF. Let Si = T ∗

i −1 + T −1 i

− I. It follows that (T ∗

i −1 − I)Ti ¯

T −1T ∗

i (T −1 i

− I) = (T ∗

i −1 − I)(T ∗ i −1 + T −1 i

− I)−1(T −1

i

− I) = (Si − T −1

i

)S−1

i

(Si − (T ∗

i )−1)

= Si − T −1

i

− (T ∗

i )−1 + T −1 i

S−1

i

(T ∗

i )−1

= −I + ¯ T −1

i

. Q.E.D. 5.1. A special decomposition. In some important applications, such as multigrid methods, a special decomposition v =

i vi may be obtained by

a sequence of linear operators. If we assume that there exist operators: Πi : V → Vi such that R(Πi − Πi−1) ⊂ Vi, ∀v ∈ V with Π0 = 0 and ΠJ = I, we have a telescopic decomposition v =

J

  • i=1

vi with vi = (Πi − Πi−1)v.

12

slide-13
SLIDE 13

Then, in view of the definition of c0, we have c0 ≤ sup

v=1 J

  • i=1

(Ti ¯ T −1

i T ∗ i wi, wi)

with (8) wi =

J

  • j=i

vj − T −1

i

vi = (I − Πi)v + (I − T −1

i

)(Πi − Πi−1)v. Furthermore c0 ≤ sup

v=1 J

  • i=1

(Ti ¯ T −1

i T ∗ i wi, wi)

≤ 2 sup

v=1 J

  • i=1
  • (Ti ¯

T −1

i T ∗ i (I − Πi)v, (I − Πi)v)

+(( ¯ T −1

i

− I)(Πi − Πi−1)v, (Πi − Πi−1)v)

  • .

An important special case is when Πi = Pi. In this case, the expression (8) is reduced to wi =

J

  • j=i

(I − T −1

i

)(Pi − Pi−1)v and (9) c0 ≤ sup

v=1 J

  • i=1

(( ¯ T −1

i

− I)(Pi − Pi−1)v, (Pi − Pi−1)v)

  • .

5.2. A multigrid method convergence proof. The estimate below estimate is very well known before and basic, but the proof here is different from those in the literature (Braess and Hackbusch and Bramble and Pasciak). We assume that Ω has been triangulated with a nested sequence of quasi-uniform triangulations Tk = {τ i

k} of size h for k = 0, . . . , j where the

quasi-uniformity constants are independent of k. These triangulations should be nested in the sense that any triangle τ l

k−1 can be written as a union

  • f triangles of {τ i

k}. We further assume that there is a constant γ > 1, independent of k, such that

hk ∼ γ−k. Associated with each Tk, a finite element space Vk ⊂ H1

0(Ω) can be defined. One has

(10) V0 ⊂ V1 ⊂ . . . ⊂ Vk ⊂ . . . ⊂ VJ.

13

slide-14
SLIDE 14

We are interested in multigrid method for solving the following finite element equation: Find uh ∈ V ≡ VJ satisfying (11) a(uh, v) = f(v) ∀v ∈ V where a(u, v) =

∇u · ∇v. Define Ai : Vi → Vi by: (Aiui, vi)0 = a(ui, vi), ∀ui, vi ∈ Vi. The subspace solver Ti in a multigrid method is often given by applying a number of, say m, smoothings using, for example, local relaxation method such as Gauss-Seidel iteration (with perhaps exception of i = 0 in which exact solver may be used, namely T0 = P0). Let Ri ≈ A−1

i

denote one such a smoothing. With m smoothings, one strategy is to apply Ri and Rt

i alternatively. Here Rt i is the adjoint of Ri with respect to (·, ·); for example, if

Ri represents forward Gauss-Seidel iteration, then Rt

i represents backward Gauss-Seidel iteration. With this kind of subspace correction, we have

Ti =

  • Pi − (K∗

i Ki)

m 2 Pi,

if m is even Pi − Ki(K∗

i Ki)

m−1 2 Pi,

if m is odd and ¯ T i = (I − Km

i,m)Pi

where Ki = I − RiAi, and K∗

i = I − Rt iAi

and Ki,m =

  • K∗

i Ki,

if m is even KiK∗

i ,

if m is odd Note that K∗

i is the adjoint of Ki with respect to a(·, ·).

In this example, we shall make two assumptions. The first assumption is on the approximation of the finite element subspaces: (12) (I − Pi−1)vi2 ≤ c1 λi a(vi, vi), ∀v ∈ Vi, where λi = ρ(Ai). The second assumption is on the smoother Ri: (13) c2 λi (v, v) ≤ ( ¯ Riv, v) ∀v ∈ Vi, where ¯ Ri is the corresponding symmetrization of Ri. We note that the first assumption is satisfied, for example, when aij are smooth and Ω is smooth or a convex Lipschitz domain and the second assumption is satisfied for Gauss-Seidel iteration. To estimate the constant c0, we consider the following decomposition v =

i vi for any v ∈ V with

(14) vi = (Pi − Pi−1)v.

14

slide-15
SLIDE 15

Since, thanks to (10), Pi−1 = Pi−1Pi = PiPi−1, (12) implies that (15) λi(vi, vi) ≤ c1a(vi, vi). We note that a( ¯ T −1

i (I − ¯

T i)vi, vi) = a((I − Km

i,m)−1Km i,mvi, vi)

= ( ¯ R−1

i

¯ RiAi(I − Km

i,m)−1Km i,mvi, vi)

≤ λi c2 ((I − Ki,m)(I − Km

i,m)−1Km i,mvi, vi)

≤ λi c2 max

t∈[0,1][(1 − tm)−1tm(1 − t)](vi, vi)

≤ λi c2m(vi, vi). Hence, by (15), we have

J

  • i=1

a( ¯ T −1

i (I − ¯

T i)vi, vi) ≤

J

  • i=1

λi c2m(vi, vi) ≤

J

  • i=1

c1 c2ma(vi, vi) = c1 c2ma(v, v). Thus, by c0 ≤ sup

v=1 J

  • i=1

(( ¯ T −1

i

− I)(Pi − Pi−1)v, (Pi − Pi−1)v)

  • .

we have c0 ≤ c1 c2m. Consequently, the method of successive subspace corrections, based on multilevel subspaces (10) with a smoother satisfying (13), has the following convergence estimate: EJ2

A ≤

c1 c1 + c2m. 5.3. More general elliptic PDEs.

  • Model variational problem: Find u ∈ H1

0(Ω), (Ω ⊂ I

Rd is Lipschitz domain), satisfying: a(u, v) = f(v), for all v ∈ H1

0(Ω),

where a(u, v) =

d

  • i,j=1

∂v ∂xi αij ∂u ∂xj dx, f(v) =

fv dx, A = (αij) is symmetric and strictly positive definite for all x ∈ Ω, A and f are smooth enough so that the variational problem is well posed.

15

slide-16
SLIDE 16

5.4. Discretization and subspace splitting.

  • Discrete variational problem: Find uh ∈ Vh ⊂ H1

0(Ω), such that

a(uh, v) = f(v), for all v ∈ Vh, where Vh is a finite dimensional subspace.

  • This results in an algebraic system of linear equations (keeping the same notation for u and f):

Au = f, A S.P.D.

  • Given u0, a sequence of iterates is obtained via: uk+1 = uk + B(f − Auk)
  • Subspaces V1 ⊂ . . . ⊂ VJ−1 ⊂ VJ = Vh.
  • Solve approximately in each of them to define the action of B.
  • Obviously V1 + . . . + VJ−1 + VJ = VJ.

5.5. Basic multigrid iteration.

  • Given u0 iterate:

uk+1 = uk + B(f − Auk)

  • Action of B ≡ BJ:

0. If k = 1 then B1g = A−1

1 g

in the 1D example 1. Pre-smoothing: x1 = ST

k g

(I − Pnh) . . . (I − P1)(u − uk) 2. Coarse grid correction: (I − PH) 1. q0 = (P k

k−1)T (g − Akx1);

2. q1 = Bk−1q0; 3. x2 = x1 + P k

k−1q1;

3. Post-smoothing: Bkg = x2 + Sk(g − Akx2). – 5.6. Basic facts from Finite element theory.

  • Elliptic projections (on Vk, ; Pkv ∈ Vk): a([Pkv], wk) = a(v, wk),

∀wk ∈ Vk

  • L2 projections (on Vk, ; Qkv ∈ Vk): ([Qkv], wk) = (v, wk),

∀wk ∈ Vk

  • Local L2 projections (Qτv /

∈ Vk): ([Qτv], wk)τ = (v, wk)τ, ∀wk ∈ Vk|τ

16

slide-17
SLIDE 17

5.7. Error estimates.

  • Error estimates and other facts:

⋆ (I − Pk)va hkv2,Ω, ⋆ (I − Qk)v0 hkva ⋆ (I − Qτ)v0,τ hkv1,τ, ⋆ (I − Pk)v0 h2

kv2,Ω, (if Aubin-Nitsche argument works).

⋆ (I − Pk)v0 hkv1,Ω, (if Aubin-Nitsche argument works). ⋆ (I − Qk)va va (Qk is energy stable) ⋆ λk = λmax(Ak) = h−2, rescaling (i.e. using ℓ2 instead of L2, gives λmin(Ak) = h2

k.

5.8. On the addtive methods. nH subdomains ∪Ωk = Ω, H1

0(Ω) = k H1 0(Ωk).

B := T = nH

k=1 ¯

Tk (B−1v, v)a ≈ (v, v)a. Theorem (Vassilevski, Wang?). Let and T = nH

i=1 ¯

  • Tk. Then the minimizer is given by

(16) vk = ¯ Tk ¯ T −1v,

  • vk = v.

and (17) min

wk=v nH

  • i=1

( ¯ T −1

i

wi, wi)a = (T −1v, v)a. A sketch of proof for ¯ Tk = Pk. inf

vi=v

  • i

|vi|2

A

= inf

ξi=0

  • i

|wi + ξi|2

A

= inf

ξi=0

  • i

(|wi|2

A + |ξi|2 A) + 2

  • i

(Awi, ξi) = inf

ξi=0

  • i

(|wi|2

A + |ξi|2 A) + 2

  • i

(AiA−1

i QiT −1v, ξi)

= inf

ξi=0

  • i

(|wi|2

A + |ξi|2 A) + 2(T −1v,

  • i

ξi) = inf

ξi=0

  • i

(|wi|2

A + |ξi|2 A).

Finally we observe that

  • i

(Awi, wi) = (Aiwi, wi) = (AiA−1

i QiT −1v, wi) = (T −1v,

  • i

wi) = (T −1v, v).

17

slide-18
SLIDE 18

5.9. Multiplicative overlapping DD-method. (1) Partition of unity: J

i=1 θi(x) ≡ 1,

suppθi ⊂ Ωi ∪ ∂Ω, 0 ≤ θi ≤ 1, maxx∈¯

Ωi |∇θi(x)| ≤ c1h−1 0 .

and L2 projection Q0 : V → V0: h−1

0 v − Q0v0,Ω + |v − Q0v|1,Ω ≤ c2|v|1,Ω. Then c0 ≤ Λ1 Λ0 C0 with C0 = C0(c1, c2) independent of aij and J: J

  • k=0

Pk

J

  • i=k+1

vi2

a,Ω ≤ v − Q0v2 a,Ω + J

  • k=1

(

J

  • i=k+1

θi)(v − Q0v))2

a,Ωk

≤ Λ1

  • |v − Q0v|2

1,Ω + J

  • k=1

max

x∈¯ Ωk

|

J

  • i=k+1

∇θi(x)|v − Q0v2

0,Ωk + |v − Q0v|2 1,Ωk

Λ1C0|v|2

1,Ω ≤ Λ1

Λ0 C0v2

a,Ω

Note: convergence rate does not depend on the possible osscilations in aij! 5.10. Convergence for finite element equations.

  • Many of divide and conquer iterative methods fall into the category of subspace correction methods.
  • More abstract problem: Find u ∈ V such that Au = f,

f ∈ V ∗

  • Space decomposition: V =

i Vi (simplest example Vi being one dimensional and Vi := span{ei}, ei, the canonical basis vector in Rn

results in Gauss-Seidel method.

  • Algorithm MSC: With u ← u0, repeat the subspace correction: u ← u + ei (for i = 1 : J) with ei ∈ Vi

ei = M−1

i

Πi(f − Au), Mi ≈ Ai. 5.11. Convergence via subspace correction framework.

  • u − uℓ = EJ(u − uℓ−1),

EJ = (I − TJ) . . . (I − T1), with Tk = M−1

k Ak.

Theorem.[XU AND Z, J. of AMS (2003)] The energy norm of the error transfer operator EJA satisfies EJ2

A = 1 −

1 c0 + 1, where c0 = sup

v=1

inf

  • k vk=v

J

  • k=1

( ¯ T −1

k T ∗ k wk, T ∗ k wk)A with wk = J

  • j=i

vj − T −1

k vk.

  • Note: ¯

Tk ≡ T ∗

k + Tk − T ∗ k Tk SPD =

⇒ c0 < ∞.

  • Assumptions: Vk = V and Tkv2

A ≤ ω(Tkv, v)A, ω ∈ (0, 2).

18

slide-19
SLIDE 19
  • An attempt to get a sharp estimate for two level method suggests that a better parameter is

K(v, ¯ Tk) = v2

A + J

  • k=1

( ¯ T −1

k T ∗ k wk, T ∗ k wk)A

  • The convergence factor then will be (1 − 1

K ), where K = supvA=1 K(v, ¯

Tk).

  • After going through some simple algebra one ends up with

K(v, ¯ Tk) =

J

  • k=1

( ¯ T −1

k (vk + T ∗ k xk), (vk + T ∗ k xk))A,

where for xk are xk =

i>k vi.

  • It is important that xk depend on vk. Otherwise we could have chosen xk = 0 and K would be κ(T), where T is the additive preconditioner:

T =

k ¯

Tk. 5.12. Further simplifications: Richardson smoother.

  • For simplicity take Tk =

1 λk AkPk and it follows that

K(v, Tk) 2 ≤ K(v, ¯ Tk) ≤ K(v, Tk)

  • Using triangle inequality we have

K sup

vA=1

inf

vk=v J

  • k=1

(T −1

k vk, vk)A + Pkxk2 A

  • By choosing a fixed decomposition vk = v, one immediately gets an estimate above on K.

5.13. Some known results.

  • A natural decomposition is vk = (Pk − Pk−1)v. Then

K(v) ≤

J

  • k=1

λk(Pk − Pk−1)v2

  • In case of nested finite element spaces, a simple application of Aubin–Nitsche duality argument gives uniform bound (regularity??).
  • Another possible decomposition is (BPWX, 91) vk = (Qk − Qk−1)v and thus

K(v) ≤

J

  • k=1

λk(Qk − Qk−1)v2

0 + J

  • k=1

(Pk − Qk)v2

A

19

slide-20
SLIDE 20

5.14. Stability and convergence.

  • The norm of the error transfer operator EJA (which we want to minimize) critically depends upon a bound on the second term in the above

expansion: K(v) ≤

J

  • k=1

λk(Qk − Qk−1)v2

0 + J

  • k=1

(Pk − Qk)v2

A.

  • It is simple (but may be not that easy) to show that (Pk − Qk)2

A = Qk2 A − 1.

  • Hence, an important quantity to be estimated (even in the two level case) is QkA, with a constant independent of the level number k. This

is true for nested FE spaces, and hence K(v) ≤

J

  • k=1

λk(Qk − Qk−1)v2

0 + J

  • k=1

(Pk − Qk)v2

A C1 + C2J.

Such estimate without any regularity assumptions gives that the convergence rate depends only logarithmically on h (J ≈ | log2 h|). BASIC REFERENCES

  • 1. D. Braess and W. Hackbusch, A new convergence proof for the multigrid method including the V -cycle, SIAM J. Numer. Anal. 20 (1983), no. 5, 967–975.

MR 714691 (85h:65233)

  • 2. James H. Bramble, Multigrid methods, Pitman Research Notes in Mathematics Series, vol. 294, Longman Scientific & Technical, Harlow, 1993. MR 1247694

(95b:65002)

  • 3. James H. Bramble, Joseph E. Pasciak, Jun Ping Wang, and Jinchao Xu, Convergence estimates for multigrid algorithms without regularity assumptions, Math.
  • Comp. 57 (1991), no. 195, 23–45. MR 1079008 (91m:65158)
  • 4. William L. Briggs, Van Emden Henson, and Steve F. McCormick, A multigrid tutorial, second ed., Society for Industrial and Applied Mathematics (SIAM),

Philadelphia, PA, 2000. MR 1774296 (2001h:65002)

  • 5. Wolfgang Hackbusch, Multigrid methods and applications, Springer Series in Computational Mathematics, vol. 4, Springer-Verlag, Berlin, 1985. MR 814495

(87e:65082)

  • 6. Oren E. Livne and Achi Brandt, Multigrid techniques: 1984 guide with applications to fluid dynamics, Revised Edition ed., Society for Industrial and Applied

Mathematics (SIAM), Philadelphia, PA, 2011.

  • 7. Andrea Toselli and Olof Widlund, Domain decomposition methods—algorithms and theory, Springer Series in Computational Mathematics, vol. 34, Springer-

Verlag, Berlin, 2005. MR 2104179 (2005g:65006)

  • 8. U. Trottenberg, C. W. Oosterlee, and A. Sch¨

uller, Multigrid, Academic Press Inc., San Diego, CA, 2001, With contributions by A. Brandt, P. Oswald and K. St¨

  • uben. MR 1807961 (2002b:65002)
  • 9. Panayot S. Vassilevski, Multilevel block factorization preconditioners, Springer, New York, 2008, Matrix-based analysis and algorithms for solving finite

element equations. MR 2427040 (2010b:65007)

  • 10. Jinchao Xu, Iterative methods by space decomposition and subspace correction, SIAM Rev. 34 (1992), no. 4, 581–613. MR 1193013 (93k:65029)
  • 11. Jinchao Xu and Ludmil Zikatanov, The method of alternating projections and the method of subspace corrections in Hilbert space, J. Amer. Math. Soc. 15

(2002), no. 3, 573–597. MR 1896233 (2003f:65095)

20