Stability of Krylov Subspace Spectral Methods James V. Lambers - - PowerPoint PPT Presentation

stability of krylov subspace spectral methods
SMART_READER_LITE
LIVE PREVIEW

Stability of Krylov Subspace Spectral Methods James V. Lambers - - PowerPoint PPT Presentation

Stability of Krylov Subspace Spectral Methods James V. Lambers Department of Energy Resources Engineering Stanford University includes joint work with Patrick Guidotti and Knut Slna, UC Irvine Margot Gerritsen, Stanford University


slide-1
SLIDE 1

Stability of Krylov Subspace Spectral Methods

James V. Lambers Department of Energy Resources Engineering Stanford University

includes joint work with Patrick Guidotti and Knut Sølna, UC Irvine Margot Gerritsen, Stanford University Harrachov ‘07 August 23, 2007

slide-2
SLIDE 2

August 23, 2007 Harrachov '07

Model Variable-coefficient Problem

where We assume

  • is positive semi-definite,

and that the coefficients are smooth

2/30

slide-3
SLIDE 3

August 23, 2007 Harrachov '07

Quick-and-dirty Solution

Let { φ φ φ φν

ν ν ν } be a set of orthonormal π-periodic

  • functions. Then, an approximate solution is

This works if the φ φ φ φν

ν ν ν are nearly eigenfunctions,

but if not, how can we compute

  • φ

φ φ φν

ν ν ν

  • as

accurately and efficiently as possible? where

3/30

slide-4
SLIDE 4

August 23, 2007 Harrachov '07

Elements of Functions of Matrices

If

  • is
  • ×

× × ×

  • and symmetric, then
  • is

given by a Riemann-Stieltjes integral provided the measure α α α α

  • λ

λ λ λ

  • which is based on

the spectral decomposition of

  • , is positive and

increasing This is the case if

  • , or if
  • is a small

perturbation of

  • 4/30
slide-5
SLIDE 5

August 23, 2007 Harrachov '07

The

≠ ≠ ≠

  • Case

For general

  • and
  • , the bilinear form
  • can be obtained by writing it as the difference

quotient where δ δ δ δ is a small constant. Both forms lead to Riemann-Stieltjes integrals with positive, increasing measures How can we approximate these integrals?

5/30

slide-6
SLIDE 6

August 23, 2007 Harrachov '07

Gaussian Quadrature

  • These two integrals can be approximated

using Gaussian quadrature rules (G. Golub and G. Meurant, '94)

  • The nodes and weights are obtained by

applying the Lanczos algorithm to

  • with

starting vectors

  • and
  • , the

tridiagonal matrix of recursion coefficients

  • The nodes are the eigenvalues of
  • , and the

weights are the products of the first components of the left and right eigenvectors

  • We want rules for each Fourier component

6/30

slide-7
SLIDE 7

August 23, 2007 Harrachov '07

What if

  • is a differential operator?
  • If
  • ω
  • ω

ω ω ω

  • , then recursion coefficients

α α α α

  • , β

β β β

  • are functions of ω

ω ω ω

  • Let
  • from before, with
  • The recursion coefficients for a 2-node

Gaussian rule are where

  • Similar formulas apply in higher dimensions

7/30

slide-8
SLIDE 8

August 23, 2007 Harrachov '07

Updating Coefficients for ≠ ≠ ≠ ≠

To produce modified recursion coefficients generated by

  • and
  • + δ

δ δ δ

  • :

This is particularly useful if the

  • , α

α α α

  • , β

β β β

  • are

parameterized families and

  • is a fixed vector

8/30

slide-9
SLIDE 9

August 23, 2007 Harrachov '07

Krylov Subspace Spectral Methods

To compute Fourier components of

  • :
  • Apply symmetric Lanczos algorithm to
  • with

starting vector

  • ω

ω ω ω

  • Use fast updating to obtain modified recursion

coefficients for starting vectors

  • ω

ω ω ω

  • ,
  • ω

ω ω ω

  • δ

δ δ δ

  • Approximate, by Gaussian quadrature,
  • Finally,

9/30

slide-10
SLIDE 10

August 23, 2007 Harrachov '07

Why Do it This Way?

  • Other, more general Krylov subspace

methods (e.g. M. Hochbruck and C. Lubich, '96) use a single Krylov subspace for each time step, with

  • as the starting vector, to

approximate

  • KSS methods obtain Fourier components from

derivatives of frequency-dependent Krylov subspace bases in the direction of

  • Thus, each component receives individual

attention, enhancing accuracy and stability

10/30

slide-11
SLIDE 11

August 23, 2007 Harrachov '07

Properties

  • They're High-Order Accurate!

Each Fourier component of

  • is

computed with local accuracy of

∆ ∆ ∆

  • ,

where

  • is the number of nodes in the

Gaussian rule

  • They're Explicit but Very Stable!

If

  • is constant and
  • is bandlimited, then

for

  • , method is unconditionally stable!

For

  • , solution operator is bounded

independently of ∆ ∆ ∆ ∆

  • and
  • 11/30
slide-12
SLIDE 12

August 23, 2007 Harrachov '07

Demonstrating Stability

Fourier, KSS methods applied to

  • , smooth initial data until
  • Contrasting Krylov subspace methods

applied to parabolic problem on different grids

They do not experience the same difficulties with stiffness as other Krylov subspace methods, or the same weak instability as the unsmoothed Fourier method

12/30

slide-13
SLIDE 13

August 23, 2007 Harrachov '07

Properties, cont’d

  • They’re efficient and scalable!

– Performance of MATLAB implementation comparable or superior to that of built-in ODE solvers – Accuracy and efficiency scale to finer grids or higher spatial dimension

13/30

slide-14
SLIDE 14

August 23, 2007 Harrachov '07

In the Limit: Derivatives of Moments!

  • Each Fourier component approximates the

Gâteaux derivative

  • node approximate solution has the form

14/30

slide-15
SLIDE 15

August 23, 2007 Harrachov '07

The Splitting Perspective

  • Derivatives of nodes and weights w.r.t. δ

δ δ δ are Fourier components of applications of pseudodifferential operators applied to

  • node approximate solution has form

where and each

  • is a constant-coefficient

pseudo-differential operator, of the same

  • rder as
  • , and positive semi-definite

15/30

slide-16
SLIDE 16

August 23, 2007 Harrachov '07

Simplest Splittings

  • The case
  • reduces to solving the

averaged-coefficient problem exactly, after applying forward Euler with the residual

  • perator
  • In the case
  • , with
  • constant, the
  • perators
  • are second-order, and yet the

approximate solution operator

∆ ∆ ∆

  • satisfies

where

  • is a bounded operator!

16/30

slide-17
SLIDE 17

August 23, 2007 Harrachov '07

General Splittings

  • Operators
  • and
  • are defined in terms of

derivatives of nodes and weights, which represent pseudo-differential operators

  • For each ω

ω ω ω, let

  • ω

ω ω ω be the

  • ×

× × ×

  • Jacobi matrix
  • utput by Lanczos, with initial vector
  • ω

ω ω ω

– Derivatives of nodes:

  • ω

ω ω ω

  • Λ

Λ Λ Λω

ω ω ω

  • , where nodes

are on diagonal of Λ Λ Λ Λω

ω ω ω, and Λ

Λ Λ Λω

ω ω ω’

  • ω

ω ω ω’

  • , where
  • ω

ω ω ω’

  • is available from fast updating algorithm

– Derivatives of weights: from solution of systems of form

  • , where
  • ω

ω ω ω(1:

(1: (1: (1:

− − −1,1: 1,1: 1,1: 1,1:

− − −1) 1) 1) 1)

  • λ

λ λ λ

  • , for
  • , with
  • a rank-one matrix

17/30

slide-18
SLIDE 18

August 23, 2007 Harrachov '07

The Wave Equation

  • The integrands are obtained from the spectral

decomposition of the propagator (P. Guidotti, JL and K. Sølna '06):

  • For each Fourier component,

∆ ∆ ∆

  • local

accuracy!

  • For
  • ,
  • constant,
  • bandlimited,

global error is 3rd order in time, and the method is unconditionally stable!

18/30

slide-19
SLIDE 19

August 23, 2007 Harrachov '07

Results for the Wave Equation

Comparison of 2-node KSS method with semi- implicit 2nd-order method of H.-O. Kreiss, N. Petersson and J. Yström, and 4th-order explicit scheme of B. Gustafsson and E. Mossberg

19/30

slide-20
SLIDE 20

August 23, 2007 Harrachov '07

Discontinuous Coefficients and Data

20/30

slide-21
SLIDE 21

August 23, 2007 Harrachov '07

The Modified Perona-Malik equation

  • The Perona-Malik equation is a nonlinear

diffusion equation used for image de-noising

  • It exhibits both forward and backward

diffusion, and so is ill-posed, but surprisingly well-behaved numerically

  • A modification proposed by P. Guidotti

weakens the nonlinearity slightly, to obtain well-posedness while still de-noising

  • KSS methods can limit effects of backward

diffusion by truncating recursion coefficients for selected frequencies

21/30

slide-22
SLIDE 22

August 23, 2007 Harrachov '07

De-Noising by Modified Perona-Malik

22/30

slide-23
SLIDE 23

August 23, 2007 Harrachov '07

Handling Discontinuities

  • In progress: other bases (e.g. wavelets, multiwavelets)

for problems with rough or discontinuous coefficients

  • Ideally, trial functions should conform to the geometry
  • f the symbol as much as possible

23/30

  • Encouraging results: Freud

reprojection (A. Gelb & J. Tanner ’06; code by A. Nelson) to deal with Gibbs phenomenon

  • Improves accuracy for heat

equation, stability for wave

slide-24
SLIDE 24

August 23, 2007 Harrachov '07

Preconditioning Through Homogenizing Transformations

  • KSS methods are most accurate when the

coefficients are smooth, so trial functions are also approximate eigenfunctions

  • Problem can be preconditioned by applying

unitary similarity transformations that homogenize coefficients

  • In 1-D, leading coefficient
  • is easily

homogenized by a change of independent variable; to make unitary, add a diagonal transformation

24/30

slide-25
SLIDE 25

August 23, 2007 Harrachov '07

Continuing the Process

  • To homogenize
  • , we use a transformation
  • f the form
  • *
  • :

where

  • is the pseudo-inverse of
  • This introduces variable coefficients of order

  • , but process can be repeated
  • Generalizes to higher dimensions, can be

applied efficiently using Fast FIO algorithms (Candes, Demanet and Ying '06)

25/30

slide-26
SLIDE 26

August 23, 2007 Harrachov '07

Smoothing

These transformations yield an operator that is nearly constant-coefficient. The gain in accuracy is comparable to that achieved by deferred correction with an analytically exact residual, which KSS methods naturally provide!

26/30

slide-27
SLIDE 27

August 23, 2007 Harrachov '07

In Progress: Systems of Equations

  • Generalization to systems of equations is

straightforward

  • For a system of the form
  • ,

where each entry of

  • is a differential
  • perator, we can use trial functions

⊕ ⊕ ⊕

  • ω

ω ω ω

  • where
  • is an eigenvector of
  • ω

ω ω ω

  • For systems arising from acoustics, in which

the solution operator can be expressed in terms of products of entries of

  • , order

∆ ∆ ∆

  • accuracy is possible
  • More on this topic at ICNAAM ‘07

27/30

slide-28
SLIDE 28

August 23, 2007 Harrachov '07

Conclusions

  • Krylov subspace spectral methods are

showing more promise as their development progresses

– Applicability to problems with rough behavior – Stability like that of implicit methods – Competitive performance and scalability

  • Through the splitting perspective, they

provide an effective means of stably extending other solution methods to variable- coefficient problems

  • Still much to do! Especially implementation

with other bases of trial functions

28/30

slide-29
SLIDE 29

August 23, 2007 Harrachov '07

References

  • JL, “Krylov Subspace Spectral Methods for Variable-

Coefficient Initial-Boundary Value Problems”, Electr.

  • Trans. Num. Anal. 20 (2005), p. 212-234
  • P. Guidotti, JL, K. Sølna, “Analysis of 1D Wave

Propagation in Inhomogeneous Media”, Num. Funct.

  • Anal. Opt. 27 (2006), p. 25-55
  • JL, “Practical Implementation of Krylov Subspace

Spectral Methods”, J. Sci. Comput. 32 (2007), p. 449- 476

  • JL, “Derivation of High-Order Spectral Methods for

Time-Dependent PDE Using Modified Moments”, submitted

29/30

slide-30
SLIDE 30

August 23, 2007 Harrachov '07

Contact Info

James Lambers lambers@stanford.edu http://www.stanford.edu/~lambers/

30/30