Reduced order modeling and numerical linear algebra Akil Narayan 1 1 - - PowerPoint PPT Presentation

reduced order modeling and numerical linear algebra
SMART_READER_LITE
LIVE PREVIEW

Reduced order modeling and numerical linear algebra Akil Narayan 1 1 - - PowerPoint PPT Presentation

Reduced order modeling and numerical linear algebra Akil Narayan 1 1 Department of Mathematics, and Scientific Computing and Imaging (SCI) Institute University of Utah February 7, 2020 ICERM A. Narayan (U. Utah SCI) NLA and ROM Continuous


slide-1
SLIDE 1

Reduced order modeling and numerical linear algebra

Akil Narayan1

1Department of Mathematics, and Scientific Computing and Imaging (SCI) Institute

University of Utah

February 7, 2020 ICERM

  • A. Narayan

(U. Utah – SCI) NLA and ROM

slide-2
SLIDE 2

Continuous Ø discrete analogies

Most standard techniques for reduced basis methods can be understood from numerical linear algebra. Kolmogorov n widths Ø Singular value decompositions Reduced basis methods Ø QR decompositions Empirical interpolation methods Ø LU decompositions

  • A. Narayan

(U. Utah – SCI) NLA and ROM

slide-3
SLIDE 3

Kolmogorov n widths are (essentially) singular values

  • A. Narayan

(U. Utah – SCI) NLA and ROM

slide-4
SLIDE 4

Singular value decompositions

Let A P ❘MˆN, with M " N. We will think of the columns of A as snapshots. A :“ ¨ ˝ a1 a2 ¨ ¨ ¨ aN ˛ ‚ The SVD of A is A “ UΣV T , where U and V are orthogonal M ˆ M and N ˆ N matrices, respectively. Σ is a diagonal matrix with non-negative entries. We’ll use the following non-standard notation for the entries in Σ: σ0 ě σ1 ě ¨ ¨ ¨ ě σN´1.

  • A. Narayan

(U. Utah – SCI) NLA and ROM

slide-5
SLIDE 5

Low-rank approximations

Among the nice properties of the SVD is its ability to form low-rank approximations, Ak :“ U kΣkV T

k ,

1 ď k ď N, where U k and V k are k-column truncations, and Σk is a k ˆ k principcal submatrix truncation. With rankpAkq “ k, then Ak “ arg min

rankpBqďk

}A ´ B}˚ , for ˚ “ 2, F.

  • A. Narayan

(U. Utah – SCI) NLA and ROM

slide-6
SLIDE 6

Low-rank approximations

Among the nice properties of the SVD is its ability to form low-rank approximations, Ak :“ U kΣkV T

k ,

1 ď k ď N, where U k and V k are k-column truncations, and Σk is a k ˆ k principcal submatrix truncation. With rankpAkq “ k, then Ak “ arg min

rankpBqďk

}A ´ B}˚ , for ˚ “ 2, F. Equivalently, Ak is the projection of the columns of A onto RpU kq: Ak “ ¨ ˝ PRpUkqa1 PRpUkqa2 ¨ ¨ ¨ PRpUkqaN ˛ ‚

  • A. Narayan

(U. Utah – SCI) NLA and ROM

slide-7
SLIDE 7

Projections onto arbitrary spaces

What if we project A onto other spaces? If V Ă ❘M is any subspace, we could consider P V A :“ ¨ ˝ P V a1 P V a2 ¨ ¨ ¨ P V aN ˛ ‚

  • A. Narayan

(U. Utah – SCI) NLA and ROM

slide-8
SLIDE 8

Projections onto arbitrary spaces

What if we project A onto other spaces? If V Ă ❘M is any subspace, we could consider P V A :“ ¨ ˝ P V a1 P V a2 ¨ ¨ ¨ P V aN ˛ ‚ And we could ask about a certain type of error committed by this approximation EpV q :“ max

}x}2“1 }Ax ´ P V Ax}2

We know V “ RpU kq does a pretty good job. What about other spaces?

  • A. Narayan

(U. Utah – SCI) NLA and ROM

slide-9
SLIDE 9

Optimal projections

For a given rank k, an “optimal” projection commits the smallest error: Ek :“ min

V Ă❘M EpV q ❘ ❘

  • A. Narayan

(U. Utah – SCI) NLA and ROM

slide-10
SLIDE 10

Optimal projections

For a given rank k, an “optimal” projection commits the smallest error: Ek :“ min

V Ă❘M EpV q

So an extremal characterization of an SVD-based low rank approximation is RpU kq “ arg min

V Ă❘N

max

}x}2“1 }Ax ´ P Ax}2 . ❘

  • A. Narayan

(U. Utah – SCI) NLA and ROM

slide-11
SLIDE 11

Optimal projections

For a given rank k, an “optimal” projection commits the smallest error: Ek :“ min

V Ă❘M EpV q

So an extremal characterization of an SVD-based low rank approximation is RpU kq “ arg min

V Ă❘N

max

}x}2“1 }Ax ´ P Ax}2 .

Or, an (unnecessarily?) pedantic alternative: Ek “ σkpAq “ min

V Ă❘N

max

}x}2“1 min vPV }Ax ´ v}2

  • A. Narayan

(U. Utah – SCI) NLA and ROM

slide-12
SLIDE 12

SVD projections

Given A P ❘MˆN, the success of a low-rank projection is dictated by the approximation numbers σkpAq “ min

V Ă❘N

max

}x}2“1 min vPV }Ax ´ v}2 .

More precisely, it is dictated by fast decay of these numbers as k increases. ❘ ❘ ❘ ❘

  • A. Narayan

(U. Utah – SCI) NLA and ROM

slide-13
SLIDE 13

SVD projections

Given A P ❘MˆN, the success of a low-rank projection is dictated by the approximation numbers σkpAq “ min

V Ă❘N

max

}x}2“1 min vPV }Ax ´ v}2 .

More precisely, it is dictated by fast decay of these numbers as k increases. These numbers are defined by our choice of metric on “output” space ❘M, and our choice of metric on “measurement” space ❘N. I.e., a generalization might look like σk ´ A; ℓp ´ ❘M¯ , ℓq ´ ❘N¯¯ “ min

dim V ďk

max

}x}q“1 min vPV }Ax ´ v}p .

  • A. Narayan

(U. Utah – SCI) NLA and ROM

slide-14
SLIDE 14

Kolmogorov n widths

σn ´ A; ℓp ´ ❘M¯ , ℓq ´ ❘N¯¯ “ min

dim V ďn

max

}x}q“1 min vPV }Ax ´ v}p .

These numbers tell us how well the columns of A are ℓp-approximated by a linear space using ℓq measurements. Another definition might be the maximum column norm error: σn ´ A; ℓp ´ ❘M¯¯ “ min

dim V ďn max iPrNs min vPV }Aei ´ v}p .

  • Great. How do we do all this with functions?
  • A. Narayan

(U. Utah – SCI) NLA and ROM

slide-15
SLIDE 15

Kolmogorov n widths

σn ´ A; ℓp ´ ❘M¯ , ℓq ´ ❘N¯¯ “ min

dim V ďn

max

}x}q“1 min vPV }Ax ´ v}p .

These numbers tell us how well the columns of A are ℓp-approximated by a linear space using ℓq measurements. Another definition might be the maximum column norm error: σn ´ A; ℓp ´ ❘M¯¯ “ min

dim V ďn max iPrNs min vPV }Aei ´ v}p .

  • Great. How do we do all this with functions?

Let A be a collection of functions in a Hilbert space H. Then one way to talk about similar concepts to (ℓ2) singular values is σn pA; Hq “ inf

dim V ďn sup aPA

inf

vPV }a ´ v}

This is called the Kolmogorov n width of A (with respect to H).

  • A. Narayan

(U. Utah – SCI) NLA and ROM

slide-16
SLIDE 16

Reduced basis methods (essentially) perform QR decompositions

  • A. Narayan

(U. Utah – SCI) NLA and ROM

slide-17
SLIDE 17

Interpolative decompositions

One disadvantage of SVD-based low rank approximations, A “ ¨ ˝ a1 a2 ¨ ¨ ¨ aN ˛ ‚“ UΣV T , is that we need information from all columns of A to define U.

  • A. Narayan

(U. Utah – SCI) NLA and ROM

slide-18
SLIDE 18

Interpolative decompositions

One disadvantage of SVD-based low rank approximations, A “ ¨ ˝ a1 a2 ¨ ¨ ¨ aN ˛ ‚“ UΣV T , is that we need information from all columns of A to define U. One alternative: Interpolative decompositions, or matrix skeletonizations. Basic idea: project all columns of A onto a subspace spanned by a few columns. A rank-n column skeletonization of A is B “ AS ´ AT

SAS

¯: AT

S

l jh n

P RpAS q

A, AS :“ A ¨ ˝ es1 es2 ¨ ¨ ¨ esn ˛ ‚, with S “ ts1, . . . snu Ă rNs.

  • A. Narayan

(U. Utah – SCI) NLA and ROM

slide-19
SLIDE 19

Choosing the columns S

The problem of choosing S that is optimal in some metric is the column subset selection problem. For metrics of interest, it’s NP-hard.

  • A. Narayan

(U. Utah – SCI) NLA and ROM

slide-20
SLIDE 20

Choosing the columns S

The problem of choosing S that is optimal in some metric is the column subset selection problem. For metrics of interest, it’s NP-hard. So let’s do something else: Let’s pick columns greedily: Given S Ă rNs of size n, we’ll add a column index via the procedure sn`1 “ arg max

jPrNs

› ›aj ´ P RpASqaj › ›

2 .

This is much cheaper since I need only to evaluate N vector norms at each step.

  • A. Narayan

(U. Utah – SCI) NLA and ROM

slide-21
SLIDE 21

Choosing the columns S

The problem of choosing S that is optimal in some metric is the column subset selection problem. For metrics of interest, it’s NP-hard. So let’s do something else: Let’s pick columns greedily: Given S Ă rNs of size n, we’ll add a column index via the procedure sn`1 “ arg max

jPrNs

› ›aj ´ P RpASqaj › ›

2 .

This is much cheaper since I need only to evaluate N vector norms at each step. There’s already a well-polished algorithm that does this: the QR decomposition.

  • A. Narayan

(U. Utah – SCI) NLA and ROM

slide-22
SLIDE 22

The QR decomposition (1/2)

The column-pivoted QR decomposition iteratively computes orthonormal vectors in the range of A. At step j, the next column is identified as the one whose projected residual is largest. P j´1 :“ Qj´1QT

j´1

sj “ arg max

jPrNs

}aj ´ P j´1aj}2 qj :“ asj }asj}2 , Qj “ “ Qj´1, qj ‰

  • A. Narayan

(U. Utah – SCI) NLA and ROM

slide-23
SLIDE 23

The QR decomposition (1/2)

The column-pivoted QR decomposition iteratively computes orthonormal vectors in the range of A. At step j, the next column is identified as the one whose projected residual is largest. P j´1 :“ Qj´1QT

j´1

sj “ arg max

jPrNs

}aj ´ P j´1aj}2 qj :“ asj }asj}2 , Qj “ “ Qj´1, qj ‰ The residual rj´1 :“ › ›asj ´ P j´1asj › ›

2,

is the largest (ℓ2-norm) column mistake we make by choosing S “ ts1, . . . , sj´1u, i.e., by replacing A Ð P V A, V :“ spantas1, . . . , asj´1u.

  • A. Narayan

(U. Utah – SCI) NLA and ROM

slide-24
SLIDE 24

The QR decomposition (2/2)

This algorithm is a greedy algorithm: instead of all-at-once optimization, we

  • ptimize one at a time.

Clearly, we don’t expect this to perform as well as the optimal SVD-based subspace. But how well does this greedy procedure work in practice?

  • A. Narayan

(U. Utah – SCI) NLA and ROM

slide-25
SLIDE 25

Discrete greedy algorithms

In some cases, this greedy algorithm performs comparably to an optimal (SVD) algorithm. In particular, σrpAq À expp´brq ù ñ sj À expp´crq, where c ă b.[Harbrecht, Peters, Schneider 2010]

  • A. Narayan

(U. Utah – SCI) NLA and ROM

slide-26
SLIDE 26

Back to the continuous world

Once more, let’s put this into appropriate language for functions. Let A be a collection of functions, parameterized by µ P ❘d, A “ ! upµq ˇ ˇ µ P Γ Ă ❘d) .

  • A. Narayan

(U. Utah – SCI) NLA and ROM

slide-27
SLIDE 27

Back to the continuous world

Once more, let’s put this into appropriate language for functions. Let A be a collection of functions, parameterized by µ P ❘d, A “ ! upµq ˇ ˇ µ P Γ Ă ❘d) . A greedy (pivoted QR!) approach to determining a low-rank space for approximation is µj “ arg max

µPΓ

}upµq ´ Pj´1upµq} , where Pj´1 is the projection operator onto spantupµ1q, . . . , upµj´1qu.

  • A. Narayan

(U. Utah – SCI) NLA and ROM

slide-28
SLIDE 28

Back to the continuous world

Once more, let’s put this into appropriate language for functions. Let A be a collection of functions, parameterized by µ P ❘d, A “ ! upµq ˇ ˇ µ P Γ Ă ❘d) . A greedy (pivoted QR!) approach to determining a low-rank space for approximation is µj “ arg max

µPΓ

}upµq ´ Pj´1upµq} , where Pj´1 is the projection operator onto spantupµ1q, . . . , upµj´1qu. This is (essentially) the reduced basis method.

  • A. Narayan

(U. Utah – SCI) NLA and ROM

slide-29
SLIDE 29

Residuals?

One disadvantage of SVD-based low rank approximations is that we need all columns of A. (“One disadvantage of Kolmogorov n-width low rank approximations is that we need all functions in A.”)

  • A. Narayan

(U. Utah – SCI) NLA and ROM

slide-30
SLIDE 30

Residuals?

One disadvantage of SVD-based low rank approximations is that we need all columns of A. (“One disadvantage of Kolmogorov n-width low rank approximations is that we need all functions in A.”) A “QR” approach still requires the residual µj “ arg max

µPΓ

}upµq ´ Pj´1upµq} , which, naively, still requires upµq.

  • A. Narayan

(U. Utah – SCI) NLA and ROM

slide-31
SLIDE 31

Residuals?

One disadvantage of SVD-based low rank approximations is that we need all columns of A. (“One disadvantage of Kolmogorov n-width low rank approximations is that we need all functions in A.”) A “QR” approach still requires the residual µj “ arg max

µPΓ

}upµq ´ Pj´1upµq} , which, naively, still requires upµq. RBM methods get around this in the same way that one can get around knowing exact solutions to linear systems: Ljaj “ bj ù ñ }aj ´ z} ď 1 σminpLjq }bj ´ Ljz}2

  • A. Narayan

(U. Utah – SCI) NLA and ROM

slide-32
SLIDE 32

RBM and QR decompositions

RBM algorithms perform snapshot-based model reduction via a QR decomposition. L pupµq; µq “ bpµq ó }upµq ´ Pj´1upµq} ď 1 “σminpLq” }bpµq ´ L pPj´1upµq; µq}2 (1)

  • A. Narayan

(U. Utah – SCI) NLA and ROM

slide-33
SLIDE 33

RBM and QR decompositions

RBM algorithms perform snapshot-based model reduction via a QR decomposition. L pupµq; µq “ bpµq ó }upµq ´ Pj´1upµq} ď 1 “σminpLq” }bpµq ´ L pPj´1upµq; µq}2 (1) This residual: can be computed without computing u if Lp¨; µq depends on µ in an affine way, provides a rigorous bound on error committed if “σminpLq” can be computed (a posteriori error estimates) Even though (1) is only an inequality, this “weak” greedy algorithm still produces a good approximation, assuming the n width decays quickly.

[Binev, Cohen, Dahmen, Devore, Petrova, Wojtaszczyk 2011], [Devore, Petrova, Wojtaszczyk 2013]

  • A. Narayan

(U. Utah – SCI) NLA and ROM

slide-34
SLIDE 34

Empirical interpolation methods (essentially) perform LU decompositions

  • A. Narayan

(U. Utah – SCI) NLA and ROM

slide-35
SLIDE 35

Affine dependence

Many times, L does not depend on µ in an affine way. In particular, L may contain functions of µ, e.g., Lpu; µq “ ´∇x ¨ pℓpx; µq∇xq u. This is affine only if ℓpx; µq “

d

ÿ

i“1

fipµqℓipxq.

  • A. Narayan

(U. Utah – SCI) NLA and ROM

slide-36
SLIDE 36

Affine dependence

Many times, L does not depend on µ in an affine way. In particular, L may contain functions of µ, e.g., Lpu; µq “ ´∇x ¨ pℓpx; µq∇xq u. This is affine only if ℓpx; µq “

d

ÿ

i“1

fipµqℓipxq. An affine approximation for L (i.e., for ℓ) is often accomplished via empirical interpolation.[Barrault, Maday, Nguyen, Patera 2004]

  • A. Narayan

(U. Utah – SCI) NLA and ROM

slide-37
SLIDE 37

Empirical interpolation

Once again, let’s understand this in the discrete setting: L “ ¨ ˝ ℓ1 ℓ2 ¨ ¨ ¨ ℓN ˛ ‚ One consequence of continuous problem practicalities: want to avoid computing column norms.

  • A. Narayan

(U. Utah – SCI) NLA and ROM

slide-38
SLIDE 38

Empirical interpolation

Once again, let’s understand this in the discrete setting: L “ ¨ ˝ ℓ1 ℓ2 ¨ ¨ ¨ ℓN ˛ ‚ One consequence of continuous problem practicalities: want to avoid computing column norms. One strategy is an “incomplete” LU factorization. A (complete-pivoting) factorization is P LQ “ ZU, where Z is lower triangular, and P and Q are permutation matrices.

  • A. Narayan

(U. Utah – SCI) NLA and ROM

slide-39
SLIDE 39

Empirical interpolation

Once again, let’s understand this in the discrete setting: L “ ¨ ˝ ℓ1 ℓ2 ¨ ¨ ¨ ℓN ˛ ‚ One consequence of continuous problem practicalities: want to avoid computing column norms. One strategy is an “incomplete” LU factorization. A (complete-pivoting) factorization is P LQ “ ZU, where Z is lower triangular, and P and Q are permutation matrices. An approximation would be an incomplete factorization: P LQ « ZdU d, where Zd (U d) is a principal d-column (-row) truncation.

  • A. Narayan

(U. Utah – SCI) NLA and ROM

slide-40
SLIDE 40

Empirical interpolation

Once again, let’s understand this in the discrete setting: L “ ¨ ˝ ℓ1 ℓ2 ¨ ¨ ¨ ℓN ˛ ‚ One consequence of continuous problem practicalities: want to avoid computing column norms. One strategy is an “incomplete” LU factorization. A (complete-pivoting) factorization is P LQ “ ZU, where Z is lower triangular, and P and Q are permutation matrices. An approximation would be an incomplete factorization: P LQ « ZdU d, where Zd (U d) is a principal d-column (-row) truncation. In the continuous setting, this is called the empirical interpolation method (EIM). P : Spatial points for interpolation Q: Parameter values defining snapshots used for spatial interpolation

  • A. Narayan

(U. Utah – SCI) NLA and ROM

slide-41
SLIDE 41

Continuous Ø discrete analogies

Kolmogorov n widths Ø Singular value decompositions Reduced basis methods Ø QR decompositions Empirical interpolation methods Ø LU decompositions

  • A. Narayan

(U. Utah – SCI) NLA and ROM

slide-42
SLIDE 42

Continuous Ø discrete analogies

Kolmogorov n widths Ø Singular value decompositions Reduced basis methods Ø QR decompositions Empirical interpolation methods Ø LU decompositions Bonus! Why do Kolmogorov n widths decay quickly? (for “nice” problems)

  • A. Narayan

(U. Utah – SCI) NLA and ROM

slide-43
SLIDE 43

Polynomial approximations

Recall some complex analysis: Suppose f : ❈ Ñ ❈ is a holomorphic function in some open disc D of the complex plane. Let Γ be a subset of D, with ¯ Γ Ă D, and dpΓ, BDq ě r.

  • A. Narayan

(U. Utah – SCI) NLA and ROM

slide-44
SLIDE 44

Polynomial approximations

Recall some complex analysis: Suppose f : ❈ Ñ ❈ is a holomorphic function in some open disc D of the complex plane. Let Γ be a subset of D, with ¯ Γ Ă D, and dpΓ, BDq ě r. Then Taylor’s theorem implies that if p is the degree-n Taylor polynomial centered around any z0 P Γ then sup

zPΓ

}fpzq ´ ppzq} À r´n.

  • A. Narayan

(U. Utah – SCI) NLA and ROM

slide-45
SLIDE 45

Polynomial approximations

Recall some complex analysis: Suppose f : ❈ Ñ ❈ is a holomorphic function in some open disc D of the complex plane. Let Γ be a subset of D, with ¯ Γ Ă D, and dpΓ, BDq ě r. Then Taylor’s theorem implies that if p is the degree-n Taylor polynomial centered around any z0 P Γ then sup

zPΓ

}fpzq ´ ppzq} À r´n. I.e., polynomial approximations are exponentially accurate for smooth functions.

  • A. Narayan

(U. Utah – SCI) NLA and ROM

slide-46
SLIDE 46

Parameterized elliptic PDEs (1/2)

Now consider the elliptic PDE ´∇x pℓpx; µq∇xq “ bpx; µq. Suppose ℓpx; µq is continuous, is µ-uniformly bounded, depends on µ in an affine way, and inf

x ℓpx, ; µq ą rmin ą 0,

uniformly for µ P Γ Ă ❘d. Let 0 P Γ. Then the solution µ ÞÑ upµq exists and is well-defined in some Hilbert space H.

  • A. Narayan

(U. Utah – SCI) NLA and ROM

slide-47
SLIDE 47

Parameterized elliptic PDEs (1/2)

Now consider the elliptic PDE ´∇x pℓpx; µq∇xq “ bpx; µq. Suppose ℓpx; µq is continuous, is µ-uniformly bounded, depends on µ in an affine way, and inf

x ℓpx, ; µq ą rmin ą 0,

uniformly for µ P Γ Ă ❘d. Let 0 P Γ. Then the solution µ ÞÑ upµq exists and is well-defined in some Hilbert space H. Under these conditions, then µ ÞÑ up¨, µq is (complex) differentiable in an open disc D, with distpΓ, BDq „ rmin. In particular, all µ-derivatives of u at µ “ 0 exist and are H-valued.

  • A. Narayan

(U. Utah – SCI) NLA and ROM

slide-48
SLIDE 48

Parameterized elliptic PDEs (2/2)

Since µ ÞÑ upµq is complex differentiable in Γ with radius rmin: Taylor’s Theorem guarantees a degree-n, d-variate polynomial approximation pn with N À nd degrees of freedom such that sup

µPO

}upµq ´ pnpzq} À r´n

min „ r´Np1{dq min

Hence, the Kolmogorov width of the manifold of solutions (in H) decays in N, but suffers the curse of dimensionality.[Cohen, Devore 2015]

  • A. Narayan

(U. Utah – SCI) NLA and ROM

slide-49
SLIDE 49

Parameterized elliptic PDEs (2/2)

Since µ ÞÑ upµq is complex differentiable in Γ with radius rmin: Taylor’s Theorem guarantees a degree-n, d-variate polynomial approximation pn with N À nd degrees of freedom such that sup

µPO

}upµq ´ pnpzq} À r´n

min „ r´Np1{dq min

Hence, the Kolmogorov width of the manifold of solutions (in H) decays in N, but suffers the curse of dimensionality.[Cohen, Devore 2015] In short, Kolmogorov widths decay quickly when u depends smoothly on the parameter, but suffer from (classical) approximation limitations.

  • A. Narayan

(U. Utah – SCI) NLA and ROM